Search results for: granular computing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1198

Search results for: granular computing

748 Dynamic Properties of Recycled Concrete Aggregate from Resonant Column Tests

Authors: Wojciech Sas, Emil Soból, Katarzyna Gabryś, Andrzej Głuchowski, Alojzy Szymański

Abstract:

Depleting of natural resources is forcing the man to look for alternative construction materials. One of them is recycled concrete aggregates (RCA). RCA from the demolition of buildings and crushed to proper gradation can be a very good replacement for natural unbound granular aggregates, gravels or sands. Physical and the mechanical properties of RCA are well known in the field of basic civil engineering applications, but to proper roads and railways design dynamic characteristic is need as well. To know maximum shear modulus (GMAX) and the minimum damping ratio (DMIN) of the RCA dynamic loads in resonant column apparatus need to be performed. The paper will contain literature revive about alternative construction materials and dynamic laboratory research technique. The article will focus on dynamic properties of RCA, but early studies conducted by the authors on physical and mechanical properties of this material also will be presented. The authors will show maximum shear modulus and minimum damping ratio. Shear modulus and damping ratio degradation curves will be shown as well. From exhibited results conclusion will be drawn at the end of the article.

Keywords: recycled concrete aggregate, shear modulus, damping ratio, resonant column

Procedia PDF Downloads 399
747 A Framework of Virtualized Software Controller for Smart Manufacturing

Authors: Pin Xiu Chen, Shang Liang Chen

Abstract:

A virtualized software controller is developed in this research to replace traditional hardware control units. This virtualized software controller transfers motion interpolation calculations from the motion control units of end devices to edge computing platforms, thereby reducing the end devices' computational load and hardware requirements and making maintenance and updates easier. The study also applies the concept of microservices, dividing the control system into several small functional modules and then deploy into a cloud data server. This reduces the interdependency among modules and enhances the overall system's flexibility and scalability. Finally, with containerization technology, the system can be deployed and started in a matter of seconds, which is more efficient than traditional virtual machine deployment methods. Furthermore, this virtualized software controller communicates with end control devices via wireless networks, making the placement of production equipment or the redesign of processes more flexible and no longer limited by physical wiring. To handle the large data flow and maintain low-latency transmission, this study integrates 5G technology, fully utilizing its high speed, wide bandwidth, and low latency features to achieve rapid and stable remote machine control. An experimental setup is designed to verify the feasibility and test the performance of this framework. This study designs a smart manufacturing site with a 5G communication architecture, serving as a field for experimental data collection and performance testing. The smart manufacturing site includes one robotic arm, three Computer Numerical Control machine tools, several Input/Output ports, and an edge computing architecture. All machinery information is uploaded to edge computing servers and cloud servers via 5G communication and the Internet of Things framework. After analysis and computation, this information is converted into motion control commands, which are transmitted back to the relevant machinery for motion control through 5G communication. The communication time intervals at each stage are calculated using the C++ chrono library to measure the time difference for each command transmission. The relevant test results will be organized and displayed in the full-text.

Keywords: 5G, MEC, microservices, virtualized software controller, smart manufacturing

Procedia PDF Downloads 82
746 Security Design of Root of Trust Based on RISC-V

Authors: Kang Huang, Wanting Zhou, Shiwei Yuan, Lei Li

Abstract:

Since information technology develops rapidly, the security issue has become an increasingly critical for computer system. In particular, as cloud computing and the Internet of Things (IoT) continue to gain widespread adoption, computer systems need to new security threats and attacks. The Root of Trust (RoT) is the foundation for providing basic trusted computing, which is used to verify the security and trustworthiness of other components. Design a reliable Root of Trust and guarantee its own security are essential for improving the overall security and credibility of computer systems. In this paper, we discuss the implementation of self-security technology based on the RISC-V Root of Trust at the hardware level. To effectively safeguard the security of the Root of Trust, researches on security safeguard technology on the Root of Trust have been studied. At first, a lightweight and secure boot framework is proposed as a secure mechanism. Secondly, two kinds of memory protection mechanism are built to against memory attacks. Moreover, hardware implementation of proposed method has been also investigated. A series of experiments and tests have been carried on to verify to effectiveness of the proposed method. The experimental results demonstrated that the proposed approach is effective in verifying the integrity of the Root of Trust’s own boot rom, user instructions, and data, ensuring authenticity and enabling the secure boot of the Root of Trust’s own system. Additionally, our approach provides memory protection against certain types of memory attacks, such as cache leaks and tampering, and ensures the security of root-of-trust sensitive information, including keys.

Keywords: root of trust, secure boot, memory protection, hardware security

Procedia PDF Downloads 215
745 The Predicted Values of the California Bearing Ratio (CBR) by Using the Measurements of the Soil Resistivity Method (DC)

Authors: Fathi Ali Swaid

Abstract:

The CBR test is widely used in the assessment of granular materials in base, subbase and subgrade layers of road and airfield pavements. Despite the success of this method, but it depends on a limited numbers of soil samples. This limitation do not adequately account for the spatial variability of soil properties. Thus, assessment is derived using these cursory soil data are likely to contain errors and thus make interpretation and soil characterization difficult. On the other hand quantitative methods of soil inventory at the field scale involve the design and adoption of sampling regimes and laboratory analysis that are time consuming and costly. In the latter case new technologies are required to efficiently sample and observe the soil in the field. This is particularly the case where soil bearing capacity is prevalent, and detailed quantitative information for determining its cause is required. In this paper, an electrical resistivity method DC is described and its application in Elg'deem Dirt road, located in Gasser Ahmad - Misurata, Libya. Results from the DC instrument were found to be correlated with the CBR values (r2 = 0.89). Finally, it is noticed that, the correlation can be used with experience for determining CBR value using basic soil electrical resistivity measurements and checked by few CBR test representing a similar range of CBR.

Keywords: California bearing ratio, basic soil electrical resistivity, CBR, soil, subgrade, new technologies

Procedia PDF Downloads 448
744 Analysis of Particulate Matter Concentration, EC, OC Emission and Elemental Composition for Biodiesel-Fuelled Diesel Engine

Authors: A. M. Ashraful, H .H. Masjuki, M. A. Kalam

Abstract:

Comparative investigations were performed on the particles matter emitted from a DI diesel engine utilizing palm biodiesel. In this experiment, palm biodiesel PB10 (90% diesel and 10% palm biodiesel), PB20 (80% diesel, 20% palm biodiesel) and diesel fuel samples exhaust were investigated at different working condition (25% and 50% load at 1500 rpm constant speed). Observation of this experiment it clearly seen that at low load condition particle matter concentration of palm biodiesel exhaust were de-creased than that of diesel fuel. At no load and 25% load condition PB10 biodiesel blend exhibited 2.2 times lower PM concentration than that of diesel fuel. On the other hand, elemental carbon (EC) and organic emission for PB10 showed decreases trend as varies 4.2% to 6.6% and 32 to 39% respectively, while elemental carbon percentage increased by 0.85 to 10% respectively. Similarly, metal composition of PB10 biodiesel blend increased by 4.8 to 26.5% respectively. SEM images for B10 and B20 demonstrated granular structure particulates with greater grain sizes compared with diesel fuel. Finally, the experimental outcomes showed that the blend composition and degree of unsaturation of the methyl ester present in biodiesel influence on the particulate matter formation.

Keywords: particulate matter, elemental carbon, organic carbon, biodiesel

Procedia PDF Downloads 391
743 Application of Griddization Management to Construction Hazard Management

Authors: Lingzhi Li, Jiankun Zhang, Tiantian Gu

Abstract:

Hazard management that can prevent fatal accidents and property losses is a fundamental process during the buildings’ construction stage. However, due to lack of safety supervision resources and operational pressures, the conduction of hazard management is poor and ineffective in China. In order to improve the quality of construction safety management, it is critical to explore the use of information technologies to ensure that the process of hazard management is efficient and effective. After exploring the existing problems of construction hazard management in China, this paper develops the griddization management model for construction hazard management. First, following the knowledge grid infrastructure, the griddization computing infrastructure for construction hazards management is designed which includes five layers: resource entity layer, information management layer, task management layer, knowledge transformation layer and application layer. This infrastructure will be as the technical support for realizing grid management. Second, this study divides the construction hazards into grids through city level, district level and construction site level according to grid principles. Last, a griddization management process including hazard identification, assessment and control is developed. Meanwhile, all stakeholders of construction safety management, such as owners, contractors, supervision organizations and government departments, should take the corresponding responsibilities in this process. Finally, a case study based on actual construction hazard identification, assessment and control is used to validate the effectiveness and efficiency of the proposed griddization management model. The advantage of this designed model is to realize information sharing and cooperative management between various safety management departments.

Keywords: construction hazard, griddization computing, grid management, process

Procedia PDF Downloads 275
742 Comparison Study between Deep Mixed Columns and Encased Sand Column for Soft Clay Soil in Egypt

Authors: Walid El Kamash

Abstract:

Sand columns (or granular piles) can be employed as soil strengthening for flexible constructions such as road embankments, oil storage tanks in addition to multistory structures. The challenge of embedding the sand columns in soft soil is that the surrounding soft soil cannot avail the enough confinement stress in order to keep the form of the sand column. Therefore, the sand columns which were installed in such soil will lose their ability to perform needed load-bearing capacity. The encasement, besides increasing the strength and stiffness of the sand column, prevents the lateral squeezing of sands when the column is installed even in extremely soft soils, thus enabling quicker and more economical installation. This paper investigates the improvement in load capacity of the sand column by encasement through a comprehensive parametric study using the 3-D finite difference analysis for the soft clay of soil in Egypt. Moreover, the study was extended to include a comparison study between encased sand column and Deep Mixed columns (DM). The study showed that confining the sand by geosynthetic resulted in an increment of shear strength. That result paid the attention to use encased sand stone rather than deep mixed columns due to relative high permeability of the first material.

Keywords: encased sand column, Deep mixed column, numerical analysis, improving soft soil

Procedia PDF Downloads 378
741 Cloud Resources Utilization and Science Teacher’s Effectiveness in Secondary Schools in Cross River State, Nigeria

Authors: Michael Udey Udam

Abstract:

Background: This study investigated the impact of cloud resources, a component of cloud computing, on science teachers’ effectiveness in secondary schools in Cross River State. Three (3) research questions and three (3) alternative hypotheses guided the study. Method: The descriptive survey design was adopted for the study. The population of the study comprised 1209 science teachers in public secondary schools of Cross River state. Sample: A sample of 487 teachers was drawn from the population using a stratified random sampling technique. The researcher-made structured questionnaire with 18 was used for data collection for the study. Research question one was answered using the Pearson Product Moment Correlation, while research question two and the hypotheses were answered using the Analysis of Variance (ANOVA) statistics in the Statistical Package for Social Sciences (SPSS) at a 0.05 level of significance. Results: The results of the study revealed that there is a positive correlation between the utilization of cloud resources in teaching and teaching effectiveness among science teachers in secondary schools in Cross River state; there is a negative correlation between gender and utilization of cloud resources among science teachers in secondary schools in Cross River state; and that there is a significant correlation between teaching experience and the utilization of cloud resources among science teachers in secondary schools in Cross River state. Conclusion: The study justifies the effectiveness of the Cross River state government policy of introducing cloud computing into the education sector. The study recommends that the policy should be sustained.

Keywords: cloud resources, science teachers, effectiveness, secondary school

Procedia PDF Downloads 75
740 An Adjoint-Based Method to Compute Derivatives with Respect to Bed Boundary Positions in Resistivity Measurements

Authors: Mostafa Shahriari, Theophile Chaumont-Frelet, David Pardo

Abstract:

Resistivity measurements are used to characterize the Earth’s subsurface. They are categorized into two different groups: (a) those acquired on the Earth’s surface, for instance, controlled source electromagnetic (CSEM) and Magnetotellurics (MT), and (b) those recorded with borehole logging instruments such as Logging-While-Drilling (LWD) devices. LWD instruments are mostly used for geo-steering purposes, i.e., to adjust dip and azimuthal angles of a well trajectory to drill along a particular geological target. Modern LWD tools measure all nine components of the magnetic field corresponding to three orthogonal transmitter and receiver orientations. In order to map the Earth’s subsurface and perform geo-steering, we invert measurements using a gradient-based method that utilizes the derivatives of the recorded measurements with respect to the inversion variables. For resistivity measurements, these inversion variables are usually the constant resistivity value of each layer and the bed boundary positions. It is well-known how to compute derivatives with respect to the constant resistivity value of each layer using semi-analytic or numerical methods. However, similar formulas for computing the derivatives with respect to bed boundary positions are unavailable. The main contribution of this work is to provide an adjoint-based formulation for computing derivatives with respect to the bed boundary positions. The key idea to obtain the aforementioned adjoint state formulations for the derivatives is to separate the tangential and normal components of the field and treat them differently. This formulation allows us to compute the derivatives faster and more accurately than with traditional finite differences approximations. In the presentation, we shall first derive a formula for computing the derivatives with respect to the bed boundary positions for the potential equation. Then, we shall extend our formulation to 3D Maxwell’s equations. Finally, by considering a 1D domain and reducing the dimensionality of the problem, which is a common practice in the inversion of resistivity measurements, we shall derive a formulation to compute the derivatives of the measurements with respect to the bed boundary positions using a 1.5D variational formulation. Then, we shall illustrate the accuracy and convergence properties of our formulations by comparing numerical results with the analytical derivatives for the potential equation. For the 1.5D Maxwell’s system, we shall compare our numerical results based on the proposed adjoint-based formulation vs those obtained with a traditional finite difference approach. Numerical results shall show that our proposed adjoint-based technique produces enhanced accuracy solutions while its cost is negligible, as opposed to the finite difference approach that requires the solution of one additional problem per derivative.

Keywords: inverse problem, bed boundary positions, electromagnetism, potential equation

Procedia PDF Downloads 178
739 Renewable Energy Potential of Diluted Poultry Manure during Ambient Anaerobic Stabilisation

Authors: Cigdem Yangin-Gomec, Aigerim Jaxybayeva, Orhan Ince

Abstract:

In this study, the anaerobic treatability of chicken manure diluted with tap water (with an influent feed ratio of 1 kg of fresh chicken manure to 6 liter of tap water) was investigated in a lab-scale anaerobic sludge bed (ASB) reactor inoculated with the granular sludge already adapted to chicken manure. The raw waste digested in this study was the manure from laying-hens having average total solids (TS) of about 30% with ca. 60% volatile content. The ASB reactor was fed semi-continuously at ambient operating temperature range (17-23C) at a HRT of 13 and 26 days for about 6 months, respectively. The respective average total and soluble chemical oxygen demand (COD) removals were ca. 90% and 75%, whereas average biomethane production rate was calculated ca. 180 lt per kg of CODremoved from the ASB reactor at an average HRT of 13 days. Moreover, total suspended solids (TSS) and volatile suspended solids (VSS) in the influent were reduced more than 97%. Hence, high removals of the organic compounds with respective biogas production made anaerobic stabilization of the diluted chicken manure by ASB reactor at ambient operating temperatures viable. By this way, external heating up to 35C (i.e. anaerobic processes have been traditionally operated at mesophilic conditions) could be avoided in the scope of this study.

Keywords: ambient anaerobic digestion, biogas recovery, poultry manure, renewable energy

Procedia PDF Downloads 420
738 Optimizing Data Transfer and Processing in Multi-Cloud Environments for Big Data Workloads

Authors: Gaurav Kumar Sinha

Abstract:

In an era defined by the proliferation of data and the utilization of cloud computing environments, the efficient transfer and processing of big data workloads across multi-cloud platforms have emerged as critical challenges. This research paper embarks on a comprehensive exploration of the complexities associated with managing and optimizing big data in a multi-cloud ecosystem.The foundation of this study is rooted in the recognition that modern enterprises increasingly rely on multiple cloud providers to meet diverse business needs, enhance redundancy, and reduce vendor lock-in. As a consequence, managing data across these heterogeneous cloud environments has become intricate, necessitating innovative approaches to ensure data integrity, security, and performance.The primary objective of this research is to investigate strategies and techniques for enhancing the efficiency of data transfer and processing in multi-cloud scenarios. It recognizes that big data workloads are characterized by their sheer volume, variety, velocity, and complexity, making traditional data management solutions insufficient for harnessing the full potential of multi-cloud architectures.The study commences by elucidating the challenges posed by multi-cloud environments in the context of big data. These challenges encompass data fragmentation, latency, security concerns, and cost optimization. To address these challenges, the research explores a range of methodologies and solutions. One of the key areas of focus is data transfer optimization. The paper delves into techniques for minimizing data movement latency, optimizing bandwidth utilization, and ensuring secure data transmission between different cloud providers. It evaluates the applicability of dedicated data transfer protocols, intelligent data routing algorithms, and edge computing approaches in reducing transfer times.Furthermore, the study examines strategies for efficient data processing across multi-cloud environments. It acknowledges that big data processing requires distributed and parallel computing capabilities that span across cloud boundaries. The research investigates containerization and orchestration technologies, serverless computing models, and interoperability standards that facilitate seamless data processing workflows.Security and data governance are paramount concerns in multi-cloud environments. The paper explores methods for ensuring data security, access control, and compliance with regulatory frameworks. It considers encryption techniques, identity and access management, and auditing mechanisms as essential components of a robust multi-cloud data security strategy.The research also evaluates cost optimization strategies, recognizing that the dynamic nature of multi-cloud pricing models can impact the overall cost of data transfer and processing. It examines approaches for workload placement, resource allocation, and predictive cost modeling to minimize operational expenses while maximizing performance.Moreover, this study provides insights into real-world case studies and best practices adopted by organizations that have successfully navigated the challenges of multi-cloud big data management. It presents a comparative analysis of various multi-cloud management platforms and tools available in the market.

Keywords: multi-cloud environments, big data workloads, data transfer optimization, data processing strategies

Procedia PDF Downloads 67
737 Yawning Computing Using Bayesian Networks

Authors: Serge Tshibangu, Turgay Celik, Zenzo Ncube

Abstract:

Road crashes kill nearly over a million people every year, and leave millions more injured or permanently disabled. Various annual reports reveal that the percentage of fatal crashes due to fatigue/driver falling asleep comes directly after the percentage of fatal crashes due to intoxicated drivers. This percentage is higher than the combined percentage of fatal crashes due to illegal/Un-Safe U-turn and illegal/Un-Safe reversing. Although a relatively small percentage of police reports on road accidents highlights drowsiness and fatigue, the importance of these factors is greater than we might think, hidden by the undercounting of their events. Some scenarios show that these factors are significant in accidents with killed and injured people. Thus the need for an automatic drivers fatigue detection system in order to considerably reduce the number of accidents owing to fatigue.This research approaches the drivers fatigue detection problem in an innovative way by combining cues collected from both temporal analysis of drivers’ faces and environment. Monotony in driving environment is inter-related with visual symptoms of fatigue on drivers’ faces to achieve fatigue detection. Optical and infrared (IR) sensors are used to analyse the monotony in driving environment and to detect the visual symptoms of fatigue on human face. Internal cues from drivers faces and external cues from environment are combined together using machine learning algorithms to automatically detect fatigue.

Keywords: intelligent transportation systems, bayesian networks, yawning computing, machine learning algorithms

Procedia PDF Downloads 455
736 FRATSAN: A New Software for Fractal Analysis of Signals

Authors: Hamidreza Namazi

Abstract:

Fractal analysis is assessing fractal characteristics of data. It consists of several methods to assign fractal characteristics to a dataset which may be a theoretical dataset or a pattern or signal extracted from phenomena including natural geometric objects, sound, market fluctuations, heart rates, digital images, molecular motion, networks, etc. Fractal analysis is now widely used in all areas of science. An important limitation of fractal analysis is that arriving at an empirically determined fractal dimension does not necessarily prove that a pattern is fractal; rather, other essential characteristics have to be considered. For this purpose a Visual C++ based software called FRATSAN (FRActal Time Series ANalyser) was developed which extract information from signals through three measures. These measures are Fractal Dimensions, Jeffrey’s Measure and Hurst Exponent. After computing these measures, the software plots the graphs for each measure. Besides computing three measures the software can classify whether the signal is fractal or no. In fact, the software uses a dynamic method of analysis for all the measures. A sliding window is selected with a value equal to 10% of the total number of data entries. This sliding window is moved one data entry at a time to obtain all the measures. This makes the computation very sensitive to slight changes in data, thereby giving the user an acute analysis of the data. In order to test the performance of this software a set of EEG signals was given as input and the results were computed and plotted. This software is useful not only for fundamental fractal analysis of signals but can be used for other purposes. For instance by analyzing the Hurst exponent plot of a given EEG signal in patients with epilepsy the onset of seizure can be predicted by noticing the sudden changes in the plot.

Keywords: EEG signals, fractal analysis, fractal dimension, hurst exponent, Jeffrey’s measure

Procedia PDF Downloads 467
735 In-situ Performance of Pre-applied Bonded Waterproofing Membranes at Contaminated Test Slabs

Authors: Ulli Heinlein, Thomas Freimann

Abstract:

Pre-applied bonded membranes are used as positive-side waterproofing on concrete basements, are installed before the concrete work, and achieve a tear-resistant and waterproof bond with the subsequently placed fresh concrete. This bond increases redundancy compared to lose waterproofing membranes by preventing lateral water migrations in the event of damage. So far, the membranes have been tested in the laboratory, but it is not yet known how they behave on construction sites in the presence of dirt, soil, cement paste or moisture. This article, therefore, conducts investigations on six construction sites using 18 test slabs where the pre-applied bonded membranes are selectively contaminated or wetted. Subsequently, cores are taken, and the influence of the contaminations on the adhesive tensile strength and waterproof bond is tested. Pre-applied bonded membranes with smooth or granular but closed surfaces show no sensitivity to wetness, whereas open-pored membranes with nonwovens do not tolerate standing water. Contaminations decline the performance of all pre-applied bonded membranes since a separating layer is formed between the bonding layer and the concrete. The influence depends on the thickness of the contamination and its mechanical properties.

Keywords: waterproofing, positive-side waterproofing, basement, pre-applied bonded waterproofing membrane, In-situ testing, lateral water migrations

Procedia PDF Downloads 186
734 Information Disclosure And Financial Sentiment Index Using a Machine Learning Approach

Authors: Alev Atak

Abstract:

In this paper, we aim to create a financial sentiment index by investigating the company’s voluntary information disclosures. We retrieve structured content from BIST 100 companies’ financial reports for the period 1998-2018 and extract relevant financial information for sentiment analysis through Natural Language Processing. We measure strategy-related disclosures and their cross-sectional variation and classify report content into generic sections using synonym lists divided into four main categories according to their liquidity risk profile, risk positions, intra-annual information, and exposure to risk. We use Word Error Rate and Cosin Similarity for comparing and measuring text similarity and derivation in sets of texts. In addition to performing text extraction, we will provide a range of text analysis options, such as the readability metrics, word counts using pre-determined lists (e.g., forward-looking, uncertainty, tone, etc.), and comparison with reference corpus (word, parts of speech and semantic level). Therefore, we create an adequate analytical tool and a financial dictionary to depict the importance of granular financial disclosure for investors to identify correctly the risk-taking behavior and hence make the aggregated effects traceable.

Keywords: financial sentiment, machine learning, information disclosure, risk

Procedia PDF Downloads 94
733 Impact of Activated Carbon and Magnetic Field in Slow Sand Filter on Water Purification for Rural Dwellers

Authors: Baiyeri R. M, Oloriegbe Y. A., Saad A. O., Yusuf, K. O.

Abstract:

Most farmers that produce food crops in Nigeria live in rural areas where potable water is not available. The farmers in some areas have problem of water borne diseases which could affect their health and could lead to death. This study was conducted to determine the impact of incorporating Granular Activated Carbon(GAC) and Magnetic Field(MF) in Slow Sand Filter(SSF) on the purification of water for rural dwellers. The SSF was developed using PVC pipe with diameter 152.4 mm and 1100 mm long, with layers of fine sand with size 0.25 mm and 350 mm depth, followed by GAC 10 mm size and 100 mm depth, fine sand 0.25mm with 500 mm depth and gravel grain size 10-14 mm and 100 mm depth. The SSF was kept moist for 21 days for biofilm layer (schmutzdecke) to fully develop, which is essential for trapping bacteria. Two SSFs fabricated consist of SSF+GAC as Filter 1, SSF+GAC+MF as Filter 2 and Control (Raw water without passing through filter. Water samples were collected from the filter and analyzed. The flow rate of Filter was 25 litres/h Total bacteria counts(TBC) for Filter 1 and Filter 2 and control were 2.4, 4.6 and 8.1 cfu/mg, respectively. Total coliform count for Filter 1 and Filter 2 and control were 1.7, 3.0 and 6.4 cfu/100mL, respectively. The filters reduced water hardness, turbidity, lead, copper, electrical conductivity and TBC by 53.13-73.44% but increased pH from 5.8 to 7.1-7.3. SSF is recommended for water purification in the rural areas.

Keywords: magnetised water, sow sand filter, portable water, activated carbon

Procedia PDF Downloads 131
732 Investigating the Dynamic Response of the Ballast

Authors: Osama Brinji, Wing Kong Chiu, Graham Tew

Abstract:

Understanding the stability of rail ballast is one of the most important aspects in the railways. An unstable track may cause some issues such as unnecessary vibration and ultimately loss of track quality. The track foundation plays an important role in the stabilization of the railway. The dynamic response of rail ballast in the vicinity of the rail sleeper can affect the stability of the rail track and this has not been studied in detail. A review of literature showed that most of the works focused on the area under the concrete sleeper. Although there are some theories about the shear (longitudinal) effect of the rail ballast, these have not properly been studied and hence are not well understood. The stability of a rail track will depend on the compactness of the ballast in its vicinity. This paper will try to determine the dynamic response of the ballast to identify its resonant behaviour. This preliminary research is one of several studies that examine the vibration response of the granular materials. The main aim is to use this information for future design of sleepers to ensure that any dynamic response of the sleeper will not compromise the state of compactness of the ballast. This paper will report on the dependence of damping and the natural frequency of the ballast as a function of depth and distance from the point of excitation introduced through a concrete block. The concrete block is used to simulate a sleeper and the ballast is simulated with gravel. In spite of these approximations, the results presented in the paper will show an agreement with theories and the assumptions that are used in study the mechanical behaviour of the rail ballast.

Keywords: ballast, dynamic response, sleeper, stability

Procedia PDF Downloads 501
731 Managing Data from One Hundred Thousand Internet of Things Devices Globally for Mining Insights

Authors: Julian Wise

Abstract:

Newcrest Mining is one of the world’s top five gold and rare earth mining organizations by production, reserves and market capitalization in the world. This paper elaborates on the data acquisition processes employed by Newcrest in collaboration with Fortune 500 listed organization, Insight Enterprises, to standardize machine learning solutions which process data from over a hundred thousand distributed Internet of Things (IoT) devices located at mine sites globally. Through the utilization of software architecture cloud technologies and edge computing, the technological developments enable for standardized processes of machine learning applications to influence the strategic optimization of mineral processing. Target objectives of the machine learning optimizations include time savings on mineral processing, production efficiencies, risk identification, and increased production throughput. The data acquired and utilized for predictive modelling is processed through edge computing by resources collectively stored within a data lake. Being involved in the digital transformation has necessitated the standardization software architecture to manage the machine learning models submitted by vendors, to ensure effective automation and continuous improvements to the mineral process models. Operating at scale, the system processes hundreds of gigabytes of data per day from distributed mine sites across the globe, for the purposes of increased improved worker safety, and production efficiency through big data applications.

Keywords: mineral technology, big data, machine learning operations, data lake

Procedia PDF Downloads 112
730 Life Cycle Analysis of Using Brick Waste in Road Technology

Authors: Mezhoud Samy, Toumi Youcef, Boukendekdji Otmane

Abstract:

Nowadays, industrial by-products and waste are increasing along with public needs increase. The engineering sector has turned to sustainable development by emphasizing the aspects of environmental and life cycle assessment as an important objective. Among this waste, the remains of the red bricks (DBR) may be an alternative worth checking out, given their availability and abundance at the construction sites. In this context, this work aims to valorize DBR in the concrete road (BR). The incorporation of DBR is carried out by the substitution of the granular fractions of mixtures from noble quarry materials. The experimental plan aims to determine the physico-mechanical performance and environmental performance of manufactured BRs from DBR with a cement content (6.5%) and compared with a control BR without DBR. The studied characteristics are proctor, resistance to compression, resistance to flexural tensile at 7 and 28 days, modulus of elasticity, and total shrinkage. The results of this experimental study showed that the characteristics of recycled aggregates (DBR) are lower than those of natural aggregates but remain acceptable with respect to regulations. Results demonstrate the mechanical performance of BR made from less DBR than the control BR without DBR but remains appreciable and encourage their jobs in the road sector. Recycled aggregates can constitute an interesting economic and ecological alternative but require elementary precautions before any use.

Keywords: life cycle assessment, brick waste, road concrete, performance

Procedia PDF Downloads 94
729 An Evolutionary Approach for QAOA for Max-Cut

Authors: Francesca Schiavello

Abstract:

This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.

Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization

Procedia PDF Downloads 60
728 A Study on FWD Deflection Bowl Parameters for Condition Assessment of Flexible Pavement

Authors: Ujjval J. Solanki, Prof.(Dr.) P.J. Gundaliya, Prof.M.D. Barasara

Abstract:

The application of Falling Weight Deflectometer is to evaluate structural performance of the flexible pavement. The exercise of back calculation is required to know the modulus of elasticity of existing in-service pavement. The process of back calculation needs in-depth field experience for the input of range of modulus of elasticity of bituminous, granular and subgrade layer, and its required number of trial to find such matching moduli with the observed FWD deflection on the field. The study carried out at Barnala-Mansa State Highway Punjab-India using FWD before and after overlay; the deflections obtained at 0 on the load cell, 300, 600, 900,1200, 1500 and 1800 mm interval from the load cell these seven deflection results used to calculate Surface Curvature Index (SCI), Base damage Index (BDI), Base curvature index (BCI). This SCI, BCI and BDI indices are useful to predict the structural performance of in-service pavement and also useful to identify homogeneous section for condition assessment. The SCI, BCI and BDI range are determined for before and after overlay the range of SCI 520 to 51 BDI 294 to 63 BCI 83 to 0.27 for old pavement and SCI 272 to 23 BDI 228 to 28, BCI 25.85 to 4.60 for new pavement. It also shows good correlation with back calculated modulus of elasticity of all the three layer.

Keywords: back calculation, base damage index, base curvature index, FWD (Falling Weight Deflectometer), surface curvature index

Procedia PDF Downloads 332
727 An Analytical Metric and Process for Critical Infrastructure Architecture System Availability Determination in Distributed Computing Environments under Infrastructure Attack

Authors: Vincent Andrew Cappellano

Abstract:

In the early phases of critical infrastructure system design, translating distributed computing requirements to an architecture has risk given the multitude of approaches (e.g., cloud, edge, fog). In many systems, a single requirement for system uptime / availability is used to encompass the system’s intended operations. However, when architected systems may perform to those availability requirements only during normal operations and not during component failure, or during outages caused by adversary attacks on critical infrastructure (e.g., physical, cyber). System designers lack a structured method to evaluate availability requirements against candidate system architectures through deep degradation scenarios (i.e., normal ops all the way down to significant damage of communications or physical nodes). This increases risk of poor selection of a candidate architecture due to the absence of insight into true performance for systems that must operate as a piece of critical infrastructure. This research effort proposes a process to analyze critical infrastructure system availability requirements and a candidate set of systems architectures, producing a metric assessing these architectures over a spectrum of degradations to aid in selecting appropriate resilient architectures. To accomplish this effort, a set of simulation and evaluation efforts are undertaken that will process, in an automated way, a set of sample requirements into a set of potential architectures where system functions and capabilities are distributed across nodes. Nodes and links will have specific characteristics and based on sampled requirements, contribute to the overall system functionality, such that as they are impacted/degraded, the impacted functional availability of a system can be determined. A machine learning reinforcement-based agent will structurally impact the nodes, links, and characteristics (e.g., bandwidth, latency) of a given architecture to provide an assessment of system functional uptime/availability under these scenarios. By varying the intensity of the attack and related aspects, we can create a structured method of evaluating the performance of candidate architectures against each other to create a metric rating its resilience to these attack types/strategies. Through multiple simulation iterations, sufficient data will exist to compare this availability metric, and an architectural recommendation against the baseline requirements, in comparison to existing multi-factor computing architectural selection processes. It is intended that this additional data will create an improvement in the matching of resilient critical infrastructure system requirements to the correct architectures and implementations that will support improved operation during times of system degradation due to failures and infrastructure attacks.

Keywords: architecture, resiliency, availability, cyber-attack

Procedia PDF Downloads 108
726 Exploring Data Stewardship in Fog Networking Using Blockchain Algorithm

Authors: Ruvaitha Banu, Amaladhithyan Krishnamoorthy

Abstract:

IoT networks today solve various consumer problems, from home automation systems to aiding in driving autonomous vehicles with the exploration of multiple devices. For example, in an autonomous vehicle environment, multiple sensors are available on roads to monitor weather and road conditions and interact with each other to aid the vehicle in reaching its destination safely and timely. IoT systems are predominantly dependent on the cloud environment for data storage, and computing needs that result in latency problems. With the advent of Fog networks, some of this storage and computing is pushed to the edge/fog nodes, saving the network bandwidth and reducing the latency proportionally. Managing the data stored in these fog nodes becomes crucial as it might also store sensitive information required for a certain application. Data management in fog nodes is strenuous because Fog networks are dynamic in terms of their availability and hardware capability. It becomes more challenging when the nodes in the network also live a short span, detaching and joining frequently. When an end-user or Fog Node wants to access, read, or write data stored in another Fog Node, then a new protocol becomes necessary to access/manage the data stored in the fog devices as a conventional static way of managing the data doesn’t work in Fog Networks. The proposed solution discusses a protocol that acts by defining sensitivity levels for the data being written and read. Additionally, a distinct data distribution and replication model among the Fog nodes is established to decentralize the access mechanism. In this paper, the proposed model implements stewardship towards the data stored in the Fog node using the application of Reinforcement Learning so that access to the data is determined dynamically based on the requests.

Keywords: IoT, fog networks, data stewardship, dynamic access policy

Procedia PDF Downloads 59
725 Design and Evaluation of Production Performance Dashboard for Achieving Oil and Gas Production Target

Authors: Ivan Ramos Sampe Immanuel, Linung Kresno Adikusumo, Liston Sitanggang

Abstract:

Achieving the production targets of oil and gas in an upstream oil and gas company represents a complex undertaking necessitating collaborative engagement from a multidisciplinary team. In addition to conducting exploration activities and executing well intervention programs, an upstream oil and gas enterprise must assess the feasibility of attaining predetermined production goals. The monitoring of production performance serves as a critical activity to ensure organizational progress towards the established oil and gas performance targets. Subsequently, decisions within the upstream oil and gas management team are informed by the received information pertaining to the respective production performance. To augment the decision-making process, the implementation of a production performance dashboard emerges as a viable solution, providing an integrated and centralized tool. The deployment of a production performance dashboard manifests as an instrumental mechanism fostering a user-friendly interface for monitoring production performance, while concurrently preserving the intrinsic characteristics of granular data. The integration of diverse data sources into a unified production performance dashboard establishes a singular veritable source, thereby enhancing the organization's capacity to uphold a consolidated and authoritative foundation for its business requisites. Additionally, the heightened accessibility of the production performance dashboard to business users constitutes a compelling substantiation of its consequential impact on facilitating the monitoring of organizational targets.

Keywords: production, performance, dashboard, data analytics

Procedia PDF Downloads 70
724 High Performance Computing Enhancement of Agent-Based Economic Models

Authors: Amit Gill, Lalith Wijerathne, Sebastian Poledna

Abstract:

This research presents the details of the implementation of high performance computing (HPC) extension of agent-based economic models (ABEMs) to simulate hundreds of millions of heterogeneous agents. ABEMs offer an alternative approach to study the economy as a dynamic system of interacting heterogeneous agents, and are gaining popularity as an alternative to standard economic models. Over the last decade, ABEMs have been increasingly applied to study various problems related to monetary policy, bank regulations, etc. When it comes to predicting the effects of local economic disruptions, like major disasters, changes in policies, exogenous shocks, etc., on the economy of the country or the region, it is pertinent to study how the disruptions cascade through every single economic entity affecting its decisions and interactions, and eventually affect the economic macro parameters. However, such simulations with hundreds of millions of agents are hindered by the lack of HPC enhanced ABEMs. In order to address this, a scalable Distributed Memory Parallel (DMP) implementation of ABEMs has been developed using message passing interface (MPI). A balanced distribution of computational load among MPI-processes (i.e. CPU cores) of computer clusters while taking all the interactions among agents into account is a major challenge for scalable DMP implementations. Economic agents interact on several random graphs, some of which are centralized (e.g. credit networks, etc.) whereas others are dense with random links (e.g. consumption markets, etc.). The agents are partitioned into mutually-exclusive subsets based on a representative employer-employee interaction graph, while the remaining graphs are made available at a minimum communication cost. To minimize the number of communications among MPI processes, real-life solutions like the introduction of recruitment agencies, sales outlets, local banks, and local branches of government in each MPI-process, are adopted. Efficient communication among MPI-processes is achieved by combining MPI derived data types with the new features of the latest MPI functions. Most of the communications are overlapped with computations, thereby significantly reducing the communication overhead. The current implementation is capable of simulating a small open economy. As an example, a single time step of a 1:1 scale model of Austria (i.e. about 9 million inhabitants and 600,000 businesses) can be simulated in 15 seconds. The implementation is further being enhanced to simulate 1:1 model of Euro-zone (i.e. 322 million agents).

Keywords: agent-based economic model, high performance computing, MPI-communication, MPI-process

Procedia PDF Downloads 127
723 Identification of Microbial Community in an Anaerobic Reactor Treating Brewery Wastewater

Authors: Abimbola M. Enitan, John O. Odiyo, Feroz M. Swalaha

Abstract:

The study of microbial ecology and their function in anaerobic digestion processes are essential to control the biological processes. This is to know the symbiotic relationship between the microorganisms that are involved in the conversion of complex organic matter in the industrial wastewater to simple molecules. In this study, diversity and quantity of bacterial community in the granular sludge taken from the different compartments of a full-scale upflow anaerobic sludge blanket (UASB) reactor treating brewery wastewater was investigated using polymerase chain reaction (PCR) and real-time quantitative PCR (qPCR). The phylogenetic analysis showed three major eubacteria phyla that belong to Proteobacteria, Firmicutes and Chloroflexi in the full-scale UASB reactor, with different groups populating different compartment. The result of qPCR assay showed high amount of eubacteria with increase in concentration along the reactor’s compartment. This study extends our understanding on the diverse, topological distribution and shifts in concentration of microbial communities in the different compartments of a full-scale UASB reactor treating brewery wastewater. The colonization and the trophic interactions among these microbial populations in reducing and transforming complex organic matter within the UASB reactors were established.

Keywords: bacteria, brewery wastewater, real-time quantitative PCR, UASB reactor

Procedia PDF Downloads 258
722 Local Homology Modules

Authors: Fatemeh Mohammadi Aghjeh Mashhad

Abstract:

In this paper, we give several ways for computing generalized local homology modules by using Gorenstein flat resolutions. Also, we find some bounds for vanishing of generalized local homology modules.

Keywords: a-adic completion functor, generalized local homology modules, Gorenstein flat modules

Procedia PDF Downloads 419
721 Heat Transfer and Diffusion Modelling

Authors: R. Whalley

Abstract:

The heat transfer modelling for a diffusion process will be considered. Difficulties in computing the time-distance dynamics of the representation will be addressed. Incomplete and irrational Laplace function will be identified as the computational issue. Alternative approaches to the response evaluation process will be provided. An illustration application problem will be presented. Graphical results confirming the theoretical procedures employed will be provided.

Keywords: heat, transfer, diffusion, modelling, computation

Procedia PDF Downloads 553
720 Optimizing Data Integration and Management Strategies for Upstream Oil and Gas Operations

Authors: Deepak Singh, Rail Kuliev

Abstract:

The abstract highlights the critical importance of optimizing data integration and management strategies in the upstream oil and gas industry. With its complex and dynamic nature generating vast volumes of data, efficient data integration and management are essential for informed decision-making, cost reduction, and maximizing operational performance. Challenges such as data silos, heterogeneity, real-time data management, and data quality issues are addressed, prompting the proposal of several strategies. These strategies include implementing a centralized data repository, adopting industry-wide data standards, employing master data management (MDM), utilizing real-time data integration technologies, and ensuring data quality assurance. Training and developing the workforce, “reskilling and upskilling” the employees and establishing robust Data Management training programs play an essential role and integral part in this strategy. The article also emphasizes the significance of data governance and best practices, as well as the role of technological advancements such as big data analytics, cloud computing, Internet of Things (IoT), and artificial intelligence (AI) and machine learning (ML). To illustrate the practicality of these strategies, real-world case studies are presented, showcasing successful implementations that improve operational efficiency and decision-making. In present study, by embracing the proposed optimization strategies, leveraging technological advancements, and adhering to best practices, upstream oil and gas companies can harness the full potential of data-driven decision-making, ultimately achieving increased profitability and a competitive edge in the ever-evolving industry.

Keywords: master data management, IoT, AI&ML, cloud Computing, data optimization

Procedia PDF Downloads 70
719 Green Technologies Developed by JSC “NIUIF”

Authors: Andrey Norov

Abstract:

In the recent years, Samoilov Research Institute for Mineral Fertilizers JSC “NIUIF”, the oldest (established in September 1919) industry-oriented institute in Russia, has developed a range of sustainable, environment-friendly, zero-waste technologies that ensure minimal consumption of materials and energy resources and fully consistent with the principles of Green Chemistry that include: - Ecofriendly energy and resource saving technology of sulfuric acid from sulfur according to DC-DA scheme (double conversion - double absorption); - Improved zero-waste technology of wet phosphoric acid (WPA) by dihydrate-hemihydrate process applicable to various types of phosphate raw materials; - Flexible, efficient, zero-waste, universal technology of NP / NPS / NPK / NPKS fertilizers with maximum heat recovery from chemical processes; - Novel, zero-waste, no-analogue technology of granular PK / PKS / NPKS fertilizers with controlled dissolution rate and nutrient supply into the soil, which allows to process a number of wastes and by-products; - Innovative resource-saving joint processing of wastes from the production of phosphogypsum and fluorosilicic acid (FSA) into ammonium sulfate with simultaneous neutralization of fluoride compounds with no lime used. - New fertilizer technology of increased environmental and agrochemical efficiency (currently under development). All listed green technologies are patented with Russian and Eurasian patents. The development of ecofriendly, safe, green technologies is ongoing in JSC “NIUIF”.

Keywords: NPKS fertilizers, FSA, sulfuric acid, WPA

Procedia PDF Downloads 94