Search results for: automated quantification
963 A Construct to Perform in Situ Deformation Measurement of Material Extrusion-Fabricated Structures
Authors: Daniel Nelson, Valeria La Saponara
Abstract:
Material extrusion is an additive manufacturing modality that continues to show great promise in the ability to create low-cost, highly intricate, and exceedingly useful structural elements. As more capable and versatile filament materials are devised, and the resolution of manufacturing systems continues to increase, the need to understand and predict manufacturing-induced warping will gain ever greater importance. The following study presents an in situ remote sensing and data analysis construct that allows for the in situ mapping and quantification of surface displacements induced by residual stresses on a specified test structure. This proof-of-concept experimental process shows that it is possible to provide designers and manufacturers with insight into the manufacturing parameters that lead to the manifestation of these deformations and a greater understanding of the behavior of these warping events over the course of the manufacturing process.Keywords: additive manufacturing, deformation, digital image correlation, fused filament fabrication, residual stress, warping
Procedia PDF Downloads 88962 Study of the Stability of the Slope Open-Pit Mines: Case of the Mine of Phosphates – Tebessa, Algeria
Authors: Mohamed Fredj, Abdallah Hafsaoui, Radouane Nakache
Abstract:
The study of the stability of the mining works in rock masses fractured is the major concern of the operating engineer. For geotechnical works in mines and quarries, it there is not today's general methodology for analysis and the quantification of the risks relating to the dangers inherent in these concrete types (falling boulders, landslides, etc.). The reasons for this are uncertainty, which weighs on available data or lack of knowledge of the values of the parameters required for this analysis type. Stability calculations must be based on reliable knowledge of the distribution of discontinuities that dissect the Rocky massif and the resistance to shear of the intact rock and discontinuities. This study is aimed to study the stability of slope of mine (Kef Sennoun - Tebessa, Algeria). The problem is analyzed using a numerical model based on the finite elements (software Plaxis 3D).Keywords: stability, discontinuities, finite elements, rock mass, open-pit mine
Procedia PDF Downloads 321961 Automated Transformation of 3D Point Cloud to BIM Model: Leveraging Algorithmic Modeling for Efficient Reconstruction
Authors: Radul Shishkov, Orlin Davchev
Abstract:
The digital era has revolutionized architectural practices, with building information modeling (BIM) emerging as a pivotal tool for architects, engineers, and construction professionals. However, the transition from traditional methods to BIM-centric approaches poses significant challenges, particularly in the context of existing structures. This research introduces a technical approach to bridge this gap through the development of algorithms that facilitate the automated transformation of 3D point cloud data into detailed BIM models. The core of this research lies in the application of algorithmic modeling and computational design methods to interpret and reconstruct point cloud data -a collection of data points in space, typically produced by 3D scanners- into comprehensive BIM models. This process involves complex stages of data cleaning, feature extraction, and geometric reconstruction, which are traditionally time-consuming and prone to human error. By automating these stages, our approach significantly enhances the efficiency and accuracy of creating BIM models for existing buildings. The proposed algorithms are designed to identify key architectural elements within point clouds, such as walls, windows, doors, and other structural components, and to translate these elements into their corresponding BIM representations. This includes the integration of parametric modeling techniques to ensure that the generated BIM models are not only geometrically accurate but also embedded with essential architectural and structural information. Our methodology has been tested on several real-world case studies, demonstrating its capability to handle diverse architectural styles and complexities. The results showcase a substantial reduction in time and resources required for BIM model generation while maintaining high levels of accuracy and detail. This research contributes significantly to the field of architectural technology by providing a scalable and efficient solution for the integration of existing structures into the BIM framework. It paves the way for more seamless and integrated workflows in renovation and heritage conservation projects, where the accuracy of existing conditions plays a critical role. The implications of this study extend beyond architectural practices, offering potential benefits in urban planning, facility management, and historic preservation.Keywords: BIM, 3D point cloud, algorithmic modeling, computational design, architectural reconstruction
Procedia PDF Downloads 63960 Design and Development of Motorized Placer for Balloon Uterine Stents in Gynecology
Authors: Metehan Mutlu, Meltem Elitas
Abstract:
This study aims to provide an automated method for placing the balloon uterine stents after hysteroscopy adhesiolysis. Currently, there are no automatized tools to place the balloon uterine stent; therefore, surgeons into the endometrial cavity manually fit it. However, it is very hard to pass the balloon stent through the cervical canal, which is roughly 10mm after the surgery. Our method aims to provide an effective and practical way of placing the stent, by automating the procedure through our designed device. Furthermore, our device does the required tasks fast compared to traditional methods, reduces the narcosis time, and decreases the bacterial contamination risks.Keywords: balloon uterine stent, endometrial cavity, hysteroscopy, motorized-tool
Procedia PDF Downloads 276959 Intelligent Recognition Tools for Industrial Automation
Authors: Amin Nazerzadeh, Afsaneh Nouri Houshyar , Azadeh Noori Hoshyar
Abstract:
With the rapid growing of information technology, the industry and manufacturing systems are becoming more automated. Therefore, achieving the highly accurate automatic systems with reliable security is becoming more critical. Biometrics that refers to identifying individual based on physiological or behavioral traits are unique identifiers provide high reliability and security in different industrial systems. As biometric cannot easily be transferred between individuals or copied, it has been receiving extensive attention. Due to the importance of security applications, this paper provides an overview on biometrics and discuss about background, types and applications of biometric as an effective tool for the industrial applications.Keywords: Industial and manufacturing applications, intelligence and security, information technology, recognition; security technology; biometrics
Procedia PDF Downloads 155958 Integration Process and Analytic Interface of different Environmental Open Data Sets with Java/Oracle and R
Authors: Pavel H. Llamocca, Victoria Lopez
Abstract:
The main objective of our work is the comparative analysis of environmental data from Open Data bases, belonging to different governments. This means that you have to integrate data from various different sources. Nowadays, many governments have the intention of publishing thousands of data sets for people and organizations to use them. In this way, the quantity of applications based on Open Data is increasing. However each government has its own procedures to publish its data, and it causes a variety of formats of data sets because there are no international standards to specify the formats of the data sets from Open Data bases. Due to this variety of formats, we must build a data integration process that is able to put together all kind of formats. There are some software tools developed in order to give support to the integration process, e.g. Data Tamer, Data Wrangler. The problem with these tools is that they need data scientist interaction to take part in the integration process as a final step. In our case we don’t want to depend on a data scientist, because environmental data are usually similar and these processes can be automated by programming. The main idea of our tool is to build Hadoop procedures adapted to data sources per each government in order to achieve an automated integration. Our work focus in environment data like temperature, energy consumption, air quality, solar radiation, speeds of wind, etc. Since 2 years, the government of Madrid is publishing its Open Data bases relative to environment indicators in real time. In the same way, other governments have published Open Data sets relative to the environment (like Andalucia or Bilbao). But all of those data sets have different formats and our solution is able to integrate all of them, furthermore it allows the user to make and visualize some analysis over the real-time data. Once the integration task is done, all the data from any government has the same format and the analysis process can be initiated in a computational better way. So the tool presented in this work has two goals: 1. Integration process; and 2. Graphic and analytic interface. As a first approach, the integration process was developed using Java and Oracle and the graphic and analytic interface with Java (jsp). However, in order to open our software tool, as second approach, we also developed an implementation with R language as mature open source technology. R is a really powerful open source programming language that allows us to process and analyze a huge amount of data with high performance. There are also some R libraries for the building of a graphic interface like shiny. A performance comparison between both implementations was made and no significant differences were found. In addition, our work provides with an Official Real-Time Integrated Data Set about Environment Data in Spain to any developer in order that they can build their own applications.Keywords: open data, R language, data integration, environmental data
Procedia PDF Downloads 315957 The Exploitation of the MOSES Project Outcomes on Supply Chain Optimisation
Authors: Reza Karimpour
Abstract:
Ports play a decisive role in the EU's external and internal trade, as about 74% of imports and exports and 37% of exchanges go through ports. Although ports, especially Deep Sea Shipping (DSS) ports, are integral nodes within multimodal logistic flows, Short Sea Shipping (SSS) and inland waterways are not so well integrated. The automated vessels and supply chain optimisations for sustainable shortsea shipping (MOSES) project aims to enhance the short sea shipping component of the European supply chain by addressing the vulnerabilities and strains related to the operation of large containerships. The MOSES concept can be shortly described as a large containership (mother-vessel) approaching a DSS port (or a large container terminal). Upon her arrival, a combined intelligent mega-system consisting of the MOSES Autonomous tugboat swarm for manoeuvring and the MOSES adapted AutoMoor system. Then, container handling processes are ready to start moving containers to their destination via hinterland connections (trucks and/or rail) or to be shipped to destinations near small ports (on the mainland or island). For the first case, containers are stored in a dedicated port area (Storage area), waiting to be moved via trucks and/or rail. For the second case, containers are stacked by existing port equipment near-dedicated berths of the DSS port. They then are loaded on the MOSES Innovative Feeder Vessel, equipped with the MOSES Robotic Container-Handling System that provides (semi-) autonomous (un) feeding of the feeder. The Robotic Container-Handling System is remotely monitored through a Shore Control Centre. When the MOSES innovative Feeder vessel approaches the small port, where her docking is achieved without tugboats, she automatically unloads the containers using the Robotic Container-Handling System on the quay or directly on trucks. As a result, ports with minimal or no available infrastructure may be effectively integrated with the container supply chain. Then, the MOSES innovative feeder vessel continues her voyage to the next small port, or she returns to the DSS port. MOSES exploitation activity mainly aims to exploit research outcomes beyond the project, facilitate utilisation of the pilot results by others, and continue the pilot service after the project ends. By the mid-lifetime of the project, the exploitation plan introduces the reader to the MOSES project and its key exploitable results. It provides a plan for delivering the MOSES innovations to the market as part of the overall exploitation plan.Keywords: automated vessels, exploitation, shortsea shipping, supply chain
Procedia PDF Downloads 110956 An Effort at Improving Reliability of Laboratory Data in Titrimetric Analysis for Zinc Sulphate Tablets Using Validated Spreadsheet Calculators
Authors: M. A. Okezue, K. L. Clase, S. R. Byrn
Abstract:
The requirement for maintaining data integrity in laboratory operations is critical for regulatory compliance. Automation of procedures reduces incidence of human errors. Quality control laboratories located in low-income economies may face some barriers in attempts to automate their processes. Since data from quality control tests on pharmaceutical products are used in making regulatory decisions, it is important that laboratory reports are accurate and reliable. Zinc Sulphate (ZnSO4) tablets is used in treatment of diarrhea in pediatric population, and as an adjunct therapy for COVID-19 regimen. Unfortunately, zinc content in these formulations is determined titrimetrically; a manual analytical procedure. The assay for ZnSO4 tablets involves time-consuming steps that contain mathematical formulae prone to calculation errors. To achieve consistency, save costs, and improve data integrity, validated spreadsheets were developed to simplify the two critical steps in the analysis of ZnSO4 tablets: standardization of 0.1M Sodium Edetate (EDTA) solution, and the complexometric titration assay procedure. The assay method in the United States Pharmacopoeia was used to create a process flow for ZnSO4 tablets. For each step in the process, different formulae were input into two spreadsheets to automate calculations. Further checks were created within the automated system to ensure validity of replicate analysis in titrimetric procedures. Validations were conducted using five data sets of manually computed assay results. The acceptance criteria set for the protocol were met. Significant p-values (p < 0.05, α = 0.05, at 95% Confidence Interval) were obtained from students’ t-test evaluation of the mean values for manual-calculated and spreadsheet results at all levels of the analysis flow. Right-first-time analysis and principles of data integrity were enhanced by use of the validated spreadsheet calculators in titrimetric evaluations of ZnSO4 tablets. Human errors were minimized in calculations when procedures were automated in quality control laboratories. The assay procedure for the formulation was achieved in a time-efficient manner with greater level of accuracy. This project is expected to promote cost savings for laboratory business models.Keywords: data integrity, spreadsheets, titrimetry, validation, zinc sulphate tablets
Procedia PDF Downloads 169955 Superparamagnetic Sensor with Lateral Flow Immunoassays as Platforms for Biomarker Quantification
Authors: M. Salvador, J. C. Martinez-Garcia, A. Moyano, M. C. Blanco-Lopez, M. Rivas
Abstract:
Biosensors play a crucial role in the detection of molecules nowadays due to their advantages of user-friendliness, high selectivity, the analysis in real time and in-situ applications. Among them, Lateral Flow Immunoassays (LFIAs) are presented among technologies for point-of-care bioassays with outstanding characteristics such as affordability, portability and low-cost. They have been widely used for the detection of a vast range of biomarkers, which do not only include proteins but also nucleic acids and even whole cells. Although the LFIA has traditionally been a positive/negative test, tremendous efforts are being done to add to the method the quantifying capability based on the combination of suitable labels and a proper sensor. One of the most successful approaches involves the use of magnetic sensors for detection of magnetic labels. Bringing together the required characteristics mentioned before, our research group has developed a biosensor to detect biomolecules. Superparamagnetic nanoparticles (SPNPs) together with LFIAs play the fundamental roles. SPMNPs are detected by their interaction with a high-frequency current flowing on a printed micro track. By means of the instant and proportional variation of the impedance of this track provoked by the presence of the SPNPs, quantitative and rapid measurement of the number of particles can be obtained. This way of detection requires no external magnetic field application, which reduces the device complexity. On the other hand, the major limitations of LFIAs are that they are only qualitative or semiquantitative when traditional gold or latex nanoparticles are used as color labels. Moreover, the necessity of always-constant ambient conditions to get reproducible results, the exclusive detection of the nanoparticles on the surface of the membrane, and the short durability of the signal are drawbacks that can be advantageously overcome with the design of magnetically labeled LFIAs. The approach followed was to coat the SPIONs with a specific monoclonal antibody which targets the protein under consideration by chemical bonds. Then, a sandwich-type immunoassay was prepared by printing onto the nitrocellulose membrane strip a second antibody against a different epitope of the protein (test line) and an IgG antibody (control line). When the sample flows along the strip, the SPION-labeled proteins are immobilized at the test line, which provides magnetic signal as described before. Preliminary results using this practical combination for the detection and quantification of the Prostatic-Specific Antigen (PSA) shows the validity and consistency of the technique in the clinical range, where a PSA level of 4.0 ng/mL is the established upper normal limit. Moreover, a LOD of 0.25 ng/mL was calculated with a confident level of 3 according to the IUPAC Gold Book definition. Its versatility has also been proved with the detection of other biomolecules such as troponin I (cardiac injury biomarker) or histamine.Keywords: biosensor, lateral flow immunoassays, point-of-care devices, superparamagnetic nanoparticles
Procedia PDF Downloads 231954 Model Updating-Based Approach for Damage Prognosis in Frames via Modal Residual Force
Authors: Gholamreza Ghodrati Amiri, Mojtaba Jafarian Abyaneh, Ali Zare Hosseinzadeh
Abstract:
This paper presents an effective model updating strategy for damage localization and quantification in frames by defining damage detection problem as an optimization issue. A generalized version of the Modal Residual Force (MRF) is employed for presenting a new damage-sensitive cost function. Then, Grey Wolf Optimization (GWO) algorithm is utilized for solving suggested inverse problem and the global extremums are reported as damage detection results. The applicability of the presented method is investigated by studying different damage patterns on the benchmark problem of the IASC-ASCE, as well as a planar shear frame structure. The obtained results emphasize good performance of the method not only in free-noise cases, but also when the input data are contaminated with different levels of noises.Keywords: frame, grey wolf optimization algorithm, modal residual force, structural damage detection
Procedia PDF Downloads 389953 The Synergistic Effects of Blockchain and AI on Enhancing Data Integrity and Decision-Making Accuracy in Smart Contracts
Authors: Sayor Ajfar Aaron, Sajjat Hossain Abir, Ashif Newaz, Mushfiqur Rahman
Abstract:
Investigating the convergence of blockchain technology and artificial intelligence, this paper examines their synergistic effects on data integrity and decision-making within smart contracts. By implementing AI-driven analytics on blockchain-based platforms, the research identifies improvements in automated contract enforcement and decision accuracy. The paper presents a framework that leverages AI to enhance transparency and trust while blockchain ensures immutable record-keeping, culminating in significantly optimized operational efficiencies in various industries.Keywords: artificial intelligence, blockchain, data integrity, smart contracts
Procedia PDF Downloads 55952 Fault Tree Analysis and Bayesian Network for Fire and Explosion of Crude Oil Tanks: Case Study
Authors: B. Zerouali, M. Kara, B. Hamaidi, H. Mahdjoub, S. Rouabhia
Abstract:
In this paper, a safety analysis for crude oil tanks to prevent undesirable events that may cause catastrophic accidents. The estimation of the probability of damage to industrial systems is carried out through a series of steps, and in accordance with a specific methodology. In this context, this work involves developing an assessment tool and risk analysis at the level of crude oil tanks system, based primarily on identification of various potential causes of crude oil tanks fire and explosion by the use of Fault Tree Analysis (FTA), then improved risk modelling by Bayesian Networks (BNs). Bayesian approach in the evaluation of failure and quantification of risks is a dynamic analysis approach. For this reason, have been selected as an analytical tool in this study. Research concludes that the Bayesian networks have a distinct and effective method in the safety analysis because of the flexibility of its structure; it is suitable for a wide variety of accident scenarios.Keywords: bayesian networks, crude oil tank, fault tree, prediction, safety
Procedia PDF Downloads 660951 Oil Pollution Analysis of the Ecuadorian Rainforest Using Remote Sensing Methods
Authors: Juan Heredia, Naci Dilekli
Abstract:
The Ecuadorian Rainforest has been polluted for almost 60 years with little to no regard to oversight, law, or regulations. The consequences have been vast environmental damage such as pollution and deforestation, as well as sickness and the death of many people and animals. The aim of this paper is to quantify and localize the polluted zones, which something that has not been conducted and is the first step for remediation. To approach this problem, multi-spectral Remote Sensing imagery was utilized using a novel algorithm developed for this study, based on four normalized indices available in the literature. The algorithm classifies the pixels in polluted or healthy ones. The results of this study include a new algorithm for pixel classification and quantification of the polluted area in the selected image. Those results were finally validated by ground control points found in the literature. The main conclusion of this work is that using hyperspectral images, it is possible to identify polluted vegetation. The future work is environmental remediation, in-situ tests, and more extensive results that would inform new policymaking.Keywords: remote sensing, oil pollution quatification, amazon forest, hyperspectral remote sensing
Procedia PDF Downloads 163950 Application of UAS in Forest Firefighting for Detecting Ignitions and 3D Fuel Volume Estimation
Authors: Artur Krukowski, Emmanouela Vogiatzaki
Abstract:
The article presents results from the AF3 project “Advanced Forest Fire Fighting” focused on Unmanned Aircraft Systems (UAS)-based 3D surveillance and 3D area mapping using high-resolution photogrammetric methods from multispectral imaging, also taking advantage of the 3D scanning techniques from the SCAN4RECO project. We also present a proprietary embedded sensor system used for the detection of fire ignitions in the forest using near-infrared based scanner with weight and form factors allowing it to be easily deployed on standard commercial micro-UAVs, such as DJI Inspire or Mavic. Results from real-life pilot trials in Greece, Spain, and Israel demonstrated added-value in the use of UAS for precise and reliable detection of forest fires, as well as high-resolution 3D aerial modeling for accurate quantification of human resources and equipment required for firefighting.Keywords: forest wildfires, surveillance, fuel volume estimation, firefighting, ignition detectors, 3D modelling, UAV
Procedia PDF Downloads 142949 E-Service and the Nigerian Banking Sector: A Review of ATM Architecture and Operations
Authors: Bashir Aliyu Yauri, Rufai Aliyu Yauri
Abstract:
With the introduction of cash-less society policy by the Central Bank of Nigeria, the concept of e-banking services has experienced a significant improvement over the years. Today quite a number of people are embracing e-banking activities especially ATM, thereby moving away from the conventional banking system. This paper presents a review of the underlying Architectural Layout of Intra-Bank and Inter-Bank ATM connectivity in Nigeria. The paper further investigates and discusses factors affecting the Intra-Bank and Inter-Bank ATM connectivity in Nigeria. And as well possible solutions to these factors affecting ATM Connectivity and Operations are proposed.Keywords: architectural layout, automated teller machine, e-services, postilion
Procedia PDF Downloads 632948 Bug Localization on Single-Line Bugs of Apache Commons Math Library
Authors: Cherry Oo, Hnin Min Oo
Abstract:
Software bug localization is one of the most costly tasks in program repair technique. Therefore, there is a high claim for automated bug localization techniques that can monitor programmers to the locations of bugs, with slight human arbitration. Spectrum-based bug localization aims to help software developers to discover bugs rapidly by investigating abstractions of the program traces to make a ranking list of most possible buggy modules. Using the Apache Commons Math library project, we study the diagnostic accuracy using our spectrum-based bug localization metric. Our outcomes show that the greater performance of a specific similarity coefficient, used to inspect the program spectra, is mostly effective on localizing of single line bugs.Keywords: software testing, bug localization, program spectra, bug
Procedia PDF Downloads 143947 Designing AI-Enabled Smart Maintenance Scheduler: Enhancing Object Reliability through Automated Management
Authors: Arun Prasad Jaganathan
Abstract:
In today's rapidly evolving technological landscape, the need for efficient and proactive maintenance management solutions has become increasingly evident across various industries. Traditional approaches often suffer from drawbacks such as reactive strategies, leading to potential downtime, increased costs, and decreased operational efficiency. In response to these challenges, this paper proposes an AI-enabled approach to object-based maintenance management aimed at enhancing reliability and efficiency. The paper contributes to the growing body of research on AI-driven maintenance management systems, highlighting the transformative impact of intelligent technologies on enhancing object reliability and operational efficiency.Keywords: AI, machine learning, predictive maintenance, object-based maintenance, expert team scheduling
Procedia PDF Downloads 58946 Integrating Computer-Aided Manufacturing and Computer-Aided Design for Streamlined Carpentry Production in Ghana
Authors: Benson Tette, Thomas Mensah
Abstract:
As a developing country, Ghana has a high potential to harness the economic value of every industry. Two of the industries that produce below capacity are handicrafts (for instance, carpentry) and information technology (i.e., computer science). To boost production and maintain competitiveness, the carpentry sector in Ghana needs more effective manufacturing procedures that are also more affordable. This issue can be resolved using computer-aided manufacturing (CAM) technology, which automates the fabrication process and decreases the amount of time and labor needed to make wood goods. Yet, the integration of CAM in carpentry-related production is rarely explored. To streamline the manufacturing process, this research investigates the equipment and technology that are currently used in the Ghanaian carpentry sector for automated fabrication. The research looks at the various CAM technologies, such as Computer Numerical Control routers, laser cutters, and plasma cutters, that are accessible to Ghanaian carpenters yet unexplored. We also investigate their potential to enhance the production process. To achieve the objective, 150 carpenters, 15 software engineers, and 10 policymakers were interviewed using structured questionnaires. The responses provided by the 175 respondents were processed to eliminate outliers and omissions were corrected using multiple imputations techniques. The processed responses were analyzed through thematic analysis. The findings showed that adaptation and integration of CAD software with CAM technologies would speed up the design-to-manufacturing process for carpenters. It must be noted that achieving such results entails first; examining the capabilities of current CAD software, then determining what new functions and resources are required to improve the software's suitability for carpentry tasks. Responses from both carpenters and computer scientists showed that it is highly practical and achievable to streamline the design-to-manufacturing process through processes such as modifying and combining CAD software with CAM technology. Making the carpentry-software integration program more useful for carpentry projects would necessitate investigating the capabilities of the current CAD software and identifying additional features in the Ghanaian ecosystem and tools that are required. In conclusion, the Ghanaian carpentry sector has a chance to increase productivity and competitiveness through the integration of CAM technology with CAD software. Carpentry companies may lower labor costs and boost production capacity by automating the fabrication process, giving them a competitive advantage. This study offers implementation-ready and representative recommendations for successful implementation as well as important insights into the equipment and technologies available for automated fabrication in the Ghanaian carpentry sector.Keywords: carpentry, computer-aided manufacturing (CAM), Ghana, information technology(IT)
Procedia PDF Downloads 98945 Localized Recharge Modeling of a Coastal Aquifer from a Dam Reservoir (Korba, Tunisia)
Authors: Nejmeddine Ouhichi, Fethi Lachaal, Radhouane Hamdi, Olivier Grunberger
Abstract:
Located in Cap Bon peninsula (Tunisia), the Lebna dam was built in 1987 to balance local water salt intrusion taking place in the coastal aquifer of Korba. The first intention was to reduce coastal groundwater over-pumping by supplying surface water to a large irrigation system. The unpredicted beneficial effect was recorded with the occurrence of a direct localized recharge to the coastal aquifer by leakage through the geological material of the southern bank of the lake. The hydrological balance of the reservoir dam gave an estimation of the annual leakage volume, but dynamic processes and sound quantification of recharge inputs are still required to understand the localized effect of the recharge in terms of piezometry and quality. Present work focused on simulating the recharge process to confirm the hypothesis, and established a sound quantification of the water supply to the coastal aquifer and extend it to multi-annual effects. A spatial frame of 30km² was used for modeling. Intensive outcrops and geophysical surveys based on 68 electrical resistivity soundings were used to characterize the aquifer 3D geometry and the limit of the Plio-quaternary geological material concerned by the underground flow paths. Permeabilities were determined using 17 pumping tests on wells and piezometers. Six seasonal piezometric surveys on 71 wells around southern reservoir dam banks were performed during the 2019-2021 period. Eight monitoring boreholes of high frequency (15min) piezometric data were used to examine dynamical aspects. Model boundary conditions were specified using the geophysics interpretations coupled with the piezometric maps. The dam-groundwater flow model was performed using Visual MODFLOW software. Firstly, permanent state calibration based on the first piezometric map of February 2019 was established to estimate the permanent flow related to the different reservoir levels. Secondly, piezometric data for the 2019-2021 period were used for transient state calibration and to confirm the robustness of the model. Preliminary results confirmed the temporal link between the reservoir level and the localized recharge flow with a strong threshold effect for levels below 16 m.a.s.l. The good agreement of computed flow through recharge cells on the southern banks and hydrological budget of the reservoir open the path to future simulation scenarios of the dilution plume imposed by the localized recharge. The dam reservoir-groundwater flow-model simulation results approve a potential for storage of up to 17mm/year in existing wells, under gravity-feed conditions during level increases on the reservoir into the three years of operation. The Lebna dam groundwater flow model characterized a spatiotemporal relation between groundwater and surface water.Keywords: leakage, MODFLOW, saltwater intrusion, surface water-groundwater interaction
Procedia PDF Downloads 138944 Evaluating 8D Reports Using Text-Mining
Authors: Benjamin Kuester, Bjoern Eilert, Malte Stonis, Ludger Overmeyer
Abstract:
Increasing quality requirements make reliable and effective quality management indispensable. This includes the complaint handling in which the 8D method is widely used. The 8D report as a written documentation of the 8D method is one of the key quality documents as it internally secures the quality standards and acts as a communication medium to the customer. In practice, however, the 8D report is mostly faulty and of poor quality. There is no quality control of 8D reports today. This paper describes the use of natural language processing for the automated evaluation of 8D reports. Based on semantic analysis and text-mining algorithms the presented system is able to uncover content and formal quality deficiencies and thus increases the quality of the complaint processing in the long term.Keywords: 8D report, complaint management, evaluation system, text-mining
Procedia PDF Downloads 315943 Mathematical Knowledge a Prerequisite for Science Education Courses in Tertiary Institution
Authors: Esther Yemisi Akinjiola
Abstract:
Mathematics has been regarded as the backbone of science and technological development, without which no nation can achieve any sustainable growth and development. Mathematics is a useful tool to simplify science by quantification of phenomena; hence physics and chemistry cannot be done without Calculus and Statistics. Mathematics is used in physical science to calculate the measurement of objects and their characteristics, as well as to show the relationship between different functions and properties. Mathematics is the building block for everything in our daily lives, including the use of mobile devices, architecture design, ancient arts, engineering sports, and. among others. Therefore the study of Mathematics is made compulsory at primary, basic, and secondary school levels. Thus, this paper discusses the concepts of Mathematics, science, and their relationships. Also, it discusses Mathematics contents needed to study science-oriented courses such as physics education, chemistry education, and biology education in the tertiary institution. The paper concluded that without adequate knowledge of Mathematics, it will be difficult, if not impossible, for science education students to cope in their field of study.Keywords: mathematical knowledge, prerequisite, science education, tertiary institution
Procedia PDF Downloads 91942 A Highly Sensitive Dip Strip for Detection of Phosphate in Water
Authors: Hojat Heidari-Bafroui, Amer Charbaji, Constantine Anagnostopoulos, Mohammad Faghri
Abstract:
Phosphorus is an essential nutrient for plant life which is most frequently found as phosphate in water. Once phosphate is found in abundance in surface water, a series of adverse effects on an ecosystem can be initiated. Therefore, a portable and reliable method is needed to monitor the phosphate concentrations in the field. In this paper, an inexpensive dip strip device with the ascorbic acid/antimony reagent dried on blotting paper along with wet chemistry is developed for the detection of low concentrations of phosphate in water. Ammonium molybdate and sulfuric acid are separately stored in liquid form so as to improve significantly the lifetime of the device and enhance the reproducibility of the device’s performance. The limit of detection and quantification for the optimized device are 0.134 ppm and 0.472 ppm for phosphate in water, respectively. The device’s shelf life, storage conditions, and limit of detection are superior to what has been previously reported for the paper-based phosphate detection devices.Keywords: phosphate detection, paper-based device, molybdenum blue method, colorimetric assay
Procedia PDF Downloads 170941 Testing of Electronic Control Unit Communication Interface
Authors: Petr Šimek, Kamil Kostruk
Abstract:
This paper deals with the problem of testing the Electronic Control Unit (ECU) for the specified function validation. Modern ECUs have many functions which need to be tested. This process requires tracking between the test and the specification. The technique discussed in this paper explores the system for automating this process. The paper focuses in its chapter IV on the introduction to the problem in general, then it describes the proposed test system concept and its principle. It looks at how the process of the ECU interface specification file for automated interface testing and test tracking works. In the end, the future possible development of the project is discussed.Keywords: electronic control unit testing, embedded system, test generate, test automation, process automation, CAN bus, ethernet
Procedia PDF Downloads 112940 Extraction of Polystyrene from Styrofoam Waste: Synthesis of Novel Chelating Resin for the Enrichment and Speciation of Cr(III)/Cr(vi) Ions in Industrial Effluents
Authors: Ali N. Siyal, Saima Q. Memon, Latif Elçi, Aydan Elçi
Abstract:
Polystyrene (PS) was extracted from Styrofoam (expanded polystyrene foam) waste, so called white pollutant. The PS was functionalized with N, N- Bis(2-aminobenzylidene)benzene-1,2-diamine (ABA) ligand through an azo spacer. The resin was characterized by FT-IR spectroscopy and elemental analysis. The PS-N=N-ABA resin was used for the enrichment and speciation of Cr(III)/Cr(VI) ions and total Cr determination in aqueous samples by Flame Atomic Absorption Spectrometry (FAAS). The separation of Cr(III)/Cr(VI) ions was achieved at pH 2. The recovery of Cr(VI) ions was achieved ≥ 95.0% at optimum parameters: pH 2; resin amount 300 mg; flow rates 2.0 mL min-1 of solution and 2.0 mL min-1 of eluent (2.0 mol L-1 HNO3). Total Cr was determined by oxidation of Cr(III) to Cr(VI) ions using H2O2. The limit of detection (LOD) and quantification (LOQ) of Cr(VI) were found to be 0.40 and 1.20 μg L-1, respectively with preconcentration factor of 250. Total saturation and breakthrough capacitates of the resin for Cr(IV) ions were found to be 0.181 and 0.531 mmol g-1, respectively. The proposed method was successfully applied for the preconcentration/speciation of Cr(III)/Cr(VI) ions and determination of total Cr in industrial effluents.Keywords: styrofoam waste, polymeric resin, preconcentration, speciation, Cr(III)/Cr(VI) ions, FAAS
Procedia PDF Downloads 294939 X-Ray Diffraction and Precision Dilatometer Study of Neutron-Irradiated Nuclear Graphite Recovery Process up to 1673K
Authors: Yuhao Jin, Zhou Zhou, Katsumi Yoshida, Zhengcao Li, Tadashi Maruyama, Toyohiko Yano
Abstract:
Four kinds of nuclear graphite, IG-110U, ETP-10, CX-2002U and IG-430U were neutron-irradiated at different fluences and temperatures, ranged from 1.38 x 1024 to 7.4 x 1025 n/m2 (E > 1.0 MeV) at 473K, 573K and 673K. To take into account the disorder in the microstructure, such as stacking faults and anisotropic coherent lengths, the X-ray diffraction patterns were interpreted using a comprehensive structural model and a refinement program CARBONXS. The deduced structural parameters show the changes of lattice parameters, coherent lengths along the c-axis and the basal plane, and the degree of turbostratic disorder as a function of the irradiation dose. Our results reveal neutron irradiation effects on the microstructure and macroscopic dimension, which are consistent with previous work. The methodology used in this work enables the quantification of the damage on the microstructure of nuclear graphite induced by neutron irradiation.Keywords: nuclear graphite, neutron irradiation, thermal annealing, recovery behavior, dimensional change, CARBONX, XRD analysis
Procedia PDF Downloads 400938 Electrokinetic Transport of Power Law Fluid through Hydrophobic Micro-Slits
Authors: Ainul Haque, Ameeye Kumar Nayak
Abstract:
Flow enhancement and species transport in a slit hydrophobic microchannel is studied for non-Newtonian fluids with the externally imposed electric field and pressure gradient. The incompressible Poisson-Nernst-Plank equations and the Navier-Stokes equations are approximated by lubrication theory to quantify the flow structure due to hydrophobic and hydrophilic surfaces. The analytical quantification of velocity and pressure of electroosmotic flow (EOF) is made with the numerical results due to the staggered grid based finite volume method for flow governing equations. The resistance force due to fluid friction and shear force along the surface are decreased by the hydrophobicity, enables the faster movement of fluid particles. The resulting flow enhancement factor Ef is increased with the low viscous fluid and provides maximum species transport. Also, the analytical comparison of EOF with pressure driven EOF justifies the flow enhancement due to hydrophobicity and shear impact on flow variation.Keywords: electroosmotic flow, hydrophobic surface, power-law fluid, shear effect
Procedia PDF Downloads 377937 Dual-use UAVs in Armed Conflicts: Opportunities and Risks for Cyber and Electronic Warfare
Authors: Piret Pernik
Abstract:
Based on strategic, operational, and technical analysis of the ongoing armed conflict in Ukraine, this paper will examine the opportunities and risks of using small commercial drones (dual-use unmanned aerial vehicles, UAV) for military purposes. The paper discusses the opportunities and risks in the information domain, encompassing both cyber and electromagnetic interference and attacks. The paper will draw conclusions on a possible strategic impact to the battlefield outcomes in the modern armed conflicts by the widespread use of dual-use UAVs. This article will contribute to filling the gap in the literature by examining based on empirical data cyberattacks and electromagnetic interference. Today, more than one hundred states and non-state actors possess UAVs ranging from low cost commodity models, widely are dual-use, available and affordable to anyone, to high-cost combat UAVs (UCAV) with lethal kinetic strike capabilities, which can be enhanced with Artificial Intelligence (AI) and Machine Learning (ML). Dual-use UAVs have been used by various actors for intelligence, reconnaissance, surveillance, situational awareness, geolocation, and kinetic targeting. Thus they function as force multipliers enabling kinetic and electronic warfare attacks and provide comparative and asymmetric operational and tactical advances. Some go as far as argue that automated (or semi-automated) systems can change the character of warfare, while others observe that the use of small drones has not changed the balance of power or battlefield outcomes. UAVs give considerable opportunities for commanders, for example, because they can be operated without GPS navigation, makes them less vulnerable and dependent on satellite communications. They can and have been used to conduct cyberattacks, electromagnetic interference, and kinetic attacks. However, they are highly vulnerable to those attacks themselves. So far, strategic studies, literature, and expert commentary have overlooked cybersecurity and electronic interference dimension of the use of dual use UAVs. The studies that link technical analysis of opportunities and risks with strategic battlefield outcomes is missing. It is expected that dual use commercial UAV proliferation in armed and hybrid conflicts will continue and accelerate in the future. Therefore, it is important to understand specific opportunities and risks related to the crowdsourced use of dual-use UAVs, which can have kinetic effects. Technical countermeasures to protect UAVs differ depending on a type of UAV (small, midsize, large, stealth combat), and this paper will offer a unique analysis of small UAVs both from the view of opportunities and risks for commanders and other actors in armed conflict.Keywords: dual-use technology, cyber attacks, electromagnetic warfare, case studies of cyberattacks in armed conflicts
Procedia PDF Downloads 102936 Effects of Additives on Thermal Decompositions of Carbon Black/High Density Polyethylene Compounds
Authors: Orathai Pornsunthorntawee, Wareerom Polrut, Nopphawan Phonthammachai
Abstract:
In the present work, the effects of additives, including contents of the added antioxidants and type of the selected metallic stearates (either calcium stearate (CaSt) or zinc stearate (ZnSt)), on the thermal stabilities of carbon black (CB)/high density polyethylene (HDPE) compounds were studied. The results showed that the AO contents played a key role in the thermal stabilities of the CB/HDPE compounds—the higher the AO content, the higher the thermal stabilities. Although the CaSt-containing compounds were slightly superior to those with ZnSt in terms of the thermal stabilities, the remaining solid residue of CaSt after heated to the temperature of 600 °C (mainly calcium carbonate (CaCO3) as characterized by the X-ray diffraction (XRD) technique) seemed to catalyze the decomposition of CB in the HDPE-based compounds. Hence, the quantification of CB in the CaSt-containing compounds with a muffle furnace gave an inaccurate CB content—much lower than actual value. However, this phenomenon was negligible in the ZnSt-containing system.Keywords: antioxidant, stearate, carbon black, polyethylene
Procedia PDF Downloads 363935 Paradigms of Assessment, Valuation and Quantification to Trade Ecosystem Services: A Review Focusing on Mangroves and Wetlands
Authors: Rama Seth, Luise Noring, Pratim Majumdar
Abstract:
Based on an extensive literature review, this paper presents distinct approaches to value, quantify and trade ecosystem services, with particular emphasis on services provided by mangroves and wetlands. Building on diverse monetary and market-based systems for the improved allocation of natural resources, such trading and exchange-based methods can help tackle the degradation of ecosystem services in a more targeted and structured manner than achievable with stand-alone policy and administrative regulations. Using various threads of literature, the paper proposes a platform that serves as the skeletal foundation for developing an efficient global market for ecosystem services trading. The paper bridges a significant research and practice gap by recommending how to establish an equilibrium in the biosphere via trading mechanisms while also discovering other research gaps and future research potential in the domain of ecosystem valuation.Keywords: environment, economics, mangroves, wetlands, markets, ESG, global capital, climate investments, valuation, ecosystem services
Procedia PDF Downloads 251934 Pharmacokinetic Monitoring of Glimepiride and Ilaprazole in Rat Plasma by High Performance Liquid Chromatography with Diode Array Detection
Authors: Anil P. Dewani, Alok S. Tripathi, Anil V. Chandewar
Abstract:
Present manuscript reports the development and validation of a quantitative high performance liquid chromatography method for the pharmacokinetic evaluation of Glimepiride (GLM) and Ilaprazole (ILA) in rat plasma. The plasma samples were involved with Solid phase extraction process (SPE). The analytes were resolved on a Phenomenex C18 column (4.6 mm× 250 mm; 5 µm particle size) using a isocratic elution mode comprising methanol:water (80:20 % v/v) with pH of water modified to 3 using Formic acid, the total run time was 10 min at 225 nm as common wavelength, the flow rate throughout was 1ml/min. The method was validated over the concentration range from 10 to 600 ng/mL for GLM and ILA, in rat plasma. Metformin (MET) was used as Internal Standard. Validation data demonstrated the method to be selective, sensitive, accurate and precise. The limit of detection was 1.54 and 4.08 and limit of quantification was 5.15 and 13.62 for GLM and ILA respectively, the method demonstrated excellent linearity with correlation coefficients (r2) 0.999. The intra and inter-day precision (RSD%) values were < 2.0% for both ILA and GLM. The method was successfully applied in pharmacokinetic studies followed by oral administration in rats.Keywords: pharmacokinetics, glimepiride, ilaprazole, HPLC, SPE
Procedia PDF Downloads 369