Search results for: artificial recharge
1251 Predictive Analysis of the Stock Price Market Trends with Deep Learning
Authors: Suraj Mehrotra
Abstract:
The stock market is a volatile, bustling marketplace that is a cornerstone of economics. It defines whether companies are successful or in spiral. A thorough understanding of it is important - many companies have whole divisions dedicated to analysis of both their stock and of rivaling companies. Linking the world of finance and artificial intelligence (AI), especially the stock market, has been a relatively recent development. Predicting how stocks will do considering all external factors and previous data has always been a human task. With the help of AI, however, machine learning models can help us make more complete predictions in financial trends. Taking a look at the stock market specifically, predicting the open, closing, high, and low prices for the next day is very hard to do. Machine learning makes this task a lot easier. A model that builds upon itself that takes in external factors as weights can predict trends far into the future. When used effectively, new doors can be opened up in the business and finance world, and companies can make better and more complete decisions. This paper explores the various techniques used in the prediction of stock prices, from traditional statistical methods to deep learning and neural networks based approaches, among other methods. It provides a detailed analysis of the techniques and also explores the challenges in predictive analysis. For the accuracy of the testing set, taking a look at four different models - linear regression, neural network, decision tree, and naïve Bayes - on the different stocks, Apple, Google, Tesla, Amazon, United Healthcare, Exxon Mobil, J.P. Morgan & Chase, and Johnson & Johnson, the naïve Bayes model and linear regression models worked best. For the testing set, the naïve Bayes model had the highest accuracy along with the linear regression model, followed by the neural network model and then the decision tree model. The training set had similar results except for the fact that the decision tree model was perfect with complete accuracy in its predictions, which makes sense. This means that the decision tree model likely overfitted the training set when used for the testing set.Keywords: machine learning, testing set, artificial intelligence, stock analysis
Procedia PDF Downloads 951250 Analysis of Friction Stir Welding Process for Joining Aluminum Alloy
Authors: A. M. Khourshid, I. Sabry
Abstract:
Friction Stir Welding (FSW), a solid state joining technique, is widely being used for joining Al alloys for aerospace, marine automotive and many other applications of commercial importance. FSW were carried out using a vertical milling machine on Al 5083 alloy pipe. These pipe sections are relatively small in diameter, 5mm, and relatively thin walled, 2 mm. In this study, 5083 aluminum alloy pipe were welded as similar alloy joints using (FSW) process in order to investigate mechanical and microstructural properties .rotation speed 1400 r.p.m and weld speed 10,40,70 mm/min. In order to investigate the effect of welding speeds on mechanical properties, metallographic and mechanical tests were carried out on the welded areas. Vickers hardness profile and tensile tests of the joints as a metallurgical feasibility of friction stir welding for joining Al 6061 aluminum alloy welding was performed on pipe with different thickness 2, 3 and 4 mm,five rotational speeds (485,710,910,1120 and 1400) rpm and a traverse speed (4, 8 and 10)mm/min was applied. This work focuses on two methods such as artificial neural networks using software (pythia) and response surface methodology (RSM) to predict the tensile strength, the percentage of elongation and hardness of friction stir welded 6061 aluminum alloy. An artificial neural network (ANN) model was developed for the analysis of the friction stir welding parameters of 6061 pipe. The tensile strength, the percentage of elongation and hardness of weld joints were predicted by taking the parameters Tool rotation speed, material thickness and travel speed as a function. A comparison was made between measured and predicted data. Response surface methodology (RSM) also developed and the values obtained for the response Tensile strengths, the percentage of elongation and hardness are compared with measured values. The effect of FSW process parameter on mechanical properties of 6061 aluminum alloy has been analyzed in detail.Keywords: friction stir welding (FSW), al alloys, mechanical properties, microstructure
Procedia PDF Downloads 4621249 Smartphone-Based Human Activity Recognition by Machine Learning Methods
Authors: Yanting Cao, Kazumitsu Nawata
Abstract:
As smartphones upgrading, their software and hardware are getting smarter, so the smartphone-based human activity recognition will be described as more refined, complex, and detailed. In this context, we analyzed a set of experimental data obtained by observing and measuring 30 volunteers with six activities of daily living (ADL). Due to the large sample size, especially a 561-feature vector with time and frequency domain variables, cleaning these intractable features and training a proper model becomes extremely challenging. After a series of feature selection and parameters adjustment, a well-performed SVM classifier has been trained.Keywords: smart sensors, human activity recognition, artificial intelligence, SVM
Procedia PDF Downloads 1441248 Hybrid Approach for Country’s Performance Evaluation
Authors: C. Slim
Abstract:
This paper presents an integrated model, which hybridized data envelopment analysis (DEA) and support vector machine (SVM) together, to class countries according to their efficiency and performance. This model takes into account aspects of multi-dimensional indicators, decision-making hierarchy and relativity of measurement. Starting from a set of indicators of performance as exhaustive as possible, a process of successive aggregations has been developed to attain an overall evaluation of a country’s competitiveness.Keywords: Artificial Neural Networks (ANN), Support vector machine (SVM), Data Envelopment Analysis (DEA), Aggregations, indicators of performance
Procedia PDF Downloads 3381247 The Integrated Water Management of the Northern Saharan Aquifer System in a Climatic Changes Context
Authors: Mohamed Redha Menani
Abstract:
The Northern Saharan aquifer system “SASS” shared by Algeria, Libya, and Tunisia, covers a surface of about 1 100 000 km². It is composed of superposed aquifers; the upper one is the “Continental terminal – CT” (Eocene calcareous formation) situated at 400 m depth in average, while the” Continental Intercalaire – CI”(clay sands from Albian to Lower Cretaceous) is generally at 1500 m depth. This aquifer system is situated in a dry zone with a very weak current recharge but with a non-renewable big volume stored, estimated between 20 000 and 31 000 km³. From 1970 to nowadays, the exploitation of the SASS has increased from 0.6 to more than 2.5 km³/year. This situation provoked risks of water salinisation, reduction of the artesianisme, an increase of drawdowns, etc. which seriously threaten the sustainable socioeconomic development engaged in the SASS zone. Face the water shortage induced by the alarming dryness noted these last years, particularly in the MENA region, the joint management of this system by the three concerned countries, engaged for many years, needs a long-term strategy of integrated water resources management to meet the expected socio-economic goals projected not only in the SASS zone but also in other places, by water transfers. The sustainable management of this extensive aquifer system, aiming to satisfy various needs not only in the areas covered by the SASS but also in other areas through hydraulic transfers, can only be considered if this management is genuinely coordinated, incorporating schemes that primarily address the major constraint of climate change, which has been observed worldwide over the past two decades and is intensifying. In this particular climate context, management schemes must necessarily target several aspects, including (i) Updating the state of water resource exploitation in the SASS. (ii) Guiding agricultural usage as the primary consumer to ensure significant water savings. (iii) Constant monitoring through a network of piezometers to control the physicochemical parameters of the exploited aquifers. (iv) Other aspects related to governance within the framework of integrated management must also be taken into consideration, particularly environmental aspects and conflict resolution. However, problems, especially political ones as currently seen in Libya, may limit or at least disrupt the prospects of coordinated and sustainable management of this aquifer system, which is vital for the three countries.Keywords: transboundary water resources, SASS, governance, climatic changes
Procedia PDF Downloads 821246 Application of Groundwater Level Data Mining in Aquifer Identification
Authors: Liang Cheng Chang, Wei Ju Huang, You Cheng Chen
Abstract:
Investigation and research are keys for conjunctive use of surface and groundwater resources. The hydrogeological structure is an important base for groundwater analysis and simulation. Traditionally, the hydrogeological structure is artificially determined based on geological drill logs, the structure of wells, groundwater levels, and so on. In Taiwan, groundwater observation network has been built and a large amount of groundwater-level observation data are available. The groundwater level is the state variable of the groundwater system, which reflects the system response combining hydrogeological structure, groundwater injection, and extraction. This study applies analytical tools to the observation database to develop a methodology for the identification of confined and unconfined aquifers. These tools include frequency analysis, cross-correlation analysis between rainfall and groundwater level, groundwater regression curve analysis, and decision tree. The developed methodology is then applied to groundwater layer identification of two groundwater systems: Zhuoshui River alluvial fan and Pingtung Plain. The abovementioned frequency analysis uses Fourier Transform processing time-series groundwater level observation data and analyzing daily frequency amplitude of groundwater level caused by artificial groundwater extraction. The cross-correlation analysis between rainfall and groundwater level is used to obtain the groundwater replenishment time between infiltration and the peak groundwater level during wet seasons. The groundwater regression curve, the average rate of groundwater regression, is used to analyze the internal flux in the groundwater system and the flux caused by artificial behaviors. The decision tree uses the information obtained from the above mentioned analytical tools and optimizes the best estimation of the hydrogeological structure. The developed method reaches training accuracy of 92.31% and verification accuracy 93.75% on Zhuoshui River alluvial fan and training accuracy 95.55%, and verification accuracy 100% on Pingtung Plain. This extraordinary accuracy indicates that the developed methodology is a great tool for identifying hydrogeological structures.Keywords: aquifer identification, decision tree, groundwater, Fourier transform
Procedia PDF Downloads 1571245 Regional Flood Frequency Analysis in Narmada Basin: A Case Study
Authors: Ankit Shah, R. K. Shrivastava
Abstract:
Flood and drought are two main features of hydrology which affect the human life. Floods are natural disasters which cause millions of rupees’ worth of damage each year in India and the whole world. Flood causes destruction in form of life and property. An accurate estimate of the flood damage potential is a key element to an effective, nationwide flood damage abatement program. Also, the increase in demand of water due to increase in population, industrial and agricultural growth, has let us know that though being a renewable resource it cannot be taken for granted. We have to optimize the use of water according to circumstances and conditions and need to harness it which can be done by construction of hydraulic structures. For their safe and proper functioning of hydraulic structures, we need to predict the flood magnitude and its impact. Hydraulic structures play a key role in harnessing and optimization of flood water which in turn results in safe and maximum use of water available. Mainly hydraulic structures are constructed on ungauged sites. There are two methods by which we can estimate flood viz. generation of Unit Hydrographs and Flood Frequency Analysis. In this study, Regional Flood Frequency Analysis has been employed. There are many methods for estimating the ‘Regional Flood Frequency Analysis’ viz. Index Flood Method. National Environmental and Research Council (NERC Methods), Multiple Regression Method, etc. However, none of the methods can be considered universal for every situation and location. The Narmada basin is located in Central India. It is drained by most of the tributaries, most of which are ungauged. Therefore it is very difficult to estimate flood on these tributaries and in the main river. As mentioned above Artificial Neural Network (ANN)s and Multiple Regression Method is used for determination of Regional flood Frequency. The annual peak flood data of 20 sites gauging sites of Narmada Basin is used in the present study to determine the Regional Flood relationships. Homogeneity of the considered sites is determined by using the Index Flood Method. Flood relationships obtained by both the methods are compared with each other, and it is found that ANN is more reliable than Multiple Regression Method for the present study area.Keywords: artificial neural network, index flood method, multi layer perceptrons, multiple regression, Narmada basin, regional flood frequency
Procedia PDF Downloads 4191244 Women Entrepreneurs’ in Nigeria: Issues and Challenges
Authors: Mohammed Mainoma, Abubakar Tijanni, Mohammed Aliyu
Abstract:
Globalization has brought a structural change in industry. It is the breaking of artificial boundaries and given way to new product, new service, new market, and new technology among others. It leads to the realization that men entrepreneurs’ alone cannot meet the demand of the teeming population. Therefore there is a need for the participation, involvement, and engagement of females in the production and distribution of goods and services. This will enhance growth and development of a nation. It is in line with the above that this paper attempt to discuss meaning of women entrepreneurs, roles, types, problems, and prospects. Also, on the basis of conclusion the paper recommended that entrepreneurship education should be introduced in all Tertiary Institutions in Nigeria.Keywords: women, entrepreneurs, issues, challenges
Procedia PDF Downloads 5181243 Revolutionizing Legal Drafting: Leveraging Artificial Intelligence for Efficient Legal Work
Authors: Shreya Poddar
Abstract:
Legal drafting and revising are recognized as highly demanding tasks for legal professionals. This paper introduces an approach to automate and refine these processes through the use of advanced Artificial Intelligence (AI). The method employs Large Language Models (LLMs), with a specific focus on 'Chain of Thoughts' (CoT) and knowledge injection via prompt engineering. This approach differs from conventional methods that depend on comprehensive training or fine-tuning of models with extensive legal knowledge bases, which are often expensive and time-consuming. The proposed method incorporates knowledge injection directly into prompts, thereby enabling the AI to generate more accurate and contextually appropriate legal texts. This approach substantially decreases the necessity for thorough model training while preserving high accuracy and relevance in drafting. Additionally, the concept of guardrails is introduced. These are predefined parameters or rules established within the AI system to ensure that the generated content adheres to legal standards and ethical guidelines. The practical implications of this method for legal work are considerable. It has the potential to markedly lessen the time lawyers allocate to document drafting and revision, freeing them to concentrate on more intricate and strategic facets of legal work. Furthermore, this method makes high-quality legal drafting more accessible, possibly reducing costs and expanding the availability of legal services. This paper will elucidate the methodology, providing specific examples and case studies to demonstrate the effectiveness of 'Chain of Thoughts' and knowledge injection in legal drafting. The potential challenges and limitations of this approach will also be discussed, along with future prospects and enhancements that could further advance legal work. The impact of this research on the legal industry is substantial. The adoption of AI-driven methods by legal professionals can lead to enhanced efficiency, precision, and consistency in legal drafting, thereby altering the landscape of legal work. This research adds to the expanding field of AI in law, introducing a method that could significantly alter the nature of legal drafting and practice.Keywords: AI-driven legal drafting, legal automation, futureoflegalwork, largelanguagemodels
Procedia PDF Downloads 651242 Maximizing the Efficiency of Knowledge Management Systems
Authors: Tori Reddy Dodla, Laura Ann Jones
Abstract:
The objective of this study was to propose strategies to improve the efficiency of Knowledge Management Systems (KMS). This study highlights best practices from various industries to create an overall summary of Knowledge Management (KM) and efficiency in organizational performance. Results indicated eleven best practices for maximizing the efficiency of organizational KMS that can be divided into four categories: Designing the KMS, Identifying Case Studies, Implementing the KMS, and Promoting adoption and usage. Our findings can be used as a foundation for scholars to conduct further research on KMS efficiency.Keywords: artificial intelligence, knowledge management efficiency, knowledge management systems, organizational performance
Procedia PDF Downloads 1131241 The Role of Twitter Bots in Political Discussion on 2019 European Elections
Authors: Thomai Voulgari, Vasilis Vasilopoulos, Antonis Skamnakis
Abstract:
The aim of this study is to investigate the effect of the European election campaigns (May 23-26, 2019) on Twitter achieving with artificial intelligence tools such as troll factories and automated inauthentic accounts. Our research focuses on the last European Parliamentary elections that took place between 23 and 26 May 2019 specifically in Italy, Greece, Germany and France. It is difficult to estimate how many Twitter users are actually bots (Echeverría, 2017). Detection for fake accounts is becoming even more complicated as AI bots are made more advanced. A political bot can be programmed to post comments on a Twitter account for a political candidate, target journalists with manipulated content or engage with politicians and artificially increase their impact and popularity. We analyze variables related to 1) the scope of activity of automated bots accounts and 2) degree of coherence and 3) degree of interaction taking into account different factors, such as the type of content of Twitter messages and their intentions, as well as the spreading to the general public. For this purpose, we collected large volumes of Twitter accounts of party leaders and MEP candidates between 10th of May and 26th of July based on content analysis of tweets based on hashtags while using an innovative network analysis tool known as MediaWatch.io (https://mediawatch.io/). According to our findings, one of the highest percentage (64.6%) of automated “bot” accounts during 2019 European election campaigns was in Greece. In general terms, political bots aim to proliferation of misinformation on social media. Targeting voters is a way that it can be achieved contribute to social media manipulation. We found that political parties and individual politicians create and promote purposeful content on Twitter using algorithmic tools. Based on this analysis, online political advertising play an important role to the process of spreading misinformation during elections campaigns. Overall, inauthentic accounts and social media algorithms are being used to manipulate political behavior and public opinion.Keywords: artificial intelligence tools, human-bot interactions, political manipulation, social networking, troll factories
Procedia PDF Downloads 1381240 Development of a Mechanical Ventilator Using A Manual Artificial Respiration Unit
Authors: Isomar Lima da Silva, Alcilene Batalha Pontes, Aristeu Jonatas Leite de Oliveira, Roberto Maia Augusto
Abstract:
Context: Mechanical ventilators are medical devices that help provide oxygen and ventilation to patients with respiratory difficulties. This equipment consists of a manual breathing unit that can be operated by a doctor or nurse and a mechanical ventilator that controls the airflow and pressure in the patient's respiratory system. This type of ventilator is commonly used in emergencies and intensive care units where it is necessary to provide breathing support to critically ill or injured patients. Objective: In this context, this work aims to develop a reliable and low-cost mechanical ventilator to meet the demand of hospitals in treating people affected by Covid-19 and other severe respiratory diseases, offering a chance of treatment as an alternative to mechanical ventilators currently available in the market. Method: The project presents the development of a low-cost auxiliary ventilator with a controlled ventilatory system assisted by integrated hardware and firmware for respiratory cycle control in non-invasive mechanical ventilation treatments using a manual artificial respiration unit. The hardware includes pressure sensors capable of identifying positive expiratory pressure, peak inspiratory flow, and injected air volume. The embedded system controls the data sent by the sensors. It ensures efficient patient breathing through the operation of the sensors, microcontroller, and actuator, providing patient data information to the healthcare professional (system operator) through the graphical interface and enabling clinical parameter adjustments as needed. Results: The test data of the developed mechanical ventilator presented satisfactory results in terms of performance and reliability, showing that the equipment developed can be a viable alternative to commercial mechanical ventilators currently available, offering a low-cost solution to meet the increasing demand for respiratory support equipment.Keywords: mechanical fans, breathing, medical equipment, COVID-19, intensive care units
Procedia PDF Downloads 701239 Challenges over Two Semantic Repositories - OWLIM and AllegroGraph
Authors: Paria Tajabor, Azin Azarbani
Abstract:
The purpose of this research study is exploring two kind of semantic repositories with regards to various factors to find the best approaches that an artificial manager can use to produce ontology in a system based on their interaction, association and research. To this end, as the best way to evaluate each system and comparing with others is analysis, several benchmarking over these two repositories were examined. These two semantic repositories: OWLIM and AllegroGraph will be the main core of this study. The general objective of this study is to be able to create an efficient and cost-effective manner reports which is required to support decision making in any large enterprise.Keywords: OWLIM, allegrograph, RDF, reasoning, semantic repository, semantic-web, SPARQL, ontology, query
Procedia PDF Downloads 2621238 Fully Autonomous Vertical Farm to Increase Crop Production
Authors: Simone Cinquemani, Lorenzo Mantovani, Aleksander Dabek
Abstract:
New technologies in agriculture are opening new challenges and new opportunities. Among these, certainly, robotics, vision, and artificial intelligence are the ones that will make a significant leap, compared to traditional agricultural techniques, possible. In particular, the indoor farming sector will be the one that will benefit the most from these solutions. Vertical farming is a new field of research where mechanical engineering can bring knowledge and know-how to transform a highly labor-based business into a fully autonomous system. The aim of the research is to develop a multi-purpose, modular, and perfectly integrated platform for crop production in indoor vertical farming. Activities will be based both on hardware development such as automatic tools to perform different activities on soil and plants, as well as research to introduce an extensive use of monitoring techniques based on machine learning algorithms. This paper presents the preliminary results of a research project of a vertical farm living lab designed to (i) develop and test vertical farming cultivation practices, (ii) introduce a very high degree of mechanization and automation that makes all processes replicable, fully measurable, standardized and automated, (iii) develop a coordinated control and management environment for autonomous multiplatform or tele-operated robots in environments with the aim of carrying out complex tasks in the presence of environmental and cultivation constraints, (iv) integrate AI-based algorithms as decision support system to improve quality production. The coordinated management of multiplatform systems still presents innumerable challenges that require a strongly multidisciplinary approach right from the design, development, and implementation phases. The methodology is based on (i) the development of models capable of describing the dynamics of the various platforms and their interactions, (ii) the integrated design of mechatronic systems able to respond to the needs of the context and to exploit the strength characteristics highlighted by the models, (iii) implementation and experimental tests performed to test the real effectiveness of the systems created, evaluate any weaknesses so as to proceed with a targeted development. To these aims, a fully automated laboratory for growing plants in vertical farming has been developed and tested. The living lab makes extensive use of sensors to determine the overall state of the structure, crops, and systems used. The possibility of having specific measurements for each element involved in the cultivation process makes it possible to evaluate the effects of each variable of interest and allows for the creation of a robust model of the system as a whole. The automation of the laboratory is completed with the use of robots to carry out all the necessary operations, from sowing to handling to harvesting. These systems work synergistically thanks to the knowledge of detailed models developed based on the information collected, which allows for deepening the knowledge of these types of crops and guarantees the possibility of tracing every action performed on each single plant. To this end, artificial intelligence algorithms have been developed to allow synergistic operation of all systems.Keywords: automation, vertical farming, robot, artificial intelligence, vision, control
Procedia PDF Downloads 391237 Studies on the Emergence Pattern of Cercariae from Fresh Water Snails (Mollusca: Gastropoda)
Authors: V. R. Kakulte, K. N. Gaikwad
Abstract:
The emergence pattern of different types of cercariae form three snail hosts Melania tuberculata, Lymnea auricularia Viviparous bengalensis has been studied in detail. In natural emerging method the snails (2 to 3 at a time) were kept in separate test tube. This was constant source of living cercariae naturally emerging from the snails. The sunlight and artificial light play an important positive role in stimulating the emergence of cercariae has been observed. The effect of light and dark on the emission pattern of cercariae has been studied.Keywords: cercariae, snail host, emergence pattern, gastropoda
Procedia PDF Downloads 3171236 Influential Parameters in Estimating Soil Properties from Cone Penetrating Test: An Artificial Neural Network Study
Authors: Ahmed G. Mahgoub, Dahlia H. Hafez, Mostafa A. Abu Kiefa
Abstract:
The Cone Penetration Test (CPT) is a common in-situ test which generally investigates a much greater volume of soil more quickly than possible from sampling and laboratory tests. Therefore, it has the potential to realize both cost savings and assessment of soil properties rapidly and continuously. The principle objective of this paper is to demonstrate the feasibility and efficiency of using artificial neural networks (ANNs) to predict the soil angle of internal friction (Φ) and the soil modulus of elasticity (E) from CPT results considering the uncertainties and non-linearities of the soil. In addition, ANNs are used to study the influence of different parameters and recommend which parameters should be included as input parameters to improve the prediction. Neural networks discover relationships in the input data sets through the iterative presentation of the data and intrinsic mapping characteristics of neural topologies. General Regression Neural Network (GRNN) is one of the powerful neural network architectures which is utilized in this study. A large amount of field and experimental data including CPT results, plate load tests, direct shear box, grain size distribution and calculated data of overburden pressure was obtained from a large project in the United Arab Emirates. This data was used for the training and the validation of the neural network. A comparison was made between the obtained results from the ANN's approach, and some common traditional correlations that predict Φ and E from CPT results with respect to the actual results of the collected data. The results show that the ANN is a very powerful tool. Very good agreement was obtained between estimated results from ANN and actual measured results with comparison to other correlations available in the literature. The study recommends some easily available parameters that should be included in the estimation of the soil properties to improve the prediction models. It is shown that the use of friction ration in the estimation of Φ and the use of fines content in the estimation of E considerable improve the prediction models.Keywords: angle of internal friction, cone penetrating test, general regression neural network, soil modulus of elasticity
Procedia PDF Downloads 4151235 Estimating Groundwater Seepage Rates: Case Study at Zegveld, Netherlands
Authors: Wondmyibza Tsegaye Bayou, Johannes C. Nonner, Joost Heijkers
Abstract:
This study aimed to identify and estimate dynamic groundwater seepage rates using four comparative methods; the Darcian approach, the water balance approach, the tracer method, and modeling. The theoretical background to these methods is put together in this study. The methodology was applied to a case study area at Zegveld following the advice of the Water Board Stichtse Rijnlanden. Data collection has been from various offices and a field campaign in the winter of 2008/09. In this complex confining layer of the study area, the location of the phreatic groundwater table is at a shallow depth compared to the piezometric water level. Data were available for the model years 1989 to 2000 and winter 2008/09. The higher groundwater table shows predominately-downward seepage in the study area. Results of the study indicated that net recharge to the groundwater table (precipitation excess) and the ditch system are the principal sources for seepage across the complex confining layer. Especially in the summer season, the contribution from the ditches is significant. Water is supplied from River Meije through a pumping system to meet the ditches' water demand. The groundwater seepage rate was distributed unevenly throughout the study area at the nature reserve averaging 0.60 mm/day for the model years 1989 to 2000 and 0.70 mm/day for winter 2008/09. Due to data restrictions, the seepage rates were mainly determined based on the Darcian method. Furthermore, the water balance approach and the tracer methods are applied to compute the flow exchange within the ditch system. The site had various validated groundwater levels and vertical flow resistance data sources. The phreatic groundwater level map compared with TNO-DINO groundwater level data values overestimated the groundwater level depth by 28 cm. The hydraulic resistance values obtained based on the 3D geological map compared with the TNO-DINO data agreed with the model values before calibration. On the other hand, the calibrated model significantly underestimated the downward seepage in the area compared with the field-based computations following the Darcian approach.Keywords: groundwater seepage, phreatic water table, piezometric water level, nature reserve, Zegveld, The Netherlands
Procedia PDF Downloads 851234 Review of Different Machine Learning Algorithms
Authors: Syed Romat Ali Shah, Bilal Shoaib, Saleem Akhtar, Munib Ahmad, Shahan Sadiqui
Abstract:
Classification is a data mining technique, which is recognizedon Machine Learning (ML) algorithm. It is used to classifythe individual articlein a knownofinformation into a set of predefinemodules or group. Web mining is also a portion of that sympathetic of data mining methods. The main purpose of this paper to analysis and compare the performance of Naïve Bayse Algorithm, Decision Tree, K-Nearest Neighbor (KNN), Artificial Neural Network (ANN)and Support Vector Machine (SVM). This paper consists of different ML algorithm and their advantages and disadvantages and also define research issues.Keywords: Data Mining, Web Mining, classification, ML Algorithms
Procedia PDF Downloads 3031233 Fibroblast Compatibility of Core-Shell Coaxially Electrospun Hybrid Poly(ε-Caprolactone)/Chitosan Scaffolds
Authors: Hilal Turkoglu Sasmazel, Ozan Ozkan, Seda Surucu
Abstract:
Tissue engineering is the field of treating defects caused by injuries, trauma or acute/chronic diseases by using artificial scaffolds that mimic the extracellular matrix (ECM), the natural biological support for the tissues and cells within the body. The main aspects of a successful artificial scaffold are (i) large surface area in order to provide multiple anchorage points for cells to attach, (ii) suitable porosity in order to achieve 3 dimensional growth of the cells within the scaffold as well as proper transport of nutrition, biosignals and waste and (iii) physical, chemical and biological compatibility of the material in order to obtain viability throughout the healing process. By hybrid scaffolds where two or more different materials were combined with advanced fabrication techniques into complex structures, it is possible to combine the advantages of individual materials into one single structure while eliminating the disadvantages of each. Adding this to the complex structure provided by advanced fabrication techniques enables obtaining the desired aspects of a successful artificial tissue scaffold. In this study, fibroblast compatibility of poly(ε-caprolactone) (PCL)/chitosan core-shell electrospun hybrid scaffolds with proper mechanical, chemical and physical properties successfully developed in our previous study was investigated. Standard 7-day cell culture was carried out with L929 fibroblast cell line. The viability of the cells cultured with the scaffolds was monitored with 3-(4,5-dimethylthiazol-2-yl)-2,5-diphenyltetrazolium bromide (MTT) viability assay for every 48 h starting with 24 h after the initial seeding. In this assay, blank commercial tissue culture polystyrene (TCPS) Petri dishes, single electrospun PCL and single electrospun chitosan mats were used as control in order to compare and contrast the performance of the hybrid scaffolds. The adhesion, proliferation, spread and growth of the cells on/within the scaffolds were observed visually on the 3rd and the 7th days of the culture period with confocal laser scanning microscopy (CSLM) and scanning electron microscopy (SEM). The viability assay showed that the hybrid scaffolds caused no toxicity for fibroblast cells and provided a steady increase in cell viability, effectively doubling the cell density for every 48 h for the course of 7 days, as compared to TCPS, single electrospun PCL or chitosan mats. The cell viability on the hybrid scaffold was ~2 fold better compared to TCPS because of its 3D ECM-like structure compared to 2D flat surface of commercially cell compatible TCPS, and the performance was ~2 fold and ~10 fold better compared to single PCL and single chitosan mats, respectively, even though both fabricated similarly with electrospinning as non-woven fibrous structures, because single PCL and chitosan mats were either too hydrophobic or too hydrophilic to maintain cell attachment points. The viability results were verified with visual images obtained with CSLM and SEM, in which cells found to achieve characteristic spindle-like fibroblast shape and spread on the surface as well within the pores successfully at high densities.Keywords: chitosan, core-shell, fibroblast, electrospinning, PCL
Procedia PDF Downloads 1761232 Impact of Interface Soil Layer on Groundwater Aquifer Behaviour
Authors: Hayder H. Kareem, Shunqi Pan
Abstract:
The geological environment where the groundwater is collected represents the most important element that affects the behaviour of groundwater aquifer. As groundwater is a worldwide vital resource, it requires knowing the parameters that affect this source accurately so that the conceptualized mathematical models would be acceptable to the broadest ranges. Therefore, groundwater models have recently become an effective and efficient tool to investigate groundwater aquifer behaviours. Groundwater aquifer may contain aquitards, aquicludes, or interfaces within its geological formations. Aquitards and aquicludes have geological formations that forced the modellers to include those formations within the conceptualized groundwater models, while interfaces are commonly neglected from the conceptualization process because the modellers believe that the interface has no effect on aquifer behaviour. The current research highlights the impact of an interface existing in a real unconfined groundwater aquifer called Dibdibba, located in Al-Najaf City, Iraq where it has a river called the Euphrates River that passes through the eastern part of this city. Dibdibba groundwater aquifer consists of two types of soil layers separated by an interface soil layer. A groundwater model is built for Al-Najaf City to explore the impact of this interface. Calibration process is done using PEST 'Parameter ESTimation' approach and the best Dibdibba groundwater model is obtained. When the soil interface is conceptualized, results show that the groundwater tables are significantly affected by that interface through appearing dry areas of 56.24 km² and 6.16 km² in the upper and lower layers of the aquifer, respectively. The Euphrates River will also leak water into the groundwater aquifer of 7359 m³/day. While these results are changed when the soil interface is neglected where the dry area became 0.16 km², the Euphrates River leakage became 6334 m³/day. In addition, the conceptualized models (with and without interface) reveal different responses for the change in the recharge rates applied on the aquifer through the uncertainty analysis test. The aquifer of Dibdibba in Al-Najaf City shows a slight deficit in the amount of water supplied by the current pumping scheme and also notices that the Euphrates River suffers from stresses applied to the aquifer. Ultimately, this study shows a crucial need to represent the interface soil layer in model conceptualization to be the intended and future predicted behaviours more reliable for consideration purposes.Keywords: Al-Najaf City, groundwater aquifer behaviour, groundwater modelling, interface soil layer, Visual MODFLOW
Procedia PDF Downloads 1831231 A Comprehensive Survey of Artificial Intelligence and Machine Learning Approaches across Distinct Phases of Wildland Fire Management
Authors: Ursula Das, Manavjit Singh Dhindsa, Kshirasagar Naik, Marzia Zaman, Richard Purcell, Srinivas Sampalli, Abdul Mutakabbir, Chung-Horng Lung, Thambirajah Ravichandran
Abstract:
Wildland fires, also known as forest fires or wildfires, are exhibiting an alarming surge in frequency in recent times, further adding to its perennial global concern. Forest fires often lead to devastating consequences ranging from loss of healthy forest foliage and wildlife to substantial economic losses and the tragic loss of human lives. Despite the existence of substantial literature on the detection of active forest fires, numerous potential research avenues in forest fire management, such as preventative measures and ancillary effects of forest fires, remain largely underexplored. This paper undertakes a systematic review of these underexplored areas in forest fire research, meticulously categorizing them into distinct phases, namely pre-fire, during-fire, and post-fire stages. The pre-fire phase encompasses the assessment of fire risk, analysis of fuel properties, and other activities aimed at preventing or reducing the risk of forest fires. The during-fire phase includes activities aimed at reducing the impact of active forest fires, such as the detection and localization of active fires, optimization of wildfire suppression methods, and prediction of the behavior of active fires. The post-fire phase involves analyzing the impact of forest fires on various aspects, such as the extent of damage in forest areas, post-fire regeneration of forests, impact on wildlife, economic losses, and health impacts from byproducts produced during burning. A comprehensive understanding of the three stages is imperative for effective forest fire management and mitigation of the impact of forest fires on both ecological systems and human well-being. Artificial intelligence and machine learning (AI/ML) methods have garnered much attention in the cyber-physical systems domain in recent times leading to their adoption in decision-making in diverse applications including disaster management. This paper explores the current state of AI/ML applications for managing the activities in the aforementioned phases of forest fire. While conventional machine learning and deep learning methods have been extensively explored for the prevention, detection, and management of forest fires, a systematic classification of these methods into distinct AI research domains is conspicuously absent. This paper gives a comprehensive overview of the state of forest fire research across more recent and prominent AI/ML disciplines, including big data, classical machine learning, computer vision, explainable AI, generative AI, natural language processing, optimization algorithms, and time series forecasting. By providing a detailed overview of the potential areas of research and identifying the diverse ways AI/ML can be employed in forest fire research, this paper aims to serve as a roadmap for future investigations in this domain.Keywords: artificial intelligence, computer vision, deep learning, during-fire activities, forest fire management, machine learning, pre-fire activities, post-fire activities
Procedia PDF Downloads 721230 Using MALDI-TOF MS to Detect Environmental Microplastics (Polyethylene, Polyethylene Terephthalate, and Polystyrene) within a Simulated Tissue Sample
Authors: Kara J. Coffman-Rea, Karen E. Samonds
Abstract:
Microplastic pollution is an urgent global threat to our planet and human health. Microplastic particles have been detected within our food, water, and atmosphere, and found within the human stool, placenta, and lung tissue. However, most spectrometric microplastic detection methods require chemical digestion which can alter or destroy microplastic particles and makes it impossible to acquire information about their in-situ distribution. MALDI TOF MS (Matrix-assisted laser desorption ionization-time of flight mass spectrometry) is an analytical method using a soft ionization technique that can be used for polymer analysis. This method provides a valuable opportunity to both acquire information regarding the in-situ distribution of microplastics and also minimizes the destructive element of chemical digestion. In addition, MALDI TOF MS allows for expanded analysis of the microplastics including detection of specific additives that may be present within them. MALDI TOF MS is particularly sensitive to sample preparation and has not yet been used to analyze environmental microplastics within their specific location (e.g., biological tissues, sediment, water). In this study, microplastics were created using polyethylene gloves, polystyrene micro-foam, and polyethylene terephthalate cable sleeving. Plastics were frozen using liquid nitrogen and ground to obtain small fragments. An artificial tissue was created using a cellulose sponge as scaffolding coated with a MaxGel Extracellular Matrix to simulate human lung tissue. Optimal preparation techniques (e.g., matrix, cationization reagent, solvent, mixing ratio, laser intensity) were first established for each specific polymer type. The artificial tissue sample was subsequently spiked with microplastics, and specific polymers were detected using MALDI-TOF-MS. This study presents a novel method for the detection of environmental polyethylene, polyethylene terephthalate, and polystyrene microplastics within a complex sample. Results of this study provide an effective method that can be used in future microplastics research and can aid in determining the potential threats to environmental and human health that they pose.Keywords: environmental plastic pollution, MALDI-TOF MS, microplastics, polymer identification
Procedia PDF Downloads 2561229 Artificial Intelligence and Development: The Missing Link
Authors: Driss Kettani
Abstract:
ICT4D actors are naturally attempted to include AI in the range of enabling technologies and tools that could support and boost the Development process, and to refer to these as AI4D. But, doing so, assumes that AI complies with the very specific features of ICT4D context, including, among others, affordability, relevance, openness, and ownership. Clearly, none of these is fulfilled, and the enthusiastic posture that AI4D is a natural part of ICT4D is not grounded and, to certain extent, does not serve the purpose of Technology for Development at all. In the context of Development, it is important to emphasize and prioritize ICT4D, in the national digital transformation strategies, instead of borrowing "trendy" waves of the IT Industry that are motivated by business considerations, with no specific care/consideration to Development.Keywords: AI, ICT4D, technology for development, position paper
Procedia PDF Downloads 881228 Chatbots in Education: Case of Development Using a Chatbot Development Platform
Authors: Dulani Jayasuriya
Abstract:
This study outlines the developmental steps of a chatbot for administrative purposes of a large undergraduate course. The chatbot is able to handle student queries about administrative details, including assessment deadlines, course documentation, how to navigate the course, group formation, etc. The development window screenshots are that of a free account on the Snatchbot platform such that this can be adopted by the wider public. While only one connection to an answer based on possible keywords is shown here, one needs to develop multiple connections leading to different answers based on different keywords for the actual chatbot to function. The overall flow of the chatbot showing connections between different interactions is depicted at the end.Keywords: chatbots, education, technology, snatch bot, artificial intelligence
Procedia PDF Downloads 1041227 Probabilistic Life Cycle Assessment of the Nano Membrane Toilet
Authors: A. Anastasopoulou, A. Kolios, T. Somorin, A. Sowale, Y. Jiang, B. Fidalgo, A. Parker, L. Williams, M. Collins, E. J. McAdam, S. Tyrrel
Abstract:
Developing countries are nowadays confronted with great challenges related to domestic sanitation services in view of the imminent water scarcity. Contemporary sanitation technologies established in these countries are likely to pose health risks unless waste management standards are followed properly. This paper provides a solution to sustainable sanitation with the development of an innovative toilet system, called Nano Membrane Toilet (NMT), which has been developed by Cranfield University and sponsored by the Bill & Melinda Gates Foundation. The particular technology converts human faeces into energy through gasification and provides treated wastewater from urine through membrane filtration. In order to evaluate the environmental profile of the NMT system, a deterministic life cycle assessment (LCA) has been conducted in SimaPro software employing the Ecoinvent v3.3 database. The particular study has determined the most contributory factors to the environmental footprint of the NMT system. However, as sensitivity analysis has identified certain critical operating parameters for the robustness of the LCA results, adopting a stochastic approach to the Life Cycle Inventory (LCI) will comprehensively capture the input data uncertainty and enhance the credibility of the LCA outcome. For that purpose, Monte Carlo simulations, in combination with an artificial neural network (ANN) model, have been conducted for the input parameters of raw material, produced electricity, NOX emissions, amount of ash and transportation of fertilizer. The given analysis has provided the distribution and the confidence intervals of the selected impact categories and, in turn, more credible conclusions are drawn on the respective LCIA (Life Cycle Impact Assessment) profile of NMT system. Last but not least, the specific study will also yield essential insights into the methodological framework that can be adopted in the environmental impact assessment of other complex engineering systems subject to a high level of input data uncertainty.Keywords: sanitation systems, nano-membrane toilet, lca, stochastic uncertainty analysis, Monte Carlo simulations, artificial neural network
Procedia PDF Downloads 2251226 Snapchat’s Scanning Feature
Authors: Reham Banwair, Lana Alshehri, Sara Hadrawi
Abstract:
The purpose of this project is to identify user satisfaction with the AI functions on Snapchat, in order to generate improvement proposals that allow its development within the app. To achieve this, a qualitative analysis was carried out through interviews to people who usually use the application, revealing their satisfaction or dissatisfaction with the usefulness of the AI. In addition, the background of the company and its introduction in these algorithms were analyzed. Furthermore, the characteristics of the three main functions of AI were explained: identify songs, solve mathematical problems, and recognize plants. As a result, it was obtained that 50% still do not know the characteristics of AI, 50% still believe song recognition is not always correct, 41.7% believe that math problems are usually accurate and 91.7% believes the plant detection tool is working properly.Keywords: artificial intelligence, scanning, Snapchat, machine learning
Procedia PDF Downloads 1341225 Design and Implementation of Wireless Syncronized AI System for Security
Authors: Saradha Priya
Abstract:
Developing virtual human is very important to meet the challenges occurred in many applications where human find difficult or risky to perform the task. A robot is a machine that can perform a task automatically or with guidance. Robotics is generally a combination of artificial intelligence and physical machines (motors). Computational intelligence involves the programmed instructions. This project proposes a robotic vehicle that has a camera, PIR sensor and text command based movement. It is specially designed to perform surveillance and other few tasks in the most efficient way. Serial communication has been occurred between a remote Base Station, GUI Application, and PC.Keywords: Zigbee, camera, pirsensor, wireless transmission, DC motor
Procedia PDF Downloads 3491224 Corrosion Interaction Between Steel and Acid Mine Drainage: Use of AI Based on Fuzzy Logic
Authors: Maria Luisa de la Torre, Javier Aroba, Jose Miguel Davila, Aguasanta M. Sarmiento
Abstract:
Steel is one of the most widely used materials in polymetallic sulfide mining installations. One of the main problems suffered by these facilities is the economic losses due to the corrosion of this material, which is accelerated and aggravated by the contact with acid waters generated in these mines when sulfides come into contact with oxygen and water. This generation of acidic water, in turn, is accelerated by the presence of acidophilic bacteria. In order to gain a more detailed understanding of this corrosion process and the interaction between steel and acidic water, a laboratory experiment was carried out in which carbon steel plates were introduced into four different solutions for 27 days: distilled water (BK), which tried to assimilate the effect produced by rain on this material, an acid solution from a mine with a high Fe2+/Fe3+ (PO) content, another acid solution of water from another mine with a high Fe3+/Fe2+ (PH) content and, finally, one that reproduced the acid mine water with a high Fe2+/Fe3+ content but in which there were no bacteria (ST). Every 24 hours, physicochemical parameters were measured, and water samples were taken to carry out an analysis of the dissolved elements. The results of these measurements were processed using an explainable AI model based on fuzzy logic. It could be seen that, in all cases, there was an increase in pH, as well as in the concentrations of Fe and, in particular, Fe(II), as a consequence of the oxidation of the steel plates. Proportionally, the increase in Fe concentration was higher in PO and ST than in PH because Fe precipitates were produced in the latter. The rise of Fe(II) was proportionally much higher in PH, especially in the first hours of exposure, because it started from a lower initial concentration of this ion. Although to a lesser extent than in PH, the greater increase in Fe(II) also occurred faster in PO than in ST, a consequence of the action of the catalytic bacteria. On the other hand, Cu concentrations decreased throughout the experiment (with the exception of distilled water, which initially had no Cu, as a result of an electrochemical process that generates a precipitation of Cu together with Fe hydroxides. This decrease is lower in PH because the high total acidity keeps it in solution for a longer time. With the application of an artificial intelligence tool, it has been possible to evaluate the effects of steel corrosion in mining environments, corroborating and extending what was obtained by means of classical statistics.Keywords: acid mine drainage, artificial intelligence, carbon steel, corrosion, fuzzy logic
Procedia PDF Downloads 81223 Duo Lingo: Learning Languages through Play
Authors: Yara Bajnaid, Malak Zaidan, Eman Dakkak
Abstract:
This research explores the use of Artificial Intelligence in Duolingo, a popular mobile application for language learning. Duolingo's success hinges on its gamified approach and adaptive learning system, both heavily reliant on AI functionalities. The research also analyzes user feedback regarding Duolingo's AI functionalities. While a significant majority (70%) consider Duolingo a reliable tool for language learning, there's room for improvement. Overall, AI plays a vital role in personalizing the learning journey and delivering interactive exercises. However, continuous improvement based on user feedback can further enhance the effectiveness of Duolingo's AI functionalities.Keywords: AI, Duolingo, language learning, application
Procedia PDF Downloads 481222 Automatic Content Curation of Visual Heritage
Authors: Delphine Ribes Lemay, Valentine Bernasconi, André Andrade, Lara DéFayes, Mathieu Salzmann, FréDéRic Kaplan, Nicolas Henchoz
Abstract:
Digitization and preservation of large heritage induce high maintenance costs to keep up with the technical standards and ensure sustainable access. Creating impactful usage is instrumental to justify the resources for long-term preservation. The Museum für Gestaltung of Zurich holds one of the biggest poster collections of the world from which 52’000 were digitised. In the process of building a digital installation to valorize the collection, one objective was to develop an algorithm capable of predicting the next poster to show according to the ones already displayed. The work presented here describes the steps to build an algorithm able to automatically create sequences of posters reflecting associations performed by curator and professional designers. The exposed challenge finds similarities with the domain of song playlist algorithms. Recently, artificial intelligence techniques and more specifically, deep-learning algorithms have been used to facilitate their generations. Promising results were found thanks to Recurrent Neural Networks (RNN) trained on manually generated playlist and paired with clusters of extracted features from songs. We used the same principles to create the proposed algorithm but applied to a challenging medium, posters. First, a convolutional autoencoder was trained to extract features of the posters. The 52’000 digital posters were used as a training set. Poster features were then clustered. Next, an RNN learned to predict the next cluster according to the previous ones. RNN training set was composed of poster sequences extracted from a collection of books from the Gestaltung Museum of Zurich dedicated to displaying posters. Finally, within the predicted cluster, the poster with the best proximity compared to the previous poster is selected. The mean square distance between features of posters was used to compute the proximity. To validate the predictive model, we compared sequences of 15 posters produced by our model to randomly and manually generated sequences. Manual sequences were created by a professional graphic designer. We asked 21 participants working as professional graphic designers to sort the sequences from the one with the strongest graphic line to the one with the weakest and to motivate their answer with a short description. The sequences produced by the designer were ranked first 60%, second 25% and third 15% of the time. The sequences produced by our predictive model were ranked first 25%, second 45% and third 30% of the time. The sequences produced randomly were ranked first 15%, second 29%, and third 55% of the time. Compared to designer sequences, and as reported by participants, model and random sequences lacked thematic continuity. According to the results, the proposed model is able to generate better poster sequencing compared to random sampling. Eventually, our algorithm is sometimes able to outperform a professional designer. As a next step, the proposed algorithm should include a possibility to create sequences according to a selected theme. To conclude, this work shows the potentiality of artificial intelligence techniques to learn from existing content and provide a tool to curate large sets of data, with a permanent renewal of the presented content.Keywords: Artificial Intelligence, Digital Humanities, serendipity, design research
Procedia PDF Downloads 184