Search results for: artificial air storage reservoir
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4583

Search results for: artificial air storage reservoir

2813 Enhancing Solar Fuel Production by CO₂ Photoreduction Using Transition Metal Oxide Catalysts in Reactors Prepared by Additive Manufacturing

Authors: Renata De Toledo Cintra, Bruno Ramos, Douglas Gouvêa

Abstract:

There is a huge global concern due to the emission of greenhouse gases, consequent environmental problems, and the increase in the average temperature of the planet, caused mainly by fossil fuels, petroleum derivatives represent a big part. One of the main greenhouse gases, in terms of volume, is CO₂. Recovering a part of this product through chemical reactions that use sunlight as an energy source and even producing renewable fuel (such as ethane, methane, ethanol, among others) is a great opportunity. The process of artificial photosynthesis, through the conversion of CO₂ and H₂O into organic products and oxygen using a metallic oxide catalyst, and incidence of sunlight, is one of the promising solutions. Therefore, this research is of great relevance. To this reaction take place efficiently, an optimized reactor was developed through simulation and prior analysis so that the geometry of the internal channel is an efficient route and allows the reaction to happen, in a controlled and optimized way, in flow continuously and offering the least possible resistance. The design of this reactor prototype can be made in different materials, such as polymers, ceramics and metals, and made through different processes, such as additive manufacturing (3D printer), CNC, among others. To carry out the photocatalysis in the reactors, different types of catalysts will be used, such as ZnO deposited by spray pyrolysis in the lighting window, probably modified ZnO, TiO₂ and modified TiO₂, among others, aiming to increase the production of organic molecules, with the lowest possible energy.

Keywords: artificial photosynthesis, CO₂ reduction, photocatalysis, photoreactor design, 3D printed reactors, solar fuels

Procedia PDF Downloads 86
2812 The Protection of Artificial Intelligence (AI)-Generated Creative Works Through Authorship: A Comparative Analysis Between the UK and Nigerian Copyright Experience to Determine Lessons to Be Learnt from the UK

Authors: Esther Ekundayo

Abstract:

The nature of AI-generated works makes it difficult to identify an author. Although, some scholars have suggested that all the players involved in its creation should be allocated authorship according to their respective contribution. From the programmer who creates and designs the AI to the investor who finances the AI and to the user of the AI who most likely ends up creating the work in question. While others suggested that this issue may be resolved by the UK computer-generated works (CGW) provision under Section 9(3) of the Copyright Designs and Patents Act 1988. However, under the UK and Nigerian copyright law, only human-created works are recognised. This is usually assessed based on their originality. This simply means that the work must have been created as a result of its author’s creative and intellectual abilities and not copied. Such works are literary, dramatic, musical and artistic works and are those that have recently been a topic of discussion with regards to generative artificial intelligence (Generative AI). Unlike Nigeria, the UK CDPA recognises computer-generated works and vests its authorship with the human who made the necessary arrangement for its creation . However, making necessary arrangement in the case of Nova Productions Ltd v Mazooma Games Ltd was interpreted similarly to the traditional authorship principle, which requires the skills of the creator to prove originality. Although, some recommend that computer-generated works complicates this issue, and AI-generated works should enter the public domain as authorship cannot be allocated to AI itself. Additionally, the UKIPO recognising these issues in line with the growing AI trend in a public consultation launched in the year 2022, considered whether computer-generated works should be protected at all and why. If not, whether a new right with a different scope and term of protection should be introduced. However, it concluded that the issue of computer-generated works would be revisited as AI was still in its early stages. Conversely, due to the recent developments in this area with regards to Generative AI systems such as ChatGPT, Midjourney, DALL-E and AIVA, amongst others, which can produce human-like copyright creations, it is therefore important to examine the relevant issues which have the possibility of altering traditional copyright principles as we know it. Considering that the UK and Nigeria are both common law jurisdictions but with slightly differing approaches to this area, this research, therefore, seeks to answer the following questions by comparative analysis: 1)Who is the author of an AI-generated work? 2)Is the UK’s CGW provision worthy of emulation by the Nigerian law? 3) Would a sui generis law be capable of protecting AI-generated works and its author under both jurisdictions? This research further examines the possible barriers to the implementation of the new law in Nigeria, such as limited technical expertise and lack of awareness by the policymakers, amongst others.

Keywords: authorship, artificial intelligence (AI), generative ai, computer-generated works, copyright, technology

Procedia PDF Downloads 97
2811 Effect of Sintering Time and Porosity on Microstructure, Mechanical and Corrosion Properties of Ti6Al15Mo Alloy for Implant Applications

Authors: Jyotsna Gupta, S. Ghosh, S. Aravindan

Abstract:

The requirement of artificial prostheses (such as hip and knee joints) has increased with time. Many researchers are working to develop new implants with improved properties such as excellent biocompatibility with no tissue reactions, corrosion resistance in body fluid, high yield strength and low elastic modulus. Further, the morphological properties of the artificial implants should also match with that of the human bone so that cell adhesion, proliferation and transportation of the minerals and nutrition through body fluid can be obtained. Present study attempts to make porous Ti6Al15Mo alloys through powder metallurgy route using space holder technique. The alloy consists of 6wt% of Al which was taken as α phase stabilizer and 15wt% Mo was taken as β phase stabilizer with theoretical density 4.708. Ammonium hydrogen carbonate is used as a space holder in order to generate the porosity. The porosity of these fabricated porous alloys was controlled by adding the 0, 50, 70 vol.% of the space holder content. Three phases were found in the microstructure: α, α_2 and β phase of titanium. Kirkendall pores are observed to be decreased with increase of holding time during sintering and parallelly compressive strength and elastic modulus value increased slightly. Compressive strength and elastic modulus of porous Ti-6Al-15Mo alloy (1.17 g/cm3 density) is found to be suitable for cancellous bone. Released ions from Ti-6Al-15Mo alloy are far below from the permissible limits in human body.

Keywords: bone implant, powder metallurgy, sintering time, Ti-6Al-15Mo

Procedia PDF Downloads 144
2810 Comparative Electrochemical Studies of Enzyme-Based and Enzyme-less Graphene Oxide-Based Nanocomposite as Glucose Biosensor

Authors: Chetna Tyagi. G. B. V. S. Lakshmi, Ambuj Tripathi, D. K. Avasthi

Abstract:

Graphene oxide provides a good host matrix for preparing nanocomposites due to the different functional groups attached to its edges and planes. Being biocompatible, it is used in therapeutic applications. As enzyme-based biosensor requires complicated enzyme purification procedure, high fabrication cost and special storage conditions, we need enzyme-less biosensors for use even in a harsh environment like high temperature, varying pH, etc. In this work, we have prepared both enzyme-based and enzyme-less graphene oxide-based biosensors for glucose detection using glucose-oxidase as enzyme and gold nanoparticles, respectively. These samples were characterized using X-ray diffraction, UV-visible spectroscopy, scanning electron microscopy, and transmission electron microscopy to confirm the successful synthesis of the working electrodes. Electrochemical measurements were performed for both the working electrodes using a 3-electrode electrochemical cell. Cyclic voltammetry curves showed the homogeneous transfer of electron on the electrodes in the scan range between -0.2V to 0.6V. The sensing measurements were performed using differential pulse voltammetry for the glucose concentration varying from 0.01 mM to 20 mM, and sensing was improved towards glucose in the presence of gold nanoparticles. Gold nanoparticles in graphene oxide nanocomposite played an important role in sensing glucose in the absence of enzyme, glucose oxidase, as evident from these measurements. The selectivity was tested by measuring the current response of the working electrode towards glucose in the presence of the other common interfering agents like cholesterol, ascorbic acid, citric acid, and urea. The enzyme-less working electrode also showed storage stability for up to 15 weeks, making it a suitable glucose biosensor.

Keywords: electrochemical, enzyme-less, glucose, gold nanoparticles, graphene oxide, nanocomposite

Procedia PDF Downloads 141
2809 Real Estate Trend Prediction with Artificial Intelligence Techniques

Authors: Sophia Liang Zhou

Abstract:

For investors, businesses, consumers, and governments, an accurate assessment of future housing prices is crucial to critical decisions in resource allocation, policy formation, and investment strategies. Previous studies are contradictory about macroeconomic determinants of housing price and largely focused on one or two areas using point prediction. This study aims to develop data-driven models to accurately predict future housing market trends in different markets. This work studied five different metropolitan areas representing different market trends and compared three-time lagging situations: no lag, 6-month lag, and 12-month lag. Linear regression (LR), random forest (RF), and artificial neural network (ANN) were employed to model the real estate price using datasets with S&P/Case-Shiller home price index and 12 demographic and macroeconomic features, such as gross domestic product (GDP), resident population, personal income, etc. in five metropolitan areas: Boston, Dallas, New York, Chicago, and San Francisco. The data from March 2005 to December 2018 were collected from the Federal Reserve Bank, FBI, and Freddie Mac. In the original data, some factors are monthly, some quarterly, and some yearly. Thus, two methods to compensate missing values, backfill or interpolation, were compared. The models were evaluated by accuracy, mean absolute error, and root mean square error. The LR and ANN models outperformed the RF model due to RF’s inherent limitations. Both ANN and LR methods generated predictive models with high accuracy ( > 95%). It was found that personal income, GDP, population, and measures of debt consistently appeared as the most important factors. It also showed that technique to compensate missing values in the dataset and implementation of time lag can have a significant influence on the model performance and require further investigation. The best performing models varied for each area, but the backfilled 12-month lag LR models and the interpolated no lag ANN models showed the best stable performance overall, with accuracies > 95% for each city. This study reveals the influence of input variables in different markets. It also provides evidence to support future studies to identify the optimal time lag and data imputing methods for establishing accurate predictive models.

Keywords: linear regression, random forest, artificial neural network, real estate price prediction

Procedia PDF Downloads 103
2808 Design and Implementation of Low-code Model-building Methods

Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu

Abstract:

This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.

Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment

Procedia PDF Downloads 29
2807 Developing a Cloud Intelligence-Based Energy Management Architecture Facilitated with Embedded Edge Analytics for Energy Conservation in Demand-Side Management

Authors: Yu-Hsiu Lin, Wen-Chun Lin, Yen-Chang Cheng, Chia-Ju Yeh, Yu-Chuan Chen, Tai-You Li

Abstract:

Demand-Side Management (DSM) has the potential to reduce electricity costs and carbon emission, which are associated with electricity used in the modern society. A home Energy Management System (EMS) commonly used by residential consumers in a down-stream sector of a smart grid to monitor, control, and optimize energy efficiency to domestic appliances is a system of computer-aided functionalities as an energy audit for residential DSM. Implementing fault detection and classification to domestic appliances monitored, controlled, and optimized is one of the most important steps to realize preventive maintenance, such as residential air conditioning and heating preventative maintenance in residential/industrial DSM. In this study, a cloud intelligence-based green EMS that comes up with an Internet of Things (IoT) technology stack for residential DSM is developed. In the EMS, Arduino MEGA Ethernet communication-based smart sockets that module a Real Time Clock chip to keep track of current time as timestamps via Network Time Protocol are designed and implemented for readings of load phenomena reflecting on voltage and current signals sensed. Also, a Network-Attached Storage providing data access to a heterogeneous group of IoT clients via Hypertext Transfer Protocol (HTTP) methods is configured to data stores of parsed sensor readings. Lastly, a desktop computer with a WAMP software bundle (the Microsoft® Windows operating system, Apache HTTP Server, MySQL relational database management system, and PHP programming language) serves as a data science analytics engine for dynamic Web APP/REpresentational State Transfer-ful web service of the residential DSM having globally-Advanced Internet of Artificial Intelligence (AI)/Computational Intelligence. Where, an abstract computing machine, Java Virtual Machine, enables the desktop computer to run Java programs, and a mash-up of Java, R language, and Python is well-suited and -configured for AI in this study. Having the ability of sending real-time push notifications to IoT clients, the desktop computer implements Google-maintained Firebase Cloud Messaging to engage IoT clients across Android/iOS devices and provide mobile notification service to residential/industrial DSM. In this study, in order to realize edge intelligence that edge devices avoiding network latency and much-needed connectivity of Internet connections for Internet of Services can support secure access to data stores and provide immediate analytical and real-time actionable insights at the edge of the network, we upgrade the designed and implemented smart sockets to be embedded AI Arduino ones (called embedded AIduino). With the realization of edge analytics by the proposed embedded AIduino for data analytics, an Arduino Ethernet shield WizNet W5100 having a micro SD card connector is conducted and used. The SD library is included for reading parsed data from and writing parsed data to an SD card. And, an Artificial Neural Network library, ArduinoANN, for Arduino MEGA is imported and used for locally-embedded AI implementation. The embedded AIduino in this study can be developed for further applications in manufacturing industry energy management and sustainable energy management, wherein in sustainable energy management rotating machinery diagnostics works to identify energy loss from gross misalignment and unbalance of rotating machines in power plants as an example.

Keywords: demand-side management, edge intelligence, energy management system, fault detection and classification

Procedia PDF Downloads 251
2806 The Fabrication of Stress Sensing Based on Artificial Antibodies to Cortisol by Molecular Imprinted Polymer

Authors: Supannika Klangphukhiew, Roongnapa Srichana, Rina Patramanon

Abstract:

Cortisol has been used as a well-known commercial stress biomarker. A homeostasis response to psychological stress is indicated by an increased level of cortisol produced in hypothalamus-pituitary-adrenal (HPA) axis. Chronic psychological stress contributing to the high level of cortisol relates to several health problems. In this study, the cortisol biosensor was fabricated that mimicked the natural receptors. The artificial antibodies were prepared using molecular imprinted polymer technique that can imitate the performance of natural anti-cortisol antibody with high stability. Cortisol-molecular imprinted polymer (cortisol-MIP) was obtained using the multi-step swelling and polymerization protocol with cortisol as a target molecule combining methacrylic acid:acrylamide (2:1) with bisacryloyl-1,2-dihydroxy-1,2-ethylenediamine and ethylenedioxy-N-methylamphetamine as cross-linkers. Cortisol-MIP was integrated to the sensor. It was coated on the disposable screen-printed carbon electrode (SPCE) for portable electrochemical analysis. The physical properties of Cortisol-MIP were characterized by means of electron microscope techniques. The binding characteristics were evaluated via covalent patterns changing in FTIR spectra which were related to voltammetry response. The performance of cortisol-MIP modified SPCE was investigated in terms of detection range, high selectivity with a detection limit of 1.28 ng/ml. The disposable cortisol biosensor represented an application of MIP technique to recognize steroids according to their structures with feasibility and cost-effectiveness that can be developed to use in point-of-care.

Keywords: stress biomarker, cortisol, molecular imprinted polymer, screen-printed carbon electrode

Procedia PDF Downloads 273
2805 pH-Responsive Carrier Based on Polymer Particle

Authors: Florin G. Borcan, Ramona C. Albulescu, Adela Chirita-Emandi

Abstract:

pH-responsive drug delivery systems are gaining more importance because these systems deliver the drug at a specific time in regards to pathophysiological necessity, resulting in improved patient therapeutic efficacy and compliance. Polyurethane materials are well-known for industrial applications (elastomers and foams used in different insulations and automotive), but they are versatile biocompatible materials with many applications in medicine, as artificial skin for the premature neonate, membrane in the hybrid artificial pancreas, prosthetic heart valves, etc. This study aimed to obtain the physico-chemical characterization of a drug delivery system based on polyurethane microparticles. The synthesis is based on a polyaddition reaction between an aqueous phase (mixture of polyethylene-glycol M=200, 1,4-butanediol and Tween® 20) and an organic phase (lysin-diisocyanate in acetone) combined with simultaneous emulsification. Different active agents (omeprazole, amoxicillin, metoclopramide) were used to verify the release profile of the macromolecular particles in different pH mediums. Zetasizer measurements were performed using an instrument based on two modules: a Vasco size analyzer and a Wallis Zeta potential analyzer (Cordouan Technol., France) in samples that were kept in various solutions with different pH and the maximum absorbance in UV-Vis spectra were collected on a UVi Line 9,400 Spectrophotometer (SI Analytics, Germany). The results of this investigation have revealed that these particles are proper for a prolonged release in gastric medium where they can assure an almost constant concentration of the active agents for 1-2 weeks, while they can be disassembled faster in a medium with neutral pHs, such as the intestinal fluid.

Keywords: lysin-diisocyanate, nanostructures, polyurethane, Zetasizer

Procedia PDF Downloads 184
2804 Comparison of Oven and Microwave Drying on Phenolic Contents and Antioxidant Activities of Red Delicious and Golden Delicious Apples

Authors: Gulcin Yildiz, Gokcen Izli

Abstract:

Drying (dehydration) is the process of removing water from food in order to preserve the food. Drying is one of the oldest methods known for the preservation of agricultural products such as fruits and vegetables. Drying of agricultural products enhances their storage life, minimizes losses during storage, and save shipping and transportation costs. Apples are considered excellent candidates for drying. The objective of this research was to investigate the effects of microwave and oven processing on the quality of selected apple products. Red delicious and golden delicious apples were washed, peeled, and sliced. Drying experiments were performed in an oven at 50, 75 and 100 °C and in a microwave at 140 W and 210 W. Quality attributes such as color, total phenolic content and antioxidant capacity of dried samples with different methods were compared with the fresh sample. A Minolta CR-300 Chroma Meter was used to examine color changes in the apples. Total phenolic content was determined using the Folin-Ciocalteu reagent. The free radical scavenging activity of the extract was determined using 1,1-diphenyl-2-picrylhydrazyl (DPPH). It was found that the phenolic contents and antioxidant capacities of dried samples under all drying conditions were decreased compared to the fresh samples. The phenolic contents of microwave dried samples at 140 W and 210 W for both red and golden delicious apples were higher than those of the oven drying at 50, 75 and 100 °C. Similarly, the antioxidant activities of microwave dried samples at 140 W and 210 W were higher than those of the oven drying at 50, 75 and 100 °C for both types of apples. All color parameters (L*, a*, b*) were changed significantly depending on the drying methods and temperatures. The closest color values to the fresh sample were found for the microwave dried samples at 140 W. Microwave drying was proven to be more effective than oven drying.

Keywords: antioxidant capacity, color, golden delicious, microwave, red delicious, total phenolic content

Procedia PDF Downloads 236
2803 Opinion Mining to Extract Community Emotions on Covid-19 Immunization Possible Side Effects

Authors: Yahya Almurtadha, Mukhtar Ghaleb, Ahmed M. Shamsan Saleh

Abstract:

The world witnessed a fierce attack from the Covid-19 virus, which affected public life socially, economically, healthily and psychologically. The world's governments tried to confront the pandemic by imposing a number of precautionary measures such as general closure, curfews and social distancing. Scientists have also made strenuous efforts to develop an effective vaccine to train the immune system to develop antibodies to combat the virus, thus reducing its symptoms and limiting its spread. Artificial intelligence, along with researchers and medical authorities, has accelerated the vaccine development process through big data processing and simulation. On the other hand, one of the most important negatives of the impact of Covid 19 was the state of anxiety and fear due to the blowout of rumors through social media, which prompted governments to try to reassure the public with the available means. This study aims to proposed using Sentiment Analysis (AKA Opinion Mining) and deep learning as efficient artificial intelligence techniques to work on retrieving the tweets of the public from Twitter and then analyze it automatically to extract their opinions, expression and feelings, negatively or positively, about the symptoms they may feel after vaccination. Sentiment analysis is characterized by its ability to access what the public post in social media within a record time and at a lower cost than traditional means such as questionnaires and interviews, not to mention the accuracy of the information as it comes from what the public expresses voluntarily.

Keywords: deep learning, opinion mining, natural language processing, sentiment analysis

Procedia PDF Downloads 171
2802 Monitoring of Sustainability of Decorated Confectionary Product 'Moskva Cake' in Order to Define the Expiration Date

Authors: Radovan Cobanovic, Milica Rankov-Sicar

Abstract:

The fresh cake is in the group of perishable food which cannot be kept a long period of time. The study of sustainability has been done in order to extend the shelf-life of the product which was 10 days. According to the plan of sustainability, it was defined that 5 samples had to be stored for 20 days at max +8°C and analyzed every 5th day from the day of reception until the 20th day. The shelf life of cake has expired during the study of sustainability in the period between 10th and 20th day of analyses. Cake samples were subjected to sensory analysis (appearance, odor, taste, color, aroma) and bacteriological analysis (Listeria monocytogenes, Salmonella spp. and Enterobacteriaceae) according to Serbian state regulation. All analysis were tested according to ISO methodology: sensory analysis ISO 6658, Listeria monocytogenes ISO 11290-1, Salmonella spp ISO 6579, and Enterobacteriaceae ISO 21258-2. Analyses showed that after ten days of storage at a temperature defined by the manufacturers and within the product's shelf life, the cake did not have any noticeable changes in sensory characteristics. Smell and taste are unaffected there was no presence of strange smell or taste. As far as microbiological analyses are concerned, neither one pathogen was detected and number of Enterobacteriaceae was at level less than 102 cfu/g. After expiry of shelf life in a period of 15th and 20th day of storage, the sensory analysis showed the presence of strange sour-milky smell and rancid taste. Concerning microbiological analyses, there still were not positive results for pathogen microorganisms but the number of Enterobacteriaceae was at level more than 103cfu/g. Reviewing the results of sensory analysis indicates that it is not recommended to extend the shelf-life of the product comparing to the already defined shelf-life because occurred changes may adversely affect the consumer desire for the choice of this product.

Keywords: confectionary product, extension of shelf life, sensory and microbiological analyses, sustainability

Procedia PDF Downloads 242
2801 Sedimentation and Morphology of the Kura River-Deltaic System in the Southern Caucasus under Anthropogenic and Sea-Level Controls

Authors: Elmira Aliyeva, Dadash Huseynov, Robert Hoogendoorn, Salomon Kroonenberg

Abstract:

The Kura River is the major water artery in the Southern Caucasus; it is a third river in the Caspian Sea basin in terms of length and size of the catchment area, the second in terms of the water budget, and the first in the volume of sediment load. Understanding of major controls on the Kura fluvial- deltaic system is valuable for efficient management of the highly populated river basin and coastal zone. We have studied grain size of sediments accumulated in the river channels and delta and dated by 210Pb method, astrophotographs, old topographic and geological maps, and archive data. At present time sediments are supplied by the Kura River to the Caspian Sea through three distributary channels oriented north-east, south-east, and south-west. The river is dominated by the suspended load - mud, silt, very fine sand. Coarse sediments are accumulated in the distributaries, levees, point bar, and delta front. The annual suspended sediment budget in the time period 1934-1952 before construction of the Mingechavir water reservoir in 1953 in the Kura River midstream area was 36 mln.t/yr. From 1953 to 1964, the suspended load has dropped to 12 mln.t/yr. After regulation of the Kura River discharge the volume of suspended load transported via north-eastern channel reduced from 35% of the total sediment amount to 4%, and through the main south-eastern channel increased from 65% to 96% with further fall to 56% due to creation of new south-western channel in 1964. Between 1967-1976 the annual sediment budget of the Kura River reached 22,5 mln. t/yr. From 1977 to 1986, the sediment load carried by the Kura River dropped to 17,6 mln.t/yr. The historical data show that between 1860 and 1907, during relatively stable Caspian Sea level two channels - N and SE, appear to have distributed an equal amount of sediments as seen from the bilateral geometry of the delta. In the time period 1907-1929, two new channels - E and NE, appeared. The growth of three delta lobes - N, NE, and SE, and rapid progradation of the delta has occurred on the background of the Caspian Sea level rise as a result of very high sediment supply. Since 1929 the Caspian Sea level decline was followed by the progradation of the delta occurring along the SE channel. The eastern and northern channels have been silted up. The slow rate of progradation at its initial stage was caused by the artificial reduction in the sediment budget. However, the continuous sea-level fall has brought to this river bed gradient increase, high erosional rate, increase in the sediment supply, and more rapid progradation. During the subsequent sea-level rise after 1977 accompanied by the decrease in the sediment budget, the southern part of the delta has turned into a complex of small, shallow channels oriented to the south. The data demonstrate that behaviour of the Kura fluvial – deltaic system and variations in the sediment budget besides anthropogenic regulation are strongly governed by the Caspian Sea level very rapid changes.

Keywords: anthropogenic control on sediment budget, Caspian sea-level variations, Kura river sediment load, morphology of the Kura river delta, sedimentation in the Kura river delta

Procedia PDF Downloads 154
2800 Hybrid Renewable Power Systems

Authors: Salman Al-Alyani

Abstract:

In line with the Kingdom’s Vision 2030, the Saudi Green initiative was announced aimed at reducing carbon emissions by more than 4% of the global contribution. The initiative included plans to generate 50% of its energy from renewables by 2030. The geographical location of Saudi Arabia makes it among the best countries in terms of solar irradiation and has good wind resources in many areas across the Kingdom. Saudi Arabia is a wide country and has many remote locations where it is not economically feasible to connect those loads to the national grid. With the improvement of battery innovation and reduction in cost, different renewable technologies (primarily wind and solar) can be integrated to meet the need for energy in a more effective and cost-effective way. Saudi Arabia is famous for high solar irradiations in which solar power generation can extend up to six (6) hours per day (25% capacity factor) in some locations. However, the net present value (NPV) falls down to negative in some locations due to distance and high installation costs. Wind generation in Saudi Arabia is a promising technology. Hybrid renewable generation will increase the net present value and lower the payback time due to additional energy generated by wind. The infrastructure of the power system can be capitalized to contain solar generation and wind generation feeding the inverter, controller, and load. Storage systems can be added to support the hours that have an absence of wind or solar energy. Also, the smart controller that can help integrate various renewable technologies primarily wind and solar, to meet demand considering load characteristics. It could be scalable for grid or off-grid applications. The objective of this paper is to study the feasibility of introducing a hybrid renewable system in remote locations and the concept for the development of a smart controller.

Keywords: battery storage systems, hybrid power generation, solar energy, wind energy

Procedia PDF Downloads 178
2799 AI Applications in Accounting: Transforming Finance with Technology

Authors: Alireza Karimi

Abstract:

Artificial Intelligence (AI) is reshaping various industries, and accounting is no exception. With the ability to process vast amounts of data quickly and accurately, AI is revolutionizing how financial professionals manage, analyze, and report financial information. In this article, we will explore the diverse applications of AI in accounting and its profound impact on the field. Automation of Repetitive Tasks: One of the most significant contributions of AI in accounting is automating repetitive tasks. AI-powered software can handle data entry, invoice processing, and reconciliation with minimal human intervention. This not only saves time but also reduces the risk of errors, leading to more accurate financial records. Pattern Recognition and Anomaly Detection: AI algorithms excel at pattern recognition. In accounting, this capability is leveraged to identify unusual patterns in financial data that might indicate fraud or errors. AI can swiftly detect discrepancies, enabling auditors and accountants to focus on resolving issues rather than hunting for them. Real-Time Financial Insights: AI-driven tools, using natural language processing and computer vision, can process documents faster than ever. This enables organizations to have real-time insights into their financial status, empowering decision-makers with up-to-date information for strategic planning. Fraud Detection and Prevention: AI is a powerful tool in the fight against financial fraud. It can analyze vast transaction datasets, flagging suspicious activities and reducing the likelihood of financial misconduct going unnoticed. This proactive approach safeguards a company's financial integrity. Enhanced Data Analysis and Forecasting: Machine learning, a subset of AI, is used for data analysis and forecasting. By examining historical financial data, AI models can provide forecasts and insights, aiding businesses in making informed financial decisions and optimizing their financial strategies. Artificial Intelligence is fundamentally transforming the accounting profession. From automating mundane tasks to enhancing data analysis and fraud detection, AI is making financial processes more efficient, accurate, and insightful. As AI continues to evolve, its role in accounting will only become more significant, offering accountants and finance professionals powerful tools to navigate the complexities of modern finance. Embracing AI in accounting is not just a trend; it's a necessity for staying competitive in the evolving financial landscape.

Keywords: artificial intelligence, accounting automation, financial analysis, fraud detection, machine learning in finance

Procedia PDF Downloads 63
2798 In Search of CO₂: Gravity and Magnetic Data for Eor Prospect Generation in Central Libya

Authors: Ahmed Saheel, Milad Ahmed Elmaradi, Tim Archer, Muammer Ahmed Aboaesha, Abdulkhaliq Abdulmajid Altoubashi

Abstract:

Enhanced oil recovery using carbon dioxide (CO₂-EOR) is a method that can increase oil production beyond what is typically achievable using conventional recovery methods by injecting and hence storing, carbon dioxide (CO₂) in the oil reservoir. In Libya, plans are underway to source a proportion of this CO₂ from subsurface geology that is known from previous drilling to contain high volumes of CO₂. But first, these subsurface volumes need to be more clearly defined and understood. Focusing on the Al-Harouj region of central Libya, ground gravity and airborne magnetic data from the LPI database and the African Magnetic Mapping Project respectively have been prepared and processed by Libyan Petroleum Institute (LPI) and Reid Geophysics Limited (RGL) to produce a range of grids and related products suitable for interpreting geological structure and to make recommendations for subsequent work that will assist CO₂ exploration for purposes of enhanced oil recovery (EOR).

Keywords: gravity anomaly, magnetic anomaly, DEDUCED lineaments, Total horizontal derivative, upward-continuation

Procedia PDF Downloads 126
2797 Viability Analysis of a Centralized Hydrogen Generation Plant for Use in Oil Refining Industry

Authors: C. Fúnez Guerra, B. Nieto Calderón, M. Jaén Caparrós, L. Reyes-Bozo, A. Godoy-Faúndez, E. Vyhmeister

Abstract:

The global energy system is experiencing a change of scenery. Unstable energy markets, an increasing focus on climate change and its sustainable development is forcing businesses to pursue new solutions in order to ensure future economic growth. This has led to the interest in using hydrogen as an energy carrier in transportation and industrial applications. As an energy carrier, hydrogen is accessible and holds a high gravimetric energy density. Abundant in hydrocarbons, hydrogen can play an important role in the shift towards low-emission fossil value chains. By combining hydrogen production by natural gas reforming with carbon capture and storage, the overall CO2 emissions are significantly reduced. In addition, the flexibility of hydrogen as an energy storage makes it applicable as a stabilizer in the renewable energy mix. The recent development in hydrogen fuel cells is also raising the expectations for a hydrogen powered transportation sector. Hydrogen value chains exist to a large extent in the industry today. The global hydrogen consumption was approximately 50 million tonnes (7.2 EJ) in 2013, where refineries, ammonia, methanol production and metal processing were main consumers. Natural gas reforming produced 48% of this hydrogen, but without carbon capture and storage (CCS). The total emissions from the production reached 500 million tonnes of CO2, hence alternative production methods with lower emissions will be necessary in future value chains. Hydrogen from electrolysis is used for a wide range of industrial chemical reactions for many years. Possibly, the earliest use was for the production of ammonia-based fertilisers by Norsk Hydro, with a test reactor set up in Notodden, Norway, in 1927. This application also claims one of the world’s largest electrolyser installations, at Sable Chemicals in Zimbabwe. Its array of 28 electrolysers consumes 80 MW per hour, producing around 21,000 Nm3/h of hydrogen. These electrolysers can compete if cheap sources of electricity are available and natural gas for steam reforming is relatively expensive. Because electrolysis of water produces oxygen as a by-product, a system of Autothermal Reforming (ATR) utilizing this oxygen has been analyzed. Replacing the air separation unit with electrolysers produces the required amount of oxygen to the ATR as well as additional hydrogen. The aim of this paper is to evaluate the technical and economic potential of large-scale production of hydrogen for oil refining industry. Sensitivity analysis of parameters such as investment costs, plant operating hours, electricity price and sale price of hydrogen and oxygen are performed.

Keywords: autothermal reforming, electrolyser, hydrogen, natural gas, steam methane reforming

Procedia PDF Downloads 211
2796 Artificial Intelligence in Bioscience: The Next Frontier

Authors: Parthiban Srinivasan

Abstract:

With recent advances in computational power and access to enough data in biosciences, artificial intelligence methods are increasingly being used in drug discovery research. These methods are essentially a series of advanced statistics based exercises that review the past to indicate the likely future. Our goal is to develop a model that accurately predicts biological activity and toxicity parameters for novel compounds. We have compiled a robust library of over 150,000 chemical compounds with different pharmacological properties from literature and public domain databases. The compounds are stored in simplified molecular-input line-entry system (SMILES), a commonly used text encoding for organic molecules. We utilize an automated process to generate an array of numerical descriptors (features) for each molecule. Redundant and irrelevant descriptors are eliminated iteratively. Our prediction engine is based on a portfolio of machine learning algorithms. We found Random Forest algorithm to be a better choice for this analysis. We captured non-linear relationship in the data and formed a prediction model with reasonable accuracy by averaging across a large number of randomized decision trees. Our next step is to apply deep neural network (DNN) algorithm to predict the biological activity and toxicity properties. We expect the DNN algorithm to give better results and improve the accuracy of the prediction. This presentation will review all these prominent machine learning and deep learning methods, our implementation protocols and discuss these techniques for their usefulness in biomedical and health informatics.

Keywords: deep learning, drug discovery, health informatics, machine learning, toxicity prediction

Procedia PDF Downloads 357
2795 Deterministic and Stochastic Modeling of a Micro-Grid Management for Optimal Power Self-Consumption

Authors: D. Calogine, O. Chau, S. Dotti, O. Ramiarinjanahary, P. Rasoavonjy, F. Tovondahiniriko

Abstract:

Mafate is a natural circus in the north-western part of Reunion Island, without an electrical grid and road network. A micro-grid concept is being experimented in this area, composed of a photovoltaic production combined with electrochemical batteries, in order to meet the local population for self-consumption of electricity demands. This work develops a discrete model as well as a stochastic model in order to reach an optimal equilibrium between production and consumptions for a cluster of houses. The management of the energy power leads to a large linearized programming system, where the time interval of interest is 24 hours The experimental data are solar production, storage energy, and the parameters of the different electrical devices and batteries. The unknown variables to evaluate are the consumptions of the various electrical services, the energy drawn from and stored in the batteries, and the inhabitants’ planning wishes. The objective is to fit the solar production to the electrical consumption of the inhabitants, with an optimal use of the energies in the batteries by satisfying as widely as possible the users' planning requirements. In the discrete model, the different parameters and solutions of the linear programming system are deterministic scalars. Whereas in the stochastic approach, the data parameters and the linear programming solutions become random variables, then the distributions of which could be imposed or established by estimation from samples of real observations or from samples of optimal discrete equilibrium solutions.

Keywords: photovoltaic production, power consumption, battery storage resources, random variables, stochastic modeling, estimations of probability distributions, mixed integer linear programming, smart micro-grid, self-consumption of electricity.

Procedia PDF Downloads 110
2794 The Importance of Visual Communication in Artificial Intelligence

Authors: Manjitsingh Rajput

Abstract:

Visual communication plays an important role in artificial intelligence (AI) because it enables machines to understand and interpret visual information, similar to how humans do. This abstract explores the importance of visual communication in AI and emphasizes the importance of various applications such as computer vision, object emphasis recognition, image classification and autonomous systems. In going deeper, with deep learning techniques and neural networks that modify visual understanding, In addition to AI programming, the abstract discusses challenges facing visual interfaces for AI, such as data scarcity, domain optimization, and interpretability. Visual communication and other approaches, such as natural language processing and speech recognition, have also been explored. Overall, this abstract highlights the critical role that visual communication plays in advancing AI capabilities and enabling machines to perceive and understand the world around them. The abstract also explores the integration of visual communication with other modalities like natural language processing and speech recognition, emphasizing the critical role of visual communication in AI capabilities. This methodology explores the importance of visual communication in AI development and implementation, highlighting its potential to enhance the effectiveness and accessibility of AI systems. It provides a comprehensive approach to integrating visual elements into AI systems, making them more user-friendly and efficient. In conclusion, Visual communication is crucial in AI systems for object recognition, facial analysis, and augmented reality, but challenges like data quality, interpretability, and ethics must be addressed. Visual communication enhances user experience, decision-making, accessibility, and collaboration. Developers can integrate visual elements for efficient and accessible AI systems.

Keywords: visual communication AI, computer vision, visual aid in communication, essence of visual communication.

Procedia PDF Downloads 95
2793 A Hybrid Simulation Approach to Evaluate Cooling Energy Consumption for Public Housings of Subtropics

Authors: Kwok W. Mui, Ling T. Wong, Chi T. Cheung

Abstract:

Cooling energy consumption in the residential sector, different from shopping mall, office or commercial buildings, is significantly subject to occupant decisions where in-depth investigations are found limited. It shows that energy consumptions could be associated with housing types. Surveys have been conducted in existing Hong Kong public housings to understand the housing characteristics, apartment electricity demands, occupant’s thermal expectations, and air–conditioning usage patterns for further cooling energy-saving assessments. The aim of this study is to develop a hybrid cooling energy prediction model, which integrated by EnergyPlus (EP) and artificial neural network (ANN) to estimate cooling energy consumption in public residential sector. Sensitivity tests are conducted to find out the energy impacts with changing building parameters regarding to external wall and window material selection, window size reduction, shading extension, building orientation and apartment size control respectively. Assessments are performed to investigate the relationships between cooling demands and occupant behavior on thermal environment criteria and air-conditioning operation patterns. The results are summarized into a cooling energy calculator for layman use to enhance the cooling energy saving awareness in their own living environment. The findings can be used as a directory framework for future cooling energy evaluation in residential buildings, especially focus on the occupant behavioral air–conditioning operation and criteria of energy-saving incentives.

Keywords: artificial neural network, cooling energy, occupant behavior, residential buildings, thermal environment

Procedia PDF Downloads 168
2792 Don't Just Guess and Slip: Estimating Bayesian Knowledge Tracing Parameters When Observations Are Scant

Authors: Michael Smalenberger

Abstract:

Intelligent tutoring systems (ITS) are computer-based platforms which can incorporate artificial intelligence to provide step-by-step guidance as students practice problem-solving skills. ITS can replicate and even exceed some benefits of one-on-one tutoring, foster transactivity in collaborative environments, and lead to substantial learning gains when used to supplement the instruction of a teacher or when used as the sole method of instruction. A common facet of many ITS is their use of Bayesian Knowledge Tracing (BKT) to estimate parameters necessary for the implementation of the artificial intelligence component, and for the probability of mastery of a knowledge component relevant to the ITS. While various techniques exist to estimate these parameters and probability of mastery, none directly and reliably ask the user to self-assess these. In this study, 111 undergraduate students used an ITS in a college-level introductory statistics course for which detailed transaction-level observations were recorded, and users were also routinely asked direct questions that would lead to such a self-assessment. Comparisons were made between these self-assessed values and those obtained using commonly used estimation techniques. Our findings show that such self-assessments are particularly relevant at the early stages of ITS usage while transaction level data are scant. Once a user’s transaction level data become available after sufficient ITS usage, these can replace the self-assessments in order to eliminate the identifiability problem in BKT. We discuss how these findings are relevant to the number of exercises necessary to lead to mastery of a knowledge component, the associated implications on learning curves, and its relevance to instruction time.

Keywords: Bayesian Knowledge Tracing, Intelligent Tutoring System, in vivo study, parameter estimation

Procedia PDF Downloads 172
2791 Centrality and Patent Impact: Coupled Network Analysis of Artificial Intelligence Patents Based on Co-Cited Scientific Papers

Authors: Xingyu Gao, Qiang Wu, Yuanyuan Liu, Yue Yang

Abstract:

In the era of the knowledge economy, the relationship between scientific knowledge and patents has garnered significant attention. Understanding the intricate interplay between the foundations of science and technological innovation has emerged as a pivotal challenge for both researchers and policymakers. This study establishes a coupled network of artificial intelligence patents based on co-cited scientific papers. Leveraging centrality metrics from network analysis offers a fresh perspective on understanding the influence of information flow and knowledge sharing within the network on patent impact. The study initially obtained patent numbers for 446,890 granted US AI patents from the United States Patent and Trademark Office’s artificial intelligence patent database for the years 2002-2020. Subsequently, specific information regarding these patents was acquired using the Lens patent retrieval platform. Additionally, a search and deduplication process was performed on scientific non-patent references (SNPRs) using the Web of Science database, resulting in the selection of 184,603 patents that cited 37,467 unique SNPRs. Finally, this study constructs a coupled network comprising 59,379 artificial intelligence patents by utilizing scientific papers co-cited in patent backward citations. In this network, nodes represent patents, and if patents reference the same scientific papers, connections are established between them, serving as edges within the network. Nodes and edges collectively constitute the patent coupling network. Structural characteristics such as node degree centrality, betweenness centrality, and closeness centrality are employed to assess the scientific connections between patents, while citation count is utilized as a quantitative metric for patent influence. Finally, a negative binomial model is employed to test the nonlinear relationship between these network structural features and patent influence. The research findings indicate that network structural features such as node degree centrality, betweenness centrality, and closeness centrality exhibit inverted U-shaped relationships with patent influence. Specifically, as these centrality metrics increase, patent influence initially shows an upward trend, but once these features reach a certain threshold, patent influence starts to decline. This discovery suggests that moderate network centrality is beneficial for enhancing patent influence, while excessively high centrality may have a detrimental effect on patent influence. This finding offers crucial insights for policymakers, emphasizing the importance of encouraging moderate knowledge flow and sharing to promote innovation when formulating technology policies. It suggests that in certain situations, data sharing and integration can contribute to innovation. Consequently, policymakers can take measures to promote data-sharing policies, such as open data initiatives, to facilitate the flow of knowledge and the generation of innovation. Additionally, governments and relevant agencies can achieve broader knowledge dissemination by supporting collaborative research projects, adjusting intellectual property policies to enhance flexibility, or nurturing technology entrepreneurship ecosystems.

Keywords: centrality, patent coupling network, patent influence, social network analysis

Procedia PDF Downloads 54
2790 Evaluating the Water Balance of Sokoto Basement Complex to Address Water Security Challenges

Authors: Murtala Gada Abubakar, Aliyu T. Umar

Abstract:

A substantial part of Nigeria is part of semi-arid areas of the world, underlain by basement complex (hard) rocks which are very poor in both transmission and storage of appreciable quantity of water. Recently, a growing attention is being paid on the need to develop water resources in these areas largely due to concerns about increasing droughts and the need to maintain water security challenges. While there is ample body of knowledge that captures the hydrological behaviours of the sedimentary part, reported research which unambiguously illustrates water distribution in the basement complex of the Sokoto basin remains sparse. Considering the growing need to meet the water requirements of those living in this region necessitated the call for accurate water balance estimations that can inform a sustainable planning and development to address water security challenges for the area. To meet this task, a one-dimensional soil water balance model was developed and utilised to assess the state of water distribution within the Sokoto basin basement complex using measured meteorological variables and information about different landscapes within the complex. The model simulated the soil water storage and rates of input and output of water in response to climate and irrigation where applicable using data from 2001 to 2010 inclusive. The results revealed areas within the Sokoto basin basement complex that are rich and deficient in groundwater resource. The high potential areas identified includes the fadama, the fractured rocks and the cultivated lands, while the low potential areas are the sealed surfaces and non-fractured rocks. This study concludes that the modelling approach is a useful tool for assessing the hydrological behaviour and for better understanding the water resource availability within a basement complex.

Keywords: basement complex, hydrological processes, Sokoto Basin, water security

Procedia PDF Downloads 319
2789 A Dam Break Analysis Using MIKE11

Authors: Oussama Derdous, Lakhdar Djemili, Hamza Bouchahed

Abstract:

The consequences of a dam breach can be devastating; both in terms of lives lost and damaged infrastructure and property. Hydraulic modeling provides a clear picture of the possible consequences of partial or complete failure of a dam, which is the key to carry out emergency planning and conduct reliable risk assessments. In this paper, the MIKE11 model developed by the Danish Hydrologic Institute (DHI) was used to simulate the flood wave propagation associated with a potential failure analysis failure of Zardezas dam located in the city of Skikda in the North East of Algeria. MIKE11 results including inundation maps and the representative channel/valley cross-sections depicting flow depth and maximal flow velocities showed that Zardezas reservoir presents a significant risk to downstream areas in the event of a dam failure. These results can be used as the basis of the development of an Emergency Action Plan (EAP).The main objective of this plan is to predict the appropriate steps to avoid or at least decrease the consequences of unexpected failure of Zardezas dam.

Keywords: MIKE11, dam break, inundation maps, emergency action plan

Procedia PDF Downloads 463
2788 Evaluation of National Research Motivation Evolution with Improved Social Influence Network Theory Model: A Case Study of Artificial Intelligence

Authors: Yating Yang, Xue Zhang, Chengli Zhao

Abstract:

In the increasingly interconnected global environment brought about by globalization, it is crucial for countries to timely grasp the development motivations in relevant research fields of other countries and seize development opportunities. Motivation, as the intrinsic driving force behind actions, is abstract in nature, making it difficult to directly measure and evaluate. Drawing on the ideas of social influence network theory, the research motivations of a country can be understood as the driving force behind the development of its science and technology sector, which is simultaneously influenced by both the country itself and other countries/regions. In response to this issue, this paper improves upon Friedkin's social influence network theory and applies it to motivation description, constructing a dynamic alliance network and hostile network centered around the United States and China, as well as a sensitivity matrix, to remotely assess the changes in national research motivations under the influence of international relations. Taking artificial intelligence as a case study, the research reveals that the motivations of most countries/regions are declining, gradually shifting from a neutral attitude to a negative one. The motivation of the United States is hardly influenced by other countries/regions and remains at a high level, while the motivation of China has been consistently increasing in recent years. By comparing the results with real data, it is found that this model can reflect, to some extent, the trends in national motivations.

Keywords: influence network theory, remote assessment, relation matrix, dynamic sensitivity matrix

Procedia PDF Downloads 68
2787 Synthesis of Highly Stable Pseudocapacitors From Secondary Resources

Authors: Samane Maroufi, Rasoul Khayyam Nekouei, Sajjad Mofarah

Abstract:

Fabrication of the state-of-the-art portable pseudocapacitors with the desired transparency, mechanical flexibility, capacitance, and durability is challenging. In most cases, the fabrication of such devices requires critical elements which are either under the crisis of depletion or their extraction from virgin mineral ores have sever environmental impacts. This urges the use of secondary resources instead of virgin resources in fabrication of advanced devices. In this research, ultrathin films of defect-rich Mn1−x−y(CexLay)O2−δ with controllable thicknesses in the range between 5 nm to 627 nm and transmittance (≈29–100%) have been fabricated via an electrochemical chronoamperometric deposition technique using an aqueous precursor derived during the selective purification of rare earth oxide (REOs) isolated from end-of-life nickel-metal hydride (Ni-MH) batteries. Intercalation/de-intercalation of anionic O2− through the atomic tunnels of the stratified Mn1−x−y(CexLay)O2−δ crystallites was found to be responsible for outstanding areal capacitance of 3.4 mF cm−2 of films with 86% transmittance. The intervalence charge transfer among interstitial Ce/La cations and Mn oxidation states within the Mn1−x−y(CexLay)O2−δ structure resulted in excellent capacitance retention of ≈90% after 16 000 cycles. The synthesised transparent flexible Mn1−x−y(CexLay)O2−δ full-cell pseudocapacitor device possessed the energy and power densities of 0.088 μWh cm⁻² and 843 µW cm⁻², respectively. These values show insignificant changes under vigorous twisting and bending to 45–180° confirming these value-added materials are intriguing alternatives for size-sensitive energy storage devices. This research confirms the feasibility of utilisation of secondary waste resources for the fabrication of high-quality pseudocapacitors with engineered defects with the desired flexibility, transparency, and cycling stability suitable for size-sensitive portable electronic devices.

Keywords: pseudocapacitors, energy storage devices, flexible and transparent, sustainability

Procedia PDF Downloads 87
2786 Artificial Insemination of Bali Cattle with Frozen-Thawed Sexed Sperm Under District AI Station Conditions in Lombok: A Preliminary Trial

Authors: Chairussyuhur Arman, Totti Tjiptosumirat, Muhammad Gunawan, Mastur, Joko Priyono, Baiq Tri Ratna Erawati

Abstract:

The present study was undertaken to synchronize oestrus of bali cattle and artificially inseminated with frozen-thawed sexed-semen. The experiment was carried out at District AI Station. Four pluriparous cows and four nulliparous heifers were used in this study and they were housed in free stall barns. The heifers fed with corn silage supplemented with UMMB, while the cows fed with green fodder. All animals were given 500 mg cloprostenolum i.m. injections PGF2α twice, 11 days apart, to synchronize the occurrence of estrus. Estrus was detected by visual observation twice a day and determined if all cattle accepted mount from other females. All animals were inseminated twice with Bali sexed-semen at 72 and 76 h after observed oestrus. Results suggested that the percentage of calving rate either for pluriparous cows or nulliparous heifers were recorded to be 75 percent. One cow and one heifer did not produce calves because of embryonic lost. Regardless the sex of calves, the mean of birth weight of calves in cows was higher than that of heifers (18.50 ± 2.60 kg vs 13.83 ± 5.20 kg). One female calf from heifer with lower birth weight (8.0 kg) was dead one day after born. In pluriparous group, two cows delivered male calves and the other delivered female calf. Conversely in nulliparous group, two heifers delivered female calves and the other male calf. It is concluded that under the conditions of this preliminary trials, the sex ratio between pluriparous and nulliparous groups was found to be 50:50 (male:female).

Keywords: artificial insemination, bali cattle, calves, sexed sperm

Procedia PDF Downloads 311
2785 Lung HRCT Pattern Classification for Cystic Fibrosis Using a Convolutional Neural Network

Authors: Parisa Mansour

Abstract:

Cystic fibrosis (CF) is one of the most common autosomal recessive diseases among whites. It mostly affects the lungs, causing infections and inflammation that account for 90% of deaths in CF patients. Because of this high variability in clinical presentation and organ involvement, investigating treatment responses and evaluating lung changes over time is critical to preventing CF progression. High-resolution computed tomography (HRCT) greatly facilitates the assessment of lung disease progression in CF patients. Recently, artificial intelligence was used to analyze chest CT scans of CF patients. In this paper, we propose a convolutional neural network (CNN) approach to classify CF lung patterns in HRCT images. The proposed network consists of two convolutional layers with 3 × 3 kernels and maximally connected in each layer, followed by two dense layers with 1024 and 10 neurons, respectively. The softmax layer prepares a predicted output probability distribution between classes. This layer has three exits corresponding to the categories of normal (healthy), bronchitis and inflammation. To train and evaluate the network, we constructed a patch-based dataset extracted from more than 1100 lung HRCT slices obtained from 45 CF patients. Comparative evaluation showed the effectiveness of the proposed CNN compared to its close peers. Classification accuracy, average sensitivity and specificity of 93.64%, 93.47% and 96.61% were achieved, indicating the potential of CNNs in analyzing lung CF patterns and monitoring lung health. In addition, the visual features extracted by our proposed method can be useful for automatic measurement and finally evaluation of the severity of CF patterns in lung HRCT images.

Keywords: HRCT, CF, cystic fibrosis, chest CT, artificial intelligence

Procedia PDF Downloads 65
2784 Weakly Solving Kalah Game Using Artificial Intelligence and Game Theory

Authors: Hiba El Assibi

Abstract:

This study aims to weakly solve Kalah, a two-player board game, by developing a start-to-finish winning strategy using an optimized Minimax algorithm with Alpha-Beta Pruning. In weakly solving Kalah, our focus is on creating an optimal strategy from the game's beginning rather than analyzing every possible position. The project will explore additional enhancements like symmetry checking and code optimizations to speed up the decision-making process. This approach is expected to give insights into efficient strategy formulation in board games and potentially help create games with a fair distribution of outcomes. Furthermore, this research provides a unique perspective on human versus Artificial Intelligence decision-making in strategic games. By comparing the AI-generated optimal moves with human choices, we can explore how seemingly advantageous moves can, in the long run, be harmful, thereby offering a deeper understanding of strategic thinking and foresight in games. Moreover, this paper discusses the evaluation of our strategy against existing methods, providing insights on performance and computational efficiency. We also discuss the scalability of our approach to the game, considering different board sizes (number of pits and stones) and rules (different variations) and studying how that affects performance and complexity. The findings have potential implications for the development of AI applications in strategic game planning, enhancing our understanding of human cognitive processes in game settings, and offer insights into creating balanced and engaging game experiences.

Keywords: minimax, alpha beta pruning, transposition tables, weakly solving, game theory

Procedia PDF Downloads 55