Search results for: pneumatic artificial muscles
1164 Smart Speed Bump
Authors: Mohammad Rahmani Rezaiyeh, Mojtaba Rahmani Rezaiyeh, Mehrdad Rahmani Rezaiyeh
Abstract:
Smart speed bump is a new invention and I am invented it. Smart speed bump is a system that can change the position of speed bumps either active or passive in necessary situations. The basic system of smart speed bumps is based on a robotic system which includes mechanic, electronic and artificial intelligence. The smart speed bump is capable of smart decision making and can change its position by anticipating the peak of terrific hours. It can be noted to the advantages of this system such as preventing the waste of petrol while crossing speed bumps, traffic management, accelerating, flowing and securing traffic, reducing accidents and judicial records.Keywords: invention, smart, robotic system, speed bump, traffic, management
Procedia PDF Downloads 4181163 Investigating Water-Oxidation Using a Ru(III) Carboxamide Water Coordinated Complex
Authors: Yosra M. Badiei, Evelyn Ortiz, Marisa Portenti, David Szalda
Abstract:
Water-oxidation half-reaction is a critical reaction that can be driven by a sustainable energy source (e.g., solar or wind) and be coupled with a chemical fuel making reaction which stores the released electrons and protons from water (e.g., H₂ or methanol). The use of molecular water-oxidation catalysts (WOC) allow the rationale design of redox active metal centers and provides a better understanding of their structure-activity-relationship. Herein, the structure of a Ru(III) complex bearing a doubly deprotonated N,N'-bis(aryl)pyridine-2,6-dicarboxamide ligand which contains a water molecule in its primary coordination sphere was elucidated by single-crystal X-ray diffraction. Further spectroscopic experimental data and pH-dependent electrochemical studies reveal its water-oxidation reactivity. Emphasis on mechanistic details for O₂ formation of this complex will be addressed.Keywords: water-oxidation, catalysis, ruthenium, artificial photosynthesis
Procedia PDF Downloads 2011162 Design and Modeling of Human Middle Ear for Harmonic Response Analysis
Authors: Shende Suraj Balu, A. B. Deoghare, K. M. Pandey
Abstract:
The human middle ear (ME) is a delicate and vital organ. It has a complex structure that performs various functions such as receiving sound pressure and producing vibrations of eardrum and propagating it to inner ear. It consists of Tympanic Membrane (TM), three auditory ossicles, various ligament structures and muscles. Incidents such as traumata, infections, ossification of ossicular structures and other pathologies may damage the ME organs. The conditions can be surgically treated by employing prosthesis. However, the suitability of the prosthesis needs to be examined in advance prior to the surgery. Few decades ago, this issue was addressed and analyzed by developing an equivalent representation either in the form of spring mass system, electrical system using R-L-C circuit or developing an approximated CAD model. But, nowadays a three-dimensional ME model can be constructed using micro X-Ray Computed Tomography (μCT) scan data. Moreover, the concern about patient specific integrity pertaining to the disease can be examined well in advance. The current research work emphasizes to develop the ME model from the stacks of μCT images which are used as input file to MIMICS Research 19.0 (Materialise Interactive Medical Image Control System) software. A stack of CT images is converted into geometrical surface model to build accurate morphology of ME. The work is further extended to understand the dynamic behaviour of Harmonic response of the stapes footplate and umbo for different sound pressure levels applied at lateral side of eardrum using finite element approach. The pathological condition Cholesteatoma of ME is investigated to obtain peak to peak displacement of stapes footplate and umbo. Apart from this condition, other pathologies, mainly, changes in the stiffness of stapedial ligament, TM thickness and ossicular chain separation and fixation are also explored. The developed model of ME for pathologies is validated by comparing the results available in the literatures and also with the results of a normal ME to calculate the percentage loss in hearing capability.Keywords: computed tomography (μCT), human middle ear (ME), harmonic response, pathologies, tympanic membrane (TM)
Procedia PDF Downloads 1751161 Leveraging Digital Transformation Initiatives and Artificial Intelligence to Optimize Readiness and Simulate Mission Performance across the Fleet
Authors: Justin Woulfe
Abstract:
Siloed logistics and supply chain management systems throughout the Department of Defense (DOD) has led to disparate approaches to modeling and simulation (M&S), a lack of understanding of how one system impacts the whole, and issues with “optimal” solutions that are good for one organization but have dramatic negative impacts on another. Many different systems have evolved to try to understand and account for uncertainty and try to reduce the consequences of the unknown. As the DoD undertakes expansive digital transformation initiatives, there is an opportunity to fuse and leverage traditionally disparate data into a centrally hosted source of truth. With a streamlined process incorporating machine learning (ML) and artificial intelligence (AI), advanced M&S will enable informed decisions guiding program success via optimized operational readiness and improved mission success. One of the current challenges is to leverage the terabytes of data generated by monitored systems to provide actionable information for all levels of users. The implementation of a cloud-based application analyzing data transactions, learning and predicting future states from current and past states in real-time, and communicating those anticipated states is an appropriate solution for the purposes of reduced latency and improved confidence in decisions. Decisions made from an ML and AI application combined with advanced optimization algorithms will improve the mission success and performance of systems, which will improve the overall cost and effectiveness of any program. The Systecon team constructs and employs model-based simulations, cutting across traditional silos of data, aggregating maintenance, and supply data, incorporating sensor information, and applying optimization and simulation methods to an as-maintained digital twin with the ability to aggregate results across a system’s lifecycle and across logical and operational groupings of systems. This coupling of data throughout the enterprise enables tactical, operational, and strategic decision support, detachable and deployable logistics services, and configuration-based automated distribution of digital technical and product data to enhance supply and logistics operations. As a complete solution, this approach significantly reduces program risk by allowing flexible configuration of data, data relationships, business process workflows, and early test and evaluation, especially budget trade-off analyses. A true capability to tie resources (dollars) to weapon system readiness in alignment with the real-world scenarios a warfighter may experience has been an objective yet to be realized to date. By developing and solidifying an organic capability to directly relate dollars to readiness and to inform the digital twin, the decision-maker is now empowered through valuable insight and traceability. This type of educated decision-making provides an advantage over the adversaries who struggle with maintaining system readiness at an affordable cost. The M&S capability developed allows program managers to independently evaluate system design and support decisions by quantifying their impact on operational availability and operations and support cost resulting in the ability to simultaneously optimize readiness and cost. This will allow the stakeholders to make data-driven decisions when trading cost and readiness throughout the life of the program. Finally, sponsors are available to validate product deliverables with efficiency and much higher accuracy than in previous years.Keywords: artificial intelligence, digital transformation, machine learning, predictive analytics
Procedia PDF Downloads 1601160 Forecasting Residential Water Consumption in Hamilton, New Zealand
Authors: Farnaz Farhangi
Abstract:
Many people in New Zealand believe that the access to water is inexhaustible, and it comes from a history of virtually unrestricted access to it. For the region like Hamilton which is one of New Zealand’s fastest growing cities, it is crucial for policy makers to know about the future water consumption and implementation of rules and regulation such as universal water metering. Hamilton residents use water freely and they do not have any idea about how much water they use. Hence, one of proposed objectives of this research is focusing on forecasting water consumption using different methods. Residential water consumption time series exhibits seasonal and trend variations. Seasonality is the pattern caused by repeating events such as weather conditions in summer and winter, public holidays, etc. The problem with this seasonal fluctuation is that, it dominates other time series components and makes difficulties in determining other variations (such as educational campaign’s effect, regulation, etc.) in time series. Apart from seasonality, a stochastic trend is also combined with seasonality and makes different effects on results of forecasting. According to the forecasting literature, preprocessing (de-trending and de-seasonalization) is essential to have more performed forecasting results, while some other researchers mention that seasonally non-adjusted data should be used. Hence, I answer the question that is pre-processing essential? A wide range of forecasting methods exists with different pros and cons. In this research, I apply double seasonal ARIMA and Artificial Neural Network (ANN), considering diverse elements such as seasonality and calendar effects (public and school holidays) and combine their results to find the best predicted values. My hypothesis is the examination the results of combined method (hybrid model) and individual methods and comparing the accuracy and robustness. In order to use ARIMA, the data should be stationary. Also, ANN has successful forecasting applications in terms of forecasting seasonal and trend time series. Using a hybrid model is a way to improve the accuracy of the methods. Due to the fact that water demand is dominated by different seasonality, in order to find their sensitivity to weather conditions or calendar effects or other seasonal patterns, I combine different methods. The advantage of this combination is reduction of errors by averaging of each individual model. It is also useful when we are not sure about the accuracy of each forecasting model and it can ease the problem of model selection. Using daily residential water consumption data from January 2000 to July 2015 in Hamilton, I indicate how prediction by different methods varies. ANN has more accurate forecasting results than other method and preprocessing is essential when we use seasonal time series. Using hybrid model reduces forecasting average errors and increases the performance.Keywords: artificial neural network (ANN), double seasonal ARIMA, forecasting, hybrid model
Procedia PDF Downloads 3371159 Gravitational Water Vortex Power Plant: Experimental-Parametric Design of a Hydraulic Structure Capable of Inducing the Artificial Formation of a Gravitational Water Vortex Appropriate for Hydroelectric Generation
Authors: Henrry Vicente Rojas Asuero, Holger Manuel Benavides Muñoz
Abstract:
Approximately 80% of the energy consumed worldwide is generated from fossil sources, which are responsible for the emission of a large volume of greenhouse gases. For this reason, the global trend, at present, is the widespread use of energy produced from renewable sources. This seeks safety and diversification of energy supply, based on social cohesion, economic feasibility and environmental protection. In this scenario, small hydropower systems (P ≤ 10MW) stand out due to their high efficiency, economic competitiveness and low environmental impact. Small hydropower systems, along with wind and solar energy, are expected to represent a significant percentage of the world's energy matrix in the near term. Among the various technologies present in the state of the art, relating to small hydropower systems, is the Gravitational Water Vortex Power Plant, a recent technology that excels because of its versatility of operation, since it can operate with jumps in the range of 0.70 m-2.00 m and flow rates from 1 m3/s to 20 m3/s. Its operating system is based on the utilization of the energy of rotation contained within a large water vortex artificially induced. This paper presents the study and experimental design of an optimal hydraulic structure with the capacity to induce the artificial formation of a gravitational water vortex trough a system of easy application and high efficiency, able to operate in conditions of very low head and minimum flow. The proposed structure consists of a channel, with variable base, vortex inductor, tangential flow generator, coupled to a circular tank with a conical transition bottom hole. In the laboratory test, the angular velocity of the water vortex was related to the geometric characteristics of the inductor channel, as well as the influence of the conical transition bottom hole on the physical characteristics of the water vortex. The results show angular velocity values of greater magnitude as a function of depth, in addition the presence of the conical transition in the bottom hole of the circular tank improves the water vortex formation conditions while increasing the angular velocity values. Thus, the proposed system is a sustainable solution for the energy supply of rural areas near to watercourses.Keywords: experimental model, gravitational water vortex power plant, renewable energy, small hydropower
Procedia PDF Downloads 2901158 Aging Behaviour of 6061 Al-15 vol% SiC Composite in T4 and T6 Treatments
Authors: Melby Chacko, Jagannath Nayak
Abstract:
The aging behaviour of 6061 Al-15 vol% SiC composite was investigated using Rockwell B hardness measurement. The composite was solutionized at 350°C and quenched in water. The composite was aged at room temperature (T4 treatment) and also at 140°C, 160°C, 180°C and 200°C (T6 treatment). The natural and artificial aging behaviour of composite was studied using aging curves determined at different temperatures. The aging period for peak aging for different temperatures was identified. The time required for attaining peak aging decreased with increase in the aging temperature. The peak hardness was found to increase with increase with aging temperature and the highest peak hardness was observed at 180ºC. Beyond 180ºC the peak hardness was found to be decreasing.Keywords: 6061 Al-SiC composite, aging curve, Rockwell B hardness, T4, T6 treatments
Procedia PDF Downloads 2671157 Developing Wearable EMG Sensor Designed for Parkinson's Disease (PD) Monitoring, and Treatment
Authors: Bulcha Belay Etana
Abstract:
Electromyography is used to measure the electrical activity of muscles for various health monitoring applications using surface electrodes or needle electrodes. Recent developments in electromyogram signal acquisition using textile electrodes open the door for wearable health monitoring which enables patients to monitor and control their health issues outside of traditional healthcare facilities. The aim of this research is therefore to develop and analyze wearable textile electrodes for the acquisition of electromyography signals for Parkinson’s patients and apply an appropriate thermal stimulus to relieve muscle cramping. In order to achieve this, textile electrodes are sewn with a silver-coated thread in an overlapping zigzag pattern into an inextensible fabric, and stainless steel knitted textile electrodes attached to a sleeve were prepared and its electrical characteristics including signal to noise ratio were compared with traditional electrodes. To relieve muscle cramping, a heating element using stainless steel conductive yarn Sewn onto a cotton fabric, coupled with a vibration system were developed. The system was integrated using a microcontroller and a Myoware muscle sensor so that when muscle cramping occurs, measured by the system activates the heating elements and vibration motors. The optimum temperature considered for treatment was 35.50c, so a Temperature measurement system was incorporated to deactivate the heating system when the temperature reaches this threshold, and the signals indicating muscle cramping have subsided. The textile electrode exhibited a signal to noise ratio of 6.38dB while the signal to noise ratio of the traditional electrode was 7.05dB. The rise time of the developed heating element was about 6 minutes to reach the optimum temperature using a 9volt power supply. The treatment of muscle cramping in Parkinson's patients using heat and muscle vibration simultaneously with a wearable electromyography signal acquisition system will improve patients’ livelihoods and enable better chronic pain management.Keywords: electromyography, heating textile, vibration therapy, parkinson’s disease, wearable electronic textile
Procedia PDF Downloads 1351156 The Application of Artificial Neural Network for Bridge Structures Design Optimization
Authors: Angga S. Fajar, A. Aminullah, J. Kiyono, R. A. Safitri
Abstract:
This paper discusses about the application of ANN for optimizing of bridge structure design. ANN has been applied in various field of science concerning prediction and optimization. The structural optimization has several benefit including accelerate structural design process, saving the structural material, and minimize self-weight and mass of structure. In this paper, there are three types of bridge structure that being optimized including PSC I-girder superstructure, composite steel-concrete girder superstructure, and RC bridge pier. The different optimization strategy on each bridge structure implement back propagation method of ANN is conducted in this research. The optimal weight and easier design process of bridge structure with satisfied error are achieved.Keywords: bridge structures, ANN, optimization, back propagation
Procedia PDF Downloads 3721155 Development of Evolutionary Algorithm by Combining Optimization and Imitation Approach for Machine Learning in Gaming
Authors: Rohit Mittal, Bright Keswani, Amit Mithal
Abstract:
This paper provides a sense about the application of computational intelligence techniques used to develop computer games, especially car racing. For the deep sense and knowledge of artificial intelligence, this paper is divided into various sections that is optimization, imitation, innovation and combining approach of optimization and imitation. This paper is mainly concerned with combining approach which tells different aspects of using fitness measures and supervised learning techniques used to imitate aspects of behavior. The main achievement of this paper is based on modelling player behaviour and evolving new game content such as racing tracks as single car racing on single track.Keywords: evolution algorithm, genetic, optimization, imitation, racing, innovation, gaming
Procedia PDF Downloads 6461154 Challenges beyond the Singapore Future-Ready School ‘LEADER’ Qualities
Authors: Zoe Boon Suan Loy
Abstract:
An exploratory research undertaken in 2000 at the beginning of the COVID-19 pandemic examined the changing roles of Singapore school leaders as they lead teachers in developing future-ready learners. While it is evident that ‘LEADER’ qualities epitomize the knowledge, competencies, and skills required, recent events in an increasing VUCA and BANI world characterized by massively disruptive Ukraine -Russian war, unabating tense US-Sino relations, issues related to sustainability, and rapid ageing will have an impact on school leadership. As an increasingly complex endeavour, this requires a relook as they lead teachers in nurturing holistically-developed future-ready students. Digitalisation, new technology, and the push for a green economy will be the key driving forces that will have an impact on job availability. Similarly, the rapid growth of artificial intelligence (AI) capabilities, including ChatGPT, will aggravate and add tremendous stress to the work of school leaders. This paper seeks to explore the key school leadership shifts required beyond the ‘LEADER’ qualities as school leaders respond to the changes, challenges, and opportunities in the 21st C new normal. The research findings for this paper are based on an exploratory qualitative study on the perceptions of 26 school leaders (vice-principals) who were attending a milestone educational leadership course at the National Institute of Education, Nanyang Technological University, Singapore. A structured questionnaire is designed to collect the data, which is then analysed using coding methodology. Broad themes on key competencies and skills of future-ready leaders in the Singapore education system are then identified. Key Findings: In undertaking their leadership roles as leaders of future-ready learners, school leaders need to demonstrate the ‘LEADER’ qualities. They need to have a long-term view, understand the educational imperatives, have a good awareness of self and the dispositions of a leader, be effective in optimizing external leverages and are clear about their role expectations. These ‘LEADER’ qualities are necessary and relevant in the post-Covid era. Beyond this, school leaders with ‘LEADER’ qualities are well supported by the Ministry of Education, which takes cognizance of emerging trends and continually review education policies to address related issues. Concluding Statement: Discussions within the education ecosystem and among other stakeholders on the implications of the use of artificial intelligence and ChatGPT on the school curriculum, including content knowledge, pedagogy, and assessment, are ongoing. This augurs well for school leaders as they undertake their responsibilities as leaders of future-ready learners.Keywords: Singapore education system, ‘LEADER’ qualities, school leadership, future-ready leaders, future-ready learners
Procedia PDF Downloads 721153 Role of Artificial Intelligence in Nano Proteomics
Authors: Mehrnaz Mostafavi
Abstract:
Recent advances in single-molecule protein identification (ID) and quantification techniques are poised to revolutionize proteomics, enabling researchers to delve into single-cell proteomics and identify low-abundance proteins crucial for biomedical and clinical research. This paper introduces a different approach to single-molecule protein ID and quantification using tri-color amino acid tags and a plasmonic nanopore device. A comprehensive simulator incorporating various physical phenomena was designed to predict and model the device's behavior under diverse experimental conditions, providing insights into its feasibility and limitations. The study employs a whole-proteome single-molecule identification algorithm based on convolutional neural networks, achieving high accuracies (>90%), particularly in challenging conditions (95–97%). To address potential challenges in clinical samples, where post-translational modifications affecting labeling efficiency, the paper evaluates protein identification accuracy under partial labeling conditions. Solid-state nanopores, capable of processing tens of individual proteins per second, are explored as a platform for this method. Unlike techniques relying solely on ion-current measurements, this approach enables parallel readout using high-density nanopore arrays and multi-pixel single-photon sensors. Convolutional neural networks contribute to the method's versatility and robustness, simplifying calibration procedures and potentially allowing protein ID based on partial reads. The study also discusses the efficacy of the approach in real experimental conditions, resolving functionally similar proteins. The theoretical analysis, protein labeler program, finite difference time domain calculation of plasmonic fields, and simulation of nanopore-based optical sensing are detailed in the methods section. The study anticipates further exploration of temporal distributions of protein translocation dwell-times and the impact on convolutional neural network identification accuracy. Overall, the research presents a promising avenue for advancing single-molecule protein identification and quantification with broad applications in proteomics research. The contributions made in methodology, accuracy, robustness, and technological exploration collectively position this work at the forefront of transformative developments in the field.Keywords: nano proteomics, nanopore-based optical sensing, deep learning, artificial intelligence
Procedia PDF Downloads 951152 Design of a Lumbar Interspinous Process Fixation Device for Minimizing Soft Tissue Removal and Operation Time
Authors: Minhyuk Heo, Jihwan Yun, Seonghun Park
Abstract:
It has been reported that intervertebral fusion surgery, which removes most of the ligaments and muscles of the spine, increases the degenerative disease in adjacent spinal segments. Therefore, it is required to develop a lumbar interspinous process fixation device that minimizes the risks and side effects from the surgery. The objective of the current study is to design an interspinous process fixation device with simple structures in order to minimize soft tissue removal and operation time during intervertebral fusion surgery. For the design concepts of a lumbar fixation device, the principle of the ratchet was first applied on the joining parts of the device in order to shorten the operation time. The coil spring structure was selected for connecting parts between the spinous processes so that a normal range of motion in spinal segments is preserved and degenerative spinal diseases are not developed in the adjacent spinal segments. The stiffness of the spring was determined not to interrupt the motion of a lumbar spine. The designed value of the spring stiffness allows the upper part of the spring to move ~10° which is higher than the range of flexion and extension for normal lumbar spine (6°-8°), when a moment of 10Nm is applied on the upper face of L1. A finite element (FE) model composed of L1 to L5 lumbar spines was generated to verify the mechanical integrity and the dynamic stability of the designed lumbar fixation device and to further optimize the lumbar fixation device. The FE model generated above produced the same pressure value on intervertebral disc and dynamic behavior as the normal intact model reported in the literature. The consistent results from this comparison validates the accuracy in the modeling of the current FE model. Currently, we are trying to generate an abnormal model with defects in one or more components of the normal FE model above. Then, the mechanical integrity and the dynamic stability of the designed lumbar fixation device will be analyzed after being installed in the abnormal model and then the lumbar fixation device will be further optimized.Keywords: lumbar interspinous process fixation device, finite element method, lumbar spine, kinematics
Procedia PDF Downloads 2281151 Portable, Noninvasive and Wireless Near Infrared Spectroscopy Device to Monitor Skeletal Muscle Metabolism during Exercise
Authors: Adkham Paiziev, Fikrat Kerimov
Abstract:
Near Infrared Spectroscopy (NIRS) is one of the biophotonic techniques which can be used to monitor oxygenation and hemodynamics in a variety of human tissues, including skeletal muscle. In the present work, we are offering tissue oximetry (OxyPrem) to measure hemodynamic parameters of skeletal muscles in rest and exercise. Purpose: - To elaborate the new wireless, portable, noninvasive, wearable NIRS device to measure skeletal muscle oxygenation during exercise. - To test this device on brachioradialis muscle of wrestler volunteers by using combined method of arterial occlusion (AO) and NIRS (AO+NIRS). Methods: Oxyprem NIRS device has been used together with AO test. AO test and Isometric brachioradialis muscle contraction experiments have been performed on one group of wrestler volunteers. ‘Accu- Measure’ caliper (USA) to measure skinfold thickness (SFT) has been used. Results: Elaborated device consists on power supply box, a sensor head and installed ‘Tubis’ software for data acquisition and to compute deoxyhemoglobin ([HHb), oxyhemoglobin ([O2Hb]), tissue oxygenation (StO2) and muscle tissue oxygen consumption (mVO2). Sensor head consists on four light sources with three light emitting diodes with nominal wavelengths of 760 nm, 805 nm, and 870 nm, and two detectors. AO and isometric voluntary forearm muscle contraction (IVFMC) on five healthy male subjects (23,2±0.84 in age, 0.43±0.05cm of SFT ) and four female subjects (22.0±1.0 in age and 0.24±0.04 cm SFT) has been measured. mVO2 for control group has been calculated (-0.65%/sec±0.07) for male and -0.69%/±0.19 for female subjects). Tissue oxygenation index for wrestlers in average about 75% whereas for control group StO2 =63%. Second experiment was connected with quality monitoring muscle activity during IVFMC at 10%,30% and 50% of MVC. It has been shown, that the concentration changes of HbO2 and HHb positively correlated to the contraction intensity. Conclusion: We have presented a portable multi-channel wireless NIRS device for real-time monitoring of muscle activity. The miniaturized NIRS sensor and the usage of wireless communication make the whole device have a compact-size, thus can be used in muscle monitoring.Keywords: skeletal muscle, oxygenation, instrumentation, near infrared spectroscopy
Procedia PDF Downloads 2751150 Life Prediction of Condenser Tubes Applying Fuzzy Logic and Neural Network Algorithms
Authors: A. Majidian
Abstract:
The life prediction of thermal power plant components is necessary to prevent the unexpected outages, optimize maintenance tasks in periodic overhauls and plan inspection tasks with their schedules. One of the main critical components in a power plant is condenser because its failure can affect many other components which are positioned in downstream of condenser. This paper deals with factors affecting life of condenser. Failure rates dependency vs. these factors has been investigated using Artificial Neural Network (ANN) and fuzzy logic algorithms. These algorithms have shown their capabilities as dynamic tools to evaluate life prediction of power plant equipments.Keywords: life prediction, condenser tube, neural network, fuzzy logic
Procedia PDF Downloads 3511149 Forecasting Solid Waste Generation in Turkey
Authors: Yeliz Ekinci, Melis Koyuncu
Abstract:
Successful planning of solid waste management systems requires successful prediction of the amount of solid waste generated in an area. Waste management planning can protect the environment and human health, hence it is tremendously important for countries. The lack of information in waste generation can cause many environmental and health problems. Turkey is a country that plans to join European Union, hence, solid waste management is one of the most significant criteria that should be handled in order to be a part of this community. Solid waste management system requires a good forecast of solid waste generation. Thus, this study aims to forecast solid waste generation in Turkey. Artificial Neural Network and Linear Regression models will be used for this aim. Many models will be run and the best one will be selected based on some predetermined performance measures.Keywords: forecast, solid waste generation, solid waste management, Turkey
Procedia PDF Downloads 5071148 Cryptographic Protocol for Secure Cloud Storage
Authors: Luvisa Kusuma, Panji Yudha Prakasa
Abstract:
Cloud storage, as a subservice of infrastructure as a service (IaaS) in Cloud Computing, is the model of nerworked storage where data can be stored in server. In this paper, we propose a secure cloud storage system consisting of two main components; client as a user who uses the cloud storage service and server who provides the cloud storage service. In this system, we propose the protocol schemes to guarantee against security attacks in the data transmission. The protocols are login protocol, upload data protocol, download protocol, and push data protocol, which implement hybrid cryptographic mechanism based on data encryption before it is sent to the cloud, so cloud storage provider does not know the user's data and cannot analysis user’s data, because there is no correspondence between data and user.Keywords: cloud storage, security, cryptographic protocol, artificial intelligence
Procedia PDF Downloads 3571147 Artificial Intelligence Based Meme Generation Technology for Engaging Audience in Social Media
Authors: Andrew Kurochkin, Kostiantyn Bokhan
Abstract:
In this study, a new meme dataset of ~650K meme instances was created, a technology of meme generation based on the state of the art deep learning technique - GPT-2 model was researched, a comparative analysis of machine-generated memes and human-created was conducted. We justified that Amazon Mechanical Turk workers can be used for the approximate estimating of users' behavior in a social network, more precisely to measure engagement. It was shown that generated memes cause the same engagement as human memes that produced low engagement in the social network (historically). Thus, generated memes are less engaging than random memes created by humans.Keywords: content generation, computational social science, memes generation, Reddit, social networks, social media interaction
Procedia PDF Downloads 1381146 Dinitrotoluene and Trinitrotoluene Measuring in Double-Base Solid Propellants
Authors: Z. H. Safari, M. Anbia, G. H. Kouzegari, R. Amirkhani
Abstract:
Toluene and Nitro derivatives are widely used in industry particularly in various defense applications. Tri-nitro-toluene derivative is a powerful basic explosive material that is a basis upon which to compare equivalent explosive power of similar materials. The aim of this paper is to measure the explosive power of these hazardous substances in fuels having different shelf-life and therefore optimizing their storage and maintenance. The methodology involves measuring the amounts of di- nitro- toluene and tri-nitro-toluene in the aged samples at 90 ° C by gas chromatography. Results show no significant difference in the concentration of the TNT compound over a given time while there was a significant difference in DNT compound over the same period. The underlying reason is attributed to the simultaneous production of the material with destruction of stabilizer.Keywords: dinitrotoluene, trinitrotoluene, double-base solid propellants, artificial aging
Procedia PDF Downloads 4031145 Enhancing AI for Global Impact: Conversations on Improvement and Societal Benefits
Authors: C. P. Chukwuka, E. V. Chukwuka, F. Ukwadi
Abstract:
This paper focuses on the advancement and societal impact of artificial intelligence (AI) systems. It explores the need for a theoretical framework in corporate governance, specifically in the context of 'hybrid' companies that have a mix of private and government ownership. The paper emphasizes the potential of AI to address challenges faced by these companies and highlights the importance of the less-explored state model in corporate governance. The aim of this research is to enhance AI systems for global impact and positive societal outcomes. It aims to explore the role of AI in refining corporate governance in hybrid companies and uncover nuanced insights into complex ownership structures. The methodology involves leveraging the capabilities of AI to address the challenges faced by hybrid companies in corporate governance. The researchers will analyze existing theoretical frameworks in corporate governance and integrate AI systems to improve problem-solving and understanding of intricate systems. The paper suggests that improved AI systems have the potential to shape a more informed and responsible corporate landscape. AI can uncover nuanced insights and navigate complex ownership structures in hybrid companies, leading to greater efficacy and positive societal outcomes. The theoretical importance of this research lies in the exploration of the role of AI in corporate governance, particularly in the context of hybrid companies. By integrating AI systems, the paper highlights the potential for improved problem-solving and understanding of intricate systems, contributing to a more informed and responsible corporate landscape. The data for this research will be collected from existing literature on corporate governance, specifically focusing on hybrid companies. Additionally, data on AI capabilities and their application in corporate governance will be collected. The collected data will be analyzed through a systematic review of existing theoretical frameworks in corporate governance. The researchers will also analyze the capabilities of AI systems and their potential application in addressing the challenges faced by hybrid companies. The findings will be synthesized and compared to identify patterns and potential improvements. The research concludes that AI systems have the potential to enhance corporate governance in hybrid companies, leading to greater efficacy and positive societal outcomes. By leveraging AI capabilities, nuanced insights can be uncovered, and complex ownership structures can be navigated, shaping a more informed and responsible corporate landscape. The findings highlight the importance of integrating AI in refining problem-solving and understanding intricate systems for global impact.Keywords: advancement, artificial intelligence, challenges, societal impact
Procedia PDF Downloads 561144 Improvement Image Summarization using Image Processing and Particle swarm optimization Algorithm
Authors: Hooman Torabifard
Abstract:
In the last few years, with the progress of technology and computers and artificial intelligence entry into all kinds of scientific and industrial fields, the lifestyles of human life have changed and in general, the way of humans live on earth has many changes and development. Until now, some of the changes has occurred in the context of digital images and image processing and still continues. However, besides all the benefits, there have been disadvantages. One of these disadvantages is the multiplicity of images with high volume and data; the focus of this paper is on improving and developing a method for summarizing and enhancing the productivity of these images. The general method used for this purpose in this paper consists of a set of methods based on data obtained from image processing and using the PSO (Particle swarm optimization) algorithm. In the remainder of this paper, the method used is elaborated in detail.Keywords: image summarization, particle swarm optimization, image threshold, image processing
Procedia PDF Downloads 1331143 Machine Learning Techniques for Estimating Ground Motion Parameters
Authors: Farid Khosravikia, Patricia Clayton
Abstract:
The main objective of this study is to evaluate the advantages and disadvantages of various machine learning techniques in forecasting ground-motion intensity measures given source characteristics, source-to-site distance, and local site condition. Intensity measures such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Estimating these variables for future earthquake events is a key step in seismic hazard assessment and potentially subsequent risk assessment of different types of structures. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as a statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The algorithms are adjusted to quantify event-to-event and site-to-site variability of the ground motions by implementing them as random effects in the proposed models to reduce the aleatory uncertainty. All the algorithms are trained using a selected database of 4,528 ground-motions, including 376 seismic events with magnitude 3 to 5.8, recorded over the hypocentral distance range of 4 to 500 km in Oklahoma, Kansas, and Texas since 2005. The main reason of the considered database stems from the recent increase in the seismicity rate of these states attributed to petroleum production and wastewater disposal activities, which necessities further investigation in the ground motion models developed for these states. Accuracy of the models in predicting intensity measures, generalization capability of the models for future data, as well as usability of the models are discussed in the evaluation process. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available.Keywords: artificial neural network, ground-motion models, machine learning, random forest, support vector machine
Procedia PDF Downloads 1221142 Efficacy of Botulinum Toxin in Alleviating Pain Syndrome in Stroke Patients with Upper Limb Spasticity
Authors: Akulov M. A., Zaharov V. O., Jurishhev P. E., Tomskij A. A.
Abstract:
Introduction: Spasticity is a severe consequence of stroke, leading to profound disability, decreased quality of life and decrease of rehabilitation efficacy [4]. Spasticity is often associated with pain syndrome, arising from joint damage of paretic limbs (postural arthropathy) or painful spasm of paretic limb muscles. It is generally accepted that injection of botulinum toxin into a cramped muscle leads to decrease of muscle tone and improves motion range in paretic limb, which is accompanied by pain alleviation. Study aim: To evaluate the change in pain syndrome intensity after incections of botulinum toxin A (Xeomin) in stroke patients with upper limb spasticity. Patients and methods. 21 patients aged 47-74 years were evaluated. Inclusion criteria were: acute stroke 4-7 months before the inclusion into the study, leading to spasticity of wrist and/or finger flexors, elbow flexor or forearm pronator, associated with severe pain syndrome. Patients received Xeomin as monotherapy 90-300 U, according to spasticity pattern. Efficacy evaluation was performed using Ashworth scale, disability assessment scale (DAS), caregiver burden scale and global treatment benefit assessment on weeks 2, 4, 8 and 12. Efficacy criterion was the decrease of pain syndrome by week 4 on PQLS and VAS. Results: The study revealed a significant improvement of measured indices after 4 weeks of treatment, which persisted until the 12 week of treatment. Xeomin is effective in reducing muscle tone of flexors of wrist, fingers and elbow, forearm pronators. By the 4th week of treatment we observed a significant improvement on DAS (р < 0,05), Ashworth scale (1-2 points) in all patients (р < 0,05), caregiver burden scale (р < 0,05). A significant decrease of pain syndrome by the 4th week of treatment on PQLS (р < 0,05) и VAS (р < 0,05) was observed. No adverse effect were registered. Conclusion: Xeomin is an effective treatment of pain syndrome in postural upper limb spasticity after stroke. Xeomin treatment leads to a significant improvement on PQLS and VAS.Keywords: botulinum toxin, pain syndrome, spasticity, stroke
Procedia PDF Downloads 3091141 Optimal Injected Current Control for Shunt Active Power Filter Using Artificial Intelligence
Authors: Brahim Berbaoui
Abstract:
In this paper, a new particle swarm optimization (PSO) based method is proposed for the implantation of optimal harmonic power flow in power systems. In this algorithm approach, proportional integral controller for reference compensating currents of active power filter is performed in order to minimize the total harmonic distortion (THD). The simulation results show that the new control method using PSO approach is not only easy to be implanted, but also very effective in reducing the unwanted harmonics and compensating reactive power. The studies carried out have been accomplished using the MATLAB Simulink Power System Toolbox.Keywords: shunt active power filter, power quality, current control, proportional integral controller, particle swarm optimization
Procedia PDF Downloads 6161140 The Impact of Artificial Intelligence on Human Rights Development
Authors: Romany Wagih Farag Zaky
Abstract:
The relationship between development and human rights has long been the subject of academic debate. To understand the dynamics between these two concepts, various principles are adopted, from the right to development to development-based human rights. Despite the initiatives taken, the relationship between development and human rights remains unclear. However, the overlap between these two views and the idea that efforts should be made in the field of human rights have increased in recent years. It is then evaluated whether the right to sustainable development is acceptable or not. The article concludes that the principles of sustainable development are directly or indirectly recognized in various human rights instruments, which is a good answer to the question posed above. This book therefore cites regional and international human rights agreements such as , as well as the jurisprudence and interpretative guidelines of human rights institutions, to prove this hypothesis.Keywords: sustainable development, human rights, the right to development, the human rights-based approach to development, environmental rights, economic development, social sustainability human rights protection, human rights violations, workers’ rights, justice, security
Procedia PDF Downloads 551139 Review on Rainfall Prediction Using Machine Learning Technique
Authors: Prachi Desai, Ankita Gandhi, Mitali Acharya
Abstract:
Rainfall forecast is mainly used for predictions of rainfall in a specified area and determining their future rainfall conditions. Rainfall is always a global issue as it affects all major aspects of one's life. Agricultural, fisheries, forestry, tourism industry and other industries are widely affected by these conditions. The studies have resulted in insufficient availability of water resources and an increase in water demand in the near future. We already have a new forecast system that uses the deep Convolutional Neural Network (CNN) to forecast monthly rainfall and climate changes. We have also compared CNN against Artificial Neural Networks (ANN). Machine Learning techniques that are used in rainfall predictions include ARIMA Model, ANN, LR, SVM etc. The dataset on which we are experimenting is gathered online over the year 1901 to 20118. Test results have suggested more realistic improvements than conventional rainfall forecasts.Keywords: ANN, CNN, supervised learning, machine learning, deep learning
Procedia PDF Downloads 2011138 Automatic Algorithm for Processing and Analysis of Images from the Comet Assay
Authors: Yeimy L. Quintana, Juan G. Zuluaga, Sandra S. Arango
Abstract:
The comet assay is a method based on electrophoresis that is used to measure DNA damage in cells and has shown important results in the identification of substances with a potential risk to the human population as innumerable physical, chemical and biological agents. With this technique is possible to obtain images like a comet, in which the tail of these refers to damaged fragments of the DNA. One of the main problems is that the image has unequal luminosity caused by the fluorescence microscope and requires different processing to condition it as well as to know how many optimal comets there are per sample and finally to perform the measurements and determine the percentage of DNA damage. In this paper, we propose the design and implementation of software using Image Processing Toolbox-MATLAB that allows the automation of image processing. The software chooses the optimum comets and measuring the necessary parameters to detect the damage.Keywords: artificial vision, comet assay, DNA damage, image processing
Procedia PDF Downloads 3101137 Innovations in the Lithium Chain Value
Authors: Fiúza A., Góis J. Leite M., Braga H., Lima A., Jorge P., Moutela P., Martins L., Futuro A.
Abstract:
Lepidolite is an important lithium mineral that, to the author’s best knowledge, has not been used to produce lithium hydroxide, necessary for energy conversion to electric vehicles. Alkaline leaching of lithium concentrates allows the establishment of a production diagram avoiding most of the environmental drawbacks that are associated with the usage of acid reagents. The tested processes involve a pretreatment by digestion at high temperatures with additives, followed by leaching at hot atmospheric pressure. The solutions obtained must be compatible with solutions from the leaching of spodumene concentrates, allowing the development of a common treatment diagram, an important accomplishment for the feasible exploitation of Portuguese resources. Statistical programming and interpretation techniques are used to minimize the laboratory effort required by conventional approaches and also allow phenomenological comprehension.Keywords: artificial intelligence, tailings free process, ferroelectric electrolyte battery, life cycle assessment
Procedia PDF Downloads 1221136 Comparative Study between Classical P-Q Method and Modern Fuzzy Controller Method to Improve the Power Quality of an Electrical Network
Authors: A. Morsli, A. Tlemçani, N. Ould Cherchali, M. S. Boucherit
Abstract:
This article presents two methods for the compensation of harmonics generated by a nonlinear load. The first is the classic method P-Q. The second is the controller by modern method of artificial intelligence specifically fuzzy logic. Both methods are applied to an Active Power Filter shunt (APFs) based on a three-phase voltage converter at five levels NPC topology. In calculating the harmonic currents of reference, we use the algorithm P-Q and pulse generation, we use the intersective PWM. For flexibility and dynamics, we use fuzzy logic. The results give us clear that the rate of Harmonic Distortion issued by fuzzy logic is better than P-Q.Keywords: fuzzy logic controller, P-Q method, pulse width modulation (PWM), shunt active power filter (sAPF), total harmonic distortion (THD)
Procedia PDF Downloads 5481135 Evotrader: Bitcoin Trading Using Evolutionary Algorithms on Technical Analysis and Social Sentiment Data
Authors: Martin Pellon Consunji
Abstract:
Due to the rise in popularity of Bitcoin and other crypto assets as a store of wealth and speculative investment, there is an ever-growing demand for automated trading tools, such as bots, in order to gain an advantage over the market. Traditionally, trading in the stock market was done by professionals with years of training who understood patterns and exploited market opportunities in order to gain a profit. However, nowadays a larger portion of market participants are at minimum aided by market-data processing bots, which can generally generate more stable signals than the average human trader. The rise in trading bot usage can be accredited to the inherent advantages that bots have over humans in terms of processing large amounts of data, lack of emotions of fear or greed, and predicting market prices using past data and artificial intelligence, hence a growing number of approaches have been brought forward to tackle this task. However, the general limitation of these approaches can still be broken down to the fact that limited historical data doesn’t always determine the future, and that a lot of market participants are still human emotion-driven traders. Moreover, developing markets such as those of the cryptocurrency space have even less historical data to interpret than most other well-established markets. Due to this, some human traders have gone back to the tried-and-tested traditional technical analysis tools for exploiting market patterns and simplifying the broader spectrum of data that is involved in making market predictions. This paper proposes a method which uses neuro evolution techniques on both sentimental data and, the more traditionally human-consumed, technical analysis data in order to gain a more accurate forecast of future market behavior and account for the way both automated bots and human traders affect the market prices of Bitcoin and other cryptocurrencies. This study’s approach uses evolutionary algorithms to automatically develop increasingly improved populations of bots which, by using the latest inflows of market analysis and sentimental data, evolve to efficiently predict future market price movements. The effectiveness of the approach is validated by testing the system in a simulated historical trading scenario, a real Bitcoin market live trading scenario, and testing its robustness in other cryptocurrency and stock market scenarios. Experimental results during a 30-day period show that this method outperformed the buy and hold strategy by over 260% in terms of net profits, even when taking into consideration standard trading fees.Keywords: neuro-evolution, Bitcoin, trading bots, artificial neural networks, technical analysis, evolutionary algorithms
Procedia PDF Downloads 123