Search results for: random early detection algorithm
10372 Electrochemical Detection of Polycyclic Aromatic Hydrocarbons in Urban Air by Exfoliated Graphite Based Electrode
Authors: A. Sacko, H. Nyoni, T. A. M. Msagati, B. Ntsendwana
Abstract:
Carbon based materials to target environmental pollutants have become increasingly recognized in science. Electrochemical methods using carbon based materials are notable methods for high sensitive detection of organic pollutants in air. It is therefore in this light that exfoliated graphite electrode was fabricated for electrochemical analysis of PAHs in urban atmospheric air. The electrochemical properties of the graphite electrode were studied using CV and EIS in the presence of acetate buffer supporting electrolyte with 2 Mm ferricyanide as a redox probe. The graphite electrode showed enhanced current response which confirms facile kinetics and enhanced sensitivity. However, the peak to peak (DE) separation increased as a function of scan rate. The EIS showed a high charger transfer resistance. The detection phenanthrene on the exfoliated graphite was studied in the presence of acetate buffer solution at PH 3.5 using DPV. The oxidation peak of phenanthrene was observed at 0.4 V. Under optimized conditions (supporting electrolyte, pH, deposition time, etc.). The detection limit observed was at 5x 10⁻⁸ M. Thus the results demonstrate with further optimization and modification lower concentration detection can be achieved.Keywords: electrochemical detection, exfoliated graphite, PAHs (polycyclic aromatic hydrocarbons), urban air
Procedia PDF Downloads 20210371 Genetic Algorithm Optimization of the Economical, Ecological and Self-Consumption Impact of the Energy Production of a Single Building
Authors: Ludovic Favre, Thibaut M. Schafer, Jean-Luc Robyr, Elena-Lavinia Niederhäuser
Abstract:
This paper presents an optimization method based on genetic algorithm for the energy management inside buildings developed in the frame of the project Smart Living Lab (SLL) in Fribourg (Switzerland). This algorithm optimizes the interaction between renewable energy production, storage systems and energy consumers. In comparison with standard algorithms, the innovative aspect of this project is the extension of the smart regulation over three simultaneous criteria: the energy self-consumption, the decrease of greenhouse gas emissions and operating costs. The genetic algorithm approach was chosen due to the large quantity of optimization variables and the non-linearity of the optimization function. The optimization process includes also real time data of the building as well as weather forecast and users habits. This information is used by a physical model of the building energy resources to predict the future energy production and needs, to select the best energetic strategy, to combine production or storage of energy in order to guarantee the demand of electrical and thermal energy. The principle of operation of the algorithm as well as typical output example of the algorithm is presented.Keywords: building's energy, control system, energy management, energy storage, genetic optimization algorithm, greenhouse gases, modelling, renewable energy
Procedia PDF Downloads 25510370 Curve Fitting by Cubic Bezier Curves Using Migrating Birds Optimization Algorithm
Authors: Mitat Uysal
Abstract:
A new met heuristic optimization algorithm called as Migrating Birds Optimization is used for curve fitting by rational cubic Bezier Curves. This requires solving a complicated multivariate optimization problem. In this study, the solution of this optimization problem is achieved by Migrating Birds Optimization algorithm that is a powerful met heuristic nature-inspired algorithm well appropriate for optimization. The results of this study show that the proposed method performs very well and being able to fit the data points to cubic Bezier Curves with a high degree of accuracy.Keywords: algorithms, Bezier curves, heuristic optimization, migrating birds optimization
Procedia PDF Downloads 33410369 Facile Synthesis of CuO Nanosheets on Cu Foil for H2O2 Detection
Authors: Yu-Kuei Hsu, Yan-Gu Lin
Abstract:
A facile and simple fabrication of copper(II) oxide (CuO) nanosheet on copper foil as nanoelectrode for H2O2 sensing application was proposed in this study. The spontaneous formation of CuO nanosheets by immersing the copper foil into 0.1 M NaOH aqueous solution for 48 hrs was carried out at room temperature. The sheet-like morphology with several ten nanometers in thickness and ~500 nm in width was observed by SEM. Those nanosheets were confirmed the monoclinic-phase CuO by the structural analysis of XRD and Raman spectra. The directly grown CuO nanosheets film is mechanically stable and offers an excellent electrochemical sensing platform. The CuO nanosheets electrode shows excellent electrocatalytic response to H2O2 with significantly lower overpotentials for its oxidation and reduction and also exhibits a fast response and high sensitivity for the amperometric detection of H2O2. The novel spontaneously grown CuO nanosheets electrode is readily applicable to other analytes and has great potential applications in the electrochemical detection.Keywords: CuO, nanosheets, H2O2 detection, Cu foil
Procedia PDF Downloads 28810368 Analysis of the Social Problems of the Early Adolescents in Northeast China
Authors: Zhidong Zhang, Zhi-Chao Zhang, Georgianna Duarte
Abstract:
The social problems of early adolescents in Northeast China were examined with the instrument of Achenbach System of Empirically Based Assessment (ASEBA). In this study, the data consisted of 2532 early adolescents. The relevant variables such as sports activities, hobbies, chores and the number of close friends, as independent variables have been included in this study. The stratified sampling method was used to collect data from 2532 participants. The analysis results indicated that sports activities, hobbies, chores and the number of close friends, as predictors can be used in a predictive model, which significantly predict the social problem T-score.Keywords: social problems, ASEBA, early adolescents, predictive Model
Procedia PDF Downloads 34710367 Hybrid Intelligent Optimization Methods for Optimal Design of Horizontal-Axis Wind Turbine Blades
Authors: E. Tandis, E. Assareh
Abstract:
Designing the optimal shape of MW wind turbine blades is provided in a number of cases through evolutionary algorithms associated with mathematical modeling (Blade Element Momentum Theory). Evolutionary algorithms, among the optimization methods, enjoy many advantages, particularly in stability. However, they usually need a large number of function evaluations. Since there are a large number of local extremes, the optimization method has to find the global extreme accurately. The present paper introduces a new population-based hybrid algorithm called Genetic-Based Bees Algorithm (GBBA). This algorithm is meant to design the optimal shape for MW wind turbine blades. The current method employs crossover and neighborhood searching operators taken from the respective Genetic Algorithm (GA) and Bees Algorithm (BA) to provide a method with good performance in accuracy and speed convergence. Different blade designs, twenty-one to be exact, were considered based on the chord length, twist angle and tip speed ratio using GA results. They were compared with BA and GBBA optimum design results targeting the power coefficient and solidity. The results suggest that the final shape, obtained by the proposed hybrid algorithm, performs better compared to either BA or GA. Furthermore, the accuracy and speed convergence increases when the GBBA is employedKeywords: Blade Design, Optimization, Genetic Algorithm, Bees Algorithm, Genetic-Based Bees Algorithm, Large Wind Turbine
Procedia PDF Downloads 31610366 Saudi Teachers’ Perceptions of Rough and Tumble Play in Early Learning
Authors: Rana Alghamdi
Abstract:
This study explored teachers’ perceptions of rough-and-tumble (R&T) play in early childhood education in Saudi Arabia. The literature on rough-and-tumble play in Saudi Arabia is limited in scope, and more research is needed to explore teachers’ perceptions on this type of play for early learners. The pertinent literature reveals that R&T play, which includes running, jumping, fighting, wrestling, chasing, pulling, pushing, and climbing, among other rough playful activities, can positively impact learning and development across psychosocial, emotional, and cognitive domains. Teachers’ understanding of R & T play is key, and the attitudes of Saudi early childhood teachers who are responsible for implementing curriculum-based play have not been fully researched. Four early childhood teachers from an urban Saudi preschool participated in the study. The data collected in this study were interpreted through a sociocultural lens. Data sources included in-depth interviews, photo-elicitation interviews, and participant-generated drawings. Three overarching themes emerged: teachers’ concerns about rough-and-tumble play, teachers’ perceptions about the benefits of rough-and-tumble play, and teachers’ expression of gender roles in R & T play as contextualized within Saudi culture. Saudi teachers’ perceptions are discussed in detail, and implications of the findings and recommendations for future research are put forth.Keywords: rough and tumble play, gender, culture, early childhood, Saudi Arabia
Procedia PDF Downloads 13010365 Walmart Sales Forecasting using Machine Learning in Python
Authors: Niyati Sharma, Om Anand, Sanjeev Kumar Prasad
Abstract:
Assuming future sale value for any of the organizations is one of the major essential characteristics of tactical development. Walmart Sales Forecasting is the finest illustration to work with as a beginner; subsequently, it has the major retail data set. Walmart uses this sales estimate problem for hiring purposes also. We would like to analyzing how the internal and external effects of one of the largest companies in the US can walk out their Weekly Sales in the future. Demand forecasting is the planned prerequisite of products or services in the imminent on the basis of present and previous data and different stages of the market. Since all associations is facing the anonymous future and we do not distinguish in the future good demand. Hence, through exploring former statistics and recent market statistics, we envisage the forthcoming claim and building of individual goods, which are extra challenging in the near future. As a result of this, we are producing the required products in pursuance of the petition of the souk in advance. We will be using several machine learning models to test the exactness and then lastly, train the whole data by Using linear regression and fitting the training data into it. Accuracy is 8.88%. The extra trees regression model gives the best accuracy of 97.15%.Keywords: random forest algorithm, linear regression algorithm, extra trees classifier, mean absolute error
Procedia PDF Downloads 14810364 Keypoint Detection Method Based on Multi-Scale Feature Fusion of Attention Mechanism
Authors: Xiaoxiao Li, Shuangcheng Jia, Qian Li
Abstract:
Keypoint detection has always been a challenge in the field of image recognition. This paper proposes a novelty keypoint detection method which is called Multi-Scale Feature Fusion Convolutional Network with Attention (MFFCNA). We verified that the multi-scale features with the attention mechanism module have better feature expression capability. The feature fusion between different scales makes the information that the network model can express more abundant, and the network is easier to converge. On our self-made street sign corner dataset, we validate the MFFCNA model with an accuracy of 97.8% and a recall of 81%, which are 5 and 8 percentage points higher than the HRNet network, respectively. On the COCO dataset, the AP is 71.9%, and the AR is 75.3%, which are 3 points and 2 points higher than HRNet, respectively. Extensive experiments show that our method has a remarkable improvement in the keypoint recognition tasks, and the recognition effect is better than the existing methods. Moreover, our method can be applied not only to keypoint detection but also to image classification and semantic segmentation with good generality.Keywords: keypoint detection, feature fusion, attention, semantic segmentation
Procedia PDF Downloads 11710363 [Keynote]: No-Trust-Zone Architecture for Securing Supervisory Control and Data Acquisition
Authors: Michael Okeke, Andrew Blyth
Abstract:
Supervisory Control And Data Acquisition (SCADA) as the state of the art Industrial Control Systems (ICS) are used in many different critical infrastructures, from smart home to energy systems and from locomotives train system to planes. Security of SCADA systems is vital since many lives depend on it for daily activities and deviation from normal operation could be disastrous to the environment as well as lives. This paper describes how No-Trust-Zone (NTZ) architecture could be incorporated into SCADA Systems in order to reduce the chances of malicious intent. The architecture is made up of two distinctive parts which are; the field devices such as; sensors, PLCs pumps, and actuators. The second part of the architecture is designed following lambda architecture, which is made up of a detection algorithm based on Particle Swarm Optimization (PSO) and Hadoop framework for data processing and storage. Apache Spark will be a part of the lambda architecture for real-time analysis of packets for anomalies detection.Keywords: industrial control system (ics, no-trust-zone (ntz), particle swarm optimisation (pso), supervisory control and data acquisition (scada), swarm intelligence (SI)
Procedia PDF Downloads 34310362 An Inviscid Compressible Flow Solver Based on Unstructured OpenFOAM Mesh Format
Authors: Utkan Caliskan
Abstract:
Two types of numerical codes based on finite volume method are developed in order to solve compressible Euler equations to simulate the flow through forward facing step channel. Both algorithms have AUSM+- up (Advection Upstream Splitting Method) scheme for flux splitting and two-stage Runge-Kutta scheme for time stepping. In this study, the flux calculations differentiate between the algorithm based on OpenFOAM mesh format which is called 'face-based' algorithm and the basic algorithm which is called 'element-based' algorithm. The face-based algorithm avoids redundant flux computations and also is more flexible with hybrid grids. Moreover, some of OpenFOAM’s preprocessing utilities can be used on the mesh. Parallelization of the face based algorithm for which atomic operations are needed due to the shared memory model, is also presented. For several mesh sizes, 2.13x speed up is obtained with face-based approach over the element-based approach.Keywords: cell centered finite volume method, compressible Euler equations, OpenFOAM mesh format, OpenMP
Procedia PDF Downloads 31810361 Plagiarism Detection for Flowchart and Figures in Texts
Authors: Ahmadu Maidorawa, Idrissa Djibo, Muhammad Tella
Abstract:
This paper presents a method for detecting flow chart and figure plagiarism based on shape of image processing and multimedia retrieval. The method managed to retrieve flowcharts with ranked similarity according to different matching sets. Plagiarism detection is well known phenomenon in the academic arena. Copying other people is considered as serious offense that needs to be checked. There are many plagiarism detection systems such as turn-it-in that has been developed to provide these checks. Most, if not all, discard the figures and charts before checking for plagiarism. Discarding the figures and charts result in look holes that people can take advantage. That means people can plagiarize figures and charts easily without the current plagiarism systems detecting it. There are very few papers which talks about flowcharts plagiarism detection. Therefore, there is a need to develop a system that will detect plagiarism in figures and charts.Keywords: flowchart, multimedia retrieval, figures similarity, image comparison, figure retrieval
Procedia PDF Downloads 46310360 Role of Molecular Changes and Immunohistochemical in Early Detection of Liver Cancer
Authors: Fatimah A. Alhomaid
Abstract:
The present study was planned to investigate the role of molecular changes and immunohistochemical in early detection of liver cancer in Saudi patients. our results were carried out on 54 patients liver cancer. We obtained our data from laboratory in King Khalid University Hospital. The specimens were taken (54) patients with liver cancer 34 male and 14 female and 2 control. The average age of varied from 37-85 years. The tumor was diagnosed as grade I in tow patients (male and female) and grade 2 in 45 patients (28 male and 17 female) while the grade 3 in 4 patients (all males). The specimens were processed for haematoxylin and eosin staining, immunohistochemical technique and flow cytometry analysis. Our study noted that most patients had adenocarcinoma which characterized by presence of signet-ring cells were very clear in advanced patients with adenocarcinoma. Our sections in adenocarcinoma in grade 2 and stage 3 had an increase in signet ring cells,an increase in the acini of glands and an increase in number of lymphocytes which spread to the muscular layer. With advancing the disease, there were haemorrhage in blood and increase in lymphocytes and increase in the number of nuclei in the tubular glands. Our study was carried on 48 patients, immunohistochemical diagnosis (CK20, PCNA, P53) and the analysis of DNA content by flow cytometry technique. Our study indicated that the presence of correlation between the immunohistochemical analysis for P53 and the grades. The reaction of P53 appeared as strong in nucleus in grades &stage 3 and appeared in other sections as dark brown pigment. Our study indicated that the absence of correlation between the immunohistochemical analysis for PCAN and the grades. In our sections there were strong reaction in the more 80% of nuclei in grade 1& stage 2. Our study indicated that the presence of correlation between the immunohistochemical analysis for CK20 and the grades. Our results indicated the presence of positive reaction in cytoplasm varied from weak to moderate in grade 3 & stage 4. Concerning the Flow cytometry technique our results indicated that the presence of correlation between the DNA and different stages of liver cancer.Keywords: cancer, CK20, DNA, cytometry analysis, liver, immunohistochemical, molecular changes, PCNA, p53
Procedia PDF Downloads 26410359 Investigation of Utilizing L-Band Horn Antenna in Landmine Detection
Authors: Ahmad H. Abdelgwad, Ahmed A. Nashat
Abstract:
Landmine detection is an important and yet challenging problem remains to be solved. Ground Penetrating Radar (GPR) is a powerful and rapidly maturing technology for subsurface threat identification. The detection methodology of GPR depends mainly on the contrast of the dielectric properties of the searched target and its surrounding soil. This contrast produces a partial reflection of the electromagnetic pulses that are being transmitted into the soil and then being collected by the GPR. One of the most critical hardware components for the performance of GPR is the antenna system. The current paper explores the design and simulation of a pyramidal horn antenna operating at L-band frequencies (1- 2 GHz) to detect a landmine. A prototype model of the GPR system setup is developed to simulate full wave analysis of the electromagnetic fields in different soil types. The contrast in the dielectric permittivity of the landmine and the sandy soil is the most important parameter to be considered for detecting the presence of landmine. L-band horn antenna is proved to be well-versed in the investigation of landmine detection.Keywords: full wave analysis, ground penetrating radar, horn antenna design, landmine detection
Procedia PDF Downloads 21710358 Heart Rate Variability Analysis for Early Stage Prediction of Sudden Cardiac Death
Authors: Reeta Devi, Hitender Kumar Tyagi, Dinesh Kumar
Abstract:
In present scenario, cardiovascular problems are growing challenge for researchers and physiologists. As heart disease have no geographic, gender or socioeconomic specific reasons; detecting cardiac irregularities at early stage followed by quick and correct treatment is very important. Electrocardiogram is the finest tool for continuous monitoring of heart activity. Heart rate variability (HRV) is used to measure naturally occurring oscillations between consecutive cardiac cycles. Analysis of this variability is carried out using time domain, frequency domain and non-linear parameters. This paper presents HRV analysis of the online dataset for normal sinus rhythm (taken as healthy subject) and sudden cardiac death (SCD subject) using all three methods computing values for parameters like standard deviation of node to node intervals (SDNN), square root of mean of the sequences of difference between adjacent RR intervals (RMSSD), mean of R to R intervals (mean RR) in time domain, very low-frequency (VLF), low-frequency (LF), high frequency (HF) and ratio of low to high frequency (LF/HF ratio) in frequency domain and Poincare plot for non linear analysis. To differentiate HRV of healthy subject from subject died with SCD, k –nearest neighbor (k-NN) classifier has been used because of its high accuracy. Results show highly reduced values for all stated parameters for SCD subjects as compared to healthy ones. As the dataset used for SCD patients is recording of their ECG signal one hour prior to their death, it is therefore, verified with an accuracy of 95% that proposed algorithm can identify mortality risk of a patient one hour before its death. The identification of a patient’s mortality risk at such an early stage may prevent him/her meeting sudden death if in-time and right treatment is given by the doctor.Keywords: early stage prediction, heart rate variability, linear and non-linear analysis, sudden cardiac death
Procedia PDF Downloads 33810357 Exploring Ways Early Childhood Teachers Integrate Information and Communication Technologies into Children's Play: Two Case Studies from the Australian Context
Authors: Caroline Labib
Abstract:
This paper reports on a qualitative study exploring the approaches teachers used to integrate computers or smart tablets into their program planning. Their aim was to integrate ICT into children’s play, thereby supporting children’s learning and development. Data was collected in preschool settings in Melbourne in 2016. Interviews with teachers, observations of teacher interactions with children and copies of teachers’ planning and observation documents informed the study. The paper looks closely at findings from two early childhood settings and focuses on exploring the differing approaches two EC teachers have adopted when integrating iPad or computers into their settings. Data analysis revealed three key approaches which have been labelled: free digital play, guided digital play and teacher-led digital use. Importantly, teacher decisions were influenced by the interplay between the opportunities that the ICT tools offered, the teachers’ prior knowledge and experience about ICT and children’s learning needs and contexts. This paper is a snapshot of two early childhood settings, and further research will encompass data from six more early childhood settings in Victoria with the aim of exploring a wide range of motivating factors for early childhood teachers trying to integrate ICT into their programs.Keywords: early childhood education (ECE), digital play, information and communication technologies (ICT), play, and teachers' interaction approaches
Procedia PDF Downloads 21010356 Automatic Content Curation of Visual Heritage
Authors: Delphine Ribes Lemay, Valentine Bernasconi, André Andrade, Lara DéFayes, Mathieu Salzmann, FréDéRic Kaplan, Nicolas Henchoz
Abstract:
Digitization and preservation of large heritage induce high maintenance costs to keep up with the technical standards and ensure sustainable access. Creating impactful usage is instrumental to justify the resources for long-term preservation. The Museum für Gestaltung of Zurich holds one of the biggest poster collections of the world from which 52’000 were digitised. In the process of building a digital installation to valorize the collection, one objective was to develop an algorithm capable of predicting the next poster to show according to the ones already displayed. The work presented here describes the steps to build an algorithm able to automatically create sequences of posters reflecting associations performed by curator and professional designers. The exposed challenge finds similarities with the domain of song playlist algorithms. Recently, artificial intelligence techniques and more specifically, deep-learning algorithms have been used to facilitate their generations. Promising results were found thanks to Recurrent Neural Networks (RNN) trained on manually generated playlist and paired with clusters of extracted features from songs. We used the same principles to create the proposed algorithm but applied to a challenging medium, posters. First, a convolutional autoencoder was trained to extract features of the posters. The 52’000 digital posters were used as a training set. Poster features were then clustered. Next, an RNN learned to predict the next cluster according to the previous ones. RNN training set was composed of poster sequences extracted from a collection of books from the Gestaltung Museum of Zurich dedicated to displaying posters. Finally, within the predicted cluster, the poster with the best proximity compared to the previous poster is selected. The mean square distance between features of posters was used to compute the proximity. To validate the predictive model, we compared sequences of 15 posters produced by our model to randomly and manually generated sequences. Manual sequences were created by a professional graphic designer. We asked 21 participants working as professional graphic designers to sort the sequences from the one with the strongest graphic line to the one with the weakest and to motivate their answer with a short description. The sequences produced by the designer were ranked first 60%, second 25% and third 15% of the time. The sequences produced by our predictive model were ranked first 25%, second 45% and third 30% of the time. The sequences produced randomly were ranked first 15%, second 29%, and third 55% of the time. Compared to designer sequences, and as reported by participants, model and random sequences lacked thematic continuity. According to the results, the proposed model is able to generate better poster sequencing compared to random sampling. Eventually, our algorithm is sometimes able to outperform a professional designer. As a next step, the proposed algorithm should include a possibility to create sequences according to a selected theme. To conclude, this work shows the potentiality of artificial intelligence techniques to learn from existing content and provide a tool to curate large sets of data, with a permanent renewal of the presented content.Keywords: Artificial Intelligence, Digital Humanities, serendipity, design research
Procedia PDF Downloads 18310355 Accurate Algorithm for Selecting Ground Motions Satisfying Code Criteria
Authors: S. J. Ha, S. J. Baik, T. O. Kim, S. W. Han
Abstract:
For computing the seismic responses of structures, current seismic design provisions permit response history analyses (RHA) that can be used without limitations in height, seismic design category, and building irregularity. In order to obtain accurate seismic responses using RHA, it is important to use adequate input ground motions. Current seismic design provisions provide criteria for selecting ground motions. In this study, the accurate and computationally efficient algorithm is proposed for accurately selecting ground motions that satisfy the requirements specified in current seismic design provisions. The accuracy of the proposed algorithm is verified using single-degree-of-freedom systems with various natural periods and yield strengths. This study shows that the mean seismic responses obtained from RHA with seven and ten ground motions selected using the proposed algorithm produce errors within 20% and 13%, respectively.Keywords: algorithm, ground motion, response history analysis, selection
Procedia PDF Downloads 28510354 Dynamic Log Parsing and Intelligent Anomaly Detection Method Combining Retrieval Augmented Generation and Prompt Engineering
Authors: Liu Linxin
Abstract:
As system complexity increases, log parsing and anomaly detection become more and more important in ensuring system stability. However, traditional methods often face the problems of insufficient adaptability and decreasing accuracy when dealing with rapidly changing log contents and unknown domains. To this end, this paper proposes an approach LogRAG, which combines RAG (Retrieval Augmented Generation) technology with Prompt Engineering for Large Language Models, applied to log analysis tasks to achieve dynamic parsing of logs and intelligent anomaly detection. By combining real-time information retrieval and prompt optimisation, this study significantly improves the adaptive capability of log analysis and the interpretability of results. Experimental results show that the method performs well on several public datasets, especially in the absence of training data, and significantly outperforms traditional methods. This paper provides a technical path for log parsing and anomaly detection, demonstrating significant theoretical value and application potential.Keywords: log parsing, anomaly detection, retrieval-augmented generation, prompt engineering, LLMs
Procedia PDF Downloads 2710353 Video Text Information Detection and Localization in Lecture Videos Using Moments
Authors: Belkacem Soundes, Guezouli Larbi
Abstract:
This paper presents a robust and accurate method for text detection and localization over lecture videos. Frame regions are classified into text or background based on visual feature analysis. However, lecture video shows significant degradation mainly related to acquisition conditions, camera motion and environmental changes resulting in low quality videos. Hence, affecting feature extraction and description efficiency. Moreover, traditional text detection methods cannot be directly applied to lecture videos. Therefore, robust feature extraction methods dedicated to this specific video genre are required for robust and accurate text detection and extraction. Method consists of a three-step process: Slide region detection and segmentation; Feature extraction and non-text filtering. For robust and effective features extraction moment functions are used. Two distinct types of moments are used: orthogonal and non-orthogonal. For orthogonal Zernike Moments, both Pseudo Zernike moments are used, whereas for non-orthogonal ones Hu moments are used. Expressivity and description efficiency are given and discussed. Proposed approach shows that in general, orthogonal moments show high accuracy in comparison to the non-orthogonal one. Pseudo Zernike moments are more effective than Zernike with better computation time.Keywords: text detection, text localization, lecture videos, pseudo zernike moments
Procedia PDF Downloads 14910352 The Whale Optimization Algorithm and Its Implementation in MATLAB
Authors: S. Adhirai, R. P. Mahapatra, Paramjit Singh
Abstract:
Optimization is an important tool in making decisions and in analysing physical systems. In mathematical terms, an optimization problem is the problem of finding the best solution from among the set of all feasible solutions. The paper discusses the Whale Optimization Algorithm (WOA), and its applications in different fields. The algorithm is tested using MATLAB because of its unique and powerful features. The benchmark functions used in WOA algorithm are grouped as: unimodal (F1-F7), multimodal (F8-F13), and fixed-dimension multimodal (F14-F23). Out of these benchmark functions, we show the experimental results for F7, F11, and F19 for different number of iterations. The search space and objective space for the selected function are drawn, and finally, the best solution as well as the best optimal value of the objective function found by WOA is presented. The algorithmic results demonstrate that the WOA performs better than the state-of-the-art meta-heuristic and conventional algorithms.Keywords: optimization, optimal value, objective function, optimization problems, meta-heuristic optimization algorithms, Whale Optimization Algorithm, implementation, MATLAB
Procedia PDF Downloads 36910351 Wireless Battery Charger with Adaptive Rapid-Charging Algorithm
Authors: Byoung-Hee Lee
Abstract:
Wireless battery charger with adaptive rapid charging algorithm is proposed. The proposed wireless charger adopts voltage regulation technique to reduce the number of power conversion steps. Moreover, based on battery models, an adaptive rapid charging algorithm for Li-ion batteries is obtained. Rapid-charging performance with the proposed wireless battery charger and the proposed rapid charging algorithm has been experimentally verified to show more than 70% charging time reduction compared to conventional constant-current constant-voltage (CC-CV) charging without the degradation of battery lifetime.Keywords: wireless, battery charger, adaptive, rapid-charging
Procedia PDF Downloads 37410350 Predicting Loss of Containment in Surface Pipeline using Computational Fluid Dynamics and Supervised Machine Learning Model to Improve Process Safety in Oil and Gas Operations
Authors: Muhammmad Riandhy Anindika Yudhy, Harry Patria, Ramadhani Santoso
Abstract:
Loss of containment is the primary hazard that process safety management is concerned within the oil and gas industry. Escalation to more serious consequences all begins with the loss of containment, starting with oil and gas release from leakage or spillage from primary containment resulting in pool fire, jet fire and even explosion when reacted with various ignition sources in the operations. Therefore, the heart of process safety management is avoiding loss of containment and mitigating its impact through the implementation of safeguards. The most effective safeguard for the case is an early detection system to alert Operations to take action prior to a potential case of loss of containment. The detection system value increases when applied to a long surface pipeline that is naturally difficult to monitor at all times and is exposed to multiple causes of loss of containment, from natural corrosion to illegal tapping. Based on prior researches and studies, detecting loss of containment accurately in the surface pipeline is difficult. The trade-off between cost-effectiveness and high accuracy has been the main issue when selecting the traditional detection method. The current best-performing method, Real-Time Transient Model (RTTM), requires analysis of closely positioned pressure, flow and temperature (PVT) points in the pipeline to be accurate. Having multiple adjacent PVT sensors along the pipeline is expensive, hence generally not a viable alternative from an economic standpoint.A conceptual approach to combine mathematical modeling using computational fluid dynamics and a supervised machine learning model has shown promising results to predict leakage in the pipeline. Mathematical modeling is used to generate simulation data where this data is used to train the leak detection and localization models. Mathematical models and simulation software have also been shown to provide comparable results with experimental data with very high levels of accuracy. While the supervised machine learning model requires a large training dataset for the development of accurate models, mathematical modeling has been shown to be able to generate the required datasets to justify the application of data analytics for the development of model-based leak detection systems for petroleum pipelines. This paper presents a review of key leak detection strategies for oil and gas pipelines, with a specific focus on crude oil applications, and presents the opportunities for the use of data analytics tools and mathematical modeling for the development of robust real-time leak detection and localization system for surface pipelines. A case study is also presented.Keywords: pipeline, leakage, detection, AI
Procedia PDF Downloads 18910349 Detection of Abnormal Process Behavior in Copper Solvent Extraction by Principal Component Analysis
Authors: Kirill Filianin, Satu-Pia Reinikainen, Tuomo Sainio
Abstract:
Frequent measurements of product steam quality create a data overload that becomes more and more difficult to handle. In the current study, plant history data with multiple variables was successfully treated by principal component analysis to detect abnormal process behavior, particularly, in copper solvent extraction. The multivariate model is based on the concentration levels of main process metals recorded by the industrial on-stream x-ray fluorescence analyzer. After mean-centering and normalization of concentration data set, two-dimensional multivariate model under principal component analysis algorithm was constructed. Normal operating conditions were defined through control limits that were assigned to squared score values on x-axis and to residual values on y-axis. 80 percent of the data set were taken as the training set and the multivariate model was tested with the remaining 20 percent of data. Model testing showed successful application of control limits to detect abnormal behavior of copper solvent extraction process as early warnings. Compared to the conventional techniques of analyzing one variable at a time, the proposed model allows to detect on-line a process failure using information from all process variables simultaneously. Complex industrial equipment combined with advanced mathematical tools may be used for on-line monitoring both of process streams’ composition and final product quality. Defining normal operating conditions of the process supports reliable decision making in a process control room. Thus, industrial x-ray fluorescence analyzers equipped with integrated data processing toolbox allows more flexibility in copper plant operation. The additional multivariate process control and monitoring procedures are recommended to apply separately for the major components and for the impurities. Principal component analysis may be utilized not only in control of major elements’ content in process streams, but also for continuous monitoring of plant feed. The proposed approach has a potential in on-line instrumentation providing fast, robust and cheap application with automation abilities.Keywords: abnormal process behavior, failure detection, principal component analysis, solvent extraction
Procedia PDF Downloads 30710348 Evaluating and Examining Pictures of Children of Five Years Old
Authors: Emine Yılmaz Bolat
Abstract:
Early childhood is a very important period in terms of identifying and developing early skills and abilities. It is likely that the child's development will be in the same direction in the future. This study was conducted with 26 children for the purpose of examining pictures of children of five years old. In the survey, children were asked to draw a picture with pastel dyes. The drawings were collected and evaluated by the researcher. At the end of the research, it was found that the children used the yellow color (N = 17, 16,34%) and the least gray color (N = 1, 0,96%). When the features of children's pictures are examined, the children's paintings have been found to have hierarchy, transparency, completion, the use of vivid colors, and the presence of vertical and horizontal painting lines.Keywords: early childhood, kindergarten, pictures of children, features of pictures
Procedia PDF Downloads 30510347 Early Outcomes and Lessons from the Implementation of a Geriatric Hip Fracture Protocol at a Level 1 Trauma Center
Authors: Peter Park, Alfonso Ayala, Douglas Saeks, Jordan Miller, Carmen Flores, Karen Nelson
Abstract:
Introduction Hip fractures account for more than 300,000 hospital admissions every year. Many present as fragility fractures in geriatric patients with multiple medical comorbidities. Standardized protocols for the multidisciplinary management of this patient population have been shown to improve patient outcomes. A hip fracture protocol was implemented at a Level I Trauma center with a focus on pre-operative medical optimization and early surgical care. This study evaluates the efficacy of that protocol, including the early transition period. Methods A retrospective review was performed of all patients ages 60 and older with isolated hip fractures who were managed surgically between 2020 and 2022. This included patients 1 year prior and 1 year following the implementation of a hip fracture protocol at a Level I Trauma center. Results 530 patients were identified: 249 patients were treated before, and 281 patients were treated after the protocol was instituted. There was no difference in mean age (p=0.35), gender (p=0.3), or Charlson Comorbidity Index (p=0.38) between the cohorts. Following the implementation of the protocol, there were observed increases in time to surgery (27.5h vs. 33.8h, p=0.01), hospital length of stay (6.3d vs. 9.7d, p<0.001), and ED LOS (5.1h vs. 6.2h, p<0.001). There were no differences in in-hospital mortality (2.01% pre vs. 3.20% post, p=0.39) and complication rates (25% pre vs 26% post, p=0.76). A trend towards improved outcomes was seen after the early transition period but failed to yield statistical significance. Conclusion Early medical management and surgical intervention are key determining factors affecting outcomes following fragility hip fractures. The implementation of a hip fracture protocol at this institution has not yet significantly affected these parameters. This could in part be due to the restrictions placed at this institution during the COVID-19 pandemic. Despite this, the time to OR pre-and post-implementation was quicker than figures reported elsewhere in literature. Further longitudinal data will be collected to determine the final influence of this protocol. Significance/Clinical Relevance Given the increasing number of elderly people and the high morbidity and mortality associated with hip fractures in this population finding cost effective ways to improve outcomes in the management of these injuries has the potential to have enormous positive impact for both patients and hospital systems.Keywords: hip fracture, geriatric, treatment algorithm, preoperative optimization
Procedia PDF Downloads 7610346 Coordinated Interference Canceling Algorithm for Uplink Massive Multiple Input Multiple Output Systems
Authors: Messaoud Eljamai, Sami Hidouri
Abstract:
Massive multiple-input multiple-output (MIMO) is an emerging technology for new cellular networks such as 5G systems. Its principle is to use many antennas per cell in order to maximize the network's spectral efficiency. Inter-cellular interference remains a fundamental problem. The use of massive MIMO will not derogate from the rule. It improves performances only when the number of antennas is significantly greater than the number of users. This, considerably, limits the networks spectral efficiency. In this paper, a coordinated detector for an uplink massive MIMO system is proposed in order to mitigate the inter-cellular interference. The proposed scheme combines the coordinated multipoint technique with an interference-cancelling algorithm. It requires the serving cell to send their received symbols, after processing, decision and error detection, to the interfered cells via a backhaul link. Each interfered cell is capable of eliminating intercellular interferences by generating and subtracting the user’s contribution from the received signal. The resulting signal is more reliable than the original received signal. This allows the uplink massive MIMO system to improve their performances dramatically. Simulation results show that the proposed detector improves system spectral efficiency compared to classical linear detectors.Keywords: massive MIMO, COMP, interference canceling algorithm, spectral efficiency
Procedia PDF Downloads 14610345 Frequency- and Content-Based Tag Cloud Font Distribution Algorithm
Authors: Ágnes Bogárdi-Mészöly, Takeshi Hashimoto, Shohei Yokoyama, Hiroshi Ishikawa
Abstract:
The spread of Web 2.0 has caused user-generated content explosion. Users can tag resources to describe and organize them. Tag clouds provide rough impression of relative importance of each tag within overall cloud in order to facilitate browsing among numerous tags and resources. The goal of our paper is to enrich visualization of tag clouds. A font distribution algorithm has been proposed to calculate a novel metric based on frequency and content, and to classify among classes from this metric based on power law distribution and percentages. The suggested algorithm has been validated and verified on the tag cloud of a real-world thesis portal.Keywords: tag cloud, font distribution algorithm, frequency-based, content-based, power law
Procedia PDF Downloads 50310344 Reliability Based Performance Evaluation of Stone Column Improved Soft Ground
Authors: A. GuhaRay, C. V. S. P. Kiranmayi, S. Rudraraju
Abstract:
The present study considers the effect of variation of different geotechnical random variables in the design of stone column-foundation systems for assessing the bearing capacity and consolidation settlement of highly compressible soil. The soil and stone column properties, spacing, diameter and arrangement of stone columns are considered as the random variables. Probability of failure (Pf) is computed for a target degree of consolidation and a target safe load by Monte Carlo Simulation (MCS). The study shows that the variation in coefficient of radial consolidation (cr) and cohesion of soil (cs) are two most important factors influencing Pf. If the coefficient of variation (COV) of cr exceeds 20%, Pf exceeds 0.001, which is unsafe following the guidelines of US Army Corps of Engineers. The bearing capacity also exceeds its safe value for COV of cs > 30%. It is also observed that as the spacing between the stone column increases, the probability of reaching a target degree of consolidation decreases. Accordingly, design guidelines, considering both consolidation and bearing capacity of improved ground, are proposed for different spacing and diameter of stone columns and geotechnical random variables.Keywords: bearing capacity, consolidation, geotechnical random variables, probability of failure, stone columns
Procedia PDF Downloads 35710343 Classifying and Analysis 8-Bit to 8-Bit S-Boxes Characteristic Using S-Box Evaluation Characteristic
Authors: Muhammad Luqman, Yusuf Kurniawan
Abstract:
S-Boxes is one of the linear parts of the cryptographic algorithm. The existence of S-Box in the cryptographic algorithm is needed to maintain non-linearity of the algorithm. Nowadays, modern cryptographic algorithms use an S-Box as a part of algorithm process. Despite the fact that several cryptographic algorithms today reuse theoretically secure and carefully constructed S-Boxes, there is an evaluation characteristic that can measure security properties of S-Boxes and hence the corresponding primitives. Analysis of an S-Box usually is done using manual mathematics calculation. Several S-Boxes are presented as a Truth Table without any mathematical background algorithm. Then, it’s rather difficult to determine the strength of Truth Table S-Box without a mathematical algorithm. A comprehensive analysis should be applied to the Truth Table S-Box to determine the characteristic. Several important characteristics should be owned by the S-Boxes, they are Nonlinearity, Balancedness, Algebraic degree, LAT, DAT, differential delta uniformity, correlation immunity and global avalanche criterion. Then, a comprehensive tool will be present to automatically calculate the characteristics of S-Boxes and determine the strength of S-Box. Comprehensive analysis is done on a deterministic process to produce a sequence of S-Boxes characteristic and give advice for a better S-Box construction.Keywords: cryptographic properties, Truth Table S-Boxes, S-Boxes characteristic, deterministic process
Procedia PDF Downloads 361