Search results for: algorithm recommendation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3962

Search results for: algorithm recommendation

1142 Design and Advancement of Hybrid Multilevel Inverter Interface with PhotoVoltaic

Authors: P.Kiruthika, K. Ramani

Abstract:

This paper presented the design and advancement of a single-phase 27-level Hybrid Multilevel DC-AC Converter interfacing with Photo Voltaic. In this context, the Multicarrier Pulse Width Modulation method can be implemented in 27-level Hybrid Multilevel Inverter for generating a switching pulse. Perturb & Observer algorithm can be used in the Maximum Power Point Tracking method for the Photo Voltaic system. By implementing Maximum Power Point Tracking with three separate solar panels as an input source to the 27-level Hybrid Multilevel Inverter. This proposed method can be simulated by using MATLAB/simulink. The result shown that the proposed method can achieve silky output wave forms, more flexibility in voltage range, and to reduce Total Harmonic Distortion in medium-voltage drives.

Keywords: Multi Carrier Pulse Width Modulation Technique (MCPWM), Multi Level Inverter (MLI), Maximum Power Point Tracking (MPPT), Perturb and Observer (P&O)

Procedia PDF Downloads 566
1141 Maximum Power Point Tracking for Small Scale Wind Turbine Using Multilayer Perceptron Neural Network Implementation without Mechanical Sensor

Authors: Piyangkun Kukutapan, Siridech Boonsang

Abstract:

The article proposes maximum power point tracking without mechanical sensor using Multilayer Perceptron Neural Network (MLPNN). The aim of article is to reduce the cost and complexity but still retain efficiency. The experimental is that duty cycle is generated maximum power, if it has suitable qualification. The measured data from DC generator, voltage (V), current (I), power (P), turnover rate of power (dP), and turnover rate of voltage (dV) are used as input for MLPNN model. The output of this model is duty cycle for driving the converter. The experiment implemented using Arduino Uno board. This diagram is compared to MPPT using MLPNN and P&O control (Perturbation and Observation control). The experimental results show that the proposed MLPNN based approach is more efficiency than P&O algorithm for this application.

Keywords: maximum power point tracking, multilayer perceptron netural network, optimal duty cycle, DC generator

Procedia PDF Downloads 312
1140 A Review of Encryption Algorithms Used in Cloud Computing

Authors: Derick M. Rakgoale, Topside E. Mathonsi, Vusumuzi Malele

Abstract:

Cloud computing offers distributed online and on-demand computational services from anywhere in the world. Cloud computing services have grown immensely over the past years, especially in the past year due to the Coronavirus pandemic. Cloud computing has changed the working environment and introduced work from work phenomenon, which enabled the adoption of technologies to fulfill the new workings, including cloud services offerings. The increased cloud computing adoption has come with new challenges regarding data privacy and its integrity in the cloud environment. Previously advanced encryption algorithms failed to reduce the memory space required for cloud computing performance, thus increasing the computational cost. This paper reviews the existing encryption algorithms used in cloud computing. In the future, artificial neural networks (ANN) algorithm design will be presented as a security solution to ensure data integrity, confidentiality, privacy, and availability of user data in cloud computing. Moreover, MATLAB will be used to evaluate the proposed solution, and simulation results will be presented.

Keywords: cloud computing, data integrity, confidentiality, privacy, availability

Procedia PDF Downloads 110
1139 A New Reliability based Channel Allocation Model in Mobile Networks

Authors: Anujendra, Parag Kumar Guha Thakurta

Abstract:

The data transmission between mobile hosts and base stations (BSs) in Mobile networks are often vulnerable to failure. Thus, efficient link connectivity, in terms of the services of both base stations and communication channels of the network, is required in wireless mobile networks to achieve highly reliable data transmission. In addition, it is observed that the number of blocked hosts is increased due to insufficient number of channels during heavy load in the network. Under such scenario, the channels are allocated accordingly to offer a reliable communication at any given time. Therefore, a reliability-based channel allocation model with acceptable system performance is proposed as a MOO problem in this paper. Two conflicting parameters known as Resource Reuse factor (RRF) and the number of blocked calls are optimized under reliability constraint in this problem. The solution to such MOO problem is obtained through NSGA-II (Non-dominated Sorting Genetic Algorithm). The effectiveness of the proposed model in this work is shown with a set of experimental results.

Keywords: base station, channel, GA, pareto-optimal, reliability

Procedia PDF Downloads 389
1138 Determination of Energy and Nutrients Composition of Potential Ready-to-Use Therapeutic Food Formulated from Locally Available Resources

Authors: Amina Sa'id Muhammad, Asmau Ishaq Alhassan, Beba Raymond, Fatima Bello

Abstract:

Severe acute malnutrition (SAM) remains a major killer of children under five years of age. Nigeria has the second highest burden of stunted children in the world, with a national prevalence rate of 32 percent of children under five. An estimated 2 million children in Nigeria suffer from severe acute malnutrition (SAM), and 3.9% of children in northwest Nigeria suffer from SAM, which is significantly higher than the national average of 2.1%. Community-Based Management of Acute Malnutrition (CMAM) has proven to be an effective intervention in the treatment of SAM in children using Ready-to-Use Therapeutic Food (RUTF). Ready-to-use therapeutic food (RUTF) is a key component for the treatment of Severe Acute Malnutrition. It contains all the energy and nutrients required for rapid catch-up growth and used particularly in the treatment of children over 6 months of age with SAM without medical complications. However, almost all RUTFs are currently imported to Nigeria from other countries. Shortages of RUTF due to logistics (shipping costs, delays, donor fatigue etc) and funding issues present a threat to the achievement of the 2030 World Health Assembly (WHA) targets for reducing malnutrition in addition to 2030 SDGs 2 (Zero Hunger), 3 (Good Health and Wellbeing), 12 (Responsible Consumption and Production), and 17 (Partnerships for the Goals), thus undermining its effectiveness in combating malnutrition On the other hand, the availability of human and material resources that will aid local production of RUTF presents an opportunity to fill in the gap in regular RUTF supply. About one thousand Nigerian children die of malnutrition-related causes every day, reaching a total of 361,000 each year. Owing to the high burden of malnutrition in Nigeria, the local production of RUTF is a logical step, that will ensure increased availability, acceptability, access, and efficiency in supply, and at lower costs. Objective(s): The objectives of this study were therefore, to formulate RUTF from locally available resources and to determine its energy and nutrients composition, incommensurate with the standard/commercial RUTF. Methods: Three samples of RUTF were formulated using locally available resources (soya beans, wheat, rice, baobab, brown-sugar, date palm and soya oil); which were subjected to various analysis to determine their energy/proximate composition, vitamin and mineral contents and organoleptic properties were also determined using sensory evaluation. Results: The energy values of the three samples of locally produced RUTF were found to be in conformity with WHO recommendation of ≥ 500 kcal per 100g. The energy values of the three RUTF samples produced in the current study were found to be 563.08, 503.67 and 528.98 kcal respectively. Sample A, B and C had protein content of 13.56% 16.71% and 14.62% respectively, which were higher than that of commercial RUTF (10.9%). Conclusions/recommendations: The locally formulated RUTF samples had energy value of more than 500 kcal per 100g; with an appreciable amount of macro and micro nutrients. The appearance, taste, flavor and general acceptability of the formulated RUTF samples were also commendable.

Keywords: energy, malnutrition, nutrients, RUTF

Procedia PDF Downloads 19
1137 Application of Association Rule Using Apriori Algorithm for Analysis of Industrial Accidents in 2013-2014 in Indonesia

Authors: Triano Nurhikmat

Abstract:

Along with the progress of science and technology, the development of the industrialized world in Indonesia took place very rapidly. This leads to a process of industrialization of society Indonesia faster with the establishment of the company and the workplace are diverse. Development of the industry relates to the activity of the worker. Where in these work activities do not cover the possibility of an impending crash on either the workers or on a construction project. The cause of the occurrence of industrial accidents was the fault of electrical damage, work procedures, and error technique. The method of an association rule is one of the main techniques in data mining and is the most common form used in finding the patterns of data collection. In this research would like to know how relations of the association between the incidence of any industrial accidents. Therefore, by using methods of analysis association rule patterns associated with combination obtained two iterations item set (2 large item set) when every factor of industrial accidents with a West Jakarta so industrial accidents caused by the occurrence of an electrical value damage = 0.2 support and confidence value = 1, and the reverse pattern with value = 0.2 support and confidence = 0.75.

Keywords: association rule, data mining, industrial accidents, rules

Procedia PDF Downloads 276
1136 The Influence of Beta Shape Parameters in Project Planning

Authors: Αlexios Kotsakis, Stefanos Katsavounis, Dimitra Alexiou

Abstract:

Networks can be utilized to represent project planning problems, using nodes for activities and arcs to indicate precedence relationship between them. For fixed activity duration, a simple algorithm calculates the amount of time required to complete a project, followed by the activities that comprise the critical path. Program Evaluation and Review Technique (PERT) generalizes the above model by incorporating uncertainty, allowing activity durations to be random variables, producing nevertheless a relatively crude solution in planning problems. In this paper, based on the findings of the relevant literature, which strongly suggests that a Beta distribution can be employed to model earthmoving activities, we utilize Monte Carlo simulation, to estimate the project completion time distribution and measure the influence of skewness, an element inherent in activities of modern technical projects. We also extract the activity criticality index, with an ultimate goal to produce more accurate planning estimations.

Keywords: beta distribution, PERT, Monte Carlo simulation, skewness, project completion time distribution

Procedia PDF Downloads 133
1135 Forced Migration and Access to Maternal Healthcare in Internally Displaced Persons Camps in North-Central Nigeria

Authors: Faith O. Olanrewaju

Abstract:

Internal displacement and the vulnerability of women are two critical aspects of forced migration that have dominated both global and local discourses. Statistics show that in November 2021, there were over 2.1 million internally displaced persons (IDPs) in Nigeria. Literature also states that displaced women and girls are more vulnerable than displaced men. They are susceptible to adversative experiences, including various forms of sexual violence and rape. As a result, the displaced women and girls are faced with psychological and physical traumas, including HIV/AIDS as well as unexpected or poorly spaced pregnancies. In addition, the poor condition of living of internally displaced women in IDP camps affects their reproductive health, pregnancy outcomes, and maternal mortality levels. Incontrovertibly, internally displaced women constitute an imperative contributor to the ills of Nigeria's maternal health status, which is the second worse globally and the worst in Africa. World Health Organisation statistics showed that approximately 536,000 girls and women die from pregnancy-related causes globally, and Nigeria accounts for 14% of the global maternal deaths. Undeniably, this supports the claims that maternal mortality remains a challenge in Nigeria and can be exacerbated by internal displacement crises. Therefore, maternal mortality remains a critical impediment to the actualisation of the 3.1 SDG target. Owing to this, concerns arise about the quality of the policy in Nigeria’s health sector. More specifically, this study is concerned with the maternal health care services displaced women receive in IDP camps in the three states affected by internal displacement in north-central Nigeria, an understudied area. The novelty of the study also lies in its comparative investigation of maternal healthcare service delivery in three different camp structures (faith-based, government, and informal IDP camps), a pattern that is absent in literature. Therefore, this study will investigate how the camp structures affect access to maternal health services in the study areas; analyse the successes and challenges in the delivery of maternal health care services to displaced women in the various camps; and recommendation and strategies for reducing maternal healthcare disparities/gaps across IDP camps in Nigeria (should they exist). It will adopt a mixed-method approach and multi-stage sampling technique. A total of 1,152 copies of the study questionnaire will be distributed to displaced pregnant and nursing mothers (PNM); nine focus group discussions will also be held with the displaced PNM; in-depth interviews will be conducted with humanitarian actors, policymakers, and health professionals. The quantitative and qualitative data will be analysed using Statistical Package for Social Science (SPSS) 21.0 and thematic analysis, respectively. The findings of the study will be used to develop a model of care that will address the fragmentations in Nigeria's healthcare system. The findings will also inform the development of best policies and practices in the maternal health of displaced women.

Keywords: forced displacement, internally displaced women, maternal healthcare, maternal mortality

Procedia PDF Downloads 154
1134 'Low Electronic Noise' Detector Technology in Computed Tomography

Authors: A. Ikhlef

Abstract:

Image noise in computed tomography, is mainly caused by the statistical noise, system noise reconstruction algorithm filters. Since last few years, low dose x-ray imaging became more and more desired and looked as a technical differentiating technology among CT manufacturers. In order to achieve this goal, several technologies and techniques are being investigated, including both hardware (integrated electronics and photon counting) and software (artificial intelligence and machine learning) based solutions. From a hardware point of view, electronic noise could indeed be a potential driver for low and ultra-low dose imaging. We demonstrated that the reduction or elimination of this term could lead to a reduction of dose without affecting image quality. Also, in this study, we will show that we can achieve this goal using conventional electronics (low cost and affordable technology), designed carefully and optimized for maximum detective quantum efficiency. We have conducted the tests using large imaging objects such as 30 cm water and 43 cm polyethylene phantoms. We compared the image quality with conventional imaging protocols with radiation as low as 10 mAs (<< 1 mGy). Clinical validation of such results has been performed as well.

Keywords: computed tomography, electronic noise, scintillation detector, x-ray detector

Procedia PDF Downloads 105
1133 Inverse Problem Method for Microwave Intrabody Medical Imaging

Authors: J. Chamorro-Servent, S. Tassani, M. A. Gonzalez-Ballester, L. J. Roca, J. Romeu, O. Camara

Abstract:

Electromagnetic and microwave imaging (MWI) have been used in medical imaging in the last years, being the most common applications of breast cancer and stroke detection or monitoring. In those applications, the subject or zone to observe is surrounded by a number of antennas, and the Nyquist criterium can be satisfied. Additionally, the space between the antennas (transmitting and receiving the electromagnetic fields) and the zone to study can be prepared in a homogeneous scenario. However, this may differ in other cases as could be intracardiac catheters, stomach monitoring devices, pelvic organ systems, liver ablation monitoring devices, or uterine fibroids’ ablation systems. In this work, we analyzed different MWI algorithms to find the most suitable method for dealing with an intrabody scenario. Due to the space limitations usually confronted on those applications, the device would have a cylindrical configuration of a maximum of eight transmitters and eight receiver antennas. This together with the positioning of the supposed device inside a body tract impose additional constraints in order to choose a reconstruction method; for instance, it inhabitants the use of well-known algorithms such as filtered backpropagation for diffraction tomography (due to the unusual configuration with probes enclosed by the imaging region). Finally, the difficulty of simulating a realistic non-homogeneous background inside the body (due to the incomplete knowledge of the dielectric properties of other tissues between the antennas’ position and the zone to observe), also prevents the use of Born and Rytov algorithms due to their limitations with a heterogeneous background. Instead, we decided to use a time-reversed algorithm (mostly used in geophysics) due to its characteristics of ignoring heterogeneities in the background medium, and of focusing its generated field onto the scatters. Therefore, a 2D time-reversed finite difference time domain was developed based on the time-reversed approach for microwave breast cancer detection. Simultaneously an in-silico testbed was also developed to compare ground-truth dielectric properties with corresponding microwave imaging reconstruction. Forward and inverse problems were computed varying: the frequency used related to a small zone to observe (7, 7.5 and 8 GHz); a small polyp diameter (5, 7 and 10 mm); two polyp positions with respect to the closest antenna (aligned or disaligned); and the (transmitters-to-receivers) antenna combination used for the reconstruction (1-1, 8-1, 8-8 or 8-3). Results indicate that when using the existent time-reversed method for breast cancer here for the different combinations of transmitters and receivers, we found false positives due to the high degrees of freedom and unusual configuration (and the possible violation of Nyquist criterium). Those false positives founded in 8-1 and 8-8 combinations, highly reduced with the 1-1 and 8-3 combination, being the 8-3 configuration de most suitable (three neighboring receivers at each time). The 8-3 configuration creates a region-of-interest reduced problem, decreasing the ill-posedness of the inverse problem. To conclude, the proposed algorithm solves the main limitations of the described intrabody application, successfully detecting the angular position of targets inside the body tract.

Keywords: FDTD, time-reversed, medical imaging, microwave imaging

Procedia PDF Downloads 108
1132 The Contribution of a Workshop Aimed at Providing Social Work Students with Practical Tools for Managing Household Finances

Authors: Ahuva Even-Zohar

Abstract:

Context: Families living in poverty often struggle with poor financial management and accumulating debts. Social workers play a crucial role in assisting these families, but they typically receive no formal training in financial management. Research Aim: The aim of this study was to assess the impact of a voluntary organization called Paamonim on social work students' attitudes toward providing material support and assistance in financial management to families living in poverty. Additionally, the study aimed to evaluate the students' perception of the workshop's contribution and their actual use of the tools acquired during the training. Methodology: The study involved 134 social work students, with 55 in a regular program and 79 in a retraining program. The first stage of the study involved the students receiving an explanation of the research, ensuring confidentiality, and signing informed consent. The students completed two questionnaires: The Student Attitudes Questionnaire and a sociodemographic questionnaire. In the second stage, three months after the workshop, the students completed the questionnaires again, along with a feedback questionnaire. Findings: The study found that there were no significant differences in the students' attitudes toward providing material support and proper financial management to families living in poverty before and after the workshop. However, their attitudes remained positive, at a medium level or higher. The students reported that the workshop provided them with tools to assist families in poverty, and they used these tools to some extent in their practical training. They found the workshop interesting and acknowledged its importance in raising awareness about budget management and enriching their knowledge of money management. The students acknowledged the workshop's effectiveness but mentioned that it was too short. Theoretical Importance: This study highlights the importance of developing programs to help families living in poverty manage their household finances. The workshop enriched the students' knowledge and skills, which are vital for their role as social workers working with families in poverty to rehabilitate their financial situations. Data Collection and Analysis Procedures: The students completed self-report questionnaires before and after the workshop, providing insights into their attitudes and perception about the role of social workers in providing material support for families living in poverty and helping them manage their household finances. Statistical analysis was conducted to compare the students' attitudes pre and post-workshop, and a feedback questionnaire assessed the students' perception of the contribution of the workshop. Question Addressed: This study addressed the question of whether a workshop on financial management can positively impact social work students' attitudes and skills in assisting families living in poverty with their household finances. Conclusion: Despite being short and consisting of only one session, the workshop proved to be valuable in enriching the students' knowledge and providing them with important tools for their role as social workers. The students reported positive attitudes toward providing material support and financial assistance to families in need. The practical recommendation is to continue offering such workshops as part of the social work curriculum to further enhance students' abilities to help families manage their finances effectively.

Keywords: financial literacy, poverty, social work students, workshop

Procedia PDF Downloads 52
1131 Machine Learning Approach for Yield Prediction in Semiconductor Production

Authors: Heramb Somthankar, Anujoy Chakraborty

Abstract:

This paper presents a classification study on yield prediction in semiconductor production using machine learning approaches. A complicated semiconductor production process is generally monitored continuously by signals acquired from sensors and measurement sites. A monitoring system contains a variety of signals, all of which contain useful information, irrelevant information, and noise. In the case of each signal being considered a feature, "Feature Selection" is used to find the most relevant signals. The open-source UCI SECOM Dataset provides 1567 such samples, out of which 104 fail in quality assurance. Feature extraction and selection are performed on the dataset, and useful signals were considered for further study. Afterward, common machine learning algorithms were employed to predict whether the signal yields pass or fail. The most relevant algorithm is selected for prediction based on the accuracy and loss of the ML model.

Keywords: deep learning, feature extraction, feature selection, machine learning classification algorithms, semiconductor production monitoring, signal processing, time-series analysis

Procedia PDF Downloads 97
1130 Local Texture and Global Color Descriptors for Content Based Image Retrieval

Authors: Tajinder Kaur, Anu Bala

Abstract:

An image retrieval system is a computer system for browsing, searching, and retrieving images from a large database of digital images a new algorithm meant for content-based image retrieval (CBIR) is presented in this paper. The proposed method combines the color and texture features which are extracted the global and local information of the image. The local texture feature is extracted by using local binary patterns (LBP), which are evaluated by taking into consideration of local difference between the center pixel and its neighbors. For the global color feature, the color histogram (CH) is used which is calculated by RGB (red, green, and blue) spaces separately. In this paper, the combination of color and texture features are proposed for content-based image retrieval. The performance of the proposed method is tested on Corel 1000 database which is the natural database. The results after being investigated show a significant improvement in terms of their evaluation measures as compared to LBP and CH.

Keywords: color, texture, feature extraction, local binary patterns, image retrieval

Procedia PDF Downloads 344
1129 Blind Super-Resolution Reconstruction Based on PSF Estimation

Authors: Osama A. Omer, Amal Hamed

Abstract:

Successful blind image Super-Resolution algorithms require the exact estimation of the Point Spread Function (PSF). In the absence of any prior information about the imagery system and the true image; this estimation is normally done by trial and error experimentation until an acceptable restored image quality is obtained. Multi-frame blind Super-Resolution algorithms often have disadvantages of slow convergence and sensitiveness to complex noises. This paper presents a Super-Resolution image reconstruction algorithm based on estimation of the PSF that yields the optimum restored image quality. The estimation of PSF is performed by the knife-edge method and it is implemented by measuring spreading of the edges in the reproduced HR image itself during the reconstruction process. The proposed image reconstruction approach is using L1 norm minimization and robust regularization based on a bilateral prior to deal with different data and noise models. A series of experiment results show that the proposed method can outperform other previous work robustly and efficiently.

Keywords: blind, PSF, super-resolution, knife-edge, blurring, bilateral, L1 norm

Procedia PDF Downloads 349
1128 Clustering Based Level Set Evaluation for Low Contrast Images

Authors: Bikshalu Kalagadda, Srikanth Rangu

Abstract:

The important object of images segmentation is to extract objects with respect to some input features. One of the important methods for image segmentation is Level set method. Generally medical images and synthetic images with low contrast of pixel profile, for such images difficult to locate interested features in images. In conventional level set function, develops irregularity during its process of evaluation of contour of objects, this destroy the stability of evolution process. For this problem a remedy is proposed, a new hybrid algorithm is Clustering Level Set Evolution. Kernel fuzzy particles swarm optimization clustering with the Distance Regularized Level Set (DRLS) and Selective Binary, and Gaussian Filtering Regularized Level Set (SBGFRLS) methods are used. The ability of identifying different regions becomes easy with improved speed. Efficiency of the modified method can be evaluated by comparing with the previous method for similar specifications. Comparison can be carried out by considering medical and synthetic images.

Keywords: segmentation, clustering, level set function, re-initialization, Kernel fuzzy, swarm optimization

Procedia PDF Downloads 335
1127 Spatial-Temporal Awareness Approach for Extensive Re-Identification

Authors: Tyng-Rong Roan, Fuji Foo, Wenwey Hseush

Abstract:

Recent development of AI and edge computing plays a critical role to capture meaningful events such as detection of an unattended bag. One of the core problems is re-identification across multiple CCTVs. Immediately following the detection of a meaningful event is to track and trace the objects related to the event. In an extensive environment, the challenge becomes severe when the number of CCTVs increases substantially, imposing difficulties in achieving high accuracy while maintaining real-time performance. The algorithm that re-identifies cross-boundary objects for extensive tracking is referred to Extensive Re-Identification, which emphasizes the issues related to the complexity behind a great number of CCTVs. The Spatial-Temporal Awareness approach challenges the conventional thinking and concept of operations which is labor intensive and time consuming. The ability to perform Extensive Re-Identification through a multi-sensory network provides the next-level insights – creating value beyond traditional risk management.

Keywords: long-short-term memory, re-identification, security critical application, spatial-temporal awareness

Procedia PDF Downloads 99
1126 Optimal Bayesian Chart for Controlling Expected Number of Defects in Production Processes

Authors: V. Makis, L. Jafari

Abstract:

In this paper, we develop an optimal Bayesian chart to control the expected number of defects per inspection unit in production processes with long production runs. We formulate this control problem in the optimal stopping framework. The objective is to determine the optimal stopping rule minimizing the long-run expected average cost per unit time considering partial information obtained from the process sampling at regular epochs. We prove the optimality of the control limit policy, i.e., the process is stopped and the search for assignable causes is initiated when the posterior probability that the process is out of control exceeds a control limit. An algorithm in the semi-Markov decision process framework is developed to calculate the optimal control limit and the corresponding average cost. Numerical examples are presented to illustrate the developed optimal control chart and to compare it with the traditional u-chart.

Keywords: Bayesian u-chart, economic design, optimal stopping, semi-Markov decision process, statistical process control

Procedia PDF Downloads 557
1125 Oil Reservoir Asphalting Precipitation Estimating during CO2 Injection

Authors: I. Alhajri, G. Zahedi, R. Alazmi, A. Akbari

Abstract:

In this paper, an Artificial Neural Network (ANN) was developed to predict Asphaltene Precipitation (AP) during the injection of carbon dioxide into crude oil reservoirs. In this study, the experimental data from six different oil fields were collected. Seventy percent of the data was used to develop the ANN model, and different ANN architectures were examined. A network with the Trainlm training algorithm was found to be the best network to estimate the AP. To check the validity of the proposed model, the model was used to predict the AP for the thirty percent of the data that was unevaluated. The Mean Square Error (MSE) of the prediction was 0.0018, which confirms the excellent prediction capability of the proposed model. In the second part of this study, the ANN model predictions were compared with modified Hirschberg model predictions. The ANN was found to provide more accurate estimates compared to the modified Hirschberg model. Finally, the proposed model was employed to examine the effect of different operating parameters during gas injection on the AP. It was found that the AP is mostly sensitive to the reservoir temperature. Furthermore, the carbon dioxide concentration in liquid phase increases the AP.

Keywords: artificial neural network, asphaltene, CO2 injection, Hirschberg model, oil reservoirs

Procedia PDF Downloads 354
1124 Research on Knowledge Graph Inference Technology Based on Proximal Policy Optimization

Authors: Yihao Kuang, Bowen Ding

Abstract:

With the increasing scale and complexity of knowledge graph, modern knowledge graph contains more and more types of entity, relationship, and attribute information. Therefore, in recent years, it has been a trend for knowledge graph inference to use reinforcement learning to deal with large-scale, incomplete, and noisy knowledge graphs and improve the inference effect and interpretability. The Proximal Policy Optimization (PPO) algorithm utilizes a near-end strategy optimization approach. This allows for more extensive updates of policy parameters while constraining the update extent to maintain training stability. This characteristic enables PPOs to converge to improved strategies more rapidly, often demonstrating enhanced performance early in the training process. Furthermore, PPO has the advantage of offline learning, effectively utilizing historical experience data for training and enhancing sample utilization. This means that even with limited resources, PPOs can efficiently train for reinforcement learning tasks. Based on these characteristics, this paper aims to obtain a better and more efficient inference effect by introducing PPO into knowledge inference technology.

Keywords: reinforcement learning, PPO, knowledge inference

Procedia PDF Downloads 218
1123 Development of Computational Approach for Calculation of Hydrogen Solubility in Hydrocarbons for Treatment of Petroleum

Authors: Abdulrahman Sumayli, Saad M. AlShahrani

Abstract:

For the hydrogenation process, knowing the solubility of hydrogen (H2) in hydrocarbons is critical to improve the efficiency of the process. We investigated the H2 solubility computation in four heavy crude oil feedstocks using machine learning techniques. Temperature, pressure, and feedstock type were considered as the inputs to the models, while the hydrogen solubility was the sole response. Specifically, we employed three different models: Support Vector Regression (SVR), Gaussian process regression (GPR), and Bayesian ridge regression (BRR). To achieve the best performance, the hyper-parameters of these models are optimized using the whale optimization algorithm (WOA). We evaluated the models using a dataset of solubility measurements in various feedstocks, and we compared their performance based on several metrics. Our results show that the WOA-SVR model tuned with WOA achieves the best performance overall, with an RMSE of 1.38 × 10− 2 and an R-squared of 0.991. These findings suggest that machine learning techniques can provide accurate predictions of hydrogen solubility in different feedstocks, which could be useful in the development of hydrogen-related technologies. Besides, the solubility of hydrogen in the four heavy oil fractions is estimated in different ranges of temperatures and pressures of 150 ◦C–350 ◦C and 1.2 MPa–10.8 MPa, respectively

Keywords: temperature, pressure variations, machine learning, oil treatment

Procedia PDF Downloads 55
1122 Solving Dimensionality Problem and Finding Statistical Constructs on Latent Regression Models: A Novel Methodology with Real Data Application

Authors: Sergio Paez Moncaleano, Alvaro Mauricio Montenegro

Abstract:

This paper presents a novel statistical methodology for measuring and founding constructs in Latent Regression Analysis. This approach uses the qualities of Factor Analysis in binary data with interpretations on Item Response Theory (IRT). In addition, based on the fundamentals of submodel theory and with a convergence of many ideas of IRT, we propose an algorithm not just to solve the dimensionality problem (nowadays an open discussion) but a new research field that promises more fear and realistic qualifications for examiners and a revolution on IRT and educational research. In the end, the methodology is applied to a set of real data set presenting impressive results for the coherence, speed and precision. Acknowledgments: This research was financed by Colciencias through the project: 'Multidimensional Item Response Theory Models for Practical Application in Large Test Designed to Measure Multiple Constructs' and both authors belong to SICS Research Group from Universidad Nacional de Colombia.

Keywords: item response theory, dimensionality, submodel theory, factorial analysis

Procedia PDF Downloads 352
1121 Representativity Based Wasserstein Active Regression

Authors: Benjamin Bobbia, Matthias Picard

Abstract:

In recent years active learning methodologies based on the representativity of the data seems more promising to limit overfitting. The presented query methodology for regression using the Wasserstein distance measuring the representativity of our labelled dataset compared to the global distribution. In this work a crucial use of GroupSort Neural Networks is made therewith to draw a double advantage. The Wasserstein distance can be exactly expressed in terms of such neural networks. Moreover, one can provide explicit bounds for their size and depth together with rates of convergence. However, heterogeneity of the dataset is also considered by weighting the Wasserstein distance with the error of approximation at the previous step of active learning. Such an approach leads to a reduction of overfitting and high prediction performance after few steps of query. After having detailed the methodology and algorithm, an empirical study is presented in order to investigate the range of our hyperparameters. The performances of this method are compared, in terms of numbers of query needed, with other classical and recent query methods on several UCI datasets.

Keywords: active learning, Lipschitz regularization, neural networks, optimal transport, regression

Procedia PDF Downloads 69
1120 Double Encrypted Data Communication Using Cryptography and Steganography

Authors: Adine Barett, Jermel Watson, Anteneh Girma, Kacem Thabet

Abstract:

In information security, secure communication of data across networks has always been a problem at the forefront. Transfer of information across networks is susceptible to being exploited by attackers engaging in malicious activity. In this paper, we leverage steganography and cryptography to create a layered security solution to protect the information being transmitted. The first layer of security leverages crypto- graphic techniques to scramble the information so that it cannot be deciphered even if the steganography-based layer is compromised. The second layer of security relies on steganography to disguise the encrypted in- formation so that it cannot be seen. We consider three cryptographic cipher methods in the cryptography layer, namely, Playfair cipher, Blowfish cipher, and Hills cipher. Then, the encrypted message is passed through the least significant bit (LSB) to the steganography algorithm for further encryption. Both encryption approaches are combined efficiently to help secure information in transit over a network. This multi-layered encryption is a solution that will benefit cloud platforms, social media platforms and networks that regularly transfer private information such as banks and insurance companies.

Keywords: cryptography, steganography, layered security, Cipher, encryption

Procedia PDF Downloads 65
1119 Efects of Data Corelation in a Sparse-View Compresive Sensing Based Image Reconstruction

Authors: Sajid Abas, Jon Pyo Hong, Jung-Ryun Le, Seungryong Cho

Abstract:

Computed tomography and laminography are heavily investigated in a compressive sensing based image reconstruction framework to reduce the dose to the patients as well as to the radiosensitive devices such as multilayer microelectronic circuit boards. Nowadays researchers are actively working on optimizing the compressive sensing based iterative image reconstruction algorithm to obtain better quality images. However, the effects of the sampled data’s properties on reconstructed the image’s quality, particularly in an insufficient sampled data conditions have not been explored in computed laminography. In this paper, we investigated the effects of two data properties i.e. sampling density and data incoherence on the reconstructed image obtained by conventional computed laminography and a recently proposed method called spherical sinusoidal scanning scheme. We have found that in a compressive sensing based image reconstruction framework, the image quality mainly depends upon the data incoherence when the data is uniformly sampled.

Keywords: computed tomography, computed laminography, compressive sending, low-dose

Procedia PDF Downloads 452
1118 Component Based Testing Using Clustering and Support Vector Machine

Authors: Iqbaldeep Kaur, Amarjeet Kaur

Abstract:

Software Reusability is important part of software development. So component based software development in case of software testing has gained a lot of practical importance in the field of software engineering from academic researcher and also from software development industry perspective. Finding test cases for efficient reuse of test cases is one of the important problems aimed by researcher. Clustering reduce the search space, reuse test cases by grouping similar entities according to requirements ensuring reduced time complexity as it reduce the search time for retrieval the test cases. In this research paper we proposed approach for re-usability of test cases by unsupervised approach. In unsupervised learning we proposed k-mean and Support Vector Machine. We have designed the algorithm for requirement and test case document clustering according to its tf-idf vector space and the output is set of highly cohesive pattern groups.

Keywords: software testing, reusability, clustering, k-mean, SVM

Procedia PDF Downloads 412
1117 Constructing White-Box Implementations Based on Threshold Shares and Composite Fields

Authors: Tingting Lin, Manfred von Willich, Dafu Lou, Phil Eisen

Abstract:

A white-box implementation of a cryptographic algorithm is a software implementation intended to resist extraction of the secret key by an adversary. To date, most of the white-box techniques are used to protect block cipher implementations. However, a large proportion of the white-box implementations are proven to be vulnerable to affine equivalence attacks and other algebraic attacks, as well as differential computation analysis (DCA). In this paper, we identify a class of block ciphers for which we propose a method of constructing white-box implementations. Our method is based on threshold implementations and operations in composite fields. The resulting implementations consist of lookup tables and few exclusive OR operations. All intermediate values (inputs and outputs of the lookup tables) are masked. The threshold implementation makes the distribution of the masked values uniform and independent of the original inputs, and the operations in composite fields reduce the size of the lookup tables. The white-box implementations can provide resistance against algebraic attacks and DCA-like attacks.

Keywords: white-box, block cipher, composite field, threshold implementation

Procedia PDF Downloads 155
1116 Model of Transhipment and Routing Applied to the Cargo Sector in Small and Medium Enterprises of Bogotá, Colombia

Authors: Oscar Javier Herrera Ochoa, Ivan Dario Romero Fonseca

Abstract:

This paper presents a design of a model for planning the distribution logistics operation. The significance of this work relies on the applicability of this fact to the analysis of small and medium enterprises (SMEs) of dry freight in Bogotá. Two stages constitute this implementation: the first one is the place where optimal planning is achieved through a hybrid model developed with mixed integer programming, which considers the transhipment operation based on a combined load allocation model as a classic transshipment model; the second one is the specific routing of that operation through the heuristics of Clark and Wright. As a result, an integral model is obtained to carry out the step by step planning of the distribution of dry freight for SMEs in Bogotá. In this manner, optimum assignments are established by utilizing transshipment centers with that purpose of determining the specific routing based on the shortest distance traveled.

Keywords: transshipment model, mixed integer programming, saving algorithm, dry freight transportation

Procedia PDF Downloads 207
1115 Osteosuture in Fixation of Displaced Lateral Third Clavicle Fractures: A Case Report

Authors: Patrícia Pires, Renata Vaz, Bárbara Teles, Marco Pato, Pedro Beckert

Abstract:

Introduction: The management of lateral third clavicle fractures can be challenging due to difficulty in distinguishing subtle variations in the fracture pattern, which may be suggestive of potential fracture instability. They occur most often in men between 30 and 50 years of age, and in individuals over 70 years of age, its distribution is equal between both men and women. These fractures account for 10%–30% of all clavicle fractures and roughly 30%–45% of all clavicle nonunion fractures. Lateral third clavicle fractures may be treated conservatively or surgically, and there is no gold standard, although the risk of nonunion or pseudoarthrosis impacts the recommendation of surgical treatment when these fractures are unstable. There are many strategies for surgical treatment, including locking plates, hook plates fixation, coracoclavicular fixation using suture anchors, devices or screws, tension band fixation with suture or wire, transacromial Kirschner wire fixation and arthroscopically assisted techniques. Whenever taking the hardware into consideration, we must not disregard that obtaining adequate lateral fixation of small fragments is a difficult task, and plates are more associated to local irritation. The aim of the appropriate treatment is to ensure fracture healing and a rapid return to preinjury activities of daily living but, as explained, definitive treatment strategies have not been established and the variety of techniques avalilable add up to the discussion of this topic. Methods and Results: We present a clinical case of a 43-year-old man with the diagnosis of a lateral third clavicle fracture (Neer IIC) in the sequence of a fall on his right shoulder after a bicycle fall. He was operated three days after the injury, and through K-wire temporary fixation and indirect reduction using a ZipTight, he underwent osteosynthesis with an interfragmentary figure-of-eight tension band with polydioxanone suture (PDS). Two weeks later, there was a good aligment. He kept the sling until 6 weeks pos-op, avoiding efforts. At 7-weeks pos-op, there was still a good aligment, starting the physiotherapy exercises. After 10 months, he had no limitation in mobility or pain and returned to work with complete recovery in strength. Conclusion: Some distal clavicle fractures may be conservatively treated, but it is widely accepted that unstable fractures require surgical treatment to obtain superior clinical outcomes. In the clinical case presented, the authors chose an osteosuture technique due to the fracture pattern, its location. Since there isn´t a consensus on the prefered fixation method, it is important for surgeons to be skilled in various techniques and decide with their patient which approach is most appropriate for them, weighting the risk-benefit of each method. For instance, with the suture technique used, there is no wire migration or breakage, and it doesn´t require a reoperation for hardware removal; there is also less tissue exposure since it requires a smaller approach in comparison to the plate fixation and avoids cuff tears like the hook plate. The good clinical outcome on this case report serves the purpose of expanding the consideration of this method has a therapeutic option.

Keywords: lateral third, clavicle, suture, fixation

Procedia PDF Downloads 56
1114 Profit-Based Artificial Neural Network (ANN) Trained by Migrating Birds Optimization: A Case Study in Credit Card Fraud Detection

Authors: Ashkan Zakaryazad, Ekrem Duman

Abstract:

A typical classification technique ranks the instances in a data set according to the likelihood of belonging to one (positive) class. A credit card (CC) fraud detection model ranks the transactions in terms of probability of being fraud. In fact, this approach is often criticized, because firms do not care about fraud probability but about the profitability or costliness of detecting a fraudulent transaction. The key contribution in this study is to focus on the profit maximization in the model building step. The artificial neural network proposed in this study works based on profit maximization instead of minimizing the error of prediction. Moreover, some studies have shown that the back propagation algorithm, similar to other gradient–based algorithms, usually gets trapped in local optima and swarm-based algorithms are more successful in this respect. In this study, we train our profit maximization ANN using the Migrating Birds optimization (MBO) which is introduced to literature recently.

Keywords: neural network, profit-based neural network, sum of squared errors (SSE), MBO, gradient descent

Procedia PDF Downloads 458
1113 Degraded Document Analysis and Extraction of Original Text Document: An Approach without Optical Character Recognition

Authors: L. Hamsaveni, Navya Prakash, Suresha

Abstract:

Document Image Analysis recognizes text and graphics in documents acquired as images. An approach without Optical Character Recognition (OCR) for degraded document image analysis has been adopted in this paper. The technique involves document imaging methods such as Image Fusing and Speeded Up Robust Features (SURF) Detection to identify and extract the degraded regions from a set of document images to obtain an original document with complete information. In case, degraded document image captured is skewed, it has to be straightened (deskew) to perform further process. A special format of image storing known as YCbCr is used as a tool to convert the Grayscale image to RGB image format. The presented algorithm is tested on various types of degraded documents such as printed documents, handwritten documents, old script documents and handwritten image sketches in documents. The purpose of this research is to obtain an original document for a given set of degraded documents of the same source.

Keywords: grayscale image format, image fusing, RGB image format, SURF detection, YCbCr image format

Procedia PDF Downloads 364