Search results for: hybrid PSO-GA algorithm and mutual information
13872 Structural Invertibility and Optimal Sensor Node Placement for Error and Input Reconstruction in Dynamic Systems
Authors: Maik Kschischo, Dominik Kahl, Philipp Wendland, Andreas Weber
Abstract:
Understanding and modelling of real-world complex dynamic systems in biology, engineering and other fields is often made difficult by incomplete knowledge about the interactions between systems states and by unknown disturbances to the system. In fact, most real-world dynamic networks are open systems receiving unknown inputs from their environment. To understand a system and to estimate the state dynamics, these inputs need to be reconstructed from output measurements. Reconstructing the input of a dynamic system from its measured outputs is an ill-posed problem if only a limited number of states is directly measurable. A first requirement for solving this problem is the invertibility of the input-output map. In our work, we exploit the fact that invertibility of a dynamic system is a structural property, which depends only on the network topology. Therefore, it is possible to check for invertibility using a structural invertibility algorithm which counts the number of node disjoint paths linking inputs and outputs. The algorithm is efficient enough, even for large networks up to a million nodes. To understand structural features influencing the invertibility of a complex dynamic network, we analyze synthetic and real networks using the structural invertibility algorithm. We find that invertibility largely depends on the degree distribution and that dense random networks are easier to invert than sparse inhomogeneous networks. We show that real networks are often very difficult to invert unless the sensor nodes are carefully chosen. To overcome this problem, we present a sensor node placement algorithm to achieve invertibility with a minimum set of measured states. This greedy algorithm is very fast and also guaranteed to find an optimal sensor node-set if it exists. Our results provide a practical approach to experimental design for open, dynamic systems. Since invertibility is a necessary condition for unknown input observers and data assimilation filters to work, it can be used as a preprocessing step to check, whether these input reconstruction algorithms can be successful. If not, we can suggest additional measurements providing sufficient information for input reconstruction. Invertibility is also important for systems design and model building. Dynamic models are always incomplete, and synthetic systems act in an environment, where they receive inputs or even attack signals from their exterior. Being able to monitor these inputs is an important design requirement, which can be achieved by our algorithms for invertibility analysis and sensor node placement.Keywords: data-driven dynamic systems, inversion of dynamic systems, observability, experimental design, sensor node placement
Procedia PDF Downloads 15013871 The Revitalization of South-south Cooperation: Evaluation of South African Direct Investment in Cameroon
Authors: Albert Herve Nkolo Mpoko
Abstract:
The Foreign Direct Investment (FDI) landscape in Cameroon has garnered significant attention from both European and Asian nations due to perceived benefits such as capital infusion, technology transfer, and potential for economic expansion. However, it is noteworthy that South Africa's investment presence remains comparatively subdued in Cameroon, lagging behind that of Europe and Asia. Equally surprising is the limited footprint of Africa's economic powerhouse within other African economies. This study delved into four specific facets of South African investment in Cameroon. Initially, it focused on identifying South African companies operating within Cameroon. Subsequently, the analysis encompassed assessing the correlation between South African investment and poverty alleviation. Additionally, the study examined the nexus between South African investment and technological advancement, and underscored the significance of investment incentives in both countries Key findings of the research shed light on several crucial points. South Africa ought to reassess its economic engagement with Francophone Africa, particularly Cameroon. Despite existing policies aimed at fostering investment, there remains substantial ground to cover in this realm. The proliferation of South African enterprises in Cameroon holds the potential to ameliorate poverty and foster employment opportunities across both nations. The advent of South African firms in Cameroon can catalyse technological advancements within the region. Data collection involved surveying 100 executives from the respective administrations and conducting ten interviews. The gathered data underwent triangulation, wherein quantitative findings were juxtaposed with qualitative insights. In conclusion, the study underscores the underutilization of Cameroon by South Africa, emphasizing the untapped potential for mutual economic growth. Furthermore, it posits that the success of South Africa's multinational corporations abroad could serve as a pivotal pillar for sustaining its domestic economy.Keywords: FDI, transfer of technology, South-South cooperation, mutual economic growth
Procedia PDF Downloads 4613870 Prediction of Physical Properties and Sound Absorption Performance of Automotive Interior Materials
Authors: Un-Hwan Park, Jun-Hyeok Heo, In-Sung Lee, Seong-Jin Cho, Tae-Hyeon Oh, Dae-Kyu Park
Abstract:
Sound absorption coefficient is considered important when designing because noise affects emotion quality of car. It is designed with lots of experiment tunings in the field because it is unreliable to predict it for multi-layer material. In this paper, we present the design of sound absorption for automotive interior material with multiple layers using estimation software of sound absorption coefficient for reverberation chamber. Additionally, we introduce the method for estimation of physical properties required to predict sound absorption coefficient of car interior materials with multiple layers too. It is calculated by inverse algorithm. It is very economical to get information about physical properties without expensive equipment. Correlation test is carried out to ensure reliability for accuracy. The data to be used for the correlation is sound absorption coefficient measured in the reverberation chamber. In this way, it is considered economical and efficient to design automotive interior materials. And design optimization for sound absorption coefficient is also easy to implement when it is designed.Keywords: sound absorption coefficient, optimization design, inverse algorithm, automotive interior material, multiple layers nonwoven, scaled reverberation chamber, sound impedance tubes
Procedia PDF Downloads 30913869 An Algorithm Based on Control Indexes to Increase the Quality of Service on Cellular Networks
Authors: Rahman Mofidi, Sina Rahimi, Farnoosh Darban
Abstract:
Communication plays a key role in today’s world, and to support it, the quality of service has the highest priority. It is very important to differentiate between traffic based on priority level. Some traffic classes should be a higher priority than other classes. It is also necessary to give high priority to customers who have more payment for better service, however, without influence on other customers. So to realize that, we will require effective quality of service methods. To ensure the optimal performance of the network in accordance with the quality of service is an important goal for all operators in the mobile network. In this work, we propose an algorithm based on control parameters which it’s based on user feedback that aims at minimizing the access to system transmit power and thus improving the network key performance indicators and increasing the quality of service. This feedback that is known as channel quality indicator (CQI) indicates the received signal level of the user. We aim at proposing an algorithm in control parameter criterion to study improving the quality of service and throughput in a cellular network at the simulated environment. In this work we tried to parameter values have close to their actual level. Simulation results show that the proposed algorithm improves the system throughput and thus satisfies users' throughput and improves service to set up a successful call.Keywords: quality of service, key performance indicators, control parameter, channel quality indicator
Procedia PDF Downloads 20313868 A Multilayer Perceptron Neural Network Model Optimized by Genetic Algorithm for Significant Wave Height Prediction
Authors: Luis C. Parra
Abstract:
The significant wave height prediction is an issue of great interest in the field of coastal activities because of the non-linear behavior of the wave height and its complexity of prediction. This study aims to present a machine learning model to forecast the significant wave height of the oceanographic wave measuring buoys anchored at Mooloolaba of the Queensland Government Data. Modeling was performed by a multilayer perceptron neural network-genetic algorithm (GA-MLP), considering Relu(x) as the activation function of the MLPNN. The GA is in charge of optimized the MLPNN hyperparameters (learning rate, hidden layers, neurons, and activation functions) and wrapper feature selection for the window width size. Results are assessed using Mean Square Error (MSE), Root Mean Square Error (RMSE), and Mean Absolute Error (MAE). The GAMLPNN algorithm was performed with a population size of thirty individuals for eight generations for the prediction optimization of 5 steps forward, obtaining a performance evaluation of 0.00104 MSE, 0.03222 RMSE, 0.02338 MAE, and 0.71163% of MAPE. The results of the analysis suggest that the MLPNNGA model is effective in predicting significant wave height in a one-step forecast with distant time windows, presenting 0.00014 MSE, 0.01180 RMSE, 0.00912 MAE, and 0.52500% of MAPE with 0.99940 of correlation factor. The GA-MLP algorithm was compared with the ARIMA forecasting model, presenting better performance criteria in all performance criteria, validating the potential of this algorithm.Keywords: significant wave height, machine learning optimization, multilayer perceptron neural networks, evolutionary algorithms
Procedia PDF Downloads 10713867 A Genetic Algorithm Based Sleep-Wake up Protocol for Area Coverage in WSNs
Authors: Seyed Mahdi Jameii, Arash Nikdel, Seyed Mohsen Jameii
Abstract:
Energy efficiency is an important issue in the field of Wireless Sensor Networks (WSNs). So, minimizing the energy consumption in this kind of networks should be an essential consideration. Sleep/wake scheduling mechanism is an efficient approach to handling this issue. In this paper, we propose a Genetic Algorithm-based Sleep-Wake up Area Coverage protocol called GA-SWAC. The proposed protocol puts the minimum of nodes in active mode and adjusts the sensing radius of each active node to decrease the energy consumption while maintaining the network’s coverage. The proposed protocol is simulated. The results demonstrate the efficiency of the proposed protocol in terms of coverage ratio, number of active nodes and energy consumption.Keywords: wireless sensor networks, genetic algorithm, coverage, connectivity
Procedia PDF Downloads 52113866 Hand Symbol Recognition Using Canny Edge Algorithm and Convolutional Neural Network
Authors: Harshit Mittal, Neeraj Garg
Abstract:
Hand symbol recognition is a pivotal component in the domain of computer vision, with far-reaching applications spanning sign language interpretation, human-computer interaction, and accessibility. This research paper discusses the approach with the integration of the Canny Edge algorithm and convolutional neural network. The significance of this study lies in its potential to enhance communication and accessibility for individuals with hearing impairments or those engaged in gesture-based interactions with technology. In the experiment mentioned, the data is manually collected by the authors from the webcam using Python codes, to increase the dataset augmentation, is applied to original images, which makes the model more compatible and advanced. Further, the dataset of about 6000 coloured images distributed equally in 5 classes (i.e., 1, 2, 3, 4, 5) are pre-processed first to gray images and then by the Canny Edge algorithm with threshold 1 and 2 as 150 each. After successful data building, this data is trained on the Convolutional Neural Network model, giving accuracy: 0.97834, precision: 0.97841, recall: 0.9783, and F1 score: 0.97832. For user purposes, a block of codes is built in Python to enable a window for hand symbol recognition. This research, at its core, seeks to advance the field of computer vision by providing an advanced perspective on hand sign recognition. By leveraging the capabilities of the Canny Edge algorithm and convolutional neural network, this study contributes to the ongoing efforts to create more accurate, efficient, and accessible solutions for individuals with diverse communication needs.Keywords: hand symbol recognition, computer vision, Canny edge algorithm, convolutional neural network
Procedia PDF Downloads 6513865 Effect of Perception on People’s Behavior in Public Space
Authors: Morteza Maleki
Abstract:
In the present study is that it tried to behave in the environment to be monitored and the respective roles of environment (assumed as a vessel) and human beings (assumed as occupants of this vessel) inevitably create effects which can be expressed as various behaviors on the part of human being. The mutual relationship between man and his environment is exhibited through perceptions, behaviors, subjective images, activities, etc. This study investigates the conceptual dimension in the form of the four components of readability, sense of place, identity, and Tenability (tenability) at the Ahmadabad Axis in Mashhad. The theoretical fundamentals and the data regarding the status quo were presented through the descriptive method and the proposed policies were derived through analyzing the available status quo information. The required data were gathered from library resources and documents related to the studied area as well as from instruments used in field methods such as questionnaires. Upon conducting the necessary investigation, the conceptual dimension within the design area was analyzed. The SWOT table was presented, and the results obtained for improving environmental perception were arranged in the form of policy-making tables and operational projects tables for improving the sense of place, creating imagery, and other investigated components.Keywords: public space, perception, environment, behavior
Procedia PDF Downloads 39213864 Predicting Daily Patient Hospital Visits Using Machine Learning
Authors: Shreya Goyal
Abstract:
The study aims to build user-friendly software to understand patient arrival patterns and compute the number of potential patients who will visit a particular health facility for a given period by using a machine learning algorithm. The underlying machine learning algorithm used in this study is the Support Vector Machine (SVM). Accurate prediction of patient arrival allows hospitals to operate more effectively, providing timely and efficient care while optimizing resources and improving patient experience. It allows for better allocation of staff, equipment, and other resources. If there's a projected surge in patients, additional staff or resources can be allocated to handle the influx, preventing bottlenecks or delays in care. Understanding patient arrival patterns can also help streamline processes to minimize waiting times for patients and ensure timely access to care for patients in need. Another big advantage of using this software is adhering to strict data protection regulations such as the Health Insurance Portability and Accountability Act (HIPAA) in the United States as the hospital will not have to share the data with any third party or upload it to the cloud because the software can read data locally from the machine. The data needs to be arranged in. a particular format and the software will be able to read the data and provide meaningful output. Using software that operates locally can facilitate compliance with these regulations by minimizing data exposure. Keeping patient data within the hospital's local systems reduces the risk of unauthorized access or breaches associated with transmitting data over networks or storing it in external servers. This can help maintain the confidentiality and integrity of sensitive patient information. Historical patient data is used in this study. The input variables used to train the model include patient age, time of day, day of the week, seasonal variations, and local events. The algorithm uses a Supervised learning method to optimize the objective function and find the global minima. The algorithm stores the values of the local minima after each iteration and at the end compares all the local minima to find the global minima. The strength of this study is the transfer function used to calculate the number of patients. The model has an output accuracy of >95%. The method proposed in this study could be used for better management planning of personnel and medical resources.Keywords: machine learning, SVM, HIPAA, data
Procedia PDF Downloads 6513863 Combination between Intrusion Systems and Honeypots
Authors: Majed Sanan, Mohammad Rammal, Wassim Rammal
Abstract:
Today, security is a major concern. Intrusion Detection, Prevention Systems and Honeypot can be used to moderate attacks. Many researchers have proposed to use many IDSs ((Intrusion Detection System) time to time. Some of these IDS’s combine their features of two or more IDSs which are called Hybrid Intrusion Detection Systems. Most of the researchers combine the features of Signature based detection methodology and Anomaly based detection methodology. For a signature based IDS, if an attacker attacks slowly and in organized way, the attack may go undetected through the IDS, as signatures include factors based on duration of the events but the actions of attacker do not match. Sometimes, for an unknown attack there is no signature updated or an attacker attack in the mean time when the database is updating. Thus, signature-based IDS fail to detect unknown attacks. Anomaly based IDS suffer from many false-positive readings. So there is a need to hybridize those IDS which can overcome the shortcomings of each other. In this paper we propose a new approach to IDS (Intrusion Detection System) which is more efficient than the traditional IDS (Intrusion Detection System). The IDS is based on Honeypot Technology and Anomaly based Detection Methodology. We have designed Architecture for the IDS in a packet tracer and then implemented it in real time. We have discussed experimental results performed: both the Honeypot and Anomaly based IDS have some shortcomings but if we hybridized these two technologies, the newly proposed Hybrid Intrusion Detection System (HIDS) is capable enough to overcome these shortcomings with much enhanced performance. In this paper, we present a modified Hybrid Intrusion Detection System (HIDS) that combines the positive features of two different detection methodologies - Honeypot methodology and anomaly based intrusion detection methodology. In the experiment, we ran both the Intrusion Detection System individually first and then together and recorded the data from time to time. From the data we can conclude that the resulting IDS are much better in detecting intrusions from the existing IDSs.Keywords: security, intrusion detection, intrusion prevention, honeypot, anomaly-based detection, signature-based detection, cloud computing, kfsensor
Procedia PDF Downloads 38313862 Computer-Aided Detection of Liver and Spleen from CT Scans using Watershed Algorithm
Authors: Belgherbi Aicha, Bessaid Abdelhafid
Abstract:
In the recent years a great deal of research work has been devoted to the development of semi-automatic and automatic techniques for the analysis of abdominal CT images. The first and fundamental step in all these studies is the semi-automatic liver and spleen segmentation that is still an open problem. In this paper, a semi-automatic liver and spleen segmentation method by the mathematical morphology based on watershed algorithm has been proposed. Our algorithm is currency in two parts. In the first, we seek to determine the region of interest by applying the morphological to extract the liver and spleen. The second step consists to improve the quality of the image gradient. In this step, we propose a method for improving the image gradient to reduce the over-segmentation problem by applying the spatial filters followed by the morphological filters. Thereafter we proceed to the segmentation of the liver, spleen. The aim of this work is to develop a method for semi-automatic segmentation liver and spleen based on watershed algorithm, improve the accuracy and the robustness of the liver and spleen segmentation and evaluate a new semi-automatic approach with the manual for liver segmentation. To validate the segmentation technique proposed, we have tested it on several images. Our segmentation approach is evaluated by comparing our results with the manual segmentation performed by an expert. The experimental results are described in the last part of this work. The system has been evaluated by computing the sensitivity and specificity between the semi-automatically segmented (liver and spleen) contour and the manually contour traced by radiological experts. Liver segmentation has achieved the sensitivity and specificity; sens Liver=96% and specif Liver=99% respectively. Spleen segmentation achieves similar, promising results sens Spleen=95% and specif Spleen=99%.Keywords: CT images, liver and spleen segmentation, anisotropic diffusion filter, morphological filters, watershed algorithm
Procedia PDF Downloads 32513861 Fast Fourier Transform-Based Steganalysis of Covert Communications over Streaming Media
Authors: Jinghui Peng, Shanyu Tang, Jia Li
Abstract:
Steganalysis seeks to detect the presence of secret data embedded in cover objects, and there is an imminent demand to detect hidden messages in streaming media. This paper shows how a steganalysis algorithm based on Fast Fourier Transform (FFT) can be used to detect the existence of secret data embedded in streaming media. The proposed algorithm uses machine parameter characteristics and a network sniffer to determine whether the Internet traffic contains streaming channels. The detected streaming data is then transferred from the time domain to the frequency domain through FFT. The distributions of power spectra in the frequency domain between original VoIP streams and stego VoIP streams are compared in turn using t-test, achieving the p-value of 7.5686E-176 which is below the threshold. The results indicate that the proposed FFT-based steganalysis algorithm is effective in detecting the secret data embedded in VoIP streaming media.Keywords: steganalysis, security, Fast Fourier Transform, streaming media
Procedia PDF Downloads 14713860 A Memetic Algorithm Approach to Clustering in Mobile Wireless Sensor Networks
Authors: Masood Ahmad, Ataul Aziz Ikram, Ishtiaq Wahid
Abstract:
Wireless sensor network (WSN) is the interconnection of mobile wireless nodes with limited energy and memory. These networks can be deployed formany critical applications like military operations, rescue management, fire detection and so on. In flat routing structure, every node plays an equal role of sensor and router. The topology may change very frequently due to the mobile nature of nodes in WSNs. The topology maintenance may produce more overhead messages. To avoid topology maintenance overhead messages, an optimized cluster based mobile wireless sensor network using memetic algorithm is proposed in this paper. The nodes in this network are first divided into clusters. The cluster leaders then transmit data to that base station. The network is validated through extensive simulation study. The results show that the proposed technique has superior results compared to existing techniques.Keywords: WSN, routing, cluster based, meme, memetic algorithm
Procedia PDF Downloads 48113859 The Changes in Motivations and the Use of Translation Strategies in Crowdsourced Translation: A Case Study on Global Voices’ Chinese Translation Project
Authors: Ya-Mei Chen
Abstract:
Online crowdsourced translation, an innovative translation practice brought by Web 2.0 technologies and the democratization of information, has become increasingly popular in the Internet era. Carried out by grass-root internet users, crowdsourced translation contains fundamentally different features from its off-line traditional counterpart, such as voluntary participation and parallel collaboration. To better understand such a participatory and collaborative nature, this paper will use the online Chinese translation project of Global Voices as a case study to investigate the following issues: (1) the changes in volunteer translators’ and reviewers’ motivations for participation, (2) translators’ and reviewers’ use of translation strategies and (3) the correlations of translators’ and reviewers’ motivations and strategies with the organizational mission, the translation style guide, the translator-reviewer interaction, the mediation of the translation platform and various types of capital within the translation field. With an aim to systematically explore the above three issues, this paper will collect both quantitative and qualitative data and then draw upon Engestrom’s activity theory and Bourdieu’s field theory as a theoretical framework to analyze the data in question. An online anonymous questionnaire will be conducted to obtain the quantitative data. The questionnaire will contain questions related to volunteer translators’ and reviewers’ backgrounds, participation motivations, translation strategies and mutual relations as well as the operation of the translation platform. Concerning the qualitative data, they will come from (1) a comparative study between some English news texts published on Global Voices and their Chinese translations, (2) an analysis of the online discussion forum associated with Global Voices’ Chinese translation project and (3) the information about the project’s translation mission and guidelines. It is hoped that this research, through a detailed sociological analysis of a cause-driven crowdsourced translation project, can enable translation researchers and practitioners to adequately meet the translation challenges appearing in the digital age.Keywords: crowdsourced translation, global voices, motivation, translation strategies
Procedia PDF Downloads 37113858 Automatic Generating CNC-Code for Milling Machine
Authors: Chalakorn Chitsaart, Suchada Rianmora, Mann Rattana-Areeyagon, Wutichai Namjaiprasert
Abstract:
G-code is the main factor in computer numerical control (CNC) machine for controlling the tool-paths and generating the profile of the object’s features. For obtaining high surface accuracy of the surface finish, non-stop operation is required for CNC machine. Recently, to design a new product, the strategy that concerns about a change that has low impact on business and does not consume lot of resources has been introduced. Cost and time for designing minor changes can be reduced since the traditional geometric details of the existing models are applied. In order to support this strategy as the alternative channel for machining operation, this research proposes the automatic generating codes for CNC milling operation. Using this technique can assist the manufacturer to easily change the size and the geometric shape of the product during the operation where the time spent for setting up or processing the machine are reduced. The algorithm implemented on MATLAB platform is developed by analyzing and evaluating the geometric information of the part. Codes are created rapidly to control the operations of the machine. Comparing to the codes obtained from CAM, this developed algorithm can shortly generate and simulate the cutting profile of the part.Keywords: geometric shapes, milling operation, minor changes, CNC Machine, G-code, cutting parameters
Procedia PDF Downloads 34913857 Self-Regenerating, Vascularizing Hybrid Scaffold-Hydrogel For Bone Tissue Engineering
Authors: Alisha Gupta
Abstract:
Osteoarthritis (OA) is the most common form of arthritis which is a degenerative joint disease causing joints to begin to break down and underlying bones to change. This “wear and tear” most frequently affects hands, hips, and knees. This is important because OA pain is considered to be a leading cause of mobility impairment in older adults, with hip and knee OA ranked 11th highest contributors to global disability. Bone tissue engineering utilizing polymer scaffolds and hydrogels is an emerging field for treating osteoarthritis. Polymer scaffolds provide a three-dimensional structure for tissue growth, and hydrogels can be used to deliver drugs and growth factors. The combination of the two materials creates a hybrid structure that can better withstand physiological and mechanical demands while also providing a more controlled environment for drug and nutrient delivery. I think using bone tissue engineering for making scaffold-hydrogel composites that are self-regenerating and vascularizing might be useful in solving this problem. Successful implementation can reconstruct healthy, simulated bone tissue on deficient applicants.Keywords: tissue engineering, regenerative medicine, scaffold-hydrogel composites, osteoarthritis
Procedia PDF Downloads 11913856 A Parallel Approach for 3D-Variational Data Assimilation on GPUs in Ocean Circulation Models
Authors: Rossella Arcucci, Luisa D'Amore, Simone Celestino, Giuseppe Scotti, Giuliano Laccetti
Abstract:
This work is the first dowel in a rather wide research activity in collaboration with Euro Mediterranean Center for Climate Changes, aimed at introducing scalable approaches in Ocean Circulation Models. We discuss designing and implementation of a parallel algorithm for solving the Variational Data Assimilation (DA) problem on Graphics Processing Units (GPUs). The algorithm is based on the fully scalable 3DVar DA model, previously proposed by the authors, which uses a Domain Decomposition approach (we refer to this model as the DD-DA model). We proceed with an incremental porting process consisting of 3 distinct stages: requirements and source code analysis, incremental development of CUDA kernels, testing and optimization. Experiments confirm the theoretic performance analysis based on the so-called scale up factor demonstrating that the DD-DA model can be suitably mapped on GPU architectures.Keywords: data assimilation, GPU architectures, ocean models, parallel algorithm
Procedia PDF Downloads 41213855 Exhaustive Study of Essential Constraint Satisfaction Problem Techniques Based on N-Queens Problem
Authors: Md. Ahsan Ayub, Kazi A. Kalpoma, Humaira Tasnim Proma, Syed Mehrab Kabir, Rakib Ibna Hamid Chowdhury
Abstract:
Constraint Satisfaction Problem (CSP) is observed in various applications, i.e., scheduling problems, timetabling problems, assignment problems, etc. Researchers adopt a CSP technique to tackle a certain problem; however, each technique follows different approaches and ways to solve a problem network. In our exhaustive study, it has been possible to visualize the processes of essential CSP algorithms from a very concrete constraint satisfaction example, NQueens Problem, in order to possess a deep understanding about how a particular constraint satisfaction problem will be dealt with by our studied and implemented techniques. Besides, benchmark results - time vs. value of N in N-Queens - have been generated from our implemented approaches, which help understand at what factor each algorithm produces solutions; especially, in N-Queens puzzle. Thus, extended decisions can be made to instantiate a real life problem within CSP’s framework.Keywords: arc consistency (AC), backjumping algorithm (BJ), backtracking algorithm (BT), constraint satisfaction problem (CSP), forward checking (FC), least constrained values (LCV), maintaining arc consistency (MAC), minimum remaining values (MRV), N-Queens problem
Procedia PDF Downloads 36413854 Attribute Analysis of Quick Response Code Payment Users Using Discriminant Non-negative Matrix Factorization
Authors: Hironori Karachi, Haruka Yamashita
Abstract:
Recently, the system of quick response (QR) code is getting popular. Many companies introduce new QR code payment services and the services are competing with each other to increase the number of users. For increasing the number of users, we should grasp the difference of feature of the demographic information, usage information, and value of users between services. In this study, we conduct an analysis of real-world data provided by Nomura Research Institute including the demographic data of users and information of users’ usages of two services; LINE Pay, and PayPay. For analyzing such data and interpret the feature of them, Nonnegative Matrix Factorization (NMF) is widely used; however, in case of the target data, there is a problem of the missing data. EM-algorithm NMF (EMNMF) to complete unknown values for understanding the feature of the given data presented by matrix shape. Moreover, for comparing the result of the NMF analysis of two matrices, there is Discriminant NMF (DNMF) shows the difference of users features between two matrices. In this study, we combine EMNMF and DNMF and also analyze the target data. As the interpretation, we show the difference of the features of users between LINE Pay and Paypay.Keywords: data science, non-negative matrix factorization, missing data, quality of services
Procedia PDF Downloads 13113853 Effect of Fibres-Chemical Treatment on the Thermal Properties of Natural Composites
Authors: J. S. S. Neto, R. A. A. Lima, D. K. K. Cavalcanti, J. P. B. Souza, R. A. A. Aguiar, M. D. Banea
Abstract:
In the last decade, investments in sustainable processes and products have gained space in several segments, such as in the civil, automobile, textile and other industries. In addition to increasing concern about the development of environmentally friendly materials that reduce, energy costs and reduces environmental impact in the production of these products, as well as reducing CO2 emissions. Natural fibers offer a great alternative to replace synthetic fibers, totally or partially, because of their low cost and their renewable source. The purpose of this research is to study the effect of surface chemical treatment on the thermal properties of hybrid fiber reinforced natural fibers (NFRC), jute + ramie, jute + sisal, jute + curauá, and jute fiber in polymer matrices. Two types of chemical treatment: alkalinization and silanization were employed, besides the condition without treatment. Differential scanning calorimetry (DSC), thermogravimetry (TG) and dynamic-mechanical analysis (DMA) were performed to explore the thermal stability and weight loss in the natural fiber reinforced composite as a function of chemical treatment.Keywords: chemical treatment, hybrid composite, jute, thermal
Procedia PDF Downloads 30813852 Estimating Evapotranspiration Irrigated Maize in Brazil Using a Hybrid Modelling Approach and Satellite Image Inputs
Authors: Ivo Zution Goncalves, Christopher M. U. Neale, Hiran Medeiros, Everardo Mantovani, Natalia Souza
Abstract:
Multispectral and thermal infrared imagery from satellite sensors coupled with climate and soil datasets were used to estimate evapotranspiration and biomass in center pivots planted to maize in Brazil during the 2016 season. The hybrid remote sensing based model named Spatial EvapoTranspiration Modelling Interface (SETMI) was applied using multispectral and thermal infrared imagery from the Landsat Thematic Mapper instrument. Field data collected by the IRRIGER center pivot management company included daily weather information such as maximum and minimum temperature, precipitation, relative humidity for estimating reference evapotranspiration. In addition, soil water content data were obtained every 0.20 m in the soil profile down to 0.60 m depth throughout the season. Early season soil samples were used to obtain water-holding capacity, wilting point, saturated hydraulic conductivity, initial volumetric soil water content, layer thickness, and saturated volumetric water content. Crop canopy development parameters and irrigation application depths were also inputs of the model. The modeling approach is based on the reflectance-based crop coefficient approach contained within the SETMI hybrid ET model using relationships developed in Nebraska. The model was applied to several fields located in Minas Gerais State in Brazil with approximate latitude: -16.630434 and longitude: -47.192876. The model provides estimates of real crop evapotranspiration (ET), crop irrigation requirements and all soil water balance outputs, including biomass estimation using multi-temporal satellite image inputs. An interpolation scheme based on the growing degree-day concept was used to model the periods between satellite inputs, filling the gaps between image dates and obtaining daily data. Actual and accumulated ET, accumulated cold temperature and water stress and crop water requirements estimated by the model were compared with data measured at the experimental fields. Results indicate that the SETMI modeling approach using data assimilation, showed reliable daily ET and crop water requirements for maize, interpolated between remote sensing observations, confirming the applicability of the SETMI model using new relationships developed in Nebraska for estimating mainly ET and water requirements in Brazil under tropical conditions.Keywords: basal crop coefficient, irrigation, remote sensing, SETMI
Procedia PDF Downloads 14013851 Arabic Light Stemmer for Better Search Accuracy
Authors: Sahar Khedr, Dina Sayed, Ayman Hanafy
Abstract:
Arabic is one of the most ancient and critical languages in the world. It has over than 250 million Arabic native speakers and more than twenty countries having Arabic as one of its official languages. In the past decade, we have witnessed a rapid evolution in smart devices, social network and technology sector which led to the need to provide tools and libraries that properly tackle the Arabic language in different domains. Stemming is one of the most crucial linguistic fundamentals. It is used in many applications especially in information extraction and text mining fields. The motivation behind this work is to enhance the Arabic light stemmer to serve the data mining industry and leverage it in an open source community. The presented implementation works on enhancing the Arabic light stemmer by utilizing and enhancing an algorithm that provides an extension for a new set of rules and patterns accompanied by adjusted procedure. This study has proven a significant enhancement for better search accuracy with an average 10% improvement in comparison with previous works.Keywords: Arabic data mining, Arabic Information extraction, Arabic Light stemmer, Arabic stemmer
Procedia PDF Downloads 30813850 Aerodynamic Interference of Propellers Group with Adjustable Mutual Position
Authors: Michal Biały, Krzysztof Skiba, Zdzislaw Kaminski
Abstract:
The research results of the influence of the adjustable mutual position of the propellers for getting optimal lift force on a specially designed bench. The bench consists of frame with electric motors and with attached propellers. Engines were arranged in a matrix of two columns and three rows. The distance between the columns averages from 0 to 20”, while the engine was placed at a height of 8”, 15.5” and 23.6”. By adjusting the tilt of an electric motor, an angle of the propeller in the range of 0° to 60°, by 15° was controlled. Propellers with a diameter of 8" and pitch of 4.5” were driven by brushless model engines Roxxy BL-Outrunner 2827/26 with a power of 110W (each). Rotational speed control of electric motors were realized parallel for all propellers. The speed adjustment was realized using an aggregate of radio-controlled regulators. Electric power supplied to the engines from zero to maximum power, by the setting for every 14W, was controlled by radio system. Measurement system was placed on a laboratory scale. The lift was measured and recorded by an electronic scale. The lift force for different configurations of propellers arrangement was recorded during the test. All propellers were driven in one rotational direction and in different directions when they were in the same pairs. Propellers were driven concurrently and contra-concurrently along one of the columns and along the selected rows. During the tests, except the lift, parameters such as: rotational speed of propellers, voltage and current to the electric engines were recorded. The main aim of the research was to show the influence of aerodynamic interference between the propellers to receive lift force depending on the drive configuration of individual propellers. The research has shown that, this interference exists. The increase of the lift force for a distance between columns above 26.6” was noticed during the driving propellers in different directions. The optimum tilt angle of the propeller was 45°. Furthermore there has been also approx. 12% increase of the lift for propellers driven alternately in column and contra-concurrently in relation to the contra-rotating drive in the row.Keywords: aerodynamic, interference, lift force, propeller, propulsion system
Procedia PDF Downloads 34413849 Developing a DNN Model for the Production of Biogas From a Hybrid BO-TPE System in an Anaerobic Wastewater Treatment Plant
Authors: Hadjer Sadoune, Liza Lamini, Scherazade Krim, Amel Djouadi, Rachida Rihani
Abstract:
Deep neural networks are highly regarded for their accuracy in predicting intricate fermentation processes. Their ability to learn from a large amount of datasets through artificial intelligence makes them particularly effective models. The primary obstacle in improving the performance of these models is to carefully choose the suitable hyperparameters, including the neural network architecture (number of hidden layers and hidden units), activation function, optimizer, learning rate, and other relevant factors. This study predicts biogas production from real wastewater treatment plant data using a sophisticated approach: hybrid Bayesian optimization with a tree-structured Parzen estimator (BO-TPE) for an optimised deep neural network (DNN) model. The plant utilizes an Upflow Anaerobic Sludge Blanket (UASB) digester that treats industrial wastewater from soft drinks and breweries. The digester has a working volume of 1574 m3 and a total volume of 1914 m3. Its internal diameter and height were 19 and 7.14 m, respectively. The data preprocessing was conducted with meticulous attention to preserving data quality while avoiding data reduction. Three normalization techniques were applied to the pre-processed data (MinMaxScaler, RobustScaler and StandardScaler) and compared with the Non-Normalized data. The RobustScaler approach has strong predictive ability for estimating the volume of biogas produced. The highest predicted biogas volume was 2236.105 Nm³/d, with coefficient of determination (R2), mean absolute error (MAE), and root mean square error (RMSE) values of 0.712, 164.610, and 223.429, respectively.Keywords: anaerobic digestion, biogas production, deep neural network, hybrid bo-tpe, hyperparameters tuning
Procedia PDF Downloads 3813848 An Experimental Study on the Effect of Heat Input on the Weld Efficiency of TIG-MIG Hybrid Welding of Type-304 Austenitic Stainless Steel
Authors: Emmanuel Ogundimu, Esther Akinlabi, Mutiu Erinosho
Abstract:
Welding is described as the process of joining metals so that bonding can be created as a result of inter-atomic penetration. This study investigated the influence of heat input on the efficiency of the welded joints of 304 stainless steel. Three welds joint were made from two similar 304 stainless steel plates of thickness 6 mm. The tensile results obtained showed that the maximum average tensile strength of 672 MPa is possessed by the sample A1 with low heat input. It was discovered that the tensile strength, % elongation and weld joint efficiency decreased with the increase in heat input into the weld. The average % elongation for the entire samples ranged from 28.4% to 36.5%. Sample A1 had the highest joint efficiency of 94.5%. However, the optimum welding current of 190 for TIG- MIG hybrid welding of type-304 austenite stainless steel can be recommended for advanced technological applications such as aircraft manufacturing, nuclear industry, automobile industry, and processing industry.Keywords: microhardness, microstructure, tensile, MIG welding, process, tensile, shear stress TIG welding, TIG-MIG welding
Procedia PDF Downloads 19913847 Optimal Allocation of PHEV Parking Lots to Minimize Dstribution System Losses
Authors: Mohsen Mazidi, Ali Abbaspour, Mahmud Fotuhi-Firuzabad, Mohamamd Rastegar
Abstract:
To tackle the air pollution issues, Plug-in Hybrid Electric Vehicles (PHEVs) are proposed as an appropriate solution. Charging a large amount of PHEV batteries, if not controlled, would have negative impacts on the distribution system. The control process of charging of these vehicles can be centralized in parking lots that may provide a chance for better coordination than the individual charging in houses. In this paper, an optimization-based approach is proposed to determine the optimum PHEV parking capacities in candidate nodes of the distribution system. In so doing, a profile for charging and discharging of PHEVs is developed in order to flatten the network load profile. Then, this profile is used in solving an optimization problem to minimize the distribution system losses. The outputs of the proposed method are the proper place for PHEV parking lots and optimum capacity for each parking. The application of the proposed method on the IEEE-34 node test feeder verifies the effectiveness of the method.Keywords: loss, plug-in hybrid electric vehicle (PHEV), PHEV parking lot, V2G
Procedia PDF Downloads 54313846 New Hybrid Method to Model Extreme Rainfalls
Authors: Youness Laaroussi, Zine Elabidine Guennoun, Amine Amar
Abstract:
Modeling and forecasting dynamics of rainfall occurrences constitute one of the major topics, which have been largely treated by statisticians, hydrologists, climatologists and many other groups of scientists. In the same issue, we propose in the present paper a new hybrid method, which combines Extreme Values and fractal theories. We illustrate the use of our methodology for transformed Emberger Index series, constructed basing on data recorded in Oujda (Morocco). The index is treated at first by Peaks Over Threshold (POT) approach, to identify excess observations over an optimal threshold u. In the second step, we consider the resulting excess as a fractal object included in one dimensional space of time. We identify fractal dimension by the box counting. We discuss the prospect descriptions of rainfall data sets under Generalized Pareto Distribution, assured by Extreme Values Theory (EVT). We show that, despite of the appropriateness of return periods given by POT approach, the introduction of fractal dimension provides accurate interpretation results, which can ameliorate apprehension of rainfall occurrences.Keywords: extreme values theory, fractals dimensions, peaks Over threshold, rainfall occurrences
Procedia PDF Downloads 36113845 Using Closed Frequent Itemsets for Hierarchical Document Clustering
Authors: Cheng-Jhe Lee, Chiun-Chieh Hsu
Abstract:
Due to the rapid development of the Internet and the increased availability of digital documents, the excessive information on the Internet has led to information overflow problem. In order to solve these problems for effective information retrieval, document clustering in text mining becomes a popular research topic. Clustering is the unsupervised classification of data items into groups without the need of training data. Many conventional document clustering methods perform inefficiently for large document collections because they were originally designed for relational database. Therefore they are impractical in real-world document clustering and require special handling for high dimensionality and high volume. We propose the FIHC (Frequent Itemset-based Hierarchical Clustering) method, which is a hierarchical clustering method developed for document clustering, where the intuition of FIHC is that there exist some common words for each cluster. FIHC uses such words to cluster documents and builds hierarchical topic tree. In this paper, we combine FIHC algorithm with ontology to solve the semantic problem and mine the meaning behind the words in documents. Furthermore, we use the closed frequent itemsets instead of only use frequent itemsets, which increases efficiency and scalability. The experimental results show that our method is more accurate than those of well-known document clustering algorithms.Keywords: FIHC, documents clustering, ontology, closed frequent itemset
Procedia PDF Downloads 39913844 Effective Affordable Housing Finance in Developing Economies: An Integration of Demand and Supply Solutions
Authors: Timothy Akinwande, Eddie Hui, Karien Dekker
Abstract:
Housing the urban poor remains a persistent challenge, despite evident research attention over many years. It is, therefore, pertinent to investigate affordable housing provision challenges with novel approaches. For innovative solutions to affordable housing constraints, it is apposite to thoroughly examine housing solutions vis a vis the key elements of the housing supply value chain (HSVC), which are housing finance, housing construction and land acquisition. A pragmatic analysis will examine affordable housing solutions from demand and supply perspectives to arrive at consolidated solutions from bilateral viewpoints. This study thoroughly examined informal housing finance strategies of the urban poor and diligently investigated expert opinion on affordable housing finance solutions. The research questions were: (1) What mutual grounds exist between informal housing finance solutions of the urban poor and housing expert solutions to affordable housing finance constraints in developing economies? (2) What are effective approaches to affordable housing finance in developing economies from an integrated demand - supply perspective? Semi-structured interviews were conducted in the 5 largest slums of Lagos, Nigeria, with 40 informal settlers for demand-oriented solutions, while focus group discussion and in-depth interviews were conducted with 12 housing experts in Nigeria for supply-oriented solutions. Following a rigorous thematic, content and descriptive analyses of data using NVivo and Excel, findings ascertained mutual solutions from both demand and supply standpoints that can be consolidated into more effective affordable housing finance solutions in Nigeria. Deliberate finance models that recognise and include the finance realities of the urban poor was found to be the most significant supply-side housing finance solution, representing 25.4% of total expert responses. Findings also show that 100% of sampled urban poor engage in vocations where they earn little irregular income or zero income, limiting their housing finance capacities and creditworthiness. Survey revealed that the urban poor are involved in community savings and employ microfinance institutions within the informal settlements to tackle their housing finance predicaments. These are informal finance models of the urban poor, revealing common grounds between demand and supply solutions for affordable housing financing. Effective, affordable housing approach will be to modify, institutionalise and incorporate the informal finance strategies of the urban poor into deliberate government policies. This consolidation of solutions from demand and supply perspectives can eliminate the persistent misalliance between affordable housing demand and affordable housing supply. This study provides insights into mutual housing solutions from demand and supply perspectives, and findings are informative for effective, affordable housing provision approaches in developing countries. This study is novel in consolidating affordable housing solutions from demand and supply viewpoints, especially in relation to housing finance as a key component of HSVC. The framework for effective, affordable housing finance in developing economies from a consolidated viewpoint generated in this study is significant for the achievement of sustainable development goals, especially goal 11 for sustainable, resilient and inclusive cities. Findings are vital for future housing studies.Keywords: affordable housing, affordable housing finance, developing economies, effective affordable housing, housing policy, urban poor, sustainable development goal, sustainable affordable housing
Procedia PDF Downloads 7013843 Vehicle Timing Motion Detection Based on Multi-Dimensional Dynamic Detection Network
Authors: Jia Li, Xing Wei, Yuchen Hong, Yang Lu
Abstract:
Detecting vehicle behavior has always been the focus of intelligent transportation, but with the explosive growth of the number of vehicles and the complexity of the road environment, the vehicle behavior videos captured by traditional surveillance have been unable to satisfy the study of vehicle behavior. The traditional method of manually labeling vehicle behavior is too time-consuming and labor-intensive, but the existing object detection and tracking algorithms have poor practicability and low behavioral location detection rate. This paper proposes a vehicle behavior detection algorithm based on the dual-stream convolution network and the multi-dimensional video dynamic detection network. In the videos, the straight-line behavior of the vehicle will default to the background behavior. The Changing lanes, turning and turning around are set as target behaviors. The purpose of this model is to automatically mark the target behavior of the vehicle from the untrimmed videos. First, the target behavior proposals in the long video are extracted through the dual-stream convolution network. The model uses a dual-stream convolutional network to generate a one-dimensional action score waveform, and then extract segments with scores above a given threshold M into preliminary vehicle behavior proposals. Second, the preliminary proposals are pruned and identified using the multi-dimensional video dynamic detection network. Referring to the hierarchical reinforcement learning, the multi-dimensional network includes a Timer module and a Spacer module, where the Timer module mines time information in the video stream and the Spacer module extracts spatial information in the video frame. The Timer and Spacer module are implemented by Long Short-Term Memory (LSTM) and start from an all-zero hidden state. The Timer module uses the Transformer mechanism to extract timing information from the video stream and extract features by linear mapping and other methods. Finally, the model fuses time information and spatial information and obtains the location and category of the behavior through the softmax layer. This paper uses recall and precision to measure the performance of the model. Extensive experiments show that based on the dataset of this paper, the proposed model has obvious advantages compared with the existing state-of-the-art behavior detection algorithms. When the Time Intersection over Union (TIoU) threshold is 0.5, the Average-Precision (MP) reaches 36.3% (the MP of baselines is 21.5%). In summary, this paper proposes a vehicle behavior detection model based on multi-dimensional dynamic detection network. This paper introduces spatial information and temporal information to extract vehicle behaviors in long videos. Experiments show that the proposed algorithm is advanced and accurate in-vehicle timing behavior detection. In the future, the focus will be on simultaneously detecting the timing behavior of multiple vehicles in complex traffic scenes (such as a busy street) while ensuring accuracy.Keywords: vehicle behavior detection, convolutional neural network, long short-term memory, deep learning
Procedia PDF Downloads 130