Search results for: total capacity algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15737

Search results for: total capacity algorithm

14987 Estimation of Optimum Parameters of Non-Linear Muskingum Model of Routing Using Imperialist Competition Algorithm (ICA)

Authors: Davood Rajabi, Mojgan Yazdani

Abstract:

Non-linear Muskingum model is an efficient method for flood routing, however, the efficiency of this method is influenced by three applied parameters. Therefore, efficiency assessment of Imperialist Competition Algorithm (ICA) to evaluate optimum parameters of non-linear Muskingum model was addressed through this study. In addition to ICA, Genetic Algorithm (GA) and Particle Swarm Optimization (PSO) were also used aiming at an available criterion to verdict ICA. In this regard, ICA was applied for Wilson flood routing; then, routing of two flood events of DoAab Samsami River was investigated. In case of Wilson flood that the target function was considered as the sum of squared deviation (SSQ) of observed and calculated discharges. Routing two other floods, in addition to SSQ, another target function was also considered as the sum of absolute deviations of observed and calculated discharge. For the first floodwater based on SSQ, GA indicated the best performance, however, ICA was on first place, based on SAD. For the second floodwater, based on both target functions, ICA indicated a better operation. According to the obtained results, it can be said that ICA could be used as an appropriate method to evaluate the parameters of Muskingum non-linear model.

Keywords: Doab Samsami river, genetic algorithm, imperialist competition algorithm, meta-exploratory algorithms, particle swarm optimization, Wilson flood

Procedia PDF Downloads 505
14986 Effect of Reinforcement Density on the Behaviour of Reinforced Sand Under a Square Footing

Authors: Dhyaalddin Bahaalddin Noori Zangana

Abstract:

This study involves the behavior of reinforced sand under a square footing. A series of bearing capacity tests were performed on a small-scale laboratory model, which filled with a poorly-graded homogenous bed of sand, which was placed in a medium dense state using sand raining technique. The sand was reinforced with 40 mm wide household aluminum foil strips. The main studied parameters was to consider the effect of reinforcing strip length, with various linear density of reinforcement, number of reinforcement layers and depth of top layer of reinforcement below the footing, on load-settlement behavior, bearing capacity ratio and settlement reduction factor. The relation of load-settlement generally showed similar trend in all the tests. Failure was defined as settlement equal to 10% of the footing width. The recommended optimum reinforcing strip length, linear density of reinforcement, number of reinforcement layers and depth of top layer of reinforcing strips that give the maximum bearing capacity improvement and minimum settlement reduction factor were presented and discussed. Different bearing capacity ration versus length of the reinforcing strips and settlement reduction factor versus length of the reinforcing strips relations at failure were showed improvement of bearing capacity ratio by a factor of 3.82 and reduction of settlement reduction factor by a factor of 0.813. The optimum length of reinforcement was found to be 7.5 times the footing width.

Keywords: square footing, relative density, linear density of reinforcement, bearing capacity ratio, load-settlement behaviour

Procedia PDF Downloads 98
14985 New Segmentation of Piecewise Moving-Average Model by Using Reversible Jump MCMC Algorithm

Authors: Suparman

Abstract:

This paper addresses the problem of the signal segmentation within a Bayesian framework by using reversible jump MCMC algorithm. The signal is modelled by piecewise constant Moving-Average (MA) model where the numbers of segments, the position of change-point, the order and the coefficient of the MA model for each segment are unknown. The reversible jump MCMC algorithm is then used to generate samples distributed according to the joint posterior distribution of the unknown parameters. These samples allow calculating some interesting features of the posterior distribution. The performance of the methodology is illustrated via several simulation results.

Keywords: piecewise, moving-average model, reversible jump MCMC, signal segmentation

Procedia PDF Downloads 227
14984 Total Quality Management in Companies Manufacturing

Authors: Malki Nadia Fatima Zahra, Kellal Cheimaa, Brahimi Houria

Abstract:

Aim of the study is to show the role of total Quality Management on firm performance; the research relied on the views of sample managers working in the Marinel pharmaceutical company. The research aims to achieve many objectives, including increasing awareness of the concepts of Total Quality Management on Firm Performance, especially in the manufacturing firm, providing a future vision of the possibility of success, and the actual application of the Principles of Total Quality Management in the manufacturing company. The research adopted a default model was built after a review and analysis of the literature review in the context of one hypothesis main points at the origin of a group of sub-hypotheses. The research presented a set of conclusions, and the most important of these conclusions was there is a relationship between the Principles of TQM and Firm Performance.

Keywords: total quality management, TQM dimension, firm performance, strategies

Procedia PDF Downloads 69
14983 Algorithmic Approach to Management of Complications of Permanent Facial Filler: A Saudi Experience

Authors: Luay Alsalmi

Abstract:

Background: Facial filler is the most common type of cosmetic surgery next to botox. Permanent filler is preferred nowadays due to the low cost brought about by non-recurring injection appointments. However, such fillers pose a higher risk for complications, with even greater adverse effects when the procedure is done using unknown dermal filler injections. AIM: This study aimed to establish an algorithm to categorize and manage patients that receive permanent fillers. Materials and Methods: Twelve participants were presented to the service through emergency or as outpatient from November 2015 to May 2021. Demographics such as age, sex, date of injection, time of onset, and types of complications were collected. After examination, all cases were managed based on an algorithm established. FACE-Q was used to measure overall satisfaction and psychological well-being. Results: The algorithm to diagnose and manage these patients effectively with a high satisfaction rate was established in this study. All participants were non-smoker females with no known medical comorbidities. The algorithm presented determined the treatment plan when faced with complications. Results revealed high appearance-related psychosocial distress was observed prior to surgery, while it significantly dropped after surgery. FACE-Q was able to establish evidence of satisfactory ratings among patients prior to and after surgery. Conclusion: This treatment algorithm can guide the surgeon in formulating a suitable plan with fewer complications and a high satisfaction rate.

Keywords: facial filler, FACE-Q, psycho-social stress, botox, treatment algorithm

Procedia PDF Downloads 84
14982 Commissioning of a Flattening Filter Free (FFF) using an Anisotropic Analytical Algorithm (AAA)

Authors: Safiqul Islam, Anamul Haque, Mohammad Amran Hossain

Abstract:

Aim: To compare the dosimetric parameters of the flattened and flattening filter free (FFF) beam and to validate the beam data using anisotropic analytical algorithm (AAA). Materials and Methods: All the dosimetric data’s (i.e. depth dose profiles, profile curves, output factors, penumbra etc.) required for the beam modeling of AAA were acquired using the Blue Phantom RFA for 6 MV, 6 FFF, 10MV & 10FFF. Progressive resolution Optimizer and Dose Volume Optimizer algorithm for VMAT and IMRT were are also configured in the beam model. Beam modeling of the AAA were compared with the measured data sets. Results: Due to the higher and lover energy component in 6FFF and 10 FFF the surface doses are 10 to 15% higher compared to flattened 6 MV and 10 MV beams. FFF beam has a lower mean energy compared to the flattened beam and the beam quality index were 6 MV 0.667, 6FFF 0.629, 10 MV 0.74 and 10 FFF 0.695 respectively. Gamma evaluation with 2% dose and 2 mm distance criteria for the Open Beam, IMRT and VMAT plans were also performed and found a good agreement between the modeled and measured data. Conclusion: We have successfully modeled the AAA algorithm for the flattened and FFF beams and achieved a good agreement with the calculated and measured value.

Keywords: commissioning of a Flattening Filter Free (FFF) , using an Anisotropic Analytical Algorithm (AAA), flattened beam, parameters

Procedia PDF Downloads 301
14981 Diesel Fault Prediction Based on Optimized Gray Neural Network

Authors: Han Bing, Yin Zhenjie

Abstract:

In order to analyze the status of a diesel engine, as well as conduct fault prediction, a new prediction model based on a gray system is proposed in this paper, which takes advantage of the neural network and the genetic algorithm. The proposed GBPGA prediction model builds on the GM (1.5) model and uses a neural network, which is optimized by a genetic algorithm to construct the error compensator. We verify our proposed model on the diesel faulty simulation data and the experimental results show that GBPGA has the potential to employ fault prediction on diesel.

Keywords: fault prediction, neural network, GM(1, 5) genetic algorithm, GBPGA

Procedia PDF Downloads 305
14980 Comparison between the Conventional Methods and PSO Based MPPT Algorithm for Photovoltaic Systems

Authors: Ramdan B. A. Koad, Ahmed F. Zobaa

Abstract:

Since the output characteristics of Photovoltaic (PV) system depends on the ambient temperature, solar radiation and load impedance, its maximum Power Point (MPP) is not constant. Under each condition PV module has a point at which it can produce its MPP. Therefore, a Maximum Power Point Tracking (MPPT) method is needed to uphold the PV panel operating at its MPP. This paper presents comparative study between the conventional MPPT methods used in (PV) system: Perturb and Observe (P&O), Incremental Conductance (IncCond), and Particle Swarm Optimization (PSO) algorithm for (MPPT) of (PV) system. To evaluate the study, the proposed PSO MPPT is implemented on a DC-DC converter and has been compared with P&O and INcond methods in terms of their tracking speed, accuracy and performance by using the Matlab tool Simulink. The simulation result shows that the proposed algorithm is simple, and is superior to the P&O and IncCond methods.

Keywords: photovoltaic systems, maximum power point tracking, perturb and observe method, incremental conductance, methods and practical swarm optimization algorithm

Procedia PDF Downloads 358
14979 Shear Strength Evaluation of Ultra-High-Performance Concrete Flexural Members Using Adaptive Neuro-Fuzzy System

Authors: Minsu Kim, Hae-Chang Cho, Jae Hoon Chung, Inwook Heo, Kang Su Kim

Abstract:

For safe design of the UHPC flexural members, accurate estimations of their shear strengths are very important. However, since the shear strengths are significantly affected by various factors such as tensile strength of concrete, shear span to depth ratio, volume ratio of steel fiber, and steel fiber factor, the accurate estimations of their shear strengths are very challenging. In this study, therefore, the Adaptive Neuro-Fuzzy System (ANFIS), which has been widely used to solve many complex problems in engineering fields, was introduced to estimate the shear strengths of UHPC flexural members. A total of 32 experimental results has been collected from previous studies for training of the ANFIS algorithm, and the well-trained ANFIS algorithm provided good estimations on the shear strengths of the UHPC test specimens. Acknowledgement: This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Science, ICT & Future Planning(NRF-2016R1A2B2010277).

Keywords: ultra-high-performance concrete, ANFIS, shear strength, flexural member

Procedia PDF Downloads 188
14978 Challenges for the Implementation of Community Led Total Sanitation in Rural Malawi

Authors: Save Kumwenda, Khumbo Kalulu, Kondwani Chidziwisano, Limbani Kalumbi, Vincent Doyle, Bagrey Ngwira

Abstract:

Introduction: The Malawi Government in partnership with Non-Governmental Organizations adopted Community Led Total Sanitation (CLTS) in 2008 as an approach in sanitation and hygiene promotion with an aim of declaring Malawi Open Defeacation Free (ODF) by 2015. While there is a significant body of research into CLTS available in public domain, there is little research done on challenges faced in implementing CLTS in Malawi. Methods: A cross-sectional qualitative study was carried out in three districts of Ntcheu, Balaka, and Phalombe. Data was collected using Focus Group Discussions (FGDs) and Key informant interviews (KII) and analysed manually. Results: In total, 96 people took part in FGDs and 9 people in KII. It was shown that choice of leaders after triggering was commonly done by chiefs, facilitators, and VHC without following CLTS principles as opposed to identifying individuals who showed leadership skills. Despite capacity building initiatives involving District Coordinating Teams, lack of resources to undertake follow-ups contributed to failure to sustain ODF in the community. It was also found that while most respondents appreciating the need for no subsidies, the elderly and those with disabilities felt the need for external support because do not have money for buying strong logs, slabs for durable toilet floor and also to hire people to build latrines for them. Conclusion: Effective implementation of CLTS requires comprehensive consideration of various issues that may affect its success.

Keywords: open defecation, community-led, sanitation, faecal matter, hygiene, Malawi

Procedia PDF Downloads 381
14977 An Assessment of the Impacts of Agro-Ecological Practices towards the Improvement of Crop Health and Yield Capacity: A Case of Mopani District, Limpopo, South Africa

Authors: Tshilidzi C. Manyanya, Nthaduleni S. Nethengwe, Edmore Kori

Abstract:

The UNFCCC, FAO, GCF, IPCC and other global structures advocate for agro-ecology do address food security and sovereignty. However, most of the expected outcomes concerning agro-ecological were not empirically tested for universal application. Agro-ecology is theorised to increase crop health over ago-ecological farms and decrease over conventional farms. Increased crop health means increased carbon sequestration and thus less CO2 in the atmosphere. This is in line with the view that global warming is anthropogenically enhanced through GHG emissions. Agro-ecology mainly affects crop health, soil carbon content and yield on the cultivated land. Economic sustainability is directly related to yield capacity, which is theorized to increase by 3-10% in a space of 3 - 10 years as a result of agro-ecological implementation. This study aimed to empirically assess the practicality and validity of these assumptions. The study utilized mainly GIS and RS techniques to assess the effectiveness of agro-ecology in crop health improvement from satellite images. The assessment involved a longitudinal study (2013 – 2015) assessing the changes that occur after a farm retrofits from conventional agriculture to agro-ecology. The assumptions guided the objectives of the study. For each objective, an agro-ecological farm was compared with a conventional farm in the same climatic conditional occupying the same general location. Crop health was assessed using satellite images analysed through ArcGIS and Erdas. This entailed the production of NDVI and Re-classified outputs of the farm area. The NDVI ranges of the entire period of study were thus compared in a stacked histogram for each farm to assess for trends. Yield capacity was calculated based on the production records acquired from the farmers and plotted in a stacked bar graph as percentages of a total for each farm. The results of the study showed decreasing crop health trends over 80% of the conventional farms and an increase over 80% of the organic farms. Yield capacity showed similar patterns to those of crop health. The study thus showed that agro-ecology is an effective strategy for crop-health improvement and yield increase.

Keywords: agro-ecosystem, conventional farm, dialectical, sustainability

Procedia PDF Downloads 216
14976 Characterization of Onion Peels Extracts and Its Utilization in a Deep Fried Snack

Authors: Nabia Siddiqui, Tahira Mohsin Ali, Tanveer Abbas, Abid Hasnain

Abstract:

The present study proposed the use of different onion peel extracts in a South Asian snacks called ‘sew’. The polyphenols extracted from peels were initially analyzed for their antimicrobial potential and bioactive components following three different extraction systems. A relatively higher level of total phenolic content (TP), total flavonoid (TF) and antioxidant activity was observed for EWE (ethanol and water based) extracts followed by EAAE (ethanol and acetic acid) and WE (water extract) sample. Onion extracts showed ability to inhibit gram-positive as well as gram-negative bacteria. The incorporation of onion peel extracts in sew showed a marked increase in bioactive components. Besides bioactivity, sensory attributes, textural characteristics and storage stability of these snacks containing onion peel extract also significantly improved during the shelf study at ambient temperature for up to two months. Thus, these results justify the utilization of these plant polyphenols in fried snacks.

Keywords: onion peels extract, South Asian snacks, antioxidant capacity, bioactivity

Procedia PDF Downloads 244
14975 Design and Implementation of a Counting and Differentiation System for Vehicles through Video Processing

Authors: Derlis Gregor, Kevin Cikel, Mario Arzamendia, Raúl Gregor

Abstract:

This paper presents a self-sustaining mobile system for counting and classification of vehicles through processing video. It proposes a counting and classification algorithm divided in four steps that can be executed multiple times in parallel in a SBC (Single Board Computer), like the Raspberry Pi 2, in such a way that it can be implemented in real time. The first step of the proposed algorithm limits the zone of the image that it will be processed. The second step performs the detection of the mobile objects using a BGS (Background Subtraction) algorithm based on the GMM (Gaussian Mixture Model), as well as a shadow removal algorithm using physical-based features, followed by morphological operations. In the first step the vehicle detection will be performed by using edge detection algorithms and the vehicle following through Kalman filters. The last step of the proposed algorithm registers the vehicle passing and performs their classification according to their areas. An auto-sustainable system is proposed, powered by batteries and photovoltaic solar panels, and the data transmission is done through GPRS (General Packet Radio Service)eliminating the need of using external cable, which will facilitate it deployment and translation to any location where it could operate. The self-sustaining trailer will allow the counting and classification of vehicles in specific zones with difficult access.

Keywords: intelligent transportation system, object detection, vehicle couting, vehicle classification, video processing

Procedia PDF Downloads 323
14974 Algorithm Optimization to Sort in Parallel by Decreasing the Number of the Processors in SIMD (Single Instruction Multiple Data) Systems

Authors: Ali Hosseini

Abstract:

Paralleling is a mechanism to decrease the time necessary to execute the programs. Sorting is one of the important operations to be used in different systems in a way that the proper function of many algorithms and operations depend on sorted data. CRCW_SORT algorithm executes ‘N’ elements sorting in O(1) time on SIMD (Single Instruction Multiple Data) computers with n^2/2-n/2 number of processors. In this article having presented a mechanism by dividing the input string by the hinge element into two less strings the number of the processors to be used in sorting ‘N’ elements in O(1) time has decreased to n^2/8-n/4 in the best state; by this mechanism the best state is when the hinge element is the middle one and the worst state is when it is minimum. The findings from assessing the proposed algorithm by other methods on data collection and number of the processors indicate that the proposed algorithm uses less processors to sort during execution than other methods.

Keywords: CRCW, SIMD (Single Instruction Multiple Data) computers, parallel computers, number of the processors

Procedia PDF Downloads 310
14973 Instructional Leadership and Competency in Capacity Development among Principals: A Mediation with Self Efficacy in Moderate Performing Schools

Authors: Mohd Ibrahim K. Azeez, Mohammed Sani Ibrahim, Rosemawati Mustapa, Maisarah A. Malik, Chandrakala Varatharajoo, Wee Akina Sia Seng Lee

Abstract:

The leadership of the principals is known to be a key indicator in development and school performance. Therefore, this study was undertaken to assess the extent of the influence of instructional leadership in the field of supervision and curriculum focus on capacity development competence in the field of communication and teamwork. In addition, this study also examines self-efficacy mediator school leadership in the field of self-improvement and self-management of school principals. The study involved 383 guest teachers from 55 secondary schools for leadership in schools. Data was analyzed using SEM aid program AMOS 21. The final result shows partial mediation model was the best model fit to obtain the best goodness of fit of (X2/df = 4.663, CFI = 0.922, GFI = 0.778, TLI = 0914, NFI = 0.903, and RMSEA = 0.098) compared to the direct effect model of the findings (X2/df = 5.319, CFI = 0.908, GFI = 0755, TLI = 0.899, NFI = 0.889, and RMSEA = 0.106). While the findings of the fully mediator model with a self-efficacy refers principals as a mediator as follows (X2/df = 4.838, CFI = 0918, GFI = 0772, TLI = 0.910, NFI = 0.899, and RMSEA = 0.100). Therefore, it can be concluded that the findings clearly demonstrate self-efficacy variables principals become a mediator in the relationship between instructional leadership capacity and competency development.

Keywords: instructional leadership, capacity development, self-efficacy, competency

Procedia PDF Downloads 725
14972 Efficient Motion Estimation by Fast Three Step Search Algorithm

Authors: S. M. Kulkarni, D. S. Bormane, S. L. Nalbalwar

Abstract:

The rapid development in the technology have dramatic impact on the medical health care field. Medical data base obtained with latest machines like CT Machine, MRI scanner requires large amount of memory storage and also it requires large bandwidth for transmission of data in telemedicine applications. Thus, there is need for video compression. As the database of medical images contain number of frames (slices), hence while coding of these images there is need of motion estimation. Motion estimation finds out movement of objects in an image sequence and gets motion vectors which represents estimated motion of object in the frame. In order to reduce temporal redundancy between successive frames of video sequence, motion compensation is preformed. In this paper three step search (TSS) block matching algorithm is implemented on different types of video sequences. It is shown that three step search algorithm produces better quality performance and less computational time compared with exhaustive full search algorithm.

Keywords: block matching, exhaustive search motion estimation, three step search, video compression

Procedia PDF Downloads 491
14971 The Parallelization of Algorithm Based on Partition Principle for Association Rules Discovery

Authors: Khadidja Belbachir, Hafida Belbachir

Abstract:

subsequently the expansion of the physical supports storage and the needs ceaseless to accumulate several data, the sequential algorithms of associations’ rules research proved to be ineffective. Thus the introduction of the new parallel versions is imperative. We propose in this paper, a parallel version of a sequential algorithm “Partition”. This last is fundamentally different from the other sequential algorithms, because it scans the data base only twice to generate the significant association rules. By consequence, the parallel approach does not require much communication between the sites. The proposed approach was implemented for an experimental study. The obtained results, shows a great reduction in execution time compared to the sequential version and Count Distributed algorithm.

Keywords: association rules, distributed data mining, partition, parallel algorithms

Procedia PDF Downloads 416
14970 Innovative Fabric Integrated Thermal Storage Systems and Applications

Authors: Ahmed Elsayed, Andrew Shea, Nicolas Kelly, John Allison

Abstract:

In northern European climates, domestic space heating and hot water represents a significant proportion of total primary total primary energy use and meeting these demands from a national electricity grid network supplied by renewable energy sources provides an opportunity for a significant reduction in EU CO2 emissions. However, in order to adapt to the intermittent nature of renewable energy generation and to avoid co-incident peak electricity usage from consumers that may exceed current capacity, the demand for heat must be decoupled from its generation. Storage of heat within the fabric of dwellings for use some hours, or days, later provides a route to complete decoupling of demand from supply and facilitates the greatly increased use of renewable energy generation into a local or national electricity network. The integration of thermal energy storage into the building fabric for retrieval at a later time requires much evaluation of the many competing thermal, physical, and practical considerations such as the profile and magnitude of heat demand, the duration of storage, charging and discharging rate, storage media, space allocation, etc. In this paper, the authors report investigations of thermal storage in building fabric using concrete material and present an evaluation of several factors that impact upon performance including heating pipe layout, heating fluid flow velocity, storage geometry, thermo-physical material properties, and also present an investigation of alternative storage materials and alternative heat transfer fluids. Reducing the heating pipe spacing from 200 mm to 100 mm enhances the stored energy by 25% and high-performance Vacuum Insulation results in heat loss flux of less than 3 W/m2, compared to 22 W/m2 for the more conventional EPS insulation. Dense concrete achieved the greatest storage capacity, relative to medium and light-weight alternatives, although a material thickness of 100 mm required more than 5 hours to charge fully. Layers of 25 mm and 50 mm thickness can be charged in 2 hours, or less, facilitating a fast response that could, aggregated across multiple dwellings, provide significant and valuable reduction in demand from grid-generated electricity in expected periods of high demand and potentially eliminate the need for additional new generating capacity from conventional sources such as gas, coal, or nuclear.

Keywords: fabric integrated thermal storage, FITS, demand side management, energy storage, load shifting, renewable energy integration

Procedia PDF Downloads 166
14969 Quality Evaluation of Grape Seed Oils of the Ionian Islands Based on GC-MS and Other Spectroscopic Techniques

Authors: I. Oikonomou, I. Lappa, D. Daferera, C. Kanakis, L. Kiokakis, K. Skordilis, A. Avramouli, E. Kalli, C. Pappas, P. A. Tarantilis, E. Skotti

Abstract:

Grape seeds are waste products of wineries and often referred to as an important agricultural and industrial waste product with the potential to be used in pharmaceutical, food, and cosmetic applications. In this study, grape seed oil from traditional Ionian varieties was examined for the determination of the quality and the characteristics of each variety. Initially, the fatty acid methyl ester (FAME) profiles were analyzed using Gas Chromatography-Mass Spectrometry, after transesterification. Furthermore, other quality parameters of the grape seed oils were determined by Spectroscopy techniques, UV-Vis and Raman included. Moreover, the antioxidant capacity of the oil was measured by 2,2'-azino-bis-3-ethylbenzothiazoline-6-sulfonic acid (ABTS) and 2,2-Diphenyl-1-picrylhydrazyl (DPPH) assays and their antioxidant capacity expressed in Trolox equivalents. K and ΔΚ indices were measured in 232, 268, 270 nm, as an oil quality index. The results indicate that the air-dried grape seed total oil content ranged from 5.26 to 8.77% w/w, which is in accordance with the other grape seed varieties tested in similar studies. The composition of grape seed oil is predominated with linoleic and oleic fatty acids, with the linoleic fatty acid ranging from 53.68 to 69.95% and both the linoleic and oleic fatty acids totaling 78-82% of FAMEs, which is analogous to the fatty acid composition of safflower oil. The antioxidant assays ABTS and DPPH scored high, exhibiting that the oils have potential in the cosmetic and culinary businesses. Above that, our results demonstrate that Ionian grape seed oils have prospects that can go further than cosmetic or culinary use, into the pharmaceuticals industry. Finally, the reclamation of grape seeds from wineries waste stream is in accordance with the bio-economy strategic framework and contributes to environmental protection.

Keywords: antioxidant capacity, fatty acid methyl esters, grape seed oil, GC-MS

Procedia PDF Downloads 204
14968 Wireless Sensor Networks Optimization by Using 2-Stage Algorithm Based on Imperialist Competitive Algorithm

Authors: Hamid R. Lashgarian Azad, Seyed N. Shetab Boushehri

Abstract:

Wireless sensor networks (WSN) have become progressively popular due to their wide range of applications. Wireless Sensor Network is made of numerous tiny sensor nodes that are battery-powered. It is a very significant problem to maximize the lifetime of wireless sensor networks. In this paper, we propose a two-stage protocol based on an imperialist competitive algorithm (2S-ICA) to solve a sensor network optimization problem. The energy of the sensors can be greatly reduced and the lifetime of the network reduced by long communication distances between the sensors and the sink. We can minimize the overall communication distance considerably, thereby extending the lifetime of the network lifetime through connecting sensors into a series of independent clusters using 2SICA. Comparison results of the proposed protocol and LEACH protocol, which is common to solving WSN problems, show that our protocol has a better performance in terms of improving network life and increasing the number of transmitted data.

Keywords: wireless sensor network, imperialist competitive algorithm, LEACH protocol, k-means clustering

Procedia PDF Downloads 103
14967 A Priority Based Imbalanced Time Minimization Assignment Problem: An Iterative Approach

Authors: Ekta Jain, Kalpana Dahiya, Vanita Verma

Abstract:

This paper discusses a priority based imbalanced time minimization assignment problem dealing with the allocation of n jobs to m < n persons in which the project is carried out in two stages, viz. Stage-I and Stage-II. Stage-I consists of n1 ( < m) primary jobs and Stage-II consists of remaining (n-n1) secondary jobs which are commenced only after primary jobs are finished. Each job is to be allocated to exactly one person, and each person has to do at least one job. It is assumed that nature of the Stage-I jobs is such that one person can do exactly one primary job whereas a person can do more than one secondary job in Stage-II. In a particular stage, all persons start doing the jobs simultaneously, but if a person is doing more than one job, he does them one after the other in any order. The aim of the proposed study is to find the feasible assignment which minimizes the total time for the two stage execution of the project. For this, an iterative algorithm is proposed, which at each iteration, solves a constrained imbalanced time minimization assignment problem to generate a pair of Stage-I and Stage-II times. For solving this constrained problem, an algorithm is developed in the current paper. Later, alternate combinations based method to solve the priority based imbalanced problem is also discussed and a comparative study is carried out. Numerical illustrations are provided in support of the theory.

Keywords: assignment, imbalanced, priority, time minimization

Procedia PDF Downloads 234
14966 Detailed Quantum Circuit Design and Evaluation of Grover's Algorithm for the Bounded Degree Traveling Salesman Problem Using the Q# Language

Authors: Wenjun Hou, Marek Perkowski

Abstract:

The Traveling Salesman problem is famous in computing and graph theory. In short, it asks for the Hamiltonian cycle of the least total weight in a given graph with N nodes. All variations on this problem, such as those with K-bounded-degree nodes, are classified as NP-complete in classical computing. Although several papers propose theoretical high-level designs of quantum algorithms for the Traveling Salesman Problem, no quantum circuit implementation of these algorithms has been created up to our best knowledge. In contrast to previous papers, the goal of this paper is not to optimize some abstract complexity measures based on the number of oracle iterations, but to be able to evaluate the real circuit and time costs of the quantum computer. Using the emerging quantum programming language Q# developed by Microsoft, which runs quantum circuits in a quantum computer simulation, an implementation of the bounded-degree problem and its respective quantum circuit were created. To apply Grover’s algorithm to this problem, a quantum oracle was designed, evaluating the cost of a particular set of edges in the graph as well as its validity as a Hamiltonian cycle. Repeating the Grover algorithm with an oracle that finds successively lower cost each time allows to transform the decision problem to an optimization problem, finding the minimum cost of Hamiltonian cycles. N log₂ K qubits are put into an equiprobablistic superposition by applying the Hadamard gate on each qubit. Within these N log₂ K qubits, the method uses an encoding in which every node is mapped to a set of its encoded edges. The oracle consists of several blocks of circuits: a custom-written edge weight adder, node index calculator, uniqueness checker, and comparator, which were all created using only quantum Toffoli gates, including its special forms, which are Feynman and Pauli X. The oracle begins by using the edge encodings specified by the qubits to calculate each node that this path visits and adding up the edge weights along the way. Next, the oracle uses the calculated nodes from the previous step and check that all the nodes are unique. Finally, the oracle checks that the calculated cost is less than the previously-calculated cost. By performing the oracle an optimal number of times, a correct answer can be generated with very high probability. The oracle of the Grover Algorithm is modified using the recalculated minimum cost value, and this procedure is repeated until the cost cannot be further reduced. This algorithm and circuit design have been verified, using several datasets, to generate correct outputs.

Keywords: quantum computing, quantum circuit optimization, quantum algorithms, hybrid quantum algorithms, quantum programming, Grover’s algorithm, traveling salesman problem, bounded-degree TSP, minimal cost, Q# language

Procedia PDF Downloads 190
14965 An Improved Approach Based on MAS Architecture and Heuristic Algorithm for Systematic Maintenance

Authors: Abdelhadi Adel, Kadri Ouahab

Abstract:

This paper proposes an improved approach based on MAS Architecture and Heuristic Algorithm for systematic maintenance to minimize makespan. We have implemented a problem-solving approach for optimizing the processing time, methods based on metaheuristics. The proposed approach is inspired by the behavior of the human body. This hybridization is between a multi-agent system and inspirations of the human body, especially genetics. The effectiveness of our approach has been demonstrated repeatedly in this paper. To solve such a complex problem, we proposed an approach which we have used advanced operators such as uniform crossover set and single point mutation. The proposed approach is applied to three preventive maintenance policies. These policies are intended to maximize the availability or to maintain a minimum level of reliability during the production chain. The results show that our algorithm outperforms existing algorithms. We assumed that the machines might be unavailable periodically during the production scheduling.

Keywords: multi-agent systems, emergence, genetic algorithm, makespan, systematic maintenance, scheduling, hybrid flow shop scheduling

Procedia PDF Downloads 301
14964 Pattern Recognition Search: An Advancement Over Interpolation Search

Authors: Shahpar Yilmaz, Yasir Nadeem, Syed A. Mehdi

Abstract:

Searching for a record in a dataset is always a frequent task for any data structure-related application. Hence, a fast and efficient algorithm for the approach has its importance in yielding the quickest results and enhancing the overall productivity of the company. Interpolation search is one such technique used to search through a sorted set of elements. This paper proposes a new algorithm, an advancement over interpolation search for the application of search over a sorted array. Pattern Recognition Search or PR Search (PRS), like interpolation search, is a pattern-based divide and conquer algorithm whose objective is to reduce the sample size in order to quicken the process and it does so by treating the array as a perfect arithmetic progression series and thereby deducing the key element’s position. We look to highlight some of the key drawbacks of interpolation search, which are accounted for in the Pattern Recognition Search.

Keywords: array, complexity, index, sorting, space, time

Procedia PDF Downloads 244
14963 A Communication Signal Recognition Algorithm Based on Holder Coefficient Characteristics

Authors: Hui Zhang, Ye Tian, Fang Ye, Ziming Guo

Abstract:

Communication signal modulation recognition technology is one of the key technologies in the field of modern information warfare. At present, communication signal automatic modulation recognition methods are mainly divided into two major categories. One is the maximum likelihood hypothesis testing method based on decision theory, the other is a statistical pattern recognition method based on feature extraction. Now, the most commonly used is a statistical pattern recognition method, which includes feature extraction and classifier design. With the increasingly complex electromagnetic environment of communications, how to effectively extract the features of various signals at low signal-to-noise ratio (SNR) is a hot topic for scholars in various countries. To solve this problem, this paper proposes a feature extraction algorithm for the communication signal based on the improved Holder cloud feature. And the extreme learning machine (ELM) is used which aims at the problem of the real-time in the modern warfare to classify the extracted features. The algorithm extracts the digital features of the improved cloud model without deterministic information in a low SNR environment, and uses the improved cloud model to obtain more stable Holder cloud features and the performance of the algorithm is improved. This algorithm addresses the problem that a simple feature extraction algorithm based on Holder coefficient feature is difficult to recognize at low SNR, and it also has a better recognition accuracy. The results of simulations show that the approach in this paper still has a good classification result at low SNR, even when the SNR is -15dB, the recognition accuracy still reaches 76%.

Keywords: communication signal, feature extraction, Holder coefficient, improved cloud model

Procedia PDF Downloads 156
14962 PointNetLK-OBB: A Point Cloud Registration Algorithm with High Accuracy

Authors: Wenhao Lan, Ning Li, Qiang Tong

Abstract:

To improve the registration accuracy of a source point cloud and template point cloud when the initial relative deflection angle is too large, a PointNetLK algorithm combined with an oriented bounding box (PointNetLK-OBB) is proposed. In this algorithm, the OBB of a 3D point cloud is used to represent the macro feature of source and template point clouds. Under the guidance of the iterative closest point algorithm, the OBB of the source and template point clouds is aligned, and a mirror symmetry effect is produced between them. According to the fitting degree of the source and template point clouds, the mirror symmetry plane is detected, and the optimal rotation and translation of the source point cloud is obtained to complete the 3D point cloud registration task. To verify the effectiveness of the proposed algorithm, a comparative experiment was performed using the publicly available ModelNet40 dataset. The experimental results demonstrate that, compared with PointNetLK, PointNetLK-OBB improves the registration accuracy of the source and template point clouds when the initial relative deflection angle is too large, and the sensitivity of the initial relative position between the source point cloud and template point cloud is reduced. The primary contribution of this paper is the use of PointNetLK to avoid the non-convex problem of traditional point cloud registration and leveraging the regularity of the OBB to avoid the local optimization problem in the PointNetLK context.

Keywords: mirror symmetry, oriented bounding box, point cloud registration, PointNetLK-OBB

Procedia PDF Downloads 150
14961 Improved Pattern Matching Applied to Surface Mounting Devices Components Localization on Automated Optical Inspection

Authors: Pedro M. A. Vitoriano, Tito. G. Amaral

Abstract:

Automated Optical Inspection (AOI) Systems are commonly used on Printed Circuit Boards (PCB) manufacturing. The use of this technology has been proven as highly efficient for process improvements and quality achievements. The correct extraction of the component for posterior analysis is a critical step of the AOI process. Nowadays, the Pattern Matching Algorithm is commonly used, although this algorithm requires extensive calculations and is time consuming. This paper will present an improved algorithm for the component localization process, with the capability of implementation in a parallel execution system.

Keywords: AOI, automated optical inspection, SMD, surface mounting devices, pattern matching, parallel execution

Procedia PDF Downloads 299
14960 Capacity Oversizing for Infrastructure Sharing Synergies: A Game Theoretic Analysis

Authors: Robin Molinier

Abstract:

Industrial symbiosis (I.S) rely on two basic modes of cooperation between organizations that are infrastructure/service sharing and resource substitution (the use of waste materials, fatal energy and recirculated utilities for production). The former consists in the intensification of use of an asset and thus requires to compare the incremental investment cost to be incurred and the stand-alone cost faced by each potential participant to satisfy its own requirements. In order to investigate the way such a cooperation mode can be implemented we formulate a game theoretic model integrating the grassroot investment decision and the ex-post access pricing problem. In the first period two actors set cooperatively (resp. non-cooperatively) a level of common (resp. individual) infrastructure capacity oversizing to attract ex-post a potential entrant with a plug-and-play offer (available capacity, tariff). The entrant’s requirement is randomly distributed and known only after investments took place. Capacity cost exhibits sub-additive property so that there is room for profitable overcapacity setting in the first period under some conditions that we derive. The entrant willingness-to-pay for the access to the infrastructure is driven by both her standalone cost and the complement cost to be incurred in case she chooses to access an infrastructure whose the available capacity is lower than her requirement level. The expected complement cost function is thus derived, and we show that it is decreasing, convex and shaped by the entrant’s requirements distribution function. For both uniform and triangular distributions optimal capacity level is obtained in the cooperative setting and equilibrium levels are determined in the non-cooperative case. Regarding the latter, we show that competition is deterred by the first period investor with the highest requirement level. Using the non-cooperative game outcomes which gives lower bounds for the profit sharing problem in the cooperative one we solve the whole game and describe situations supporting sharing agreements.

Keywords: capacity, cooperation, industrial symbiosis, pricing

Procedia PDF Downloads 440
14959 Effect of Size and Soil Characteristic on Contribution of Side and Tip Resistance of the Drilled Shafts Axial Load Carrying Capacity

Authors: Mehrak Zargaryaeghoubi, Masood Hajali

Abstract:

Drilled shafts are the most popular of deep foundations, because they have the capability that one single shaft can easily carry the entire load of a large column from a bridge or tall building. Drilled shaft may be an economical alternative to pile foundations because a pile cap is not needed, which not only reduces that expense, but also provides a rough surface in the border of soil and concrete to carry a more axial load. Due to the larger construction sizes of drilled shafts, they have an excellent axial load carrying capacity. Part of the axial load carrying capacity of the drilled shaft is resisted by the soil below the tip of the shaft which is tip resistance and the other part is resisted by the friction developed around the drilled shaft which is side resistance. The condition at the bottom of the excavation can affect the end bearing capacity of the drilled shaft. Also, type of the soil and size of the drilled shaft can affect the frictional resistance. The main loads applied on the drilled shafts are axial compressive loads. It is important to know how many percent of the maximum applied load will be shed inside friction and how much will be transferred to the base. The axial capacity of the drilled shaft foundation is influenced by the size of the drilled shaft, and soil characteristics. In this study, the effect of the size and soil characteristic will be investigated on the contribution of side resistance and end-bearing capacity. Also, the study presents a three-dimensional finite element modeling of a drilled shaft subjected to axial load using ANSYS. The top displacement and settlement of the drilled shaft are verified with analytical results. The soil profile is considered as Table 1 and for a drilled shaft with 7 ft diameter and 95 ft length the stresses in z-direction are calculated through the length of the shaft. From the stresses in z-direction through the length of the shaft the side resistance can be calculated and with the z-direction stress at the tip, the tip resistance can be calculated. The result of the side and tip resistance for this drilled shaft are compared with the analytical results.

Keywords: Drilled Shaft Foundation, size and soil characteristic, axial load capacity, Finite Element

Procedia PDF Downloads 379
14958 Integration of Hydropower and Solar Photovoltaic Generation into Distribution System: Case of South Sudan

Authors: Ater Amogpai

Abstract:

Hydropower and solar photovoltaic (PV) generation are crucial in sustainability and transitioning from fossil fuel to clean energy. Integrating renewable energy sources such as hydropower and solar photovoltaic (PV) into the distributed networks contributes to achieving energy balance, pollution mitigation, and cost reduction. Frequent power outages and a lack of load reliability characterize the current South Sudan electricity distribution system. The country’s electricity demand is 300MW; however, the installed capacity is around 212.4M. Insufficient funds to build new electricity facilities and expand generation are the reasons for the gap in installed capacity. The South Sudan Ministry of Energy and Dams gave a contract to an Egyptian Elsewedy Electric Company that completed the construction of a solar PV plant in 2023. The plant has a 35 MWh battery storage and 20 MW solar PV system capacity. The construction of Juba Solar PV Park started in 2022 to increase the current installed capacity in Juba City to 53 MW. The plant will begin serving 59000 residents in Juba and save 10,886.2t of carbon dioxide (CO2) annually.

Keywords: renewable energy, hydropower, solar energy, photovoltaic, South Sudan

Procedia PDF Downloads 142