Search results for: total capacity algorithm
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15363

Search results for: total capacity algorithm

14733 A Comparative Analysis of Heuristics Applied to Collecting Used Lubricant Oils Generated in the City of Pereira, Colombia

Authors: Diana Fajardo, Sebastián Ortiz, Oscar Herrera, Angélica Santis

Abstract:

Currently, in Colombia is arising a problem related to collecting used lubricant oils which are generated by the increment of the vehicle fleet. This situation does not allow a proper disposal of this type of waste, which in turn results in a negative impact on the environment. Therefore, through the comparative analysis of various heuristics, the best solution to the VRP (Vehicle Routing Problem) was selected by comparing costs and times for the collection of used lubricant oils in the city of Pereira, Colombia; since there is no presence of management companies engaged in the direct administration of the collection of this pollutant. To achieve this aim, six proposals of through methods of solution of two phases were discussed. First, the assignment of the group of generator points of the residue was made (previously identified). Proposals one and four of through methods are based on the closeness of points. The proposals two and five are using the scanning method and the proposals three and six are considering the restriction of the capacity of collection vehicle. Subsequently, the routes were developed - in the first three proposals by the Clarke and Wright's savings algorithm and in the following proposals by the Traveling Salesman optimization mathematical model. After applying techniques, a comparative analysis of the results was performed and it was determined which of the proposals presented the most optimal values in terms of the distance, cost and travel time.

Keywords: Heuristics, optimization Model, savings algorithm, used vehicular oil, V.R.P.

Procedia PDF Downloads 400
14732 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 175
14731 In Vitro Antioxidant and Free Radical Scavenging Activity of Phyllanthus Emblica L. Extract

Authors: Benyapa Suksuwan

Abstract:

Introduction: Oxidative stress is identified as the root cause of the development and progression of several diseases as the disproportion of free radicals in the body leads to tissue or cell damage. Polyphenols are the most common antioxidant found in plants and are efficient in capturing oxidative free radicals. Aim of the Study: This study focused on the antioxidant activity of polyphenols extracted from Phyllanthus Emblica L. as oxidative stress plays a vital role in developing and progressing many diseases, including cardiovascular diseases and cancer. Materials and Methods: The plant was extracted using a mixture solvent (ethyl alcohol: water in ratio 8:2). The total phenolic content of P. Emblica extract was determined using the Folin-Cioucalteu method and calculated as gallic acid equivalents (GAE) and various antioxidant assays DPPH and ABTS radical scavenging capacity assays. Results and Discussion: The findings exhibited a strong correlation between antioxidant activity and the total phenol contents. In addition, the IC₅₀ of P. Emblica extract via DPPH and ABTS assays were 68.10 μg/mL ± 0.455, and 49.24 μg/mL ± 0.716, respectively. Furthermore, P. Emblica extract showed antioxidant activities in a concentration-dependent manner. Vitamin C was used as a positive control in the DPPH assay, while Trolox was used as a positive control in the ABTS assay. Conclusions: In conclusion, P. Emblica extract consisted of a high amount of total phenolic content, which possesses potent antioxidant activity. However, further antioxidant activity assays using human cell lines such as SOD, ROS, and RNS scavenging assays and in vitro antioxidant experiments should be performed in order.

Keywords: antioxidant, ABTS scavenging, DPPH scavenging assay, total phenol contents assay, Phyllanthus Emblica L

Procedia PDF Downloads 181
14730 From Problem Space to Executional Architecture: The Development of a Simulator to Examine the Effect of Autonomy on Mainline Rail Capacity

Authors: Emily J. Morey, Kevin Galvin, Thomas Riley, R. Eddie Wilson

Abstract:

The key challenges faced by integrating autonomous rail operations into the existing mainline railway environment have been identified through the understanding and framing of the problem space and stakeholder analysis. This was achieved through the completion of the first four steps of Soft Systems Methodology, where the problem space has been expressed via conceptual models. Having identified these challenges, we investigated one of them, namely capacity, via the use of models and simulation. This paper examines the approach used to move from the conceptual models to a simulation which can determine whether the integration of autonomous trains can plausibly increase capacity. Within this approach, we developed an architecture and converted logical models into physical resource models and associated design features which were used to build a simulator. From this simulator, we are able to analyse mixtures of legacy-autonomous operations and produce fundamental diagrams and trajectory plots to describe the dynamic behaviour of mixed mainline railway operations.

Keywords: autonomy, executable architecture, modelling and simulation, railway capacity

Procedia PDF Downloads 58
14729 BTEX Removal from Water: A Comparative Analysis of Efficiency of Low Cost Adsorbents and Granular Activated Carbon

Authors: Juliet Okoli

Abstract:

The removal of BTEX (Benzene, toluene, Ethylbenzene and p-Xylene) from water by orange peel and eggshell compared to GAC were investigated. The influence of various factors such as contact time, dosage and pH on BTEX removal by virgin orange peel and egg shell were accessed using the batch adsorption set-up. These were also compared to that of GAC which serves as a benchmark for this study. Further modification (preparation of Activated carbon) of these virgin low-cost adsorbents was also carried out. The batch adsorption result showed that the optimum contact time, dosage and pH for BTEX removal by virgin LCAs were 180 minutes, 0.5g and 7 and that of GAC was 30mintues, 0.2g and 7. The maximum adsorption capacity for total BTEX showed by orange peel and egg shell were 42mg/g and 59mg/g respectively while that of GAC was 864mg/g. The adsorbent preference for adsorbate were in order of X>E>T>B. A comparison of batch and column set-up showed that the batch set-up was more efficient than the column set-up. The isotherm data for the virgin LCA and GAC prove to fit the Freundlich isotherm better than the Langmuir model, which produced n values >1 in case of GAC and n< 1 in case of virgin LCAs; indicating a more appropriate adsorption of BTEX onto the GAC. The adsorption kinetics for the three studied adsorbents were described well by the pseudo-second order, suggesting chemisorption as the rate limiting step. This was further confirmed by desorption study, as low levels of BTEX (<10%) were recovered from the spent adsorbents especially for GAC (<3%). Further activation of the LCAs which was compared to the virgin LCAs, revealed that the virgin LCAs had minor higher adsorption capacity than the activated LCAs. Economic analysis revealed that the total cost required to clean-up 9,600m3 of BTEX contaminated water using LCA was just 2.8% lesser than GAC, a difference which could be considered negligible. However, this area still requires a more detailed cost-benefit analysis, and if similar conclusions are reached; a low-cost adsorbent, easy to obtain are still promising adsorbents for BTEX removal from aqueous solution; however, the GAC are still more superior to these materials.

Keywords: activated carbon, BTEX removal, low cost adsorbents, water treatment

Procedia PDF Downloads 238
14728 Discriminant Analysis as a Function of Predictive Learning to Select Evolutionary Algorithms in Intelligent Transportation System

Authors: Jorge A. Ruiz-Vanoye, Ocotlán Díaz-Parra, Alejandro Fuentes-Penna, Daniel Vélez-Díaz, Edith Olaco García

Abstract:

In this paper, we present the use of the discriminant analysis to select evolutionary algorithms that better solve instances of the vehicle routing problem with time windows. We use indicators as independent variables to obtain the classification criteria, and the best algorithm from the generic genetic algorithm (GA), random search (RS), steady-state genetic algorithm (SSGA), and sexual genetic algorithm (SXGA) as the dependent variable for the classification. The discriminant classification was trained with classic instances of the vehicle routing problem with time windows obtained from the Solomon benchmark. We obtained a classification of the discriminant analysis of 66.7%.

Keywords: Intelligent Transportation Systems, data-mining techniques, evolutionary algorithms, discriminant analysis, machine learning

Procedia PDF Downloads 447
14727 The Effect of Porous Alkali Activated Material Composition on Buffer Capacity in Bioreactors

Authors: Girts Bumanis, Diana Bajare

Abstract:

With demand for primary energy continuously growing, search for renewable and efficient energy sources has been high on agenda of our society. One of the most promising energy sources is biogas technology. Residues coming from dairy industry and milk processing could be used in biogas production; however, low efficiency and high cost impede wide application of such technology. One of the main problems is management and conversion of organic residues through the anaerobic digestion process which is characterized by acidic environment due to the low whey pH (<6) whereas additional pH control system is required. Low buffering capacity of whey is responsible for the rapid acidification in biological treatments; therefore alkali activated material is a promising solution of this problem. Alkali activated material is formed using SiO2 and Al2O3 rich materials under highly alkaline solution. After material structure forming process is completed, free alkalis remain in the structure of materials which are available for leaching and could provide buffer capacity potential. In this research porous alkali activated material was investigated. Highly porous material structure ensures gradual leaching of alkalis during time which is important in biogas digestion process. Research of mixture composition and SiO2/Na2O and SiO2/Al2O ratio was studied to test the buffer capacity potential of alkali activated material. This research has proved that by changing molar ratio of components it is possible to obtain a material with different buffer capacity, and this novel material was seen to have considerable potential for using it in processes where buffer capacity and pH control is vitally important.

Keywords: alkaline material, buffer capacity, biogas production, bioreactors

Procedia PDF Downloads 233
14726 Handling Missing Data by Using Expectation-Maximization and Expectation-Maximization with Bootstrapping for Linear Functional Relationship Model

Authors: Adilah Abdul Ghapor, Yong Zulina Zubairi, A. H. M. R. Imon

Abstract:

Missing value problem is common in statistics and has been of interest for years. This article considers two modern techniques in handling missing data for linear functional relationship model (LFRM) namely the Expectation-Maximization (EM) algorithm and Expectation-Maximization with Bootstrapping (EMB) algorithm using three performance indicators; namely the mean absolute error (MAE), root mean square error (RMSE) and estimated biased (EB). In this study, we applied the methods of imputing missing values in two types of LFRM namely the full model of LFRM and in LFRM when the slope is estimated using a nonparametric method. Results of the simulation study suggest that EMB algorithm performs much better than EM algorithm in both models. We also illustrate the applicability of the approach in a real data set.

Keywords: expectation-maximization, expectation-maximization with bootstrapping, linear functional relationship model, performance indicators

Procedia PDF Downloads 439
14725 An Introduction to E-Content Producing Algorithm for Screen-Recorded Videos

Authors: Jamileh Darsareh, Mohammad Nikafrooz

Abstract:

Some teachers and e-content producers, based on their experiences, try to produce educational videos using screen recording software. There are many challenges that they may encounter while producing screen-recorded videos. These are in the domains of technical and pedagogical challenges like designing the roadmap, preparing the screen, setting the recording software and recording the screen, editing, etc. This study is a descriptive study and tries to present some procedures for producing acceptable and well-made videos. These procedures are presented in the form of an algorithm for producing screen-recorded video. This algorithm presents the main producing phases, including design, pre-production, production, post-production, and distribution. These phases consist of some steps which are supported by several technical and pedagogical considerations. Following these phases and steps according to the suggested order helps the producers to produce their intended and desired video by saving time and also facing fewer technical problems. It is expected that by using this algorithm, e-content producers and teachers gain better performance in producing educational videos.

Keywords: e-content producing algorithm, screen-recorded videos, screen recording software, technical and pedagogical considerations

Procedia PDF Downloads 181
14724 Performance Comparison of Prim’s and Ant Colony Optimization Algorithm to Select Shortest Path in Case of Link Failure

Authors: Rimmy Yadav, Avtar Singh

Abstract:

—Ant Colony Optimization (ACO) is a promising modern approach to the unused combinatorial optimization. Here ACO is applied to finding the shortest during communication link failure. In this paper, the performances of the prim’s and ACO algorithm are made. By comparing the time complexity and program execution time as set of parameters, we demonstrate the pleasant performance of ACO in finding excellent solution to finding shortest path during communication link failure.

Keywords: ant colony optimization, link failure, prim’s algorithm, shortest path

Procedia PDF Downloads 378
14723 3D Reconstruction of Human Body Based on Gender Classification

Authors: Jiahe Liu, Hongyang Yu, Feng Qian, Miao Luo

Abstract:

SMPL-X was a powerful parametric human body model that included male, neutral, and female models, with significant gender differences between these three models. During the process of 3D human body reconstruction, the correct selection of standard templates was crucial for obtaining accurate results. To address this issue, we developed an efficient gender classification algorithm to automatically select the appropriate template for 3D human body reconstruction. The key to this gender classification algorithm was the precise analysis of human body features. By using the SMPL-X model, the algorithm could detect and identify gender features of the human body, thereby determining which standard template should be used. The accuracy of this algorithm made the 3D reconstruction process more accurate and reliable, as it could adjust model parameters based on individual gender differences. SMPL-X and the related gender classification algorithm have brought important advancements to the field of 3D human body reconstruction. By accurately selecting standard templates, they have improved the accuracy of reconstruction and have broad potential in various application fields. These technologies continue to drive the development of the 3D reconstruction field, providing us with more realistic and accurate human body models.

Keywords: gender classification, joint detection, SMPL-X, 3D reconstruction

Procedia PDF Downloads 47
14722 Analyzing Test Data Generation Techniques Using Evolutionary Algorithms

Authors: Arslan Ellahi, Syed Amjad Hussain

Abstract:

Software Testing is a vital process in software development life cycle. We can attain the quality of software after passing it through software testing phase. We have tried to find out automatic test data generation techniques that are a key research area of software testing to achieve test automation that can eventually decrease testing time. In this paper, we review some of the approaches presented in the literature which use evolutionary search based algorithms like Genetic Algorithm, Particle Swarm Optimization (PSO), etc. to validate the test data generation process. We also look into the quality of test data generation which increases or decreases the efficiency of testing. We have proposed test data generation techniques for model-based testing. We have worked on tuning and fitness function of PSO algorithm.

Keywords: search based, evolutionary algorithm, particle swarm optimization, genetic algorithm, test data generation

Procedia PDF Downloads 164
14721 Comparative Performance of Artificial Bee Colony Based Algorithms for Wind-Thermal Unit Commitment

Authors: P. K. Singhal, R. Naresh, V. Sharma

Abstract:

This paper presents the three optimization models, namely New Binary Artificial Bee Colony (NBABC) algorithm, NBABC with Local Search (NBABC-LS), and NBABC with Genetic Crossover (NBABC-GC) for solving the Wind-Thermal Unit Commitment (WTUC) problem. The uncertain nature of the wind power is incorporated using the Weibull probability density function, which is used to calculate the overestimation and underestimation costs associated with the wind power fluctuation. The NBABC algorithm utilizes a mechanism based on the dissimilarity measure between binary strings for generating the binary solutions in WTUC problem. In NBABC algorithm, an intelligent scout bee phase is proposed that replaces the abandoned solution with the global best solution. The local search operator exploits the neighboring region of the current solutions, whereas the integration of genetic crossover with the NBABC algorithm increases the diversity in the search space and thus avoids the problem of local trappings encountered with the NBABC algorithm. These models are then used to decide the units on/off status, whereas the lambda iteration method is used to dispatch the hourly load demand among the committed units. The effectiveness of the proposed models is validated on an IEEE 10-unit thermal system combined with a wind farm over the planning period of 24 hours.

Keywords: artificial bee colony algorithm, economic dispatch, unit commitment, wind power

Procedia PDF Downloads 359
14720 A Comparative Study on Behavior Among Different Types of Shear Connectors using Finite Element Analysis

Authors: Mohd Tahseen Islam Talukder, Sheikh Adnan Enam, Latifa Akter Lithi, Soebur Rahman

Abstract:

Composite structures have made significant advances in construction applications during the last few decades. Composite structures are composed of structural steel shapes and reinforced concrete combined with shear connectors, which benefit each material's unique properties. Significant research has been conducted on different types of connectors’ behavior and shear capacity. Moreover, the AISC 360-16 “Specification for Steel Structural Buildings” consists of a formula for channel shear connectors' shear capacity. This research compares the behavior of C type and L type shear connectors using Finite Element Analysis. Experimental results from published literature are used to validate the finite element models. The 3-D Finite Element Model (FEM) was built using ABAQUS 2017 to investigate non-linear capabilities and the ultimate load-carrying potential of the connectors using push-out tests. The changes in connector dimensions were analyzed using this non-linear model in parametric investigations. The parametric study shows that by increasing the length of the shear connector by 10 mm, its shear strength increases by 21%. Shear capacity increased by 13% as the height was increased by 10 mm. The thickness of the specimen was raised by 1 mm, resulting in a 2% increase in shear capacity. However, the shear capacity of channel connectors was reduced by 21% due to an increase of thickness by 2 mm.

Keywords: finite element method, channel shear connector, angle shear connector, ABAQUS, composite structure, shear connector, parametric study, ultimate shear capacity, push-out test

Procedia PDF Downloads 101
14719 Cooperative Spectrum Sensing Using Hybrid IWO/PSO Algorithm in Cognitive Radio Networks

Authors: Deepa Das, Susmita Das

Abstract:

Cognitive Radio (CR) is an emerging technology to combat the spectrum scarcity issues. This is achieved by consistently sensing the spectrum, and detecting the under-utilized frequency bands without causing undue interference to the primary user (PU). In soft decision fusion (SDF) based cooperative spectrum sensing, various evolutionary algorithms have been discussed, which optimize the weight coefficient vector for maximizing the detection performance. In this paper, we propose the hybrid invasive weed optimization and particle swarm optimization (IWO/PSO) algorithm as a fast and global optimization method, which improves the detection probability with a lesser sensing time. Then, the efficiency of this algorithm is compared with the standard invasive weed optimization (IWO), particle swarm optimization (PSO), genetic algorithm (GA) and other conventional SDF based methods on the basis of convergence and detection probability.

Keywords: cognitive radio, spectrum sensing, soft decision fusion, GA, PSO, IWO, hybrid IWO/PSO

Procedia PDF Downloads 441
14718 Experimental Study of Geotextile Effect on Improving Soil Bearing Capacity in Aggregate Surfaced Roads

Authors: Mahdi Taghipour Masoumi, Ali Abdi Kordani, Mahmoud Nazirizad

Abstract:

Geosynthetics utilization plays an important role in the construction of highways with no additive layers, such as asphalt concrete or cement concrete, or in a subgrade layer which affects the bearing capacity of unbounded layers. This laboratory experimental study was carried out to evaluate changes in the load bearing capacity of reinforced soil with these materials in highway roadbed with regard to geotextile properties. California Bearing Ratio (CBR) test samples were prepared with two types of soil: Clayey and sandy containing non-reinforced and reinforced soil. The samples comprised three types of geotextiles with different characteristics (150, 200, 300 g/m2) and depths (H= 5, 10, 20, 30, 50, 100 mm), and were grouped into two forms, one-layered and two-layered, based on the sample materials in order to perform defined tests. The results showed that the soil bearing characteristics increased when one layer of geotextile was used in clayey and sandy samples reinforced by geotextile. However, the bearing capacity of the soil, in the presence of a geotextile layer material with depth of more than 30 mm, had no remarkable effect. Furthermore, when the two-layered geotextile was applied in material samples, although it increased the soil resistance, it also showed that through the addition of a number or weights of geotextile into samples, the natural composition of the soil changed and the results are unreliable.

Keywords: reinforced soil, geosynthetics, geotextile, transportation capacity, CBR experiments

Procedia PDF Downloads 270
14717 LiTa2PO8-based Composite Solid Polymer Electrolytes for High-Voltage Cathodes in Lithium-Metal Batteries

Authors: Kumlachew Zelalem Walle, Chun-Chen Yang

Abstract:

Solid-state Lithium metal batteries (SSLMBs) that contain polymer and ceramic solid electrolytes have received considerable attention as an alternative to substitute liquid electrolytes in lithium metal batteries (LMBs) for highly safe, excellent energy storage performance and stability under elevated temperature situations. Here, a novel fast Li-ion conducting material, LiTa₂PO₈ (LTPO), was synthesized and electrochemical performance of as-prepared powder and LTPO-incorporated composite solid polymer electrolyte (LTPO-CPE) membrane were investigated. The as-prepared LTPO powder was homogeneously dispersed in polymer matrices, and a hybrid solid electrolyte membrane was synthesized via a simple solution-casting method. The room temperature total ionic conductivity (σt) of the LTPO pellet and LTPO-CPE membrane were 0.14 and 0.57 mS cm-1, respectively. A coin battery with NCM811 cathode is cycled under 1C between 2.8 to 4.5 V at room temperature, achieving a Coulombic efficiency of 99.3% with capacity retention of 74.1% after 300 cycles. Similarly, the LFP cathode also delivered an excellent performance at 0.5C with an average Coulombic efficiency of 100% without virtually capacity loss (the maximum specific capacity is at 27th: 138 mAh g−1 and 500th: 131.3 mAh g−1). These results demonstrates the feasibility of a high Li-ion conductor LTPO as a filler, and the developed polymer/ceramic hybrid electrolyte has potential to be a high-performance electrolyte for high-voltage cathodes, which may provide a fresh platform for developing more advanced solid-state electrolytes.

Keywords: li-ion conductor, lithium-metal batteries, composite solid electrolytes, liTa2PO8, high-voltage cathode

Procedia PDF Downloads 45
14716 Earthquake Effect in Micro Hydro Sector: Case Study of Dulakha District, Nepal

Authors: Keshav Raj Dhakal, Jit Bahadur Rokaya Chhetri

Abstract:

The Micro Hydro (MH) is one of the successful technology in Rural Nepal. Out of 75 district, 59 districts have installed 1287 MH projects with a total capacity of 24 Mega Watt (MW). Now, the challenge is how to sustain them. Dolakha is a prominent district for sustainable endues of power to sustain the MH projects. A total of 37 MH projects have been constructed with producing 886 Kilo Watt (KW) of energy in the district. This study traces out the impact of earthquake in MH sector in Dolakha district. It shows that 59 % of projects have been affected by devastating earthquake in April and May, 2015 where 29 % are completely damaged. Most of the damages are in civil structures like Penstock, forebay, power house, Canal, Intake. Transmission and distribution line have been partially damaged. This paper analysis failure of the civil structural component of MH projects and its financial consequence to the community. This study recommends that a disaster impact assessment is essential before construction of MH projects.

Keywords: micro hydro, earthquake, structural failure, financial consequence

Procedia PDF Downloads 186
14715 Increasing the Capacity of Plant Bottlenecks by Using of Improving the Ratio of Mean Time between Failures to Mean Time to Repair

Authors: Jalal Soleimannejad, Mohammad Asadizeidabadi, Mahmoud Koorki, Mojtaba Azarpira

Abstract:

A significant percentage of production costs is the maintenance costs, and analysis of maintenance costs could to achieve greater productivity and competitiveness. With this is mind, the maintenance of machines and installations is considered as an essential part of organizational functions and applying effective strategies causes significant added value in manufacturing activities. Organizations are trying to achieve performance levels on a global scale with emphasis on creating competitive advantage by different methods consist of RCM (Reliability-Center-Maintenance), TPM (Total Productivity Maintenance) etc. In this study, increasing the capacity of Concentration Plant of Golgohar Iron Ore Mining & Industrial Company (GEG) was examined by using of reliability and maintainability analyses. The results of this research showed that instead of increasing the number of machines (in order to solve the bottleneck problems), the improving of reliability and maintainability would solve bottleneck problems in the best way. It should be mention that in the abovementioned study, the data set of Concentration Plant of GEG as a case study, was applied and analyzed.

Keywords: bottleneck, golgohar iron ore mining & industrial company, maintainability, maintenance costs, reliability

Procedia PDF Downloads 332
14714 3D Human Body Reconstruction Based on Multiple Viewpoints

Authors: Jiahe Liu, HongyangYu, Feng Qian, Miao Luo

Abstract:

The aim of this study was to improve the effects of human body 3D reconstruction. The MvP algorithm was adopted to obtain key point information from multiple perspectives. This algorithm allowed the capture of human posture and joint positions from multiple angles, providing more comprehensive and accurate data. The study also incorporated the SMPL-X model, which has been widely used for human body modeling, to achieve more accurate 3D reconstruction results. The use of the MvP algorithm made it possible to observe the reconstructed object from multiple angles, thus reducing the problems of blind spots and missing information. This algorithm was able to effectively capture key point information, including the position and rotation angle of limbs, providing key data for subsequent 3D reconstruction. Compared with traditional single-view methods, the method of multi-view fusion significantly improved the accuracy and stability of reconstruction. By combining the MvP algorithm with the SMPL-X model, we successfully achieved better human body 3D reconstruction effects. The SMPL-X model is highly scalable and can generate highly realistic 3D human body models, thus providing more detail and shape information.

Keywords: 3D human reconstruction, multi-view, joint point, SMPL-X

Procedia PDF Downloads 47
14713 Bi-Criteria Objective Network Design Model for Multi Period Multi Product Green Supply Chain

Authors: Shahul Hamid Khan, S. Santhosh, Abhinav Kumar Sharma

Abstract:

Environmental performance along with social performance is becoming vital factors for industries to achieve global standards. With a good environmental policy global industries are differentiating them from their competitors. This paper concentrates on multi stage, multi product and multi period manufacturing network. Bi-objective mathematical models for total cost and total emission for the entire forward supply chain are considered. Here five different problems are considered by varying the number of suppliers, manufacturers, and environmental levels, for illustrating the taken mathematical model. GA, and Random search are used for finding the optimal solution. The input parameters of the optimal solution are used to find the tradeoff between the initial investment by the industry and the long term benefit of the environment.

Keywords: closed loop supply chain, genetic algorithm, random search, green supply chain

Procedia PDF Downloads 528
14712 Fuzzy Population-Based Meta-Heuristic Approaches for Attribute Reduction in Rough Set Theory

Authors: Mafarja Majdi, Salwani Abdullah, Najmeh S. Jaddi

Abstract:

One of the global combinatorial optimization problems in machine learning is feature selection. It concerned with removing the irrelevant, noisy, and redundant data, along with keeping the original meaning of the original data. Attribute reduction in rough set theory is an important feature selection method. Since attribute reduction is an NP-hard problem, it is necessary to investigate fast and effective approximate algorithms. In this paper, we proposed two feature selection mechanisms based on memetic algorithms (MAs) which combine the genetic algorithm with a fuzzy record to record travel algorithm and a fuzzy controlled great deluge algorithm to identify a good balance between local search and genetic search. In order to verify the proposed approaches, numerical experiments are carried out on thirteen datasets. The results show that the MAs approaches are efficient in solving attribute reduction problems when compared with other meta-heuristic approaches.

Keywords: rough set theory, attribute reduction, fuzzy logic, memetic algorithms, record to record algorithm, great deluge algorithm

Procedia PDF Downloads 431
14711 A Hybrid Genetic Algorithm for Assembly Line Balancing In Automotive Sector

Authors: Qazi Salman Khalid, Muhammad Khalid, Shahid Maqsood

Abstract:

This paper presents a solution for optimizing the cycle time in an assembly line with human-robot collaboration and diverse operators. A genetic algorithm with tailored parameters is used to address the assembly line balancing problem in the automobile sector. A mathematical model is developed, depicting the problem. Currently, the firm runs on the largest candidate rule; however, it causes a lag in orders, which ultimately gets penalized. The results of the study show that the proposed GA is effective in providing efficient solutions and that the cycle time has significantly impacted productivity.

Keywords: line balancing, cycle time, genetic algorithm, productivity

Procedia PDF Downloads 117
14710 Return to Work Following Knee Arthroplasty: A Retrospective Review in Urban Asian Population

Authors: Fiona Tan, Cheryl Tan, Thomas Wong, Remesh

Abstract:

Background: An increasing number of working adults undergo knee arthroplasty in Singapore. There is limited data concerning Southeast Asian patients returning to work (RTW) following knee replacement surgery. Our aim was to identify and study factors influencing patients' RTW following total knee arthroplasty (TKA) or unicompartmental knee arthroplasty (UKA). Methods: Patients who underwent TKA or UKA between August 2017 to March 2020 in our center were included in this study. Outcomes include RTW and duration prior to RTW. Results: 441 patients underwent TKA (295 women, 146 men, mean age 67.3 years), and 69 underwent UKA (48 women, 21 men, mean age 61.1 years). Patients who underwent TKA returned to work earlier (mean 83.7 ± 27.1 days) compared to UKA (mean 94.4 ± 42.3 days). 90.0% of TKA patients RTW compared to 95.5% who underwent UKA. Of patients who RTW, 94.3% of the TKA group returned to employment of the same nature compared to 92.9% of UKA patients. Patients who RTW were of a younger age (p = 0.03), white-collared workers (p = 0.04), and had independent preoperative ambulatory status (p <0.01). Conclusion: Younger and independently ambulating patients may have a better capacity for rehabilitation and RTW post-arthroplasty surgery.

Keywords: return to work, total knee arthroplasty, unilateral knee arthroplasty, employment

Procedia PDF Downloads 19
14709 Comparisons of Surveying with Terrestrial Laser Scanner and Total Station for Volume Determination of Overburden and Coal Excavations in Large Open-Pit Mine

Authors: B. Keawaram, P. Dumrongchai

Abstract:

The volume of overburden and coal excavations in open-pit mine is generally determined by conventional survey such as total station. This study aimed to evaluate the accuracy of terrestrial laser scanner (TLS) used to measure overburden and coal excavations, and to compare TLS survey data sets with the data of the total station. Results revealed that, the reference points measured with the total station showed 0.2 mm precision for both horizontal and vertical coordinates. When using TLS on the same points, the standard deviations of 4.93 cm and 0.53 cm for horizontal and vertical coordinates, respectively, were achieved. For volume measurements covering the mining areas of 79,844 m2, TLS yielded the mean difference of about 1% and the surface error margin of 6 cm at the 95% confidence level when compared to the volume obtained by total station.

Keywords: mine, survey, terrestrial laser scanner, total station

Procedia PDF Downloads 361
14708 Real-Time Detection of Space Manipulator Self-Collision

Authors: Zhang Xiaodong, Tang Zixin, Liu Xin

Abstract:

In order to avoid self-collision of space manipulators during operation process, a real-time detection method is proposed in this paper. The manipulator is fitted into a cylinder enveloping surface, and then the detection algorithm of collision between cylinders is analyzed. The collision model of space manipulator self-links can be detected by using this algorithm in real-time detection during the operation process. To ensure security of the operation, a safety threshold is designed. The simulation and experiment results verify the effectiveness of the proposed algorithm for a 7-DOF space manipulator.

Keywords: space manipulator, collision detection, self-collision, the real-time collision detection

Procedia PDF Downloads 444
14707 Lipid from Activated Sludge as a Feedstock for the Production of Biodiesel

Authors: Ifeanyichukwu Edeh, Tim Overton, Steve Bowra

Abstract:

There is increasing interest in utilising low grade or waste biomass for the production of renewable bioenergy vectors i.e. waste to energy. In this study we have chosen to assess, activated sludge, which is a microbial biomass generated during the second stage of waste water treatment as a source of lipid for biodiesel production. To date a significant proportion of biodiesel is produced from used cooking oil and animal fats. It was reasoned that if activated sludge proved a viable feedstock it has the potential to support increase biodiesel production capacity. Activated sludge was obtained at different times of the year and from two different sewage treatment works in the UK. The biomass within the activated sludge slurry was recovered by filtration and the total weight of material calculated by combining the dry weight of the total suspended solid (TSS) and the total dissolved solid (TDS) fractions. Total lipids were extracted from the TSS and TDS using solvent extraction (Folch methods). The classes of lipids within the total lipid extract were characterised using high performance thin layer chromatography (HPTLC) by referencing known standards. The fatty acid profile and content of the lipid extract were determined using acid mediated-methanolysis to obtain fatty acid methyl esters (FAMEs) which were analysed by gas chromatography and HPTLC. The results showed that there were differences in the total biomass content in the activated sludge collected from different sewage works. Lipid yields from TSS obtained from both sewage treatment works differed according to the time of year (between 3.0 and 7.4 wt. %). The lipid yield varied slightly within the same source of biomass but more widely between the two sewage treatment works. The neutral lipid classes identified were acylglycerols, free fatty acids, sterols and wax esters while the phospholipid class included phosphatidylcholine, lysophosphatidycholine, phosphatidylethanolamine and phosphatidylinositol. The fatty acid profile revealed the presence of palmitic acid, palmitoleic acid, linoleic acid, oleic acid and stearic acid and that unsaturated fatty acids were the most abundant. Following optimisation, the FAME yield was greater than 10 wt. % which was required to have an economic advantage in biodiesel production.

Keywords: activated sludge, biodiesel, lipid, methanolysis

Procedia PDF Downloads 455
14706 Measuring Science and Technology Innovation Capacity in Developing Countries: From a National Innovation System

Authors: Haeng A. Seo, Changseok Oh, Seung Jun Yoo

Abstract:

This study attempts to examine the disparities in S&T innovation capacity from 14 developing countries to discuss how to support specific features in national innovation systems. It includes East-Asian, Middle-Asian, Central American and African countries. Here, we particularly focus on five dimensions- resources, activities, network, environment and performance- with 37 indicators. They were derived as structuring components of the relevant diagnostic model, which encompasses the whole process of S&T innovation from the input of resources to the output of economically valuable results. For many developing nations, economic industries remain weaker than actual S&T capabilities, and relevant regulatory authorities may not exist. This paper will be helpful to provide basic evidence and to set directions for better national S&T Innovation capacities and toward national competitiveness.

Keywords: developing countries, measurement, NIS, S&T innovation capacity

Procedia PDF Downloads 264
14705 Examining the Performance of Three Multiobjective Evolutionary Algorithms Based on Benchmarking Problems

Authors: Konstantinos Metaxiotis, Konstantinos Liagkouras

Abstract:

The objective of this study is to examine the performance of three well-known multiobjective evolutionary algorithms for solving optimization problems. The first algorithm is the Non-dominated Sorting Genetic Algorithm-II (NSGA-II), the second one is the Strength Pareto Evolutionary Algorithm 2 (SPEA-2), and the third one is the Multiobjective Evolutionary Algorithms based on decomposition (MOEA/D). The examined multiobjective algorithms are analyzed and tested on the ZDT set of test functions by three performance metrics. The results indicate that the NSGA-II performs better than the other two algorithms based on three performance metrics.

Keywords: MOEAs, multiobjective optimization, ZDT test functions, evolutionary algorithms

Procedia PDF Downloads 448
14704 Incorporating Information Gain in Regular Expressions Based Classifiers

Authors: Rosa L. Figueroa, Christopher A. Flores, Qing Zeng-Treitler

Abstract:

A regular expression consists of sequence characters which allow describing a text path. Usually, in clinical research, regular expressions are manually created by programmers together with domain experts. Lately, there have been several efforts to investigate how to generate them automatically. This article presents a text classification algorithm based on regexes. The algorithm named REX was designed, and then, implemented as a simplified method to create regexes to classify Spanish text automatically. In order to classify ambiguous cases, such as, when multiple labels are assigned to a testing example, REX includes an information gain method Two sets of data were used to evaluate the algorithm’s effectiveness in clinical text classification tasks. The results indicate that the regular expression based classifier proposed in this work performs statically better regarding accuracy and F-measure than Support Vector Machine and Naïve Bayes for both datasets.

Keywords: information gain, regular expressions, smith-waterman algorithm, text classification

Procedia PDF Downloads 299