Search results for: automated teller machines (atm)
646 Mage Fusion Based Eye Tumor Detection
Authors: Ahmed Ashit
Abstract:
Image fusion is a significant and efficient image processing method used for detecting different types of tumors. This method has been used as an effective combination technique for obtaining high quality images that combine anatomy and physiology of an organ. It is the main key in the huge biomedical machines for diagnosing cancer such as PET-CT machine. This thesis aims to develop an image analysis system for the detection of the eye tumor. Different image processing methods are used to extract the tumor and then mark it on the original image. The images are first smoothed using median filtering. The background of the image is subtracted, to be then added to the original, results in a brighter area of interest or tumor area. The images are adjusted in order to increase the intensity of their pixels which lead to clearer and brighter images. once the images are enhanced, the edges of the images are detected using canny operators results in a segmented image comprises only of the pupil and the tumor for the abnormal images, and the pupil only for the normal images that have no tumor. The images of normal and abnormal images are collected from two sources: “Miles Research” and “Eye Cancer”. The computerized experimental results show that the developed image fusion based eye tumor detection system is capable of detecting the eye tumor and segment it to be superimposed on the original image.Keywords: image fusion, eye tumor, canny operators, superimposed
Procedia PDF Downloads 361645 Analyzing the Impact of Code Commenting on Software Quality
Authors: Thulya Premathilake, Tharushi Perera, Hansi Thathsarani, Tharushi Nethmini, Dilshan De Silva, Piyumika Samarasekara
Abstract:
One of the most efficient ways to assist developers in grasping the source code is to make use of comments, which can be found throughout the code. When working in fields such as software development, having comments in your code that are of good quality is a fundamental requirement. Tackling software problems while making use of programs that have already been built. It is essential for the intention of the source code to be made crystal apparent in the comments that are added to the code. This assists programmers in better comprehending the programs they are working on and enables them to complete software maintenance jobs in a more timely manner. In spite of the fact that comments and documentation are meant to improve readability and maintainability, the vast majority of programmers place the majority of their focus on the actual code that is being written. This study provides a complete and comprehensive overview of the previous research that has been conducted on the topic of code comments. The study focuses on four main topics, including automated comment production, comment consistency, comment classification, and comment quality rating. One is able to get the knowledge that is more complete for use in following inquiries if they conduct an analysis of the proper approaches that were used in this study issue.Keywords: code commenting, source code, software quality, quality assurance
Procedia PDF Downloads 82644 An Intelligent Traffic Management System Based on the WiFi and Bluetooth Sensing
Authors: Hamed Hossein Afshari, Shahrzad Jalali, Amir Hossein Ghods, Bijan Raahemi
Abstract:
This paper introduces an automated clustering solution that applies to WiFi/Bluetooth sensing data and is later used for traffic management applications. The paper initially summarizes a number of clustering approaches and thereafter shows their performance for noise removal. In this context, clustering is used to recognize WiFi and Bluetooth MAC addresses that belong to passengers traveling by a public urban transit bus. The main objective is to build an intelligent system that automatically filters out MAC addresses that belong to persons located outside the bus for different routes in the city of Ottawa. The proposed intelligent system alleviates the need for defining restrictive thresholds that however reduces the accuracy as well as the range of applicability of the solution for different routes. This paper moreover discusses the performance benefits of the presented clustering approaches in terms of the accuracy, time and space complexity, and the ease of use. Note that results of clustering can further be used for the purpose of the origin-destination estimation of individual passengers, predicting the traffic load, and intelligent management of urban bus schedules.Keywords: WiFi-Bluetooth sensing, cluster analysis, artificial intelligence, traffic management
Procedia PDF Downloads 241643 Performance Optimization of Low-Cost Solar Dryer Using Modified PI Controller
Authors: Rajesh Kondareddy, Prakash Kumar Nayak, Maunash Das, Vrinatri Velentina Boro
Abstract:
Today, there is a huge global concern for sustainable development which would include minimizing the consumption of non-renewable energies without affecting the basic global economy. Solar drying is one of the important processes used for extending the shelf life of agricultural products. The performance of a low cost automated solar dryer fitted with cascade control scheme and modified PI controller for drying chilli was investigated. The dryer was composed of designed solar collector (air heater) fitted with cylindrical pipes to improve the air velocity and a solar drying chamber containing rack of two cheese cloth (net) trays both being integrated together. The air allowed in through air inlet is heated up in the solar collector and channelled through the drying chamber where it is utilized in drying (removing the moisture content from the food substance or agricultural produce loaded). Here, to maintain the temperature in the heating chambers and to improve performance, a modified PI (Proportional–Integral) controller was used due its simplicity and robustness. Drying time for drying chilli from the initial moisture content of 88.5% (wb) to 7.3% (wb) was estimated to be 14 hours in solar dryer whereas 32 h was observed in the open sun drying.Keywords: cascade control, chilli, PI controller, solar dryer
Procedia PDF Downloads 286642 Biodsorption as an Efficient Technology for the Removal of Phosphate, Nitrate and Sulphate Anions in Industrial Wastewater
Authors: Angel Villabona-Ortíz, Candelaria Tejada-Tovar, Andrea Viera-Devoz
Abstract:
Wastewater treatment is an issue of vital importance in these times where the impacts of human activities are most evident, which have become essential tasks for the normal functioning of society. However, they put entire ecosystems at risk by time destroying the possibility of sustainable development. Various conventional technologies are used to remove pollutants from water. Agroindustrial waste is the product with the potential to be used as a renewable raw material for the production of energy and chemical products, and their use is beneficial since products with added value are generated from materials that were not used before. Considering the benefits that the use of residual biomass brings, this project proposes the use of agro-industrial residues from corn crops for the production of natural adsorbents whose purpose is aimed at the remediation of contaminated water bodies with large loads of nutrients. The adsorption capacity of two biomaterials obtained from the processing of corn stalks was evaluated by batch system tests. Biochar impregnated with sulfuric acid and thermally activated was synthesized. On the other hand, the cellulose was extracted from the corn stalks and chemically modified with cetyltrimethylammonium chloride in order to quaternize the surface of the adsorbent. The adsorbents obtained were characterized by thermogravimetric analysis (TGA), scanning electron microscopy (SEM), infrared spectrometry with Fourier Transform (FTIR), analysis by Brunauer, Emmett and Teller method (BET) and X-ray Diffraction analysis ( XRD), which showed favorable characteristics for the cellulose extraction process. Higher adsorption capacities of the nutrients were obtained with the use of biochar, with phosphate being the anion with the best removal percentages. The effect of the initial adsorbate concentration was evaluated, with which it was shown that the Freundlich isotherm better describes the adsorption process in most systems. The adsorbent-phosphate / nitrate systems fit better to the Pseudo Primer Order kinetic model, while the adsorbent-sulfate systems showed a better fit to the Pseudo second-order model, which indicates that there are both physical and chemical interactions in the process. Multicomponent adsorption tests revealed that phosphate anions have a higher affinity for both adsorbents. On the other hand, the thermodynamic parameters standard enthalpy (ΔH °) and standard entropy (ΔS °) with negative results indicate the exothermic nature of the process, whereas the ascending values of standard Gibbs free energy (ΔG °). The adsorption process of anions with biocarbon and modified cellulose is spontaneous and exothermic. The use of the evaluated biomateriles is recommended for the treatment of industrial effluents contaminated with sulfate, nitrate and phosphate anions.Keywords: adsorption, biochar, modified cellulose, corn stalks
Procedia PDF Downloads 180641 A Study on Solutions to Connect Distribution Power Grid up to Renewable Energy Sources at KEPCO
Authors: Seung Yoon Hyun, Hyeong Seung An, Myeong Ho Choi, Sung Hwan Bae, Yu Jong Sim
Abstract:
In 2015, the southern part of the Korean Peninsula has 8.6 million poles, 1.25 million km power lines, and 2 million transformers, etc. It is the massive amount of distribution equipments which could cover a round-trip distance from the earth to the moon and 11 turns around the earth. These distribution equipments are spread out like capillaries and supplying power to every corner of the Korean Peninsula. In order to manage these huge power facility efficiently, KEPCO use DAS (Distribution Automation System) to operate distribution power system since 1997. DAS is integrated system that enables to remotely supervise and control breakers and switches on distribution network. Using DAS, we can reduce outage time and power loss. KEPCO has about 160,000 switches, 50%(about 80,000) of switches are automated, and 41 distribution center monitoring&control these switches 24-hour 365 days to get the best efficiency of distribution networks. However, the rapid increasing renewable energy sources become the problem in the efficient operation of distributed power system. (currently 2,400 MW, 75,000 generators operate in distribution power system). In this paper, it suggests the way to interconnect between renewable energy source and distribution power system.Keywords: distribution, renewable, connect, DAS (Distribution Automation System)
Procedia PDF Downloads 620640 Analysis and Modeling of Vibratory Signals Based on LMD for Rolling Bearing Fault Diagnosis
Authors: Toufik Bensana, Slimane Mekhilef, Kamel Tadjine
Abstract:
The use of vibration analysis has been established as the most common and reliable method of analysis in the field of condition monitoring and diagnostics of rotating machinery. Rolling bearings cover a broad range of rotary machines and plays a crucial role in the modern manufacturing industry. Unfortunately, the vibration signals collected from a faulty bearing are generally non-stationary, nonlinear and with strong noise interference, so it is essential to obtain the fault features correctly. In this paper, a novel numerical analysis method based on local mean decomposition (LMD) is proposed. LMD decompose the signal into a series of product functions (PFs), each of which is the product of an envelope signal and a purely frequency modulated FM signal. The envelope of a PF is the instantaneous amplitude (IA) and the derivative of the unwrapped phase of a purely flat frequency demodulated (FM) signal is the IF. After that, the fault characteristic frequency of the roller bearing can be extracted by performing spectrum analysis to the instantaneous amplitude of PF component containing dominant fault information. the results show the effectiveness of the proposed technique in fault detection and diagnosis of rolling element bearing.Keywords: fault diagnosis, local mean decomposition, rolling element bearing, vibration analysis
Procedia PDF Downloads 406639 Application of Chemical Tests for the Inhibition of Scaling From Hamma Hard Waters
Authors: Samira Ghizellaoui, Manel Boumagoura
Abstract:
Calcium carbonate precipitation is a widespread problem, especially in hard water systems. The main water supply that supplies the city of Constantine with drinking water is underground water called Hamma water. This water has a very high hardness of around 590 mg/L CaCO₃. This leads to the formation of scale, consisting mainly of calcium carbonate, which can be responsible for the clogging of valves and the deterioration of equipment (water heaters, washing machines and encrustations in the pipes). Plant extracts used as scale inhibitors have attracted the attention of several researchers. In recent years, green inhibitors have attracted great interest because they are biodegradable, non-toxic and do not affect the environment. The aim of our work is to evaluate the effectiveness of a chemical antiscale treatment in the presence of three green inhibitors: gallicacid; quercetin; alginate, and three mixtures: (gallic acid-quercetin); (quercetin-alginate); (gallic acid-alginate). The results show that the inhibitory effect is manifested from an addition of 1mg/L of gallic acid, 10 mg/L of quercetin, 0.2 mg/L of alginate, 0.4mg/L of (gallic acid-quercetin), 2mg/L of (quercetin-alginate) and 0.4 mg/L of (gallic acid-alginate). On the other hand, 100 mg/L (Drinking water standard) of Ca2+is reached for partial softening at 4 mg/L of gallic acid, 40 mg/L of quercetin, 0.6mg/L of alginate, 4mg/L of (gallic acid-quercetin), 10mg/L of (quercetin-alginate) and 1.6 mg/L of (gallic acid-alginate).Keywords: water, scaling, calcium carbonate, green inhibitor
Procedia PDF Downloads 66638 Early Detection of Breast Cancer in Digital Mammograms Based on Image Processing and Artificial Intelligence
Authors: Sehreen Moorat, Mussarat Lakho
Abstract:
A method of artificial intelligence using digital mammograms data has been proposed in this paper for detection of breast cancer. Many researchers have developed techniques for the early detection of breast cancer; the early diagnosis helps to save many lives. The detection of breast cancer through mammography is effective method which detects the cancer before it is felt and increases the survival rate. In this paper, we have purposed image processing technique for enhancing the image to detect the graphical table data and markings. Texture features based on Gray-Level Co-Occurrence Matrix and intensity based features are extracted from the selected region. For classification purpose, neural network based supervised classifier system has been used which can discriminate between benign and malignant. Hence, 68 digital mammograms have been used to train the classifier. The obtained result proved that automated detection of breast cancer is beneficial for early diagnosis and increases the survival rates of breast cancer patients. The proposed system will help radiologist in the better interpretation of breast cancer.Keywords: medical imaging, cancer, processing, neural network
Procedia PDF Downloads 257637 Auto-Tuning of CNC Parameters According to the Machining Mode Selection
Authors: Jenq-Shyong Chen, Ben-Fong Yu
Abstract:
CNC(computer numerical control) machining centers have been widely used for machining different metal components for various industries. For a specific CNC machine, its everyday job is assigned to cut different products with quite different attributes such as material type, workpiece weight, geometry, tooling, and cutting conditions. Theoretically, the dynamic characteristics of the CNC machine should be properly tuned match each machining job in order to get the optimal machining performance. However, most of the CNC machines are set with only a standard set of CNC parameters. In this study, we have developed an auto-tuning system which can automatically change the CNC parameters and in hence change the machine dynamic characteristics according to the selection of machining modes which are set by the mixed combination of three machine performance indexes: the HO (high surface quality) index, HP (high precision) index and HS (high speed) index. The acceleration, jerk, corner error tolerance, oscillation and dynamic bandwidth of machine’s feed axes have been changed according to the selection of the machine performance indexes. The proposed auto-tuning system of the CNC parameters has been implemented on a PC-based CNC controller and a three-axis machining center. The measured experimental result have shown the promising of our proposed auto-tuning system.Keywords: auto-tuning, CNC parameters, machining mode, high speed, high accuracy, high surface quality
Procedia PDF Downloads 378636 Health Monitoring and Failure Detection of Electronic and Structural Components in Small Unmanned Aerial Vehicles
Authors: Gopi Kandaswamy, P. Balamuralidhar
Abstract:
Fully autonomous small Unmanned Aerial Vehicles (UAVs) are increasingly being used in many commercial applications. Although a lot of research has been done to develop safe, reliable and durable UAVs, accidents due to electronic and structural failures are not uncommon and pose a huge safety risk to the UAV operators and the public. Hence there is a strong need for an automated health monitoring system for UAVs with a view to minimizing mission failures thereby increasing safety. This paper describes our approach to monitoring the electronic and structural components in a small UAV without the need for additional sensors to do the monitoring. Our system monitors data from four sources; sensors, navigation algorithms, control inputs from the operator and flight controller outputs. It then does statistical analysis on the data and applies a rule based engine to detect failures. This information can then be fed back into the UAV and a decision to continue or abort the mission can be taken automatically by the UAV and independent of the operator. Our system has been verified using data obtained from real flights over the past year from UAVs of various sizes that have been designed and deployed by us for various applications.Keywords: fault detection, health monitoring, unmanned aerial vehicles, vibration analysis
Procedia PDF Downloads 260635 Evaluation of Environmental Impact Assessment of Dam Using GIS/Remote Sensing-Review
Authors: Ntungamili Kenosi, Moatlhodi W. Letshwenyo
Abstract:
Negative environmental impacts due to construction of large projects such as dams have become an important aspect of land degradation. This paper will review the previous literature on the previous researches or study in the same area of study in the other parts of the world. After dam has been constructed, the actual environmental impacts are investigated and compared to the predicted results of the carried out Environmental Impact Assessment. GIS and Remote Sensing, play an important role in generating automated spatial data sets and in establishing spatial relationships. Results from other sources shows that the normalized vegetation index (NDVI) analysis was used to detect the spatial and temporal change of vegetation biomass in the study area. The result indicated that the natural vegetation biomass is declining. This is mainly due to the expansion of agricultural land and escalating human made structures in the area. Urgent environmental conservation is necessary when adjoining projects site. Less study on the evaluation of EIA on dam has been conducted in Botswana hence there is a need for the same study to be conducted and then it will be easy to be compared to other studies around the world.Keywords: Botswana, dam, environmental impact assessment, GIS, normalized vegetation index (NDVI), remote sensing
Procedia PDF Downloads 403634 Accuracy Improvement of Traffic Participant Classification Using Millimeter-Wave Radar by Leveraging Simulator Based on Domain Adaptation
Authors: Tokihiko Akita, Seiichi Mita
Abstract:
A millimeter-wave radar is the most robust against adverse environments, making it an essential environment recognition sensor for automated driving. However, the reflection signal is sparse and unstable, so it is difficult to obtain the high recognition accuracy. Deep learning provides high accuracy even for them in recognition, but requires large scale datasets with ground truth. Specially, it takes a lot of cost to annotate for a millimeter-wave radar. For the solution, utilizing a simulator that can generate an annotated huge dataset is effective. Simulation of the radar is more difficult to match with real world data than camera image, and recognition by deep learning with higher-order features using the simulator causes further deviation. We have challenged to improve the accuracy of traffic participant classification by fusing simulator and real-world data with domain adaptation technique. Experimental results with the domain adaptation network created by us show that classification accuracy can be improved even with a few real-world data.Keywords: millimeter-wave radar, object classification, deep learning, simulation, domain adaptation
Procedia PDF Downloads 90633 Assertion-Driven Test Repair Based on Priority Criteria
Authors: Ruilian Zhao, Shukai Zhang, Yan Wang, Weiwei Wang
Abstract:
Repairing broken test cases is an expensive and challenging task in evolving software systems. Although an automated repair technique with intent preservation has been proposed, but it does not take into account the association between test repairs and assertions, leading to a large number of irrelevant candidates and decreasing the repair capability. This paper proposes an assertion-driven test repair approach. Furthermore, an intent-oriented priority criterion is raised to guide the repair candidate generation, making the repairs closer to the intent of the test. In more detail, repair targets are determined through post-dominance relations between assertions and the methods that directly cause compilation errors. Then, test repairs are generated from the target in a bottom-up way, guided by the intent-oriented priority criteria. Finally, the generated repair candidates are prioritized to match the original test intent. The approach is implemented and evaluated on the benchmark of 4 open-source programs and 91 broken test cases. The result shows that the approach can fix 89% (81/91) of broken test cases, which is more effective than the existing intentpreserved test repair approach, and our intent-oriented priority criteria work well.Keywords: test repair, test intent, software test, test case evolution
Procedia PDF Downloads 128632 A Conceptual Framework for the Adoption of Information and Communication Technology for Anti-Corruption in the DR Congo
Authors: Itulelo Matiyabu Imaja, Patrick Ndayizigamiye, Manoj Maharaj
Abstract:
There are many catalysts of corruption. These include amongst others, lack of effective control measures to deter or detect corrupt behaviour. Literature suggests that ICT could assist in curbing corruption through the implementation of automated systems, citizens engagement through e-government and online media to name a few. In the Democratic Republic of Congo, lack of transparency and accountability in public funds collection and allocation contribute to corruption in funds mismanagement. Using the accountability theory and available literature, this paper analyses how Democratic Republic of Congo (DRC) institutions could be strengthened through ICT in order to deter instances of corruption. Findings reveal that DRC lacks reliable control, monitoring and evaluation mechanisms that could identify potentially corrupt behavior. In addition, citizens and civil society organizations who are meant to hold the institutions accountable are not given secure platform to express their views and potentially flag any corrupt behavior. Hence, the paper presents a preliminary conceptual framework that depicts how ICT could be used to strengthen current institutions to potentially deter corrupt behavior in public funds management in Congo.Keywords: corruption, ICT adoption, transparency, DR Congo
Procedia PDF Downloads 184631 A Speeded up Robust Scale-Invariant Feature Transform Currency Recognition Algorithm
Authors: Daliyah S. Aljutaili, Redna A. Almutlaq, Suha A. Alharbi, Dina M. Ibrahim
Abstract:
All currencies around the world look very different from each other. For instance, the size, color, and pattern of the paper are different. With the development of modern banking services, automatic methods for paper currency recognition become important in many applications like vending machines. One of the currency recognition architecture’s phases is Feature detection and description. There are many algorithms that are used for this phase, but they still have some disadvantages. This paper proposes a feature detection algorithm, which merges the advantages given in the current SIFT and SURF algorithms, which we call, Speeded up Robust Scale-Invariant Feature Transform (SR-SIFT) algorithm. Our proposed SR-SIFT algorithm overcomes the problems of both the SIFT and SURF algorithms. The proposed algorithm aims to speed up the SIFT feature detection algorithm and keep it robust. Simulation results demonstrate that the proposed SR-SIFT algorithm decreases the average response time, especially in small and minimum number of best key points, increases the distribution of the number of best key points on the surface of the currency. Furthermore, the proposed algorithm increases the accuracy of the true best point distribution inside the currency edge than the other two algorithms.Keywords: currency recognition, feature detection and description, SIFT algorithm, SURF algorithm, speeded up and robust features
Procedia PDF Downloads 234630 IT Perspective of Service-Oriented e-Government Enterprise
Authors: Anu Paul, Varghese Paul
Abstract:
The focal aspire of e-Government (eGovt) is to offer citizen-centered service delivery. Accordingly, the citizenry consumes services from multiple government agencies through national portal. Thus, eGovt is an enterprise with the primary business motive of transparent, efficient and effective public services to its citizenry and its logical structure is the eGovernment Enterprise Architecture (eGEA). Since eGovt is IT oriented multifaceted service-centric system, EA doesn’t do much on an automated enterprise other than the business artifacts. Service-Oriented Architecture (SOA) manifestation led some governments to pertain this in their eGovts, but it limits the source of business artifacts. The concurrent use of EA and SOA in eGovt executes interoperability and integration and leads to Service-Oriented e-Government Enterprise (SOeGE). Consequently, agile eGovt system becomes a reality. As an IT perspective eGovt comprises of centralized public service artifacts with the existing application logics belong to various departments at central, state and local level. The eGovt is renovating to SOeGE by apply the Service-Orientation (SO) principles in the entire system. This paper explores IT perspective of SOeGE in India which encompasses the public service models and illustrated with a case study the Passport service of India.Keywords: enterprise architecture, service-oriented e-Government enterprise, service interface layer, service model
Procedia PDF Downloads 519629 SPR Immunosensor for the Detection of Staphylococcus aureus
Authors: Muhammad Ali Syed, Arshad Saleem Bhatti, Chen-zhong Li, Habib Ali Bokhari
Abstract:
Surface plasmon resonance (SPR) biosensors have emerged as a promising technique for bioanalysis as well as microbial detection and identification. Real time, sensitive, cost effective, and label free detection of biomolecules from complex samples is required for early and accurate diagnosis of infectious diseases. Like many other types of optical techniques, SPR biosensors may also be successfully utilized for microbial detection for accurate, point of care, and rapid results. In the present study, we have utilized a commercially available automated SPR biosensor of BI company to study the microbial detection form water samples spiked with different concentration of Staphylococcus aureus bacterial cells. The gold thin film sensor surface was functionalized to react with proteins such as protein G, which was used for directed immobilization of monoclonal antibodies against Staphylococcus aureus. The results of our work reveal that this immunosensor can be used to detect very small number of bacterial cells with higher sensitivity and specificity. In our case 10^3 cells/ml of water have been successfully detected. Therefore, it may be concluded that this technique has a strong potential to be used in microbial detection and identification.Keywords: surface plasmon resonance (SPR), Staphylococcus aureus, biosensors, microbial detection
Procedia PDF Downloads 473628 Automated Ultrasound Carotid Artery Image Segmentation Using Curvelet Threshold Decomposition
Authors: Latha Subbiah, Dhanalakshmi Samiappan
Abstract:
In this paper, we propose denoising Common Carotid Artery (CCA) B mode ultrasound images by a decomposition approach to curvelet thresholding and automatic segmentation of the intima media thickness and adventitia boundary. By decomposition, the local geometry of the image, its direction of gradients are well preserved. The components are combined into a single vector valued function, thus removes noise patches. Double threshold is applied to inherently remove speckle noise in the image. The denoised image is segmented by active contour without specifying seed points. Combined with level set theory, they provide sub regions with continuous boundaries. The deformable contours match to the shapes and motion of objects in the images. A curve or a surface under constraints is developed from the image with the goal that it is pulled into the necessary features of the image. Region based and boundary based information are integrated to achieve the contour. The method treats the multiplicative speckle noise in objective and subjective quality measurements and thus leads to better-segmented results. The proposed denoising method gives better performance metrics compared with other state of art denoising algorithms.Keywords: curvelet, decomposition, levelset, ultrasound
Procedia PDF Downloads 339627 Automatic Product Identification Based on Deep-Learning Theory in an Assembly Line
Authors: Fidel Lòpez Saca, Carlos Avilés-Cruz, Miguel Magos-Rivera, José Antonio Lara-Chávez
Abstract:
Automated object recognition and identification systems are widely used throughout the world, particularly in assembly lines, where they perform quality control and automatic part selection tasks. This article presents the design and implementation of an object recognition system in an assembly line. The proposed shapes-color recognition system is based on deep learning theory in a specially designed convolutional network architecture. The used methodology involve stages such as: image capturing, color filtering, location of object mass centers, horizontal and vertical object boundaries, and object clipping. Once the objects are cut out, they are sent to a convolutional neural network, which automatically identifies the type of figure. The identification system works in real-time. The implementation was done on a Raspberry Pi 3 system and on a Jetson-Nano device. The proposal is used in an assembly course of bachelor’s degree in industrial engineering. The results presented include studying the efficiency of the recognition and processing time.Keywords: deep-learning, image classification, image identification, industrial engineering.
Procedia PDF Downloads 157626 Analysis of Noodle Production Process at Yan Hu Food Manufacturing: Basis for Production Improvement
Authors: Rhadinia Tayag-Relanes, Felina C. Young
Abstract:
This study was conducted to analyze the noodle production process at Yan Hu Food Manufacturing for the basis of production improvement. The study utilized the PDCA approach and record review in the gathering of data for the calendar year 2019 from August to October data of the noodle products miki, canton, and misua. Causal-comparative research was used in this study; it attempts to establish cause-effect relationships among the variables such as descriptive statistics and correlation, both were used to compute the data gathered. The study found that miki, canton, and misua production has different cycle time sets for each production and has different production outputs in every set of its production process and a different number of wastages. The company has not yet established its allowable rejection rate/ wastage; instead, this paper used a 1% wastage limit. The researcher recommended the following: machines used for each process of the noodle product must be consistently maintained and monitored; an assessment of all the production operators by checking their performance statistically based on the output and the machine performance; a root cause analysis for finding the solution must be conducted; and an improvement on the recording system of the input and output of the production process of noodle product should be established to eliminate the poor recording of data.Keywords: continuous improvement, process, operations, PDCA
Procedia PDF Downloads 69625 Application of Principle Component Analysis for Classification of Random Doppler-Radar Targets during the Surveillance Operations
Authors: G. C. Tikkiwal, Mukesh Upadhyay
Abstract:
During the surveillance operations at war or peace time, the Radar operator gets a scatter of targets over the screen. This may be a tracked vehicle like tank vis-à-vis T72, BMP etc, or it may be a wheeled vehicle like ALS, TATRA, 2.5Tonne, Shaktiman or moving army, moving convoys etc. The Radar operator selects one of the promising targets into Single Target Tracking (STT) mode. Once the target is locked, the operator gets a typical audible signal into his headphones. With reference to the gained experience and training over the time, the operator then identifies the random target. But this process is cumbersome and is solely dependent on the skills of the operator, thus may lead to misclassification of the object. In this paper we present a technique using mathematical and statistical methods like Fast Fourier Transformation (FFT) and Principal Component Analysis (PCA) to identify the random objects. The process of classification is based on transforming the audible signature of target into music octave-notes. The whole methodology is then automated by developing suitable software. This automation increases the efficiency of identification of the random target by reducing the chances of misclassification. This whole study is based on live data.Keywords: radar target, fft, principal component analysis, eigenvector, octave-notes, dsp
Procedia PDF Downloads 345624 Resource-Constrained Assembly Line Balancing Problems with Multi-Manned Workstations
Authors: Yin-Yann Chen, Jia-Ying Li
Abstract:
Assembly line balancing problems can be categorized into one-sided, two-sided, and multi-manned ones by using the number of operators deployed at workstations. This study explores the balancing problem of a resource-constrained assembly line with multi-manned workstations. Resources include machines or tools in assembly lines such as jigs, fixtures, and hand tools. A mathematical programming model was developed to carry out decision-making and planning in order to minimize the numbers of workstations, resources, and operators for achieving optimal production efficiency. To improve the solution-finding efficiency, a genetic algorithm (GA) and a simulated annealing algorithm (SA) were designed and developed in this study to be combined with a practical case in car making. Results of the GA/SA and mathematics programming were compared to verify their validity. Finally, analysis and comparison were conducted in terms of the target values, production efficiency, and deployment combinations provided by the algorithms in order for the results of this study to provide references for decision-making on production deployment.Keywords: heuristic algorithms, line balancing, multi-manned workstation, resource-constrained
Procedia PDF Downloads 206623 Multi-Agent System for Irrigation Using Fuzzy Logic Algorithm and Open Platform Communication Data Access
Authors: T. Wanyama, B. Far
Abstract:
Automatic irrigation systems usually conveniently protect landscape investment. While conventional irrigation systems are known to be inefficient, automated ones have the potential to optimize water usage. In fact, there is a new generation of irrigation systems that are smart in the sense that they monitor the weather, soil conditions, evaporation and plant water use, and automatically adjust the irrigation schedule. In this paper, we present an agent based smart irrigation system. The agents are built using a mix of commercial off the shelf software, including MATLAB, Microsoft Excel and KEPServer Ex5 OPC server, and custom written code. The Irrigation Scheduler Agent uses fuzzy logic to integrate the information that affect the irrigation schedule. In addition, the Multi-Agent system uses Open Platform Connectivity (OPC) technology to share data. OPC technology enables the Irrigation Scheduler Agent to communicate over the Internet, making the system scalable to a municipal or regional agent based water monitoring, management, and optimization system. Finally, this paper presents simulation and pilot installation test result that show the operational effectiveness of our system.Keywords: community water usage, fuzzy logic, irrigation, multi-agent system
Procedia PDF Downloads 296622 Distributed Processing for Content Based Lecture Video Retrieval on Hadoop Framework
Authors: U. S. N. Raju, Kothuri Sai Kiran, Meena G. Kamal, Vinay Nikhil Pabba, Suresh Kanaparthi
Abstract:
There is huge amount of lecture video data available for public use, and many more lecture videos are being created and uploaded every day. Searching for videos on required topics from this huge database is a challenging task. Therefore, an efficient method for video retrieval is needed. An approach for automated video indexing and video search in large lecture video archives is presented. As the amount of video lecture data is huge, it is very inefficient to do the processing in a centralized computation framework. Hence, Hadoop Framework for distributed computing for Big Video Data is used. First, step in the process is automatic video segmentation and key-frame detection to offer a visual guideline for the video content navigation. In the next step, we extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames. The OCR and detected slide text line types are adopted for keyword extraction, by which both video- and segment-level keywords are extracted for content-based video browsing and search. The performance of the indexing process can be improved for a large database by using distributed computing on Hadoop framework.Keywords: video lectures, big video data, video retrieval, hadoop
Procedia PDF Downloads 532621 Efficiency of Google Translate and Bing Translator in Translating Persian-to-English Texts
Authors: Samad Sajjadi
Abstract:
Machine translation is a new subject increasingly being used by academic writers, especially students and researchers whose native language is not English. There are numerous studies conducted on machine translation, but few investigations have assessed the accuracy of machine translation from Persian to English at lexical, semantic, and syntactic levels. Using Groves and Mundt’s (2015) Model of error taxonomy, the current study evaluated Persian-to-English translations produced by two famous online translators, Google Translate and Bing Translator. A total of 240 texts were randomly selected from different academic fields (law, literature, medicine, and mass media), and 60 texts were considered for each domain. All texts were rendered by the two translation systems and then by four human translators. All statistical analyses were applied using SPSS. The results indicated that Google translations were more accurate than the translations produced by the Bing Translator, especially in the domains of medicine (lexis: 186 vs. 225; semantic: 44 vs. 48; syntactic: 148 vs. 264 errors) and mass media (lexis: 118 vs. 149; semantic: 25 vs. 32; syntactic: 110 vs. 220 errors), respectively. Nonetheless, both machines are reasonably accurate in Persian-to-English translation of lexicons and syntactic structures, particularly from mass media and medical texts.Keywords: machine translations, accuracy, human translation, efficiency
Procedia PDF Downloads 73620 Intelligent System and Renewable Energy: A Farming Platform in Precision Agriculture
Authors: Ryan B. Escorial, Elmer A. Maravillas, Chris Jordan G. Aliac
Abstract:
This study presents a small-scale water pumping system utilizing a fuzzy logic inference system attached to a renewable energy source. The fuzzy logic controller was designed and simulated in MATLAB fuzzy logic toolbox to examine the properties and characteristics of the input and output variables. The result of the simulation was implemented in a microcontroller, together with sensors, modules, and photovoltaic cells. The study used a grand rapid variety of lettuce, organic substrates, and foliar for observation of the capability of the device to irrigate crops. Two plant boxes intended for manual and automated irrigation were prepared with each box having 48 heads of lettuce. The observation of the system took 22-31 days, which is one harvest period of the crop. Results showed a 22.55% increase in agricultural productivity compared to manual irrigation. Aside from reducing human effort, and time, the smart irrigation system could help lessen some of the shortcomings of manual irrigations. It could facilitate the economical utilization of water, reducing consumption by 25%. The use of renewable energy could also help farmers reduce the cost of production by minimizing the use of diesel and gasoline.Keywords: fuzzy logic, intelligent system, precision agriculture, renewable energy
Procedia PDF Downloads 127619 Healthcare Big Data Analytics Using Hadoop
Authors: Chellammal Surianarayanan
Abstract:
Healthcare industry is generating large amounts of data driven by various needs such as record keeping, physician’s prescription, medical imaging, sensor data, Electronic Patient Record(EPR), laboratory, pharmacy, etc. Healthcare data is so big and complex that they cannot be managed by conventional hardware and software. The complexity of healthcare big data arises from large volume of data, the velocity with which the data is accumulated and different varieties such as structured, semi-structured and unstructured nature of data. Despite the complexity of big data, if the trends and patterns that exist within the big data are uncovered and analyzed, higher quality healthcare at lower cost can be provided. Hadoop is an open source software framework for distributed processing of large data sets across clusters of commodity hardware using a simple programming model. The core components of Hadoop include Hadoop Distributed File System which offers way to store large amount of data across multiple machines and MapReduce which offers way to process large data sets with a parallel, distributed algorithm on a cluster. Hadoop ecosystem also includes various other tools such as Hive (a SQL-like query language), Pig (a higher level query language for MapReduce), Hbase(a columnar data store), etc. In this paper an analysis has been done as how healthcare big data can be processed and analyzed using Hadoop ecosystem.Keywords: big data analytics, Hadoop, healthcare data, towards quality healthcare
Procedia PDF Downloads 413618 Development and Automation of Medium-Scale NFT Hydroponic Systems: Design Methodology and State of the Art Review
Authors: Oscar Armando González-Marin, Jhon F. Rodríguez-León, Oscar Mota-Pérez, Jorge Pineda-Piñón, Roberto S. Velázquez-González., Julio C. Sosa-Savedra
Abstract:
Over the past six years, the World Meteorological Organization (WMO) has recorded the warmest years since 1880, primarily attributed to climate change. In addition, the overexploitation of agricultural lands, combined with food and water scarcity, has highlighted the urgent need for sustainable cultivation methods. Hydroponics has emerged as a sustainable farming technique that enables plant cultivation using nutrient solutions without the requirement for traditional soil. Among hydroponic methods, the Nutrient Film Technique (NFT) facilitates plant growth by circulating a nutrient solution continuously. This approach allows the monitoring and precise control of nutritional parameters, with potential for automation and technological integration. This study aims to present the state of the art of automated NFT hydroponic systems, discussing their design methodologies and considerations for implementation. Moreover, a medium-scale NFT system developed at CICATA-QRO is introduced, detailing its current manual management and progress toward automation.Keywords: automation, hydroponics, nutrient film technique, sustainability
Procedia PDF Downloads 33617 Development of K-Factor for Road Geometric Design: A Case Study of North Coast Road in Java
Authors: Edwin Hidayat, Redi Yulianto, Disi Hanafiah
Abstract:
On the one hand, parameters which are used for determining the number of lane on the new road construction are average annual average daily traffic (AADT) and peak hour factor (K-factor). On the other hand, the value of K-factor listed in the guidelines and manual for road planning in Indonesia is a value of adoption or adaptation from foreign guidelines or manuals. Thus, the value is less suitable for Indonesian condition due to differences in road conditions, vehicle type, and driving behavior. The purpose of this study is to provide an example on how to determine k-factor values at a road segment with particular conditions in north coast road, West Java. The methodology is started with collecting traffic volume data for 24 hours over 365 days using PLATO (Automated Traffic Counter) with the approach of video image processing. Then, the traffic volume data is divided into per hour and analyzed by comparing the peak traffic volume in the 30th hour (or other) with the AADT in the same year. The analysis has resulted that for the 30th peak hour the K-factor is 0.97. This value can be used for planning road geometry or evaluating the road capacity performance for the 4/2D interurban road.Keywords: road geometry, K-factor, annual average daily traffic, north coast road
Procedia PDF Downloads 158