Search results for: automated teller machines (atm)
1276 ARABEX: Automated Dotted Arabic Expiration Date Extraction using Optimized Convolutional Autoencoder and Custom Convolutional Recurrent Neural Network
Authors: Hozaifa Zaki, Ghada Soliman
Abstract:
In this paper, we introduced an approach for Automated Dotted Arabic Expiration Date Extraction using Optimized Convolutional Autoencoder (ARABEX) with bidirectional LSTM. This approach is used for translating the Arabic dot-matrix expiration dates into their corresponding filled-in dates. A custom lightweight Convolutional Recurrent Neural Network (CRNN) model is then employed to extract the expiration dates. Due to the lack of available dataset images for the Arabic dot-matrix expiration date, we generated synthetic images by creating an Arabic dot-matrix True Type Font (TTF) matrix to address this limitation. Our model was trained on a realistic synthetic dataset of 3287 images, covering the period from 2019 to 2027, represented in the format of yyyy/mm/dd. We then trained our custom CRNN model using the generated synthetic images to assess the performance of our model (ARABEX) by extracting expiration dates from the translated images. Our proposed approach achieved an accuracy of 99.4% on the test dataset of 658 images, while also achieving a Structural Similarity Index (SSIM) of 0.46 for image translation on our dataset. The ARABEX approach demonstrates its ability to be applied to various downstream learning tasks, including image translation and reconstruction. Moreover, this pipeline (ARABEX+CRNN) can be seamlessly integrated into automated sorting systems to extract expiry dates and sort products accordingly during the manufacturing stage. By eliminating the need for manual entry of expiration dates, which can be time-consuming and inefficient for merchants, our approach offers significant results in terms of efficiency and accuracy for Arabic dot-matrix expiration date recognition.Keywords: computer vision, deep learning, image processing, character recognition
Procedia PDF Downloads 811275 CompleX-Machine: An Automated Testing Tool Using X-Machine Theory
Authors: E. K. A. Ogunshile
Abstract:
This paper is aimed at creating an Automatic Java X-Machine testing tool for software development. The nature of software development is changing; thus, the type of software testing tools required is also changing. Software is growing increasingly complex and, in part due to commercial impetus for faster software releases with new features and value, increasingly in danger of containing faults. These faults can incur huge cost for software development organisations and users; Cambridge Judge Business School’s research estimated the cost of software bugs to the global economy is $312 billion. Beyond the cost, faster software development methodologies and increasing expectations on developers to become testers is driving demand for faster, automated, and effective tools to prevent potential faults as early as possible in the software development lifecycle. Using X-Machine theory, this paper will explore a new tool to address software complexity, changing expectations on developers, faster development pressures and methodologies, with a view to reducing the huge cost of fixing software bugs.Keywords: conformance testing, finite state machine, software testing, x-machine
Procedia PDF Downloads 2671274 Automated Tracking and Statistics of Vehicles at the Signalized Intersection
Authors: Qiang Zhang, Xiaojian Hu1
Abstract:
Intersection is the place where vehicles and pedestrians must pass through, turn and evacuate. Obtaining the motion data of vehicles near the intersection is of great significance for transportation research. Since there are usually many targets and there are more conflicts between targets, this makes it difficult to obtain vehicle motion parameters in traffic videos of intersections. According to the characteristics of traffic videos, this paper applies video technology to realize the automated track, count and trajectory extraction of vehicles to collect traffic data by roadside surveillance cameras installed near the intersections. Based on the video recognition method, the vehicles in each lane near the intersection are tracked with extracting trajectory and counted respectively in various degrees of occlusion and visibility. The performances are compared with current recognized CPU-based algorithms of real-time tracking-by-detection. The speed of the presented system is higher than the others and the system has a better real-time performance. The accuracy of direction has reached about 94.99% on average, and the accuracy of classification and statistics has reached about 75.12% on average.Keywords: tracking and statistics, vehicle, signalized intersection, motion parameter, trajectory
Procedia PDF Downloads 2191273 Wind Energy Potential of Southern Sindh, Pakistan for Power Generation
Authors: M. Akhlaque Ahmed, Maliha Afshan Siddiqui
Abstract:
A study has been carried out to see the prospect of wind power potential of southern Sindh namely Karachi, Hawksbay, Norriabad, Hyderabad, Ketibander and Shahbander using local wind speed data. The monthly average wind speed for these area ranges from 4.5m/sec to 8.5m/sec at 30m height from ground. Extractable wind power, wind energy and Weibul parameter for above mentioned areas have been examined. Furthermore, the power output using fast and slow wind machine using different blade diameter along with the 4Kw and 20 Kw aero-generator were examined to see the possible use for deep well pumping and electricity supply to remote villages. The analysis reveals that in this wind corridor of southern Sindh Hawksbay, Ketibander and Shahbander belongs to wind power class-3 Hyderabad and Nooriabad belongs to wind power class-5 and Karachi belongs to wind power class-2. The result shows that the that higher wind speed values occur between June till August. It was found that considering maximum wind speed location, Hawksbay,Noriabad are the best location for setting up wind machines for power generation.Keywords: wind energy generation, Southern Sindh, seasonal change, Weibull parameter, wind machines
Procedia PDF Downloads 1471272 Modelling of Hydric Behaviour of Textiles
Authors: A. Marolleau, F. Salaun, D. Dupont, H. Gidik, S. Ducept.
Abstract:
The goal of this study is to analyze the hydric behaviour of textiles which can impact significantly the comfort of the wearer. Indeed, fabrics can be adapted for different climate if hydric and thermal behaviors are known. In this study, fabrics are only submitted to hydric variations. Sorption and desorption isotherms obtained from the dynamic vapour sorption apparatus (DVS) are fitted with the parallel exponential kinetics (PEK), the Hailwood-Horrobin (HH) and the Brunauer-Emmett-Teller (BET) models. One of the major finding is the relationship existing between PEK and HH models. During slow and fast processes, the sorption of water molecules on the polymer can be in monolayer and multilayer form. According to the BET model, moisture regain, a physical property of textiles, show a linear correlation with the total amount of water taken in monolayer. This study provides potential information of the end uses of these fabrics according to the selected activity level.Keywords: comfort, hydric properties, modelling, underwears
Procedia PDF Downloads 1471271 A Greedy Alignment Algorithm Supporting Medication Reconciliation
Authors: David Tresner-Kirsch
Abstract:
Reconciling patient medication lists from multiple sources is a critical task supporting the safe delivery of patient care. Manual reconciliation is a time-consuming and error-prone process, and recently attempts have been made to develop efficiency- and safety-oriented automated support for professionals performing the task. An important capability of any such support system is automated alignment – finding which medications from a list correspond to which medications from a different source, regardless of misspellings, naming differences (e.g. brand name vs. generic), or changes in treatment (e.g. switching a patient from one antidepressant class to another). This work describes a new algorithmic solution to this alignment task, using a greedy matching approach based on string similarity, edit distances, concept extraction and normalization, and synonym search derived from the RxNorm nomenclature. The accuracy of this algorithm was evaluated against a gold-standard corpus of 681 medication records; this evaluation found that the algorithm predicted alignments with 99% precision and 91% recall. This performance is sufficient to support decision support applications for medication reconciliation.Keywords: clinical decision support, medication reconciliation, natural language processing, RxNorm
Procedia PDF Downloads 2831270 An Analysis of Packaging Materials for an Energy-Efficient Wrapping System
Authors: John Sweeney, Martin Leeming, Raj Thaker, Cristina L. Tuinea-Bobe
Abstract:
Shrink wrapping is widely used as a method for secondary packaging to assemble individual items, such as cans or other consumer products, into single packages. This method involves conveying the packages into heated tunnels and so has the disadvantages that it is energy-intensive, and, in the case of aerosol products, potentially hazardous. We are developing an automated packaging system that uses stretch wrapping to address both these problems, by using a mechanical rather than a thermal process. In this study, we present a comparative study of shrink wrapping and stretch wrapping materials to assess the relative capability of candidate stretch wrap polymer film in terms of mechanical response. The stretch wrap materials are of oriented polymer and therefore elastically anisotropic. We are developing material constitutive models that include both anisotropy and nonlinearity. These material models are to be incorporated into computer simulations of the automated stretch wrapping system. We present results showing the validity of these models and the feasibility of applying them in the simulations.Keywords: constitutive model, polymer, mechanical testing, wrapping system
Procedia PDF Downloads 2921269 Meta Mask Correction for Nuclei Segmentation in Histopathological Image
Authors: Jiangbo Shi, Zeyu Gao, Chen Li
Abstract:
Nuclei segmentation is a fundamental task in digital pathology analysis and can be automated by deep learning-based methods. However, the development of such an automated method requires a large amount of data with precisely annotated masks which is hard to obtain. Training with weakly labeled data is a popular solution for reducing the workload of annotation. In this paper, we propose a novel meta-learning-based nuclei segmentation method which follows the label correction paradigm to leverage data with noisy masks. Specifically, we design a fully conventional meta-model that can correct noisy masks by using a small amount of clean meta-data. Then the corrected masks are used to supervise the training of the segmentation model. Meanwhile, a bi-level optimization method is adopted to alternately update the parameters of the main segmentation model and the meta-model. Extensive experimental results on two nuclear segmentation datasets show that our method achieves the state-of-the-art result. In particular, in some noise scenarios, it even exceeds the performance of training on supervised data.Keywords: deep learning, histopathological image, meta-learning, nuclei segmentation, weak annotations
Procedia PDF Downloads 1381268 Design and Field Programmable Gate Array Implementation of Radio Frequency Identification for Boosting up Tag Data Processing
Authors: G. Rajeshwari, V. D. M. Jabez Daniel
Abstract:
Radio Frequency Identification systems are used for automated identification in various applications such as automobiles, health care and security. It is also called as the automated data collection technology. RFID readers are placed in any area to scan large number of tags to cover a wide distance. The placement of the RFID elements may result in several types of collisions. A major challenge in RFID system is collision avoidance. In the previous works the collision was avoided by using algorithms such as ALOHA and tree algorithm. This work proposes collision reduction and increased throughput through reading enhancement method with tree algorithm. The reading enhancement is done by improving interrogation procedure and increasing the data handling capacity of RFID reader with parallel processing. The work is simulated using Xilinx ISE 14.5 verilog language. By implementing this in the RFID system, we can able to achieve high throughput and avoid collision in the reader at a same instant of time. The overall system efficiency will be increased by implementing this.Keywords: antenna, anti-collision protocols, data management system, reader, reading enhancement, tag
Procedia PDF Downloads 3041267 Automated Heart Sound Classification from Unsegmented Phonocardiogram Signals Using Time Frequency Features
Authors: Nadia Masood Khan, Muhammad Salman Khan, Gul Muhammad Khan
Abstract:
Cardiologists perform cardiac auscultation to detect abnormalities in heart sounds. Since accurate auscultation is a crucial first step in screening patients with heart diseases, there is a need to develop computer-aided detection/diagnosis (CAD) systems to assist cardiologists in interpreting heart sounds and provide second opinions. In this paper different algorithms are implemented for automated heart sound classification using unsegmented phonocardiogram (PCG) signals. Support vector machine (SVM), artificial neural network (ANN) and cartesian genetic programming evolved artificial neural network (CGPANN) without the application of any segmentation algorithm has been explored in this study. The signals are first pre-processed to remove any unwanted frequencies. Both time and frequency domain features are then extracted for training the different models. The different algorithms are tested in multiple scenarios and their strengths and weaknesses are discussed. Results indicate that SVM outperforms the rest with an accuracy of 73.64%.Keywords: pattern recognition, machine learning, computer aided diagnosis, heart sound classification, and feature extraction
Procedia PDF Downloads 2601266 Increasing the Capacity of Plant Bottlenecks by Using of Improving the Ratio of Mean Time between Failures to Mean Time to Repair
Authors: Jalal Soleimannejad, Mohammad Asadizeidabadi, Mahmoud Koorki, Mojtaba Azarpira
Abstract:
A significant percentage of production costs is the maintenance costs, and analysis of maintenance costs could to achieve greater productivity and competitiveness. With this is mind, the maintenance of machines and installations is considered as an essential part of organizational functions and applying effective strategies causes significant added value in manufacturing activities. Organizations are trying to achieve performance levels on a global scale with emphasis on creating competitive advantage by different methods consist of RCM (Reliability-Center-Maintenance), TPM (Total Productivity Maintenance) etc. In this study, increasing the capacity of Concentration Plant of Golgohar Iron Ore Mining & Industrial Company (GEG) was examined by using of reliability and maintainability analyses. The results of this research showed that instead of increasing the number of machines (in order to solve the bottleneck problems), the improving of reliability and maintainability would solve bottleneck problems in the best way. It should be mention that in the abovementioned study, the data set of Concentration Plant of GEG as a case study, was applied and analyzed.Keywords: bottleneck, golgohar iron ore mining & industrial company, maintainability, maintenance costs, reliability
Procedia PDF Downloads 3621265 Energy Efficient Resource Allocation and Scheduling in Cloud Computing Platform
Authors: Shuen-Tai Wang, Ying-Chuan Chen, Yu-Ching Lin
Abstract:
There has been renewal of interest in the relation between Green IT and cloud computing in recent years. Cloud computing has to be a highly elastic environment which provides stable services to users. The growing use of cloud computing facilities has caused marked energy consumption, putting negative pressure on electricity cost of computing center or data center. Each year more and more network devices, storages and computers are purchased and put to use, but it is not just the number of computers that is driving energy consumption upward. We could foresee that the power consumption of cloud computing facilities will double, triple, or even more in the next decade. This paper aims at resource allocation and scheduling technologies that are short of or have not well developed yet to reduce energy utilization in cloud computing platform. In particular, our approach relies on recalling services dynamically onto appropriate amount of the machines according to user’s requirement and temporarily shutting down the machines after finish in order to conserve energy. We present initial work on integration of resource and power management system that focuses on reducing power consumption such that they suffice for meeting the minimizing quality of service required by the cloud computing platform.Keywords: cloud computing, energy utilization, power consumption, resource allocation
Procedia PDF Downloads 3381264 On the Network Packet Loss Tolerance of SVM Based Activity Recognition
Authors: Gamze Uslu, Sebnem Baydere, Alper K. Demir
Abstract:
In this study, data loss tolerance of Support Vector Machines (SVM) based activity recognition model and multi activity classification performance when data are received over a lossy wireless sensor network is examined. Initially, the classification algorithm we use is evaluated in terms of resilience to random data loss with 3D acceleration sensor data for sitting, lying, walking and standing actions. The results show that the proposed classification method can recognize these activities successfully despite high data loss. Secondly, the effect of differentiated quality of service performance on activity recognition success is measured with activity data acquired from a multi hop wireless sensor network, which introduces high data loss. The effect of number of nodes on the reliability and multi activity classification success is demonstrated in simulation environment. To the best of our knowledge, the effect of data loss in a wireless sensor network on activity detection success rate of an SVM based classification algorithm has not been studied before.Keywords: activity recognition, support vector machines, acceleration sensor, wireless sensor networks, packet loss
Procedia PDF Downloads 4751263 Improving Urban Mobility: Analyzing Impacts of Connected and Automated Vehicles on Traffic and Emissions
Authors: Saad Roustom, Hajo Ribberink
Abstract:
In most cities in the world, traffic has increased strongly over the last decades, causing high levels of congestion and deteriorating inner-city air quality. This study analyzes the impact of connected and automated vehicles (CAVs) on traffic performance and greenhouse gas (GHG) emissions under different CAV penetration rates in mixed fleet environments of CAVs and driver-operated vehicles (DOVs) and under three different traffic demand levels. Utilizing meso-scale traffic simulations of the City of Ottawa, Canada, the research evaluates the traffic performance of three distinct CAV driving behaviors—Cautious, Normal, and Aggressive—at penetration rates of 25%, 50%, 75%, and 100%, across three different traffic demand levels. The study employs advanced correlation models to estimate GHG emissions. The results reveal that Aggressive and Normal CAVs generally reduce traffic congestion and GHG emissions, with their benefits being more pronounced at higher penetration rates (50% to 100%) and elevated traffic demand levels. On the other hand, Cautious CAVs exhibit an increase in both traffic congestion and GHG emissions. However, results also show deteriorated traffic flow conditions when introducing 25% penetration rates of any type of CAVs. Aggressive CAVs outperform all other driving at improving traffic flow conditions and reducing GHG emissions. The findings of this study highlight the crucial role CAVs can play in enhancing urban traffic performance and mitigating the adverse impact of transportation on the environment. This research advocates for the adoption of effective CAV-related policies by regulatory bodies to optimize traffic flow and reduce GHG emissions. By providing insights into the impact of CAVs, this study aims to inform strategic decision-making and stimulate the development of sustainable urban mobility solutions.Keywords: connected and automated vehicles, congestion, GHG emissions, mixed fleet environment, traffic performance, traffic simulations
Procedia PDF Downloads 881262 Statistical Manufacturing Cell/Process Qualification Sample Size Optimization
Authors: Angad Arora
Abstract:
In production operations/manufacturing, a cell or line is typically a bunch of similar machines (computer numerical control (CNCs), advanced cutting, 3D printing or special purpose machines. For qualifying a typical manufacturing line /cell / new process, Ideally, we need a sample of parts that can be flown through the process and then we make a judgment on the health of the line/cell. However, with huge volumes and mass production scope, such as in the mobile phone industry, for example, the actual cells or lines can go in thousands and to qualify each one of them with statistical confidence means utilizing samples that are very large and eventually add to product /manufacturing cost + huge waste if the parts are not intended to be customer shipped. To solve this, we come up with 2 steps statistical approach. We start with a small sample size and then objectively evaluate whether the process needs additional samples or not. For example, if a process is producing bad parts and we saw those samples early, then there is a high chance that the process will not meet the desired yield and there is no point in keeping adding more samples. We used this hypothesis and came up with 2 steps binomial testing approach. Further, we also prove through results that we can achieve an 18-25% reduction in samples while keeping the same statistical confidence.Keywords: statistics, data science, manufacturing process qualification, production planning
Procedia PDF Downloads 921261 Ophthalmic Hashing Based Supervision of Glaucoma and Corneal Disorders Imposed on Deep Graphical Model
Authors: P. S. Jagadeesh Kumar, Yang Yung, Mingmin Pan, Xianpei Li, Wenli Hu
Abstract:
Glaucoma is impelled by optic nerve mutilation habitually represented as cupping and visual field injury frequently with an arcuate pattern of mid-peripheral loss, subordinate to retinal ganglion cell damage and death. Glaucoma is the second foremost cause of blindness and the chief cause of permanent blindness worldwide. Consequently, all-embracing study into the analysis and empathy of glaucoma is happening to escort deep learning based neural network intrusions to deliberate this substantial optic neuropathy. This paper advances an ophthalmic hashing based supervision of glaucoma and corneal disorders preeminent on deep graphical model. Ophthalmic hashing is a newly proposed method extending the efficacy of visual hash-coding to predict glaucoma corneal disorder matching, which is the faster than the existing methods. Deep graphical model is proficient of learning interior explications of corneal disorders in satisfactory time to solve hard combinatoric incongruities using deep Boltzmann machines.Keywords: corneal disorders, deep Boltzmann machines, deep graphical model, glaucoma, neural networks, ophthalmic hashing
Procedia PDF Downloads 2491260 Optimizing Water Consumption of a Washer-Dryer Which Contains Water Condensation Technology under a Constraint of Energy Consumption and Drying Performance
Authors: Aysegul Sarac
Abstract:
Washer-dryers are the machines which can either wash the laundries or can dry them. In other words, we can define a washer-dryer as a washing machine and a dryer in one machine. Washing machines are characterized by the loading capacity, cabinet depth and spin speed. Dryers are characterized by the drying technology. On the other hand, energy efficiency, water consumption, and noise levels are main characteristics that influence customer decisions to buy washers. Water condensation technology is the most common drying technology existing in the washer-dryer market. Water condensation technology uses water to dry the laundry inside the machine. Thus, in this type of the drying technology water consumption is at high levels comparing other technologies. Water condensation technology sprays cold water in the drum to condense the humidity of hot weather in order to dry the laundry inside. Thus, water consumption influences the drying performance. The scope of this study is to optimize water consumption during drying process under a constraint of energy consumption and drying performance. We are using 6-Sigma methodology to find the optimum water consumption by comparing drying performances of different drying algorithms.Keywords: optimization, 6-Sigma methodology, washer-dryers, water condensation technology
Procedia PDF Downloads 3581259 Automated Building Internal Layout Design Incorporating Post-Earthquake Evacuation Considerations
Authors: Sajjad Hassanpour, Vicente A. González, Yang Zou, Jiamou Liu
Abstract:
Earthquakes pose a significant threat to both structural and non-structural elements in buildings, putting human lives at risk. Effective post-earthquake evacuation is critical for ensuring the safety of building occupants. However, current design practices often neglect the integration of post-earthquake evacuation considerations into the early-stage architectural design process. To address this gap, this paper presents a novel automated internal architectural layout generation tool that optimizes post-earthquake evacuation performance. The tool takes an initial plain floor plan as input, along with specific requirements from the user/architect, such as minimum room dimensions, corridor width, and exit lengths. Based on these inputs, firstly, the tool randomly generates different architectural layouts. Secondly, the human post-earthquake evacuation behaviour will be thoroughly assessed for each generated layout using the advanced Agent-Based Building Earthquake Evacuation Simulation (AB2E2S) model. The AB2E2S prototype is a post-earthquake evacuation simulation tool that incorporates variables related to earthquake intensity, architectural layout, and human factors. It leverages a hierarchical agent-based simulation approach, incorporating reinforcement learning to mimic human behaviour during evacuation. The model evaluates different layout options and provides feedback on evacuation flow, time, and possible casualties due to earthquake non-structural damage. By integrating the AB2E2S model into the automated layout generation tool, architects and designers can obtain optimized architectural layouts that prioritize post-earthquake evacuation performance. Through the use of the tool, architects and designers can explore various design alternatives, considering different minimum room requirements, corridor widths, and exit lengths. This approach ensures that evacuation considerations are embedded in the early stages of the design process. In conclusion, this research presents an innovative automated internal architectural layout generation tool that integrates post-earthquake evacuation simulation. By incorporating evacuation considerations into the early-stage design process, architects and designers can optimize building layouts for improved post-earthquake evacuation performance. This tool empowers professionals to create resilient designs that prioritize the safety of building occupants in the face of seismic events.Keywords: agent-based simulation, automation in design, architectural layout, post-earthquake evacuation behavior
Procedia PDF Downloads 1031258 Segmentation Using Multi-Thresholded Sobel Images: Application to the Separation of Stuck Pollen Grains
Authors: Endrick Barnacin, Jean-Luc Henry, Jimmy Nagau, Jack Molinie
Abstract:
Being able to identify biological particles such as spores, viruses, or pollens is important for health care professionals, as it allows for appropriate therapeutic management of patients. Optical microscopy is a technology widely used for the analysis of these types of microorganisms, because, compared to other types of microscopy, it is not expensive. The analysis of an optical microscope slide is a tedious and time-consuming task when done manually. However, using machine learning and computer vision, this process can be automated. The first step of an automated microscope slide image analysis process is segmentation. During this step, the biological particles are localized and extracted. Very often, the use of an automatic thresholding method is sufficient to locate and extract the particles. However, in some cases, the particles are not extracted individually because they are stuck to other biological elements. In this paper, we propose a stuck particles separation method based on the use of the Sobel operator and thresholding. We illustrate it by applying it to the separation of 813 images of adjacent pollen grains. The method correctly separated 95.4% of these images.Keywords: image segmentation, stuck particles separation, Sobel operator, thresholding
Procedia PDF Downloads 1281257 Improvement in Ni (II) Adsorption Capacity by Using Fe-Nano Zeolite
Authors: Pham-Thi Huong, Byeong-Kyu Lee, Jitae Kim, Chi-Hyeon Lee
Abstract:
Fe-nano zeolite adsorbent was used for removal of Ni (II) ions from aqueous solution. The adsorbent was characterized by Fourier transform infrared spectroscopy (FT-IR), scanning electron microscopy (SEM) and the surface area Brunauer–Emmett–Teller (BET) using for analysis of functional groups, morphology and surface area. Bath adsorption experiments were analyzed on the effect of pH, time, adsorbent doses and initial Ni (II) concentration. The optimum pH for Ni (II) removal using Fe-nano zeolite was found at 5.0 and 90 min of reaction time. The maximum adsorption capacity of Ni (II) was 231.68 mg/g based on the Langmuir isotherm. The kinetics data for the adsorption process was fitted with the pseudo-second-order model. The desorption of Ni (II) from Ni-loaded Fe-nano zeolite was analyzed and even after 10 cycles 72 % desorption was achieved. These finding supported that Fe-nano zeolite with high adsorption capacity, high reuse ability would be utilized for Ni (II) removal from water.Keywords: Fe-nano zeolite, adsorption, Ni (II) removal, regeneration
Procedia PDF Downloads 2301256 Advanced Particle Characterisation of Suspended Sediment in the Danube River Using Automated Imaging and Laser Diffraction
Authors: Flóra Pomázi, Sándor Baranya, Zoltán Szalai
Abstract:
A harmonized monitoring of the suspended sediment transport along such a large river as the world’s most international river, the Danube River, is a rather challenging task. The traditional monitoring method in Hungary is obsolete but using indirect measurement devices and techniques like optical backscatter sensors (OBS), laser diffraction or acoustic backscatter sensors (ABS) could provide a fast and efficient alternative option of direct methods. However, these methods are strongly sensitive to the particle characteristics (i.e. particle shape, particle size and mineral composition). The current method does not provide sufficient information about particle size distribution, mineral analysis is rarely done, and the shape of the suspended sediment particles have not been examined yet. The aims of the study are (1) to determine the particle characterisation of suspended sediment in the Danube River using advanced particle characterisation methods as laser diffraction and automated imaging, and (2) to perform a sensitivity analysis of the indirect methods in order to determine the impact of suspended particle characteristics. The particle size distribution is determined by laser diffraction. The particle shape and mineral composition analysis is done by the Morphologi G3ID image analyser. The investigated indirect measurement devices are the LISST-Portable|XR, the LISST-ABS (Sequoia Inc.) and the Rio Grande 1200 kHz ADCP (Teledyne Marine). The major findings of this study are (1) the statistical shape of the suspended sediment particle - this is the first research in this context, (2) the actualised particle size distribution – that can be compared to historical information, so that the morphological changes can be tracked, (3) the actual mineral composition of the suspended sediment in the Danube River, and (4) the reliability of the tested indirect methods has been increased – based on the results of the sensitivity analysis and the previous findings.Keywords: advanced particle characterisation, automated imaging, indirect methods, laser diffraction, mineral composition, suspended sediment
Procedia PDF Downloads 1441255 A Technique for Image Segmentation Using K-Means Clustering Classification
Authors: Sadia Basar, Naila Habib, Awais Adnan
Abstract:
The paper presents the Technique for Image Segmentation Using K-Means Clustering Classification. The presented algorithms were specific, however, missed the neighboring information and required high-speed computerized machines to run the segmentation algorithms. Clustering is the process of partitioning a group of data points into a small number of clusters. The proposed method is content-aware and feature extraction method which is able to run on low-end computerized machines, simple algorithm, required low-quality streaming, efficient and used for security purpose. It has the capability to highlight the boundary and the object. At first, the user enters the data in the representation of the input. Then in the next step, the digital image is converted into groups clusters. Clusters are divided into many regions. The same categories with same features of clusters are assembled within a group and different clusters are placed in other groups. Finally, the clusters are combined with respect to similar features and then represented in the form of segments. The clustered image depicts the clear representation of the digital image in order to highlight the regions and boundaries of the image. At last, the final image is presented in the form of segments. All colors of the image are separated in clusters.Keywords: clustering, image segmentation, K-means function, local and global minimum, region
Procedia PDF Downloads 3711254 HcDD: The Hybrid Combination of Disk Drives in Active Storage Systems
Authors: Shu Yin, Zhiyang Ding, Jianzhong Huang, Xiaojun Ruan, Xiaomin Zhu, Xiao Qin
Abstract:
Since large-scale and data-intensive applications have been widely deployed, there is a growing demand for high-performance storage systems to support data-intensive applications. Compared with traditional storage systems, next-generation systems will embrace dedicated processor to reduce computational load of host machines and will have hybrid combinations of different storage devices. The advent of flash- memory-based solid state disk has become a critical role in revolutionizing the storage world. However, instead of simply replacing the traditional magnetic hard disk with the solid state disk, it is believed that finding a complementary approach to corporate both of them is more challenging and attractive. This paper explores an idea of active storage, an emerging new storage configuration, in terms of the architecture and design, the parallel processing capability, the cooperation of other machines in cluster computing environment, and a disk configuration, the hybrid combination of different types of disk drives. Experimental results indicate that the proposed HcDD achieves better I/O performance and longer storage system lifespan.Keywords: arallel storage system, hybrid storage system, data inten- sive, solid state disks, reliability
Procedia PDF Downloads 4471253 First-Person Point of View in Contrast to Globalisation in Somerset Maugham’s ‘Mr. Know-All’
Authors: Armel Mbon
Abstract:
This paper discusses the first-person point of view in Maugham's 'Mr. Know-All.' It particularly analyses the narrator's position in relation to the story told in this short story, with the intention of disclosing the latter's prejudice against Mr. Kelada, the protagonist, and, consequently, its hindrance to globalisation. It thus underlines the fact that this protagonist and other travellers are different colours, but one person on this ship epitomises globalisation. The general attitude of readers is that they are inclined to easily believe the narrator while forgetting that fiction is the work of a taler, a teller, but, first and foremost, a liar. The audience, whether it is disconnected from the setting or not, also tends to forget that "travellers from afar can lie with impunity." In fact, the nameless narrator in Maugham's short story has a persona that leaves a lot to be desired. He is prejudiced against Mr. Kelada, known as Mr. Know-All, as will be evidenced by the scrutiny of his diction. This paper finally purports to show that those who proclaim globalisation loudly are not ready to live together.Keywords: narrator, persona, point of view, diction, contrast, globalisation
Procedia PDF Downloads 901252 Hydrothermal Synthesis of Mesoporous Carbon Nanospheres and Their Electrochemical Properties for Glucose Detection
Authors: Ali Akbar Kazemi Asl, Mansour Rahsepar
Abstract:
Mesoporous carbon nanospheres (MCNs) with uniform particle size distribution having an average of 290 nm and large specific surface area (274.4 m²/g) were synthesized by a one-step hydrothermal method followed by the calcination process and then utilized as an enzyme-free glucose biosensor. Morphology, crystal structure, and porous nature of the synthesized nanospheres were characterized by scanning electron microscopy (SEM), X-Ray diffraction (XRD), and Brunauer–Emmett–Teller (BET) analysis, respectively. Also, the electrochemical performance of the MCNs@GCE electrode for the measurement of glucose concentration in alkaline media was investigated by electrochemical impedance spectroscopy (EIS), cyclic voltammetry (CV), and chronoamperometry (CA). MCNs@GCE electrode shows good sensing performance, including a rapid glucose oxidation response within 3.1 s, a wide linear range of 0.026-12 mM, a sensitivity of 212.34 μA.mM⁻¹.cm⁻², and a detection limit of 25.7 μM with excellent selectivity.Keywords: biosensor, electrochemical, glucose, mesoporous carbon, non-enzymatic
Procedia PDF Downloads 1871251 Data Security and Privacy Challenges in Cloud Computing
Authors: Amir Rashid
Abstract:
Cloud Computing frameworks empower organizations to cut expenses by outsourcing computation resources on-request. As of now, customers of Cloud service providers have no methods for confirming the privacy and ownership of their information and data. To address this issue we propose the platform of a trusted cloud computing program (TCCP). TCCP empowers Infrastructure as a Service (IaaS) suppliers, for example, Amazon EC2 to give a shout box execution condition that ensures secret execution of visitor virtual machines. Also, it permits clients to bear witness to the IaaS supplier and decide if the administration is secure before they dispatch their virtual machines. This paper proposes a Trusted Cloud Computing Platform (TCCP) for guaranteeing the privacy and trustworthiness of computed data that are outsourced to IaaS service providers. The TCCP gives the deliberation of a shut box execution condition for a client's VM, ensuring that no cloud supplier's authorized manager can examine or mess up with its data. Furthermore, before launching the VM, the TCCP permits a client to dependably and remotely acknowledge that the provider at backend is running a confided in TCCP. This capacity extends the verification of whole administration, and hence permits a client to confirm the data operation in secure mode.Keywords: cloud security, IaaS, cloud data privacy and integrity, hybrid cloud
Procedia PDF Downloads 2971250 A Case Study of Deep Learning for Disease Detection in Crops
Authors: Felipe A. Guth, Shane Ward, Kevin McDonnell
Abstract:
In the precision agriculture area, one of the main tasks is the automated detection of diseases in crops. Machine Learning algorithms have been studied in recent decades for such tasks in view of their potential for improving economic outcomes that automated disease detection may attain over crop fields. The latest generation of deep learning convolution neural networks has presented significant results in the area of image classification. In this way, this work has tested the implementation of an architecture of deep learning convolution neural network for the detection of diseases in different types of crops. A data augmentation strategy was used to meet the requirements of the algorithm implemented with a deep learning framework. Two test scenarios were deployed. The first scenario implemented a neural network under images extracted from a controlled environment while the second one took images both from the field and the controlled environment. The results evaluated the generalisation capacity of the neural networks in relation to the two types of images presented. Results yielded a general classification accuracy of 59% in scenario 1 and 96% in scenario 2.Keywords: convolutional neural networks, deep learning, disease detection, precision agriculture
Procedia PDF Downloads 2561249 Shared Versus Pooled Automated Vehicles: Exploring Behavioral Intentions Towards On-Demand Automated Vehicles
Authors: Samira Hamiditehrani
Abstract:
Automated vehicles (AVs) are emerging technologies that could potentially offer a wide range of opportunities and challenges for the transportation sector. The advent of AV technology has also resulted in new business models in shared mobility services where many ride hailing and car sharing companies are developing on-demand AVs including shared automated vehicles (SAVs) and pooled automated vehicles (Pooled AVs). SAVs and Pooled AVs could provide alternative shared mobility services which encourage sustainable transport systems, mitigate traffic congestion, and reduce automobile dependency. However, the success of on-demand AVs in addressing major transportation policy issues depends on whether and how the public adopts them as regular travel modes. To identify conditions under which individuals may adopt on-demand AVs, previous studies have applied human behavior and technology acceptance theories, where Theory of Planned Behavior (TPB) has been validated and is among the most tested in on-demand AV research. In this respect, this study has three objectives: (a) to propose and validate a theoretical model for behavioral intention to use SAVs and Pooled AVs by extending the original TPB model; (b) to identify the characteristics of early adopters of SAVs, who prefer to have a shorter and private ride, versus prospective users of Pooled AVs, who choose more affordable but longer and shared trips; and (c) to investigate Canadians’ intentions to adopt on-demand AVs for regular trips. Toward this end, this study uses data from an online survey (n = 3,622) of workers or adult students (18 to 75 years old) conducted in October and November 2021 for six major Canadian metropolitan areas: Toronto, Vancouver, Ottawa, Montreal, Calgary, and Hamilton. To accomplish the goals of this study, a base bivariate ordered probit model, in which both SAV and Pooled AV adoptions are estimated as ordered dependent variables, alongside a full structural equation modeling (SEM) system are estimated. The findings of this study indicate that affective motivations such as attitude towards AV technology, perceived privacy, and subjective norms, matter more than sociodemographic and travel behavior characteristic in adopting on-demand AVs. Also, the results of second objective provide evidence that although there are a few affective motivations, such as subjective norms and having ample knowledge, that are common between early adopters of SAVs and PooledAVs, many examined motivations differ among SAV and Pooled AV adoption factors. In other words, motivations influencing intention to use on-demand AVs differ among the service types. Likewise, depending on the types of on-demand AVs, the sociodemographic characteristics of early adopters differ significantly. In general, findings paint a complex picture with respect to the application of constructs from common technology adoption models to the study of on-demand AVs. Findings from the final objective suggest that policymakers, planners, the vehicle and technology industries, and the public at large should moderate their expectations that on-demand AVs may suddenly transform the entire transportation sector. Instead, this study suggests that SAVs and Pooled AVs (when they entire the Canadian market) are likely to be adopted as supplementary mobility tools rather than substitutions for current travel modesKeywords: automated vehicles, Canadian perception, theory of planned behavior, on-demand AVs
Procedia PDF Downloads 721248 Analysis Mechanized Boring (TBM) of Tehran Subway Line 7
Authors: Shahin Shabani, Pouya Pourmadadi
Abstract:
Tunnel boring machines (TBMs) have been used for the construction of various tunnels for mining projects for the purpose of access, conveyance of ore and waste, drainage, exploration, water supply and water diversion. Several mining projects have seen the successful and economic beneficial use of TBMs, and there is an increasing awareness of the benefits of TBMs for mining projects. Key technical considerations for the use of TBMs for the construction of tunnels for mining projects include geological issues (rock type, rock alteration, rock strength, rock abrasivity, durability, ground water inflows), depth of cover and the potential for overstressing/rockbursts, site access and terrain, portal locations, TBM constraints, minimum tunnel size, tunnel support requirements, contractor and labor experience, and project schedule demands. This study focuses on tunnelling mining, with the goal to develop methods and tools to be used to gain understanding of these processes, and to analyze metro of Tehran. The Metro Line 7 of Tehran is one of the Longest (26 Km) and deepest (27m) of projects that’s under implementation. Because of major differences like passing under all geotechnical layers of the town and encountering part of it with underground water table and also using mechanized excavation system, is one of special metro projects.Keywords: TBM, tunnel boring machines economic, metro, line 7
Procedia PDF Downloads 3831247 Designing of Induction Motor Efficiency Monitoring System
Authors: Ali Mamizadeh, Ires Iskender, Saeid Aghaei
Abstract:
Energy is one of the important issues with high priority property in the world. Energy demand is rapidly increasing depending on the growing population and industry. The useable energy sources in the world will be insufficient to meet the need for energy. Therefore, the efficient and economical usage of energy sources is getting more importance. In a survey conducted among electric consuming machines, the electrical machines are consuming about 40% of the total electrical energy consumed by electrical devices and 96% of this consumption belongs to induction motors. Induction motors are the workhorses of industry and have very large application areas in industry and urban systems like water pumping and distribution systems, steel and paper industries and etc. Monitoring and the control of the motors have an important effect on the operating performance of the motor, driver selection and replacement strategy management of electrical machines. The sensorless monitoring system for monitoring and calculating efficiency of induction motors are studied in this study. The equivalent circuit of IEEE is used in the design of this study. The terminal current and voltage of induction motor are used in this motor to measure the efficiency of induction motor. The motor nameplate information and the measured current and voltage are used in this system to calculate accurately the losses of induction motor to calculate its input and output power. The efficiency of the induction motor is monitored online in the proposed method without disconnecting the motor from the driver and without adding any additional connection at the motor terminal box. The proposed monitoring system measure accurately the efficiency by including all losses without using torque meter and speed sensor. The monitoring system uses embedded architecture and does not need to connect to a computer to measure and log measured data. The conclusion regarding the efficiency, the accuracy and technical and economical benefits of the proposed method are presented. The experimental verification has been obtained on a 3 phase 1.1 kW, 2-pole induction motor. The proposed method can be used for optimal control of induction motors, efficiency monitoring and motor replacement strategy.Keywords: induction motor, efficiency, power losses, monitoring, embedded design
Procedia PDF Downloads 346