Search results for: automated teller machines
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1536

Search results for: automated teller machines

936 The Nature of Intelligence and Its Forms: An Ontological-Modeling Approach

Authors: Husam El-Asfour, Fateh Adhnouss, Kenneth McIsaac, Abdul Mutalib Wahaishi, Raafat Aburukba, Idris El-Feghia

Abstract:

Although intelligence is commonly referred to as the observable behavior in various fields and domains, it must also be shown how it develops by exhibiting multiple forms and without observing the inherent behavior. There have been several official and informal definitions of intelligence in various areas; however, no scientific agreement on a definition has been agreed upon. There must be a single definition, structure, and precise modeling for articulating how intelligence is perceived by people and machines in order to comprehend intelligence. Another key challenge is defining the different environment types based on the integral elements (agents) and their possible interactions. On the basis of conceptualization, this paper proposes a formal model for defining and developing intelligence. Forms of intelligence are derived from an ontological view, and thus intelligence is defined, described, and modeled based on the various types of environments.

Keywords: intelligence, forms, transformation, conceptualization, ontological view

Procedia PDF Downloads 112
935 Automated Detection of Related Software Changes by Probabilistic Neural Networks Model

Authors: Yuan Huang, Xiangping Chen, Xiaonan Luo

Abstract:

Current software are continuously updating. The change between two versions usually involves multiple program entities (e.g., packages, classes, methods, attributes) with multiple purposes (e.g., changed requirements, bug fixing). It is hard for developers to understand which changes are made for the same purpose. Whether two changes are related is not decided by the relationship between this two entities in the program. In this paper, we summarized 4 coupling rules(16 instances) and 4 state-combination types at the class, method and attribute levels for software change. Related Change Vector (RCV) are defined based on coupling rules and state-combination types, and applied to classify related software changes by using Probabilistic Neural Network during a software updating.

Keywords: PNN, related change, state-combination, logical coupling, software entity

Procedia PDF Downloads 415
934 EhfadHaya (SaveLife) / AateHayah (GiveLife) Blood Donor Website

Authors: Sameer Muhammad Aslam, Nura Said Mohsin Al-Saifi

Abstract:

This research shows the process of creating a blood donation website for Oman. Blood donation is a widespread, crucial, ongoing process, so it is important that this website is easy to use. Several automated blood management systems are available, but none provides an effective algorithm that takes into account variables such as frequency of donation, donation date, and gender. In Oman, the Ministry of Health maintains a blood bank and keeps donors informed about the need for blood through a website. They also inform donors and the wider public where and when is their next blood donation event. The website's main goals are to educate the community about the benefits of blood donation. It also manages donor and receiver documentation and encourages voluntary blood donation by providing easy access to information about blood types and blood distribution in various hospitals in Oman, based on hospital needs.

Keywords: Oman, blood bank, blood donors, donor website

Procedia PDF Downloads 194
933 Design and Control Algorithms for Power Electronic Converters for EV Applications

Authors: Ilya Kavalchuk, Mehdi Seyedmahmoudian, Ben Horan, Aman Than Oo, Alex Stojcevski

Abstract:

The power electronic components within Electric Vehicles (EV) need to operate in several important modes. Some modes directly influence safety, while others influence vehicle performance. Given the variety of functions and operational modes required of the power electronics, it needs to meet efficiency requirements to minimize power losses. Another challenge in the control and construction of such systems is the ability to support bidirectional power flow. This paper considers the construction, operation, and feasibility of available converters for electric vehicles with feasible configurations of electrical buses and loads. This paper describes logic and control signals for the converters for different operations conditions based on the efficiency and energy usage bases.

Keywords: electric vehicles, electrical machines control, power electronics, powerflow regulations

Procedia PDF Downloads 539
932 The Operating Behaviour of Unbalanced Unpaced Merging Assembly Lines

Authors: S. Shaaban, T. McNamara, S. Hudson

Abstract:

This paper reports on the performance of deliberately unbalanced, reliable, non-automated and assembly lines that merge, whose workstations differ in terms of their mean operation times. Simulations are carried out on 5- and 8-station lines with 1, 2 and 4 buffer capacity units, % degrees of line imbalance of 2, 5 and 12, and 24 different patterns of means imbalance. Data on two performance measures, namely throughput and average buffer level were gathered, statistically analysed and compared to a merging balanced line counterpart. It was found that the best configurations are a balanced line arrangement and a monotone decreasing order for each of the parallel merging lines, with the first generally resulting in a lower throughput and the second leading to a lower average buffer level than those of a balanced line.

Keywords: average buffer level, merging lines, simulation, throughput, unbalanced

Procedia PDF Downloads 301
931 Computer Aided Assembly Attributes Retrieval Methods for Automated Assembly Sequence Generation

Authors: M. V. A. Raju Bahubalendruni, Bibhuti Bhusan Biswal, B. B. V. L. Deepak

Abstract:

Achieving an appropriate assembly sequence needs deep verification for its physical feasibility. For this purpose, industrial engineers use several assembly predicates; namely, liaison, geometric feasibility, stability and mechanical feasibility. However, testing an assembly sequence for these predicates requires huge assembly information. Extracting such assembly information from an assembled product is a time consuming and highly skillful task with complex reasoning methods. In this paper, computer aided methods are proposed to extract all the necessary assembly information from computer aided design (CAD) environment in order to perform the assembly sequence planning efficiently. These methods use preliminary capabilities of three-dimensional solid modelling and assembly modelling methods used in CAD software considering equilibrium laws of physical bodies.

Keywords: assembly automation, assembly attributes, assembly, CAD

Procedia PDF Downloads 279
930 From User's Requirements to UML Class Diagram

Authors: Zeineb Ben Azzouz, Wahiba Ben Abdessalem Karaa

Abstract:

The automated extraction of UML class diagram from natural language requirements is a highly challenging task. Many approaches, frameworks and tools have been presented in this field. Nonetheless, the experiments of these tools have shown that there is no approach that can work best all the time. In this context, we propose a new accurate approach to facilitate the automatic mapping from textual requirements to UML class diagram. Our new approach integrates the best properties of statistical Natural Language Processing (NLP) techniques to reduce ambiguity when analysing natural language requirements text. In addition, our approach follows the best practices defined by conceptual modelling experts to determine some patterns indispensable for the extraction of basic elements and concepts of the class diagram. Once the relevant information of class diagram is captured, a XMI document is generated and imported with a CASE tool to build the corresponding UML class diagram.

Keywords: class diagram, user’s requirements, XMI, software engineering

Procedia PDF Downloads 450
929 Predicting Groundwater Areas Using Data Mining Techniques: Groundwater in Jordan as Case Study

Authors: Faisal Aburub, Wael Hadi

Abstract:

Data mining is the process of extracting useful or hidden information from a large database. Extracted information can be used to discover relationships among features, where data objects are grouped according to logical relationships; or to predict unseen objects to one of the predefined groups. In this paper, we aim to investigate four well-known data mining algorithms in order to predict groundwater areas in Jordan. These algorithms are Support Vector Machines (SVMs), Naïve Bayes (NB), K-Nearest Neighbor (kNN) and Classification Based on Association Rule (CBA). The experimental results indicate that the SVMs algorithm outperformed other algorithms in terms of classification accuracy, precision and F1 evaluation measures using the datasets of groundwater areas that were collected from Jordanian Ministry of Water and Irrigation.

Keywords: classification, data mining, evaluation measures, groundwater

Procedia PDF Downloads 259
928 Evotrader: Bitcoin Trading Using Evolutionary Algorithms on Technical Analysis and Social Sentiment Data

Authors: Martin Pellon Consunji

Abstract:

Due to the rise in popularity of Bitcoin and other crypto assets as a store of wealth and speculative investment, there is an ever-growing demand for automated trading tools, such as bots, in order to gain an advantage over the market. Traditionally, trading in the stock market was done by professionals with years of training who understood patterns and exploited market opportunities in order to gain a profit. However, nowadays a larger portion of market participants are at minimum aided by market-data processing bots, which can generally generate more stable signals than the average human trader. The rise in trading bot usage can be accredited to the inherent advantages that bots have over humans in terms of processing large amounts of data, lack of emotions of fear or greed, and predicting market prices using past data and artificial intelligence, hence a growing number of approaches have been brought forward to tackle this task. However, the general limitation of these approaches can still be broken down to the fact that limited historical data doesn’t always determine the future, and that a lot of market participants are still human emotion-driven traders. Moreover, developing markets such as those of the cryptocurrency space have even less historical data to interpret than most other well-established markets. Due to this, some human traders have gone back to the tried-and-tested traditional technical analysis tools for exploiting market patterns and simplifying the broader spectrum of data that is involved in making market predictions. This paper proposes a method which uses neuro evolution techniques on both sentimental data and, the more traditionally human-consumed, technical analysis data in order to gain a more accurate forecast of future market behavior and account for the way both automated bots and human traders affect the market prices of Bitcoin and other cryptocurrencies. This study’s approach uses evolutionary algorithms to automatically develop increasingly improved populations of bots which, by using the latest inflows of market analysis and sentimental data, evolve to efficiently predict future market price movements. The effectiveness of the approach is validated by testing the system in a simulated historical trading scenario, a real Bitcoin market live trading scenario, and testing its robustness in other cryptocurrency and stock market scenarios. Experimental results during a 30-day period show that this method outperformed the buy and hold strategy by over 260% in terms of net profits, even when taking into consideration standard trading fees.

Keywords: neuro-evolution, Bitcoin, trading bots, artificial neural networks, technical analysis, evolutionary algorithms

Procedia PDF Downloads 101
927 Accuracy of Computed Tomography Dose Monitor Values: A Multicentric Study in India

Authors: Adhimoolam Saravana Kumar, K. N. Govindarajan, B. Devanand, R. Rajakumar

Abstract:

The quality of Computed Tomography (CT) procedures has improved in recent years due to technological developments and increased diagnostic ability of CT scanners. Due to the fact that CT doses are the peak among diagnostic radiology practices, it is of great significance to be aware of patient’s CT radiation dose whenever a CT examination is preferred. CT radiation dose delivered to patients in the form of volume CT dose index (CTDIvol) values, is displayed on scanner monitors at the end of each examination and it is an important fact to assure that this information is accurate. The objective of this study was to estimate the CTDIvol values for great number of patients during the most frequent CT examinations, to study the comparison between CT dose monitor values and measured ones, as well as to highlight the fluctuation of CTDIvol values for the same CT examination at different centres and scanner models. The output CT dose indices measurements were carried out on single and multislice scanners for available kV, 5 mm slice thickness, 100 mA and FOV combination used. The 100 CT scanners were involved in this study. Data with regard to 15,000 examinations in patients, who underwent routine head, chest and abdomen CT were collected using a questionnaire sent to a large number of hospitals. Out of the 15,000 examinations, 5000 were head CT examinations, 5000 were chest CT examinations and 5000 were abdominal CT examinations. Comprehensive quality assurance (QA) was performed for all the machines involved in this work. Followed by QA, CT phantom dose measurements were carried out in South India using actual scanning parameters used clinically by the hospitals. From this study, we have measured the mean divergence between the measured and displayed CTDIvol values were 5.2, 8.4, and -5.7 for selected head, chest and abdomen procedures for protocols as mentioned above, respectively. Thus, this investigation revealed an observable change in CT practices, with a much wider range of studies being performed currently in South India. This reflects the improved capacity of CT scanners to scan longer scan lengths and at finer resolutions as permitted by helical and multislice technology. Also, some of the CT scanners have used smaller slice thickness for routine CT procedures to achieve better resolution and image quality. It leads to an increase in the patient radiation dose as well as the measured CTDIv, so it is suggested that such CT scanners should select appropriate slice thickness and scanning parameters in order to reduce the patient dose. If these routine scan parameters for head, chest and abdomen procedures are optimized than the dose indices would be optimal and lead to the lowering of the CT doses. In South Indian region all the CT machines were routinely tested for QA once in a year as per AERB requirements.

Keywords: CT dose index, weighted CTDI, volumetric CTDI, radiation dose

Procedia PDF Downloads 234
926 Updating Stochastic Hosting Capacity Algorithm for Voltage Optimization Programs and Interconnect Standards

Authors: Nicholas Burica, Nina Selak

Abstract:

The ADHCAT (Automated Distribution Hosting Capacity Assessment Tool) was designed to run Hosting Capacity Analysis on the ComEd system via a stochastic DER (Distributed Energy Resource) placement on multiple power flow simulations against a set of violation criteria. The violation criteria in the initial version of the tool captured a limited amount of issues that individual departments design against for DER interconnections. Enhancements were made to the tool to further align with individual department violation and operation criteria, as well as the addition of new modules for use for future load profile analysis. A reporting engine was created for future analytical use based on the simulations and observations in the tool.

Keywords: distributed energy resources, hosting capacity, interconnect, voltage optimization

Procedia PDF Downloads 164
925 Preparation and Characterization of Biosorbent from Cactus (Opuntia ficus-indica) cladodes and its Application for Dye Removal from Aqueous Solution

Authors: Manisha Choudhary, Sudarsan Neogi

Abstract:

Malachite green (MG), an organic basic dye, has been widely used for the dyeing purpose, as well as a fungicide and antiseptic in aquaculture industry to control fish parasites and disease. However, MG has now turned out to be an extremely controversial compound due to its adverse impact on living beings. Due to high toxicity, proper treatment of wastewater containing MG is utmost important. Among different available technologies, adsorption process is one of the most efficient and cost-effective treatment method due to its simplicity of design, ease of operation and regeneration of used materials. Nonetheless, commercial activated carbon is expensive leading the researchers to focus on utilizing natural resources. In the present work, a species of cactus, Opuntia ficus-indica (OFI), was used to develop a highly efficient, low-cost powdered activated carbon by chemical activation using NaOH. The biosorbent was characterized by Fourier-transform infrared spectroscopy, field emission scanning electron microscope, energy-dispersive X-ray spectroscopy, Brunauer–Emmett–Teller (BET) and X-ray diffraction analysis. Batch adsorption studies were performed to remove MG from an aqueous solution as a function of contact time, initial solution pH, initial dye concentration, biosorbent dosages, the presence of salt and temperature. By increasing the initial dye concentration from 100 to 500 mg/l, adsorption capacity increased from 165.45 to 831.58 mg/g. The adsorption kinetics followed the pseudo-second-order model and the chemisorption mechanisms were revealed. The electrostatic attractions and chemical interactions were observed between amino and hydroxyl groups of the biosorbent and amine groups of the dye. The adsorption was solely controlled by film diffusion. Different isotherm models were used to fit the adsorption data. The excellent recovery of adsorption efficiency after the regeneration of biosorbent indicated the high potential of this adsorbent to remove MG from aqueous solution and an excellent cost-effective biosorbent for wide application in wastewater treatment.

Keywords: adsorption, biosorbent, cactus, malachite green

Procedia PDF Downloads 352
924 The Implementation of Human Resource Information System in the Public Sector: An Exploratory Study of Perceived Benefits and Challenges

Authors: Aneeqa Suhail, Shabana Naveed

Abstract:

The public sector (in both developed and developing countries) has gone through various waves of radical reforms in recent decades. In Pakistan, under the influence of New Public Management(NPM) Reforms; best practices of private sector are introduced in the public sector to modernize public organizations. Human Resource Information System (HRIS) has been popular in the private sector and proven to be a successful system, therefore it is being adopted in the public sector too. However, implementation of private business practices in public organizations us very challenging due to differences in context. This implementation gets further critical in Pakistan due to a centralizing tendency and lack of autonomy in public organizations. Adoption of HRIS by public organizations in Pakistan raises several questions: What challenges are faced by public organizations in implementation of HRIS? Are benefits of HRIS such as efficiency, process integration and cost reduction achieved? How is the previous system improved with this change and what are the impacts? Yet, it is an under-researched topic, especially in public enterprises. This study contributes to the existing body of knowledge by empirically exploring benefits and challenges of implementation of HRIS in public organizations. The research adopts a case study approach and uses qualitative data based on in-depth interviews conducted at various levels in the hierarchy including top management, departmental heads and employees. The unit of analysis is LESCO, the Lahore Electric Supply Company, a state-owned entity that generates, transmits and distributes electricity to 4 big cities in Punjab, Pakistan. The findings of the study show that LESCO has not achieved the benefits of HRIS as established in literature. The implementation process remained quite slow and costly. Various functions of HR are still in isolation and integration is a big challenge for the organization. Although the data is automated, the previous system of manually record maintenance and paperwork is still in work, resulting in the presence of parallel practices. The findings also identified resistance to change from top management and labor workforce, lack of commitment and technical knowledge, and costly vendors as major barriers that affect the effective implementation of HRIS. The paper suggests some potential actions to overcome these barriers and to enhance effective implementation of HR-technology. The findings are explained in light of an institutional logics perspective. HRIS’ new logic of automated and integrated HR system is in sharp contrast with the prevailing logic of process-oriented manual data maintenance, leading to resistance to change and deadlock.

Keywords: human resource information system, technological changes, state-owned enterprise, implementation challenges

Procedia PDF Downloads 131
923 Trajectory Planning Algorithms for Autonomous Agricultural Vehicles

Authors: Caner Koc, Dilara Gerdan Koc, Mustafa Vatandas

Abstract:

The fundamental components of autonomous agricultural robot design, such as having a working understanding of coordinates, correctly constructing the desired route, and sensing environmental elements, are the most important. A variety of sensors, hardware, and software are employed by agricultural robots to find these systems.These enable the fully automated driving system of an autonomous vehicle to simulate how a human-driven vehicle would respond to changing environmental conditions. To calculate the vehicle's motion trajectory using data from the sensors, this automation system typically consists of a sophisticated software architecture based on object detection and driving decisions. In this study, the software architecture of an autonomous agricultural vehicle is compared to the trajectory planning techniques.

Keywords: agriculture 5.0, computational intelligence, motion planning, trajectory planning

Procedia PDF Downloads 53
922 Conception of a Predictive Maintenance System for Forest Harvesters from Multiple Data Sources

Authors: Lazlo Fauth, Andreas Ligocki

Abstract:

For cost-effective use of harvesters, expensive repairs and unplanned downtimes must be reduced as far as possible. The predictive detection of failing systems and the calculation of intelligent service intervals, necessary to avoid these factors, require in-depth knowledge of the machines' behavior. Such know-how needs permanent monitoring of the machine state from different technical perspectives. In this paper, three approaches will be presented as they are currently pursued in the publicly funded project PreForst at Ostfalia University of Applied Sciences. These include the intelligent linking of workshop and service data, sensors on the harvester, and a special online hydraulic oil condition monitoring system. Furthermore the paper shows potentials as well as challenges for the use of these data in the conception of a predictive maintenance system.

Keywords: predictive maintenance, condition monitoring, forest harvesting, forest engineering, oil data, hydraulic data

Procedia PDF Downloads 118
921 A Review on Various Approaches for Energy Conservation in Green Cloud Computing

Authors: Sumati Manchanda

Abstract:

Cloud computing is one of the most recent developing engineering and is consistently utilized as a part of different IT firms so as to make benefits like expense sparing or financial minimization, it must be eco cordial also. In this manner, Green Cloud Computing is the need of the today's current situation. It is an innovation that is rising as data correspondence engineering. This paper surveys the unequivocal endeavors made by different specialists to make Cloud Computing more vitality preserving, to break down its vitality utilization focused around sorts of administrations gave furthermore to diminish the carbon foot shaped impression rate by colossal methodologies furthermore edify virtualization idea alongside different diverse methodologies which utilize virtual machines scheduling and migration. The summary of the proposed work by various authors that we have reviewed is also presented in the paper.

Keywords: cloud computing, green cloud computing, scheduling, migration, virtualization, energy efficiency

Procedia PDF Downloads 374
920 Meeting the Challanges of Regulating Artificial Intelligence

Authors: Abdulrahman S. Shryan Aldossary

Abstract:

Globally, artificial intelligence (AI) is already performing legitimate tasks on behalf of humans. In Saudi Arabia, large-scale national projects, primarily based on AI technologies and receiving billions of dollars of funding, are projected for completion by 2030. However, the legal aspect of these projects is seriously vulnerable, given AI’s unprecedented ability to self-learn and act independently. This paper, therefore, identifies the critical legal aspects of AI that authorities and policymakers should be aware of, specifically whether AI can possess identity and be liable for the risk of public harm. The article begins by identifying the problematic characteristics of AI and what should be considered by legal experts when dealing with it. Also discussed are the possible competent institutions that could regulate AI in Saudi Arabia. Finally, a procedural proposal is presented for controlling AI, focused on Saudi Arabia but potentially of interest to other jurisdictions facing similar concerns about AI safety.

Keywords: regulation, artificial intelligence, tech law, automated systems

Procedia PDF Downloads 153
919 Automated Recognition of Still’s Murmur in Children

Authors: Sukryool Kang, James McConnaughey, Robin Doroshow, Raj Shekhar

Abstract:

Still’s murmur, a vibratory heart murmur, is the most common normal innocent murmur of childhood. Many children with this murmur are unnecessarily referred for cardiology consultation and testing, which exacts a high cost financially and emotionally on the patients and their parents. Pediatricians to date are not successful at distinguishing Still’s murmur from murmurs of true heart disease. In this paper, we present a new algorithmic approach to distinguish Still’s murmur from pathological murmurs in children. We propose two distinct features, spectral width and signal power, which describe the sharpness of the spectrum and the signal intensity of the murmur, respectively. Seventy pediatric heart sound recordings of 41 Still’s and 29 pathological murmurs were used to develop and evaluate our algorithm that achieved a true positive rate of 97% and false positive rate of 0%. This approach would meet clinical standards in recognizing Still’s murmur.

Keywords: AR modeling, auscultation, heart murmurs, Still's murmur

Procedia PDF Downloads 346
918 Power Quality Audit Using Fluke Analyzer

Authors: N. Ravikumar, S. Krishnan, B. Yokeshkumar

Abstract:

In present days, the power quality issues are increases due to non-linear loads like fridge, AC, washing machines, induction motor, etc. This power quality issues will affects the output voltages, output current, and output power of the total performance of the generator. This paper explains how to test the generator using the Fluke 435 II series power quality analyser. This Fluke 435 II series power quality analyser is used to measure the voltage, current, power, energy, total harmonic distortion (THD), current harmonics, voltage harmonics, power factor, and frequency. The Fluke 435 II series power quality analyser have several advantages. They are i) it will records output in analog and digital format. ii) the fluke analyzer will records at every 0.25 sec. iii) it will also measure all the electrical parameter at a time.

Keywords: THD, harmonics, power quality, TNEB, Fluke 435

Procedia PDF Downloads 161
917 An Automated Sensor System for Cochlear Implants Electrode Array Insertion

Authors: Lei Hou, Xinli Du, Nikolaos Boulgouris

Abstract:

A cochlear implant, referred to as a CI, is a small electronic device that can provide direct electrical stimulation to the auditory nerve. During cochlear implant surgery, atraumatic electrode array insertion is considered to be a crucial step. However, during implantation, the mechanical behaviour of an electrode array inside the cochlea is not known. The behaviour of an electrode array inside of the cochlea is hardly identified by regular methods. In this study, a CI electrode array capacitive sensor system is proposed. It is able to automatically determine the array state as a result of the capacitance variations. Instead of applying sensors to the electrode array, the capacitance information from the electrodes will be gathered and analysed. Results reveal that this sensing method is capable of recognising different states when fed into a pre-shaped model.

Keywords: cochlear implant, electrode, hearing preservation, insertion force, capacitive sensing

Procedia PDF Downloads 214
916 Rapid and Culture-Independent Detection of Staphylococcus Aureus by PCR Based Protocols

Authors: V. Verma, Syed Riyaz-ul-Hassan

Abstract:

Staphylococcus aureus is one of the most commonly found pathogenic bacteria and is hard to eliminate from the human environment. It is responsible for many nosocomial infections, besides being the main causative agent of food intoxication by virtue of its variety of enterotoxins. Routine detection of S. aureus in food is usually carried out by traditional methods based on morphological and biochemical characterization. These methods are time-consuming and tedious. In addition, misclassifications with automated susceptibility testing systems or commercially available latex agglutination kits have been reported by several workers. Consequently, there is a need for methods to specifically discriminate S. aureus from other staphylococci as quickly as possible. Data on protocols developed using molecular means like PCR technology will be presented for rapid and specific detection of this pathogen in food, clinical and environmental samples, especially milk.

Keywords: food Pathogens, PCR technology, rapid and specific detection, staphylococcus aureus

Procedia PDF Downloads 499
915 Automatic Method for Classification of Informative and Noninformative Images in Colonoscopy Video

Authors: Nidhal K. Azawi, John M. Gauch

Abstract:

Colorectal cancer is one of the leading causes of cancer death in the US and the world, which is why millions of colonoscopy examinations are performed annually. Unfortunately, noise, specular highlights, and motion artifacts corrupt many images in a typical colonoscopy exam. The goal of our research is to produce automated techniques to detect and correct or remove these noninformative images from colonoscopy videos, so physicians can focus their attention on informative images. In this research, we first automatically extract features from images. Then we use machine learning and deep neural network to classify colonoscopy images as either informative or noninformative. Our results show that we achieve image classification accuracy between 92-98%. We also show how the removal of noninformative images together with image alignment can aid in the creation of image panoramas and other visualizations of colonoscopy images.

Keywords: colonoscopy classification, feature extraction, image alignment, machine learning

Procedia PDF Downloads 238
914 Seersucker Fabrics Development Using Single Warp Beam

Authors: Khubab Shaker, Yasir Nawab, Muhammad Usman Javed, Muhammad Umair, Muhammad Maqsood

Abstract:

Seersucker is a thin and puckered fabric commonly striped or chequered, used to make clothing for spring and woven in such a way that some threads bunch together, giving the fabric a wrinkled appearance in places. Due to use of two warp beams, such fabrics were not possible to weave on conventional weaving machines. Objective of this study was to weave a seersucker fabric on conventional looms using single warp beam. This objective was achieved using two types of yarns, forming stripes in weft: one being 100% cotton yarn and the other core spun elastane yarn with sheath of cotton (95.7% cotton and 4.3% elastane). Stress-strain behaviour of the produced fabric samples were tested and explained.

Keywords: seersucker fabrics, elastane yarns, single warp beam, weaving

Procedia PDF Downloads 502
913 OILU Tag: A Projective Invariant Fiducial System

Authors: Youssef Chahir, Messaoud Mostefai, Salah Khodja

Abstract:

This paper presents the development of a 2D visual marker, derived from a recent patented work in the field of numbering systems. The proposed fiducial uses a group of projective invariant straight-line patterns, easily detectable and remotely recognizable. Based on an efficient data coding scheme, the developed marker enables producing a large panel of unique real time identifiers with highly distinguishable patterns. The proposed marker Incorporates simultaneously decimal and binary information, making it readable by both humans and machines. This important feature opens up new opportunities for the development of efficient visual human-machine communication and monitoring protocols. Extensive experiment tests validate the robustness of the marker against acquisition and geometric distortions.

Keywords: visual markers, projective invariants, distance map, level sets

Procedia PDF Downloads 136
912 LaPEA: Language for Preprocessing of Edge Applications in Smart Factory

Authors: Masaki Sakai, Tsuyoshi Nakajima, Kazuya Takahashi

Abstract:

In order to improve the productivity of a factory, it is often the case to create an inference model by collecting and analyzing operational data off-line and then to develop an edge application (EAP) that evaluates the quality of the products or diagnoses machine faults in real-time. To accelerate this development cycle, an edge application framework for the smart factory is proposed, which enables to create and modify EAPs based on prepared inference models. In the framework, the preprocessing component is the key part to make it work. This paper proposes a language for preprocessing of edge applications, called LaPEA, which can flexibly process several sensor data from machines into explanatory variables for an inference model, and proves that it meets the requirements for the preprocessing.

Keywords: edge application framework, edgecross, preprocessing language, smart factory

Procedia PDF Downloads 127
911 New Hybrid Process for Converting Small Structural Parts from Metal to CFRP

Authors: Yannick Willemin

Abstract:

Carbon fibre-reinforced plastic (CFRP) offers outstanding value. However, like all materials, CFRP also has its challenges. Many forming processes are largely manual and hard to automate, making it challenging to control repeatability and reproducibility (R&R); they generate significant scrap and are too slow for high-series production; fibre costs are relatively high and subject to supply and cost fluctuations; the supply chain is fragmented; many forms of CFRP are not recyclable, and many materials have yet to be fully characterized for accurate simulation; shelf life and outlife limitations add cost; continuous-fibre forms have design limitations; many materials are brittle; and small and/or thick parts are costly to produce and difficult to automate. A majority of small structural parts are metal due to high CFRP fabrication costs for the small-size class. The fact that CFRP manufacturing processes that produce the highest performance parts also tend to be the slowest and least automated is another reason CFRP parts are generally higher in cost than comparably performing metal parts, which are easier to produce. Fortunately, business is in the midst of a major manufacturing evolution—Industry 4.0— one technology seeing rapid growth is additive manufacturing/3D printing, thanks to new processes and materials, plus an ability to harness Industry 4.0 tools. No longer limited to just prototype parts, metal-additive technologies are used to produce tooling and mold components for high-volume manufacturing, and polymer-additive technologies can incorporate fibres to produce true composites and be used to produce end-use parts with high aesthetics, unmatched complexity, mass customization opportunities, and high mechanical performance. A new hybrid manufacturing process combines the best capabilities of additive—high complexity, low energy usage and waste, 100% traceability, faster to market—and post-consolidation—tight tolerances, high R&R, established materials, and supply chains—technologies. The platform was developed by Zürich-based 9T Labs AG and is called Additive Fusion Technology (AFT). It consists of a design software offering the possibility to determine optimal fibre layup, then exports files back to check predicted performance—plus two pieces of equipment: a 3d-printer—which lays up (near)-net-shape preforms using neat thermoplastic filaments and slit, roll-formed unidirectional carbon fibre-reinforced thermoplastic tapes—and a post-consolidation module—which consolidates then shapes preforms into final parts using a compact compression press fitted with a heating unit and matched metal molds. Matrices—currently including PEKK, PEEK, PA12, and PPS, although nearly any high-quality commercial thermoplastic tapes and filaments can be used—are matched between filaments and tapes to assure excellent bonding. Since thermoplastics are used exclusively, larger assemblies can be produced by bonding or welding together smaller components, and end-of-life parts can be recycled. By combining compression molding with 3D printing, higher part quality with very-low voids and excellent surface finish on A and B sides can be produced. Tight tolerances (min. section thickness=1.5mm, min. section height=0.6mm, min. fibre radius=1.5mm) with high R&R can be cost-competitively held in production volumes of 100 to 10,000 parts/year on a single set of machines.

Keywords: additive manufacturing, composites, thermoplastic, hybrid manufacturing

Procedia PDF Downloads 77
910 A Laundry Algorithm for Colored Textiles

Authors: H. E. Budak, B. Arslan-Ilkiz, N. Cakmakci, I. Gocek, U. K. Sahin, H. Acikgoz-Tufan, M. H. Arslan

Abstract:

The aim of this study is to design a novel laundry algorithm for colored textiles which have significant decoloring problem. During the experimental work, bleached knitted single jersey fabric made of 100% cotton and dyed with reactive dyestuff was utilized, since according to a conducted survey textiles made of cotton are the most demanded textile products in the textile market by the textile consumers and for coloration of textiles reactive dyestuffs are the ones that are the most commonly used in the textile industry for dyeing cotton-made products. Therefore, the fabric used in this study was selected and purchased in accordance with the survey results. The fabric samples cut out of this fabric were dyed with different dyeing parameters by using Remazol Brilliant Red 3BS dyestuff in Gyrowash machine at laboratory conditions. From the alternative reactive-dyed cotton fabric samples, the ones that have high tendency to color loss were determined and examined. Accordingly, the parameters of the dyeing process used for these fabric samples were evaluated and the dyeing process which was chosen to be used for causing high tendency to color loss for the cotton fabrics was determined in order to reveal the level of improvement in color loss during this study clearly. Afterwards, all of the untreated fabric samples cut out of the fabric purchased were dyed with the dyeing process selected. When dyeing process was completed, an experimental design was created for the laundering process by using Minitab® program considering temperature, time and mechanical action as parameters. All of the washing experiments were performed in domestic washing machine. 16 washing experiments were performed with 8 different experimental conditions and 2 repeats for each condition. After each of the washing experiments, water samples of the main wash of the laundering process were measured with UV spectrophotometer. The values obtained were compared with the calibration curve of the materials used for the dyeing process. The results of the washing experiments were statistically analyzed with Minitab® program. According to the results, the most suitable washing algorithm to be used in terms of the parameters temperature, time and mechanical action for domestic washing machines for minimizing fabric color loss was chosen. The laundry algorithm proposed in this study have the ability of minimalizing the problem of color loss of colored textiles in washing machines by eliminating the negative effects of the parameters of laundering process on color of textiles without compromising the fundamental effects of basic cleaning action being performed properly. Therefore, since fabric color loss is minimized with this washing algorithm, dyestuff residuals will definitely be lower in the grey water released from the laundering process. In addition to this, with this laundry algorithm it is possible to wash and clean other types of textile products with proper cleaning effect and minimized color loss.

Keywords: color loss, laundry algorithm, textiles, domestic washing process

Procedia PDF Downloads 329
909 An Automated R-Peak Detection Method Using Common Vector Approach

Authors: Ali Kirkbas

Abstract:

R peaks in an electrocardiogram (ECG) are signs of cardiac activity in individuals that reveal valuable information about cardiac abnormalities, which can lead to mortalities in some cases. This paper examines the problem of detecting R-peaks in ECG signals, which is a two-class pattern classification problem in fact. To handle this problem with a reliable high accuracy, we propose to use the common vector approach which is a successful machine learning algorithm. The dataset used in the proposed method is obtained from MIT-BIH, which is publicly available. The results are compared with the other popular methods under the performance metrics. The obtained results show that the proposed method shows good performance than that of the other. methods compared in the meaning of diagnosis accuracy and simplicity which can be operated on wearable devices.

Keywords: ECG, R-peak classification, common vector approach, machine learning

Procedia PDF Downloads 39
908 Seamless MATLAB® to Register-Transfer Level Design Methodology Using High-Level Synthesis

Authors: Petri Solanti, Russell Klein

Abstract:

Many designers are asking for an automated path from an abstract mathematical MATLAB model to a high-quality Register-Transfer Level (RTL) hardware description. Manual transformations of MATLAB or intermediate code are needed, when the design abstraction is changed. Design conversion is problematic as it is multidimensional and it requires many different design steps to translate the mathematical representation of the desired functionality to an efficient hardware description with the same behavior and configurability. Yet, a manual model conversion is not an insurmountable task. Using currently available design tools and an appropriate design methodology, converting a MATLAB model to efficient hardware is a reasonable effort. This paper describes a simple and flexible design methodology that was developed together with several design teams.

Keywords: design methodology, high-level synthesis, MATLAB, verification

Procedia PDF Downloads 119
907 Analysis on Cyber Threat Actors Targeting Automated Border Security Systems

Authors: Mirko Sailio

Abstract:

Border crossing automatization reduces required human resources in handling people crossing borders. As technology replaces and augments the work done by border officers, new cyber threats arise to threaten border security. This research analyses the current cyber threat actors and their capabilities. The analysis is conducted by gathering the threat actor data from a wide range of public sources. A model for a general border automatization system is presented, and its most significant cyber-security attributes are then compared to threat actor activity and capabilities in order to predict priorities in securing such systems. Organized crime and nation-state actors present the clearest threat to border cyber-security, and additional focus is given to their motivations and activities.

Keywords: border automation, cyber-security, threat actors, border cyber-security

Procedia PDF Downloads 181