Search results for: array signal processing
3477 Evaluation of Low Power Wi-Fi Modules in Simulated Ocean Environments
Authors: Gabriel Chenevert, Abhilash Arora, Zeljko Pantic
Abstract:
The major problem underwater acoustic communication faces is the low data rate due to low signal frequency. By contrast, the Wi-Fi communication protocol offers high throughput but limited operating range due to the attenuation effect of the sea and ocean medium. However, short-range near-field underwater wireless power transfer systems offer an environment where Wi-Fi communication can be effectively integrated to collect data and deliver instructions to sensors in underwater sensor networks. In this paper, low-power, low-cost off-the-shelf Wi-Fi modules are explored experimentally for four selected parameters for different distances between units and water salinities. The results reveal a shorter operating range and stronger dependence on water salinity than reported so far for high-end Wi-Fi modules.Keywords: Wi-Fi, wireless power transfer, underwater communications, ESP
Procedia PDF Downloads 1163476 Assessment of Image Databases Used for Human Skin Detection Methods
Authors: Saleh Alshehri
Abstract:
Human skin detection is a vital step in many applications. Some of the applications are critical especially those related to security. This leverages the importance of a high-performance detection algorithm. To validate the accuracy of the algorithm, image databases are usually used. However, the suitability of these image databases is still questionable. It is suggested that the suitability can be measured mainly by the span the database covers of the color space. This research investigates the validity of three famous image databases.Keywords: image databases, image processing, pattern recognition, neural networks
Procedia PDF Downloads 2713475 A New Approach to the Digital Implementation of Analog Controllers for a Power System Control
Authors: G. Shabib, Esam H. Abd-Elhameed, G. Magdy
Abstract:
In this paper, a comparison of discrete time PID, PSS controllers is presented through small signal stability of power system comprising of one machine connected to infinite bus system. This comparison achieved by using a new approach of discretization which converts the S-domain model of analog controllers to a Z-domain model to enhance the damping of a single machine power system. The new method utilizes the Plant Input Mapping (PIM) algorithm. The proposed algorithm is stable for any sampling rate, as well as it takes the closed loop characteristic into consideration. On the other hand, the traditional discretization methods such as Tustin’s method is produce satisfactory results only; when the sampling period is sufficiently low.Keywords: PSS, power system stabilizer PID, proportional-integral-derivative PIM, plant input mapping
Procedia PDF Downloads 5053474 Data Analytics in Energy Management
Authors: Sanjivrao Katakam, Thanumoorthi I., Antony Gerald, Ratan Kulkarni, Shaju Nair
Abstract:
With increasing energy costs and its impact on the business, sustainability today has evolved from a social expectation to an economic imperative. Therefore, finding methods to reduce cost has become a critical directive for Industry leaders. Effective energy management is the only way to cut costs. However, Energy Management has been a challenge because it requires a change in old habits and legacy systems followed for decades. Today exorbitant levels of energy and operational data is being captured and stored by Industries, but they are unable to convert these structured and unstructured data sets into meaningful business intelligence. It must be noted that for quick decisions, organizations must learn to cope with large volumes of operational data in different formats. Energy analytics not only helps in extracting inferences from these data sets, but also is instrumental in transformation from old approaches of energy management to new. This in turn assists in effective decision making for implementation. It is the requirement of organizations to have an established corporate strategy for reducing operational costs through visibility and optimization of energy usage. Energy analytics play a key role in optimization of operations. The paper describes how today energy data analytics is extensively used in different scenarios like reducing operational costs, predicting energy demands, optimizing network efficiency, asset maintenance, improving customer insights and device data insights. The paper also highlights how analytics helps transform insights obtained from energy data into sustainable solutions. The paper utilizes data from an array of segments such as retail, transportation, and water sectors.Keywords: energy analytics, energy management, operational data, business intelligence, optimization
Procedia PDF Downloads 3643473 Reduction of Fermentation Duration of Cassava to Remove Hydrogen Cyanide
Authors: Jean Paul Hategekimana, Josiane Irakoze, Eugene Niyonzima, Annick Ndekezi
Abstract:
Cassava (Manihot esculenta Crantz) is a root crop comprising an anti-nutritive factor known as cyanide. The compound can be removed by numerous processing methods such as boiling, fermentation, blanching, and sun drying to avoid the possibility of cyanide poisoning. Inappropriate processing mean can lead to disease and death. Cassava-based dishes are consumed in different ways, where cassava is cultivated according to their culture and preference. However, they have been shown to be unsafe based on high cyanide levels. The current study targeted to resolve the problem of high cyanide in cassava consumed in Rwanda. This study was conducted to determine the effect of slicing, blanching, and soaking time to reduce the fermentation duration of cassava for hydrogen cyanide (HCN) in mg/g removal. Cassava was sliced into three different portions (1cm, 2cm, and 5cm). The first portions were naturally fermented for seven days, where each portion was removed every 24 hours from soaking tanks and then oven dried at a temperature of 60°C and then milled to obtain naturally fermented cassava flours. Other portions of 1cm, 2cm, and 5cm were blanched for 2, 5, 10 min, respectively, and each similarly dried at 60°C and milled to produce blanched cassava flour. Other blanched portions were used to follow the previous fermentation steps. The last portions, which formed the control, were simply chopped. Cyanide content and starch content in mg/100g were investigated. According to the conducted analysis on different cassava treatments for detoxification, found that usual fermentation can be used, but for sliced portions aimed to size reduction for the easy hydrogen cyanide diffuse out and it takes four days to complete fermentation, which has reduced at 94.44% with significantly different (p<0.05)of total hydrogen cyanide contained in cassava to safe level of consumption, and what is recommended as more effective is to apply blanching combined with fermentation due to the fact that, it takes three days to complete hydrogen cyanide removal at 95.56% on significantly different (p<0.05) of reduction to the safe level of consumption.Keywords: cassava, cyanide, blanching, drying, fermentation
Procedia PDF Downloads 683472 Deep Learning Based Text to Image Synthesis for Accurate Facial Composites in Criminal Investigations
Authors: Zhao Gao, Eran Edirisinghe
Abstract:
The production of an accurate sketch of a suspect based on a verbal description obtained from a witness is an essential task for most criminal investigations. The criminal investigation system employs specifically trained professional artists to manually draw a facial image of the suspect according to the descriptions of an eyewitness for subsequent identification. Within the advancement of Deep Learning, Recurrent Neural Networks (RNN) have shown great promise in Natural Language Processing (NLP) tasks. Additionally, Generative Adversarial Networks (GAN) have also proven to be very effective in image generation. In this study, a trained GAN conditioned on textual features such as keywords automatically encoded from a verbal description of a human face using an RNN is used to generate photo-realistic facial images for criminal investigations. The intention of the proposed system is to map corresponding features into text generated from verbal descriptions. With this, it becomes possible to generate many reasonably accurate alternatives to which the witness can use to hopefully identify a suspect from. This reduces subjectivity in decision making both by the eyewitness and the artist while giving an opportunity for the witness to evaluate and reconsider decisions. Furthermore, the proposed approach benefits law enforcement agencies by reducing the time taken to physically draw each potential sketch, thus increasing response times and mitigating potentially malicious human intervention. With publically available 'CelebFaces Attributes Dataset' (CelebA) and additionally providing verbal description as training data, the proposed architecture is able to effectively produce facial structures from given text. Word Embeddings are learnt by applying the RNN architecture in order to perform semantic parsing, the output of which is fed into the GAN for synthesizing photo-realistic images. Rather than the grid search method, a metaheuristic search based on genetic algorithms is applied to evolve the network with the intent of achieving optimal hyperparameters in a fraction the time of a typical brute force approach. With the exception of the ‘CelebA’ training database, further novel test cases are supplied to the network for evaluation. Witness reports detailing criminals from Interpol or other law enforcement agencies are sampled on the network. Using the descriptions provided, samples are generated and compared with the ground truth images of a criminal in order to calculate the similarities. Two factors are used for performance evaluation: The Structural Similarity Index (SSIM) and the Peak Signal-to-Noise Ratio (PSNR). A high percentile output from this performance matrix should attribute to demonstrating the accuracy, in hope of proving that the proposed approach can be an effective tool for law enforcement agencies. The proposed approach to criminal facial image generation has potential to increase the ratio of criminal cases that can be ultimately resolved using eyewitness information gathering.Keywords: RNN, GAN, NLP, facial composition, criminal investigation
Procedia PDF Downloads 1613471 Determining Coordinates of Ultra-Light Drones Based on the Time Difference of Arrival (TDOA) Method
Authors: Nguyen Huy Hoang, Do Thanh Quan, Tran Vu Kien
Abstract:
The use of the active radar to measure the coordinates of ultra-light drones is frequently difficult due to long-distance, absolutely small radar cross-section (RCS) and obstacles. Since ultra-light drones are usually controlled by the Time Difference of Arrival (RF), the paper proposed a method to measure the coordinates of ultra-light drones in the space based on the arrival time of the signal at receiving antennas and the time difference of arrival (TDOA). The experimental results demonstrate that the proposed method is really potential and highly accurate.Keywords: ultra-light drone, TDOA, radar cross-section (RCS), RF
Procedia PDF Downloads 2083470 The Role of Artificial Intelligence Algorithms in Psychiatry: Advancing Diagnosis and Treatment
Authors: Netanel Stern
Abstract:
Artificial intelligence (AI) algorithms have emerged as powerful tools in the field of psychiatry, offering new possibilities for enhancing diagnosis and treatment outcomes. This article explores the utilization of AI algorithms in psychiatry, highlighting their potential to revolutionize patient care. Various AI algorithms, including machine learning, natural language processing (NLP), reinforcement learning, clustering, and Bayesian networks, are discussed in detail. Moreover, ethical considerations and future directions for research and implementation are addressed.Keywords: AI, software engineering, psychiatry, neuroimaging
Procedia PDF Downloads 1163469 Development of a Real Time Axial Force Measurement System and IoT-Based Monitoring for Smart Bearing
Authors: Hassam Ahmed, Yuanzhi Liu, Yassine Selami, Wei Tao, Hui Zhao
Abstract:
The purpose of this research is to develop a real time axial force measurement system for a smart bearing through the use of strain-gauges, whereby the data acquisition is performed by an Arduino microcontroller due to its easy manipulation and low-cost. The measured signal is acquired and then discretized using a Wheatstone Bridge and an Analog-Digital Converter (ADC) respectively. For bearing monitoring, a real time monitoring system based on Internet of things (IoT) and Bluetooth were developed. Experimental tests were performed on a bearing within a force range up to 600 kN. The experimental results show that there is a proportional linear relationship between the applied force and the output voltage, and the error R squared is within 0.9878 based on the regression analysis.Keywords: bearing, force measurement, IoT, strain gauge
Procedia PDF Downloads 1423468 Integrated Geotechnical and Geophysical Investigation of a Proposed Construction Site at Mowe, Southwestern Nigeria
Authors: Kayode Festus Oyedele, Sunday Oladele, Adaora Chibundu Nduka
Abstract:
The subsurface of a proposed site for building development in Mowe, Nigeria, using Standard Penetration Test (SPT) and Cone Penetrometer Test (CPT) supplemented with Horizontal Electrical Profiling (HEP) was investigated with the aim of evaluating the suitability of the strata for foundation materials. Four SPT and CPT were implemented using 10 tonnes hammer. HEP utilizing Wenner array were performed with inter-electrode spacing of 10 – 60 m along four traverses coincident with each of the SPT and CPT. The HEP data were processed using DIPRO software and textural filtering of the resulting resistivity sections was implemented to enable delineation of hidden layers. Sandy lateritic clay, silty lateritic clay, clay, clayey sand and sand horizons were delineated. The SPT “N” value defined very soft to soft sandy lateritic (<4), stiff silty lateritic clay (7 – 12), very stiff silty clay (12 - 15), clayey sand (15- 20) and sand (27 – 37). Sandy lateritic clay (5-40 kg/cm2) and silty lateritic clay (25 - 65 kg/cm2) were defined from the CPT response. Sandy lateritic clay (220-750 Ωm), clay (< 50 Ωm) and sand (415-5359 Ωm) were delineated from the resistivity sections with two thin layers of silty lateritic clay and clayey sand defined in the texturally filtered resistivity sections. This study concluded that the presence of incompetent thick clayey materials (18 m) beneath the study area makes it unsuitable for shallow foundation. Deep foundation involving piling through the clayey layers to the competent sand at 20 m depth was recommended.Keywords: cone penetrometer, foundation, lithologic texture, resistivity section, standard penetration test
Procedia PDF Downloads 2653467 Femtocell Stationed Flawless Handover in High Agility Trains
Authors: S. Dhivya, M. Abirami, M. Farjana Parveen, M. Keerthiga
Abstract:
The development of high-speed railway makes people’s lives more and more convenient; meanwhile, handover is the major problem on high-speed railway communication services. In order to overcome that drawback the architecture of Long-Term Evolution (LTE) femtocell networks is used to improve network performance, and the deployment of a femtocell is a key for bandwidth limitation and coverage issues in conventional mobile network system. To increase the handover performance this paper proposed a multiple input multiple output (MIMO) assisted handoff (MAHO) algorithm. It is a technique used in mobile telecom to transfer a mobile phone to a new radio channel with stronger signal strength and improved channel quality.Keywords: flawless handover, high-speed train, home evolved Node B, LTE, mobile femtocell, RSS
Procedia PDF Downloads 4733466 Geological Mapping of Gabel Humr Akarim Area, Southern Eastern Desert, Egypt: Constrain from Remote Sensing Data, Petrographic Description and Field Investigation
Authors: Doaa Hamdi, Ahmed Hashem
Abstract:
The present study aims at integrating the ASTER data and Landsat 8 data to discriminate and map alteration and/or mineralization zones in addition to delineating different lithological units of Humr Akarim Granites area. The study area is located at 24º9' to 24º13' N and 34º1' to 34º2'45"E., covering a total exposed surface area of about 17 km². The area is characterized by rugged topography with low to moderate relief. Geologic fieldwork and petrographic investigations revealed that the basement complex of the study area is composed of metasediments, mafic dikes, older granitoids, and alkali-feldspar granites. Petrographic investigations revealed that the secondary minerals in the study area are mainly represented by chlorite, epidote, clay minerals and iron oxides. These minerals have specific spectral signatures in the region of visible near-infrared and short-wave infrared (0.4 to 2.5 µm). So that the ASTER imagery processing was concentrated on VNIR-SWIR spectrometric data in order to achieve the purposes of this study (geologic mapping of hydrothermal alteration zones and delineate possible radioactive potentialities). Mapping of hydrothermal alterations zones in addition to discriminating the lithological units in the study area are achieved through the utilization of some different image processing, including color band composites (CBC) and data transformation techniques such as band ratios (BR), band ratio codes (BRCs), principal component analysis(PCA), Crosta Technique and minimum noise fraction (MNF). The field verification and petrographic investigation confirm the results of ASTER imagery and Landsat 8 data, proposing a geological map (scale 1:50000).Keywords: remote sensing, petrography, mineralization, alteration detection
Procedia PDF Downloads 1643465 High Performance Computing and Big Data Analytics
Authors: Branci Sarra, Branci Saadia
Abstract:
Because of the multiplied data growth, many computer science tools have been developed to process and analyze these Big Data. High-performance computing architectures have been designed to meet the treatment needs of Big Data (view transaction processing standpoint, strategic, and tactical analytics). The purpose of this article is to provide a historical and global perspective on the recent trend of high-performance computing architectures especially what has a relation with Analytics and Data Mining.Keywords: high performance computing, HPC, big data, data analysis
Procedia PDF Downloads 5203464 Image Based Landing Solutions for Large Passenger Aircraft
Authors: Thierry Sammour Sawaya, Heikki Deschacht
Abstract:
In commercial aircraft operations, almost half of the accidents happen during approach or landing phases. Automatic guidance and automatic landings have proven to bring significant safety value added for this challenging landing phase. This is why Airbus and ScioTeq have decided to work together to explore the capability of image-based landing solutions as additional landing aids to further expand the possibility to perform automatic approach and landing to runways where the current guiding systems are either not fitted or not optimum. Current systems for automated landing often depend on radio signals provided by airport ground infrastructure on the airport or satellite coverage. In addition, these radio signals may not always be available with the integrity and performance required for safe automatic landing. Being independent from these radio signals would widen the operations possibilities and increase the number of automated landings. Airbus and ScioTeq are joining their expertise in the field of Computer Vision in the European Program called Clean Sky 2 Large Passenger Aircraft, in which they are leading the IMBALS (IMage BAsed Landing Solutions) project. The ultimate goal of this project is to demonstrate, develop, validate and verify a certifiable automatic landing system guiding an airplane during the approach and landing phases based on an onboard camera system capturing images, enabling automatic landing independent from radio signals and without precision instrument for landing. In the frame of this project, ScioTeq is responsible for the development of the Image Processing Platform (IPP), while Airbus is responsible for defining the functional and system requirements as well as the testing and integration of the developed equipment in a Large Passenger Aircraft representative environment. The aim of this paper will be to describe the system as well as the associated methods and tools developed for validation and verification.Keywords: aircraft landing system, aircraft safety, autoland, avionic system, computer vision, image processing
Procedia PDF Downloads 1013463 Optimal MRO Process Scheduling with Rotable Inventory to Minimize Total Earliness
Authors: Murat Erkoc, Kadir Ertogral
Abstract:
Maintenance, repair and overhauling (MRO) of high cost equipment used in many industries such as transportation, military and construction are typically subject to regulations set by local governments or international agencies. Aircrafts are prime examples for this kind of equipment. Such equipment must be overhauled at certain intervals for continuing permission of use. As such, the overhaul must be completed by strict deadlines, which often times cannot be exceeded. Due to the fact that the overhaul is typically a long process, MRO companies carry so called rotable inventory for exchange of expensive modules in the overhaul process of the equipment so that the equipment continue its services with minimal interruption. The extracted module is overhauled and returned back to the inventory for future exchange, hence the name rotable inventory. However, since the rotable inventory and overhaul capacity are limited, it may be necessary to carry out some of the exchanges earlier than their deadlines in order to produce a feasible overhaul schedule. An early exchange results with a decrease in the equipment’s cycle time in between overhauls and as such, is not desired by the equipment operators. This study introduces an integer programming model for the optimal overhaul and exchange scheduling. We assume that there is certain number of rotables at hand at the beginning of the planning horizon for a single type module and there are multiple demands with known deadlines for the exchange of the modules. We consider an MRO system with identical parallel processing lines. The model minimizes total earliness by generating optimal overhaul start times for rotables on parallel processing lines and exchange timetables for orders. We develop a fast exact solution algorithm for the model. The algorithm employs full-delay scheduling approach with backward allocation and can easily be used for overhaul scheduling problems in various MRO settings with modular rotable items. The proposed procedure is demonstrated by a case study from the aerospace industry.Keywords: rotable inventory, full-delay scheduling, maintenance, overhaul, total earliness
Procedia PDF Downloads 5443462 Development of a Computer Aided Diagnosis Tool for Brain Tumor Extraction and Classification
Authors: Fathi Kallel, Abdulelah Alabd Uljabbar, Abdulrahman Aldukhail, Abdulaziz Alomran
Abstract:
The brain is an important organ in our body since it is responsible about the majority actions such as vision, memory, etc. However, different diseases such as Alzheimer and tumors could affect the brain and conduct to a partial or full disorder. Regular diagnosis are necessary as a preventive measure and could help doctors to early detect a possible trouble and therefore taking the appropriate treatment, especially in the case of brain tumors. Different imaging modalities are proposed for diagnosis of brain tumor. The powerful and most used modality is the Magnetic Resonance Imaging (MRI). MRI images are analyzed by doctor in order to locate eventual tumor in the brain and describe the appropriate and needed treatment. Diverse image processing methods are also proposed for helping doctors in identifying and analyzing the tumor. In fact, a large Computer Aided Diagnostic (CAD) tools including developed image processing algorithms are proposed and exploited by doctors as a second opinion to analyze and identify the brain tumors. In this paper, we proposed a new advanced CAD for brain tumor identification, classification and feature extraction. Our proposed CAD includes three main parts. Firstly, we load the brain MRI. Secondly, a robust technique for brain tumor extraction is proposed. This technique is based on both Discrete Wavelet Transform (DWT) and Principal Component Analysis (PCA). DWT is characterized by its multiresolution analytic property, that’s why it was applied on MRI images with different decomposition levels for feature extraction. Nevertheless, this technique suffers from a main drawback since it necessitates a huge storage and is computationally expensive. To decrease the dimensions of the feature vector and the computing time, PCA technique is considered. In the last stage, according to different extracted features, the brain tumor is classified into either benign or malignant tumor using Support Vector Machine (SVM) algorithm. A CAD tool for brain tumor detection and classification, including all above-mentioned stages, is designed and developed using MATLAB guide user interface.Keywords: MRI, brain tumor, CAD, feature extraction, DWT, PCA, classification, SVM
Procedia PDF Downloads 2493461 Strategies of Risk Management for Smallholder Farmers in South Africa: A Case Study on Pigeonpea (Cajanus cajan) Production
Authors: Sanari Chalin Moriri, Kwabena Kingsley Ayisi, Alina Mofokeng
Abstract:
Dryland smallholder farmers in South Africa are vulnerable to all kinds of risks, and it negatively affects crop productivity and profit. Pigeonpea is a leguminous and multipurpose crop that provides food, fodder, and wood for smallholder farmers. The majority of these farmers are still growing pigeonpea from traditional unimproved seeds, which comprise a mixture of genotypes. The objectives of the study were to identify the key risk factors that affect pigeonpea productivity and to develop management strategies on how to alleviate the risk factors in pigeonpea production. The study was conducted in two provinces (Limpopo and Mpumalanga) of South Africa in six municipalities during the 2020/2021 growing seasons. The non-probability sampling method using purposive and snowball sampling techniques were used to collect data from the farmers through a structured questionnaire. A total of 114 pigeonpea producers were interviewed individually using a questionnaire. Key stakeholders in each municipality were also identified, invited, and interviewed to verify the information given by farmers. Data collected were subjected to SPSS statistical software 25 version. The findings of the study were that majority of farmers affected by risk factors were women, subsistence, and old farmers resulted in low food production. Drought, unavailability of improved pigeonpea seeds for planting, access to information, and processing equipment were found to be the main risk factors contributing to low crop productivity in farmer’s fields. Above 80% of farmers lack knowledge on the improvement of the crop and also on the processing techniques to secure high prices during the crop off-season. Market availability, pricing, and incidence of pests and diseases were found to be minor risk factors which were triggered by the major risk factors. The minor risk factors can be corrected only if the major risk factors are first given the necessary attention. About 10% of the farmers found to use the crop as a mulch to reduce soil temperatures and to improve soil fertility. The study revealed that most of the farmers were unaware of its utilisation as fodder, much, medicinal, nitrogen fixation, and many more. The risk of frequent drought in dry areas of South Africa where farmers solely depend on rainfall poses a serious threat to crop productivity. The majority of these risk factors are caused by climate change due to unrealistic, low rainfall with extreme temperatures poses a threat to food security, water, and the environment. The use of drought-tolerant, multipurpose legume crops such as pigeonpea, access to new information, provision of processing equipment, and support from all stakeholders will help in addressing food security for smallholder farmers. Policies should be revisited to address the prevailing risk factors faced by farmers and involve them in addressing the risk factors. Awareness should be prioritized in promoting the crop to improve its production and commercialization in the dryland farming system of South Africa.Keywords: management strategies, pigeonpea, risk factors, smallholder farmers
Procedia PDF Downloads 2133460 Knowledge Transfer among Cross-Functional Teams as a Continual Improvement Process
Authors: Sergio Mauricio Pérez López, Luis Rodrigo Valencia Pérez, Juan Manuel Peña Aguilar, Adelina Morita Alexander
Abstract:
The culture of continuous improvement in organizations is very important as it represents a source of competitive advantage. This article discusses the transfer of knowledge between companies which formed cross-functional teams and used a dynamic model for knowledge creation as a framework. In addition, the article discusses the structure of cognitive assets in companies and the concept of "stickiness" (which is defined as an obstacle to the transfer of knowledge). The purpose of this analysis is to show that an improvement in the attitude of individual members of an organization creates opportunities, and that an exchange of information and knowledge leads to generating continuous improvements in the company as a whole. This article also discusses the importance of creating the proper conditions for sharing tacit knowledge. By narrowing gaps between people, mutual trust can be created and thus contribute to an increase in sharing. The concept of adapting knowledge to new environments will be highlighted, as it is essential for companies to translate and modify information so that such information can fit the context of receiving organizations. Adaptation will ensure that the transfer process is carried out smoothly by preventing "stickiness". When developing the transfer process on cross-functional teams (as opposed to working groups), the team acquires the flexibility and responsiveness necessary to meet objectives. These types of cross-functional teams also generate synergy due to the array of different work backgrounds of their individuals. When synergy is established, a culture of continuous improvement is created.Keywords: knowledge transfer, continuous improvement, teamwork, cognitive assets
Procedia PDF Downloads 3243459 Third Party Logistics (3PL) Selection Criteria for an Indian Heavy Industry Using SEM
Authors: Nadama Kumar, P. Parthiban, T. Niranjan
Abstract:
In the present paper, we propose an incorporated approach for 3PL supplier choice that suits the distinctive strategic needs of the outsourcing organization in southern part of India. Four fundamental criteria have been used in particular Performance, IT, Service and Intangible. These are additionally subdivided into fifteen sub-criteria. The proposed strategy coordinates Structural Equation Modeling (SEM) and Non-additive Fuzzy Integral strategies. The presentation of fluffiness manages the unclearness of human judgments. The SEM approach has been used to approve the determination criteria for the proposed show though the Non-additive Fuzzy Integral approach uses the SEM display contribution to assess a supplier choice score. The case organization has a exclusive vertically integrated assembly that comprises of several companies focusing on a slight array of the value chain. To confirm manufacturing and logistics proficiency, it significantly relies on 3PL suppliers to attain supply chain superiority. However, 3PL supplier selection is an intricate decision-making procedure relating multiple selection criteria. The goal of this work is to recognize the crucial 3PL selection criteria by using the non-additive fuzzy integral approach. Unlike the outmoded multi criterion decision-making (MCDM) methods which frequently undertake independence among criteria and additive importance weights, the nonadditive fuzzy integral is an effective method to resolve the dependency among criteria, vague information, and vital fuzziness of human judgment. In this work, we validate an empirical case that engages the nonadditive fuzzy integral to assess the importance weight of selection criteria and indicate the most suitable 3PL supplier.Keywords: 3PL, non-additive fuzzy integral approach, SEM, fuzzy
Procedia PDF Downloads 2803458 Dynamic Corrosion Prevention through Magneto-Responsive Nanostructure with Controllable Hydrophobicity
Authors: Anne McCarthy, Anna Kim, Yin Song, Kyoo Jo, Donald Cropek, Sungmin Hong
Abstract:
Corrosion prevention remains an indispensable concern across a spectrum of industries, demanding inventive and adaptable methodologies to effectively tackle the ever-evolving obstacles presented by corrosive surroundings. This abstract introduces a pioneering approach to corrosion prevention that amalgamates the distinct attributes of magneto-responsive polymers with finely adjustable hydrophobicity inspired by the structure of cicada wings, effectively deterring bacterial proliferation and biofilm formation. The proposed strategy entails the creation of an innovative array of magneto-responsive nanostructures endowed with the capacity to dynamically modulate their hydrophobic characteristics. This dynamic control over hydrophobicity facilitates active repulsion of water and corrosive agents on demand. Additionally, the cyclic motion generated by magnetic activation prevents the biofilms formation and rejection. Thus, the synergistic interplay between magneto-active nanostructures and hydrophobicity manipulation establishes a versatile defensive mechanism against diverse corrosive agents. This study introduces a novel method for corrosion prevention, harnessing the advantages of magneto-active nanostructures and the precision of hydrophobicity adjustment, resulting in water-repellency, effective biofilm removal, and offering a promising solution to handle corrosion-related challenges. We believe that the combined effect will significantly contribute to extending asset lifespan, improving safety, and reducing maintenance costs in the face of corrosion threats.Keywords: magneto-active material, nanoimprinting, corrosion prevention, hydrophobicity
Procedia PDF Downloads 653457 Comparison of Back-Projection with Non-Uniform Fast Fourier Transform for Real-Time Photoacoustic Tomography
Authors: Moung Young Lee, Chul Gyu Song
Abstract:
Photoacoustic imaging is the imaging technology that combines the optical imaging and ultrasound. This provides the high contrast and resolution due to optical imaging and ultrasound imaging, respectively. We developed the real-time photoacoustic tomography (PAT) system using linear-ultrasound transducer and digital acquisition (DAQ) board. There are two types of algorithm for reconstructing the photoacoustic signal. One is back-projection algorithm, the other is FFT algorithm. Especially, we used the non-uniform FFT algorithm. To evaluate the performance of our system and algorithms, we monitored two wires that stands at interval of 2.89 mm and 0.87 mm. Then, we compared the images reconstructed by algorithms. Finally, we monitored the two hairs crossed and compared between these algorithms.Keywords: back-projection, image comparison, non-uniform FFT, photoacoustic tomography
Procedia PDF Downloads 4343456 Subdued Electrodermal Response to Empathic Induction Task in Intimate Partner Violence (IPV) Perpetrators
Authors: Javier Comes Fayos, Isabel Rodríguez Moreno, Sara Bressanutti, Marisol Lila, Angel Romero Martínez, Luis Moya Albiol
Abstract:
Empathy is a cognitive-affective capacity whose deterioration is associated with aggressive behaviour. Deficient affective processing is one of the predominant risk factors in men convicted of intimate partner violence (IPV perpetrators), since it makes their capacity to empathize very difficult. The objective of this study is to compare the response of electrodermal activity (EDA), as an indicator of emotionality, to an empathic induction task, between IPV perpetrators and men without a history of violence. The sample was composed of 51 men who attended the CONTEXTO program, with penalties for gender violence under two years, and 47 men with no history of violence. Empathic induction was achieved through the visualization of 4 negative emotional-eliciting videos taken from an emotional induction battery of videos validated for the Spanish population. The participants were asked to actively empathize with the video characters (previously pointed out). The psychophysiological recording of the EDA was accomplished by the "Vrije Universiteit Ambulatory Monitoring System (VU-AMS)." An analysis of repeated measurements was carried out with 10 intra-subject measurements (time) and "group" (IPV perpetrators and non-violent perpetrators) as the inter-subject factor. First, there were no significant differences between groups in the baseline AED levels. Yet, a significant interaction between the “time” and “group” was found with IPV perpetrators exhibiting lower EDA response than controls after the empathic induction task. These findings provide evidence of a subdued EDA response after an empathic induction task in IPV perpetrators with respect to men without a history of violence. Therefore, the lower psychophysiological activation would be indicative of difficulties in the emotional processing and response, functions that are necessary for the empathic function. Consequently, the importance of addressing possible empathic difficulties in IPV perpetrator psycho-educational programs is reinforced, putting special emphasis on the affective dimension that could hinder the empathic function.Keywords: electrodermal activity, emotional induction, empathy, intimate partner violence
Procedia PDF Downloads 2003455 Isolation of Clitorin and Manghaslin from Carica papaya L. Leaves by CPC and Its Quantitative Analysis by QNMR
Authors: Norazlan Mohmad Misnan, Maizatul Hasyima Omar, Mohd Isa Wasiman
Abstract:
Papaya (Carica papaya L., Caricaceae) is a tree which mainly cultivated for its fruits in many tropical regions including Australia, Brazil, China, Hawaii, and Malaysia. Beside of fruits, its leaves, seeds, and latex have also been traditionally used for treating diseases, which also reported to possess anti-cancer and anti- malaria properties. Its leaves have been reported to consist of various chemical compounds such as alkaloids, flavonoids and phenolics. Clitorin and manghaslin are among major flavonoids presence. Thus, the aim of this study is to quantify the purity of these isolated compounds (clitorin and manghsalin) by using quantitative Nuclear Magnetic Resonance (qNMR) analysis. Only fresh C. papaya leaves were used for juice extraction procedure and subsequently was freeze-dried to obtain a dark green powdered form of the extract prior to Centrifugal Partition Chromatography (CPC) separation. The CPC experiments were performed using a two-phase solvent system comprising ethyl acetate/butanol/water (1:4:5, v/v/v/v) solvent. The upper organic phase was used as the stationary phase, and the lower aqueous phase was employed as the mobile phase. Ten fractions were obtained after an hour runtime analysis. Fraction 6 and fraction 8 has been identified as clitorin (m/z 739.21 [M-H]-) and manghaslin (m/z 755.21 [M-H]-), respectively, based on LCMS data and full analysis of NMR (1H NMR, 13C NMR, HMBC, and HSQC). The 1H-qNMR measurements were carried out using a 400 MHz NMR spectrometer (JEOL ECS 400MHz, Japan) and deuterated methanol was used as a solvent. Quantification was performed using the AQARI method (Accurate Quantitative NMR) with deuterated 1,4-Bis(trimethylsilyl)benzene (BTMSB) as an internal reference substances. This AQARI protocol includes not only NMR measurement but also sample preparation that provide highest precision and accuracy than other qNMR methods. The 90° pulse length and the T1 relaxation times for compounds and BTMSB were determined prior to the quantification to give the best signal-to-noise ratio. Regions containing the two downfield signals from aromatic part (6.00–6.89 ppm), and the singlet signal, (18H) arising from BTMSB (0.63-1.05ppm) were selected for integration. The purity of clitorin and manghaslin were calculated to be 52.22% and 43.36%, respectively. Further purification is needed in order to increase its purity. This finding has demonstrated the use of qNMR for quality control and standardization of various plant extracts and which can be applied for NMR fingerprinting of other plant-based products with good reproducibility and in the case where commercial standards is not readily available.Keywords: Carica papaya, clitorin, manghaslin, quantitative Nuclear Magnetic Resonance, Centrifugal Partition Chromatography
Procedia PDF Downloads 4963454 HRV Analysis Based Arrhythmic Beat Detection Using kNN Classifier
Authors: Onder Yakut, Oguzhan Timus, Emine Dogru Bolat
Abstract:
Health diseases have a vital significance affecting human being's life and life quality. Sudden death events can be prevented owing to early diagnosis and treatment methods. Electrical signals, taken from the human being's body using non-invasive methods and showing the heart activity is called Electrocardiogram (ECG). The ECG signal is used for following daily activity of the heart by clinicians. Heart Rate Variability (HRV) is a physiological parameter giving the variation between the heart beats. ECG data taken from MITBIH Arrhythmia Database is used in the model employed in this study. The detection of arrhythmic heart beats is aimed utilizing the features extracted from the HRV time domain parameters. The developed model provides a satisfactory performance with ~89% accuracy, 91.7 % sensitivity and 85% specificity rates for the detection of arrhythmic beats.Keywords: arrhythmic beat detection, ECG, HRV, kNN classifier
Procedia PDF Downloads 3523453 Dual Mode Mobile Based Detection of Endogenous Hydrogen Sulfide for Determination of Live and Antibiotic Resistant Bacteria
Authors: Shashank Gahlaut, Chandrashekhar Sharan, J. P. Singh
Abstract:
Increasing incidence of antibiotic-resistant bacteria is a big concern for the treatment of pathogenic diseases. The effect of treatment of patients with antibiotics often leads to the evolution of antibiotic resistance in the pathogens. The detection of antibiotic or antimicrobial resistant bacteria (microbes) is quite essential as it is becoming one of the big threats globally. Here we propose a novel technique to tackle this problem. We are taking a step forward to prevent the infections and diseases due to drug resistant microbes. This detection is based on some unique features of silver (a noble metal) nanorods (AgNRs) which are fabricated by a physical deposition method called thermal glancing angle deposition (GLAD). Silver nanorods are found to be highly sensitive and selective for hydrogen sulfide (H2S) gas. Color and water wetting (contact angle) of AgNRs are two parameters what are effected in the presence of this gas. H₂S is one of the major gaseous products evolved in the bacterial metabolic process. It is also known as gasotransmitter that transmits some biological singles in living systems. Nitric Oxide (NO) and Carbon mono oxide (CO) are two another members of this family. Orlowski (1895) observed the emission of H₂S by the bacteria for the first time. Most of the microorganism produce these gases. Here we are focusing on H₂S gas evolution to determine live/dead and antibiotic-resistant bacteria. AgNRs array has been used for the detection of H₂S from micro-organisms. A mobile app is also developed to make it easy, portable, user-friendly, and cost-effective.Keywords: antibiotic resistance, hydrogen sulfide, live and dead bacteria, mobile app
Procedia PDF Downloads 1453452 Additive White Gaussian Noise Filtering from ECG by Wiener Filter and Median Filter: A Comparative Study
Authors: Hossein Javidnia, Salehe Taheri
Abstract:
The Electrocardiogram (ECG) is the recording of the heart’s electrical potential versus time. ECG signals are often contaminated with noise such as baseline wander and muscle noise. As these signals have been widely used in clinical studies to detect heart diseases, it is essential to filter these noises. In this paper we compare performance of Wiener Filtering and Median Filtering methods to filter Additive White Gaussian (AWG) noise with the determined signal to noise ratio (SNR) ranging from 3 to 5 dB applied to long-term ECG recordings samples. Root mean square error (RMSE) and coefficient of determination (R2) between the filtered ECG and original ECG was used as the filter performance indicator. Experimental results show that Wiener filter has better noise filtering performance than Median filter.Keywords: ECG noise filtering, Wiener filtering, median filtering, Gaussian noise, filtering performance
Procedia PDF Downloads 5293451 Reinforcement Learning For Agile CNC Manufacturing: Optimizing Configurations And Sequencing
Authors: Huan Ting Liao
Abstract:
In a typical manufacturing environment, computer numerical control (CNC) machining is essential for automating production through precise computer-controlled tool operations, significantly enhancing efficiency and ensuring consistent product quality. However, traditional CNC production lines often rely on manual loading and unloading, limiting operational efficiency and scalability. Although automated loading systems have been developed, they frequently lack sufficient intelligence and configuration efficiency, requiring extensive setup adjustments for different products and impacting overall productivity. This research addresses the job shop scheduling problem (JSSP) in CNC machining environments, aiming to minimize total completion time (makespan) and maximize CNC machine utilization. We propose a novel approach using reinforcement learning (RL), specifically the Q-learning algorithm, to optimize scheduling decisions. The study simulates the JSSP, incorporating robotic arm operations, machine processing times, and work order demand allocation to determine optimal processing sequences. The Q-learning algorithm enhances machine utilization by dynamically balancing workloads across CNC machines, adapting to varying job demands and machine states. This approach offers robust solutions for complex manufacturing environments by automating decision-making processes for job assignments. Additionally, we evaluate various layout configurations to identify the most efficient setup. By integrating RL-based scheduling optimization with layout analysis, this research aims to provide a comprehensive solution for improving manufacturing efficiency and productivity in CNC-based job shops. The proposed method's adaptability and automation potential promise significant advancements in tackling dynamic manufacturing challenges.Keywords: job shop scheduling problem, reinforcement learning, operations sequence, layout optimization, q-learning
Procedia PDF Downloads 243450 Blind Speech Separation Using SRP-PHAT Localization and Optimal Beamformer in Two-Speaker Environments
Authors: Hai Quang Hong Dam, Hai Ho, Minh Hoang Le Ngo
Abstract:
This paper investigates the problem of blind speech separation from the speech mixture of two speakers. A voice activity detector employing the Steered Response Power - Phase Transform (SRP-PHAT) is presented for detecting the activity information of speech sources and then the desired speech signals are extracted from the speech mixture by using an optimal beamformer. For evaluation, the algorithm effectiveness, a simulation using real speech recordings had been performed in a double-talk situation where two speakers are active all the time. Evaluations show that the proposed blind speech separation algorithm offers a good interference suppression level whilst maintaining a low distortion level of the desired signal.Keywords: blind speech separation, voice activity detector, SRP-PHAT, optimal beamformer
Procedia PDF Downloads 2833449 Time Temperature Dependence of Long Fiber Reinforced Polypropylene Manufactured by Direct Long Fiber Thermoplastic Process
Authors: K. A. Weidenmann, M. Grigo, B. Brylka, P. Elsner, T. Böhlke
Abstract:
In order to reduce fuel consumption, the weight of automobiles has to be reduced. Fiber reinforced polymers offer the potential to reach this aim because of their high stiffness to weight ratio. Additionally, the use of fiber reinforced polymers in automotive applications has to allow for an economic large-scale production. In this regard, long fiber reinforced thermoplastics made by direct processing offer both mechanical performance and processability in injection moulding and compression moulding. The work presented in this contribution deals with long glass fiber reinforced polypropylene directly processed in compression moulding (D-LFT). For the use in automotive applications both the temperature and the time dependency of the materials properties have to be investigated to fulfill performance requirements during crash or the demands of service temperatures ranging from -40 °C to 80 °C. To consider both the influence of temperature and time, quasistatic tensile tests have been carried out at different temperatures. These tests have been complemented by high speed tensile tests at different strain rates. As expected, the increase in strain rate results in an increase of the elastic modulus which correlates to an increase of the stiffness with decreasing service temperature. The results are in good accordance with results determined by dynamic mechanical analysis within the range of 0.1 to 100 Hz. The experimental results from different testing methods were grouped and interpreted by using different time temperature shift approaches. In this regard, Williams-Landel-Ferry and Arrhenius approach based on kinetics have been used. As the theoretical shift factor follows an arctan function, an empirical approach was also taken into consideration. It could be shown that this approach describes best the time and temperature superposition for glass fiber reinforced polypropylene manufactured by D-LFT processing.Keywords: composite, dynamic mechanical analysis, long fibre reinforced thermoplastics, mechanical properties, time temperature superposition
Procedia PDF Downloads 1993448 A CORDIC Based Design Technique for Efficient Computation of DCT
Authors: Deboraj Muchahary, Amlan Deep Borah Abir J. Mondal, Alak Majumder
Abstract:
A discrete cosine transform (DCT) is described and a technique to compute it using fast Fourier transform (FFT) is developed. In this work, DCT of a finite length sequence is obtained by incorporating CORDIC methodology in radix-2 FFT algorithm. The proposed methodology is simple to comprehend and maintains a regular structure, thereby reducing computational complexity. DCTs are used extensively in the area of digital processing for the purpose of pattern recognition. So the efficient computation of DCT maintaining a transparent design flow is highly solicited.Keywords: DCT, DFT, CORDIC, FFT
Procedia PDF Downloads 478