Search results for: deep drawing process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17190

Search results for: deep drawing process

16650 Guidelines for the Management Process Development of Research Journals in Order to Develop Suan Sunandha Rajabhat University to International Standards

Authors: Araya Yordchim, Rosjana Chandhasa, Suwaree Yordchim

Abstract:

This research aims to study guidelines on the development of management process for research journals in order to develop Suan Sunandha Rajabhat University to international standards. This research investigated affecting elements ranging from the format of the article, evaluation form for research article quality, the process of creating a scholarly journal, satisfaction level of those with knowledge and competency to conduct research, arisen problems, and solutions. Drawing upon the sample size of 40 persons who had knowledge and competency in conducting research and creating scholarly journal articles at an international level, the data for this research were collected using questionnaires as a tool. Through the usage of computer software, data were analyzed by using the statistics in the forms of frequency, percentage, mean, standard deviation, and multiple regression analysis. The majority of participants were civil servants with a doctorate degree, followed by civil servants with a master's degree. Among them, the suitability of the article format was rated at a good level while the evaluation form for research articles quality was assessed at a good level. Based on participants' viewpoints, the process of creating scholarly journals was at a good level, while the satisfaction of those who had knowledge and competency in conducting research was at a satisfactory level. The problems encountered were the difficulty in accessing the website. The solution to the problem was to develop a website with user-friendly accessibility, including setting up a Google scholar profile for the purpose of references counting and the articles being used for reference in real-time. Research article format influenced the level of satisfaction of those who had the knowledge and competency to conduct research with statistical significance at the 0.01 level. The research article quality assessment form (preface section, research article writing section, preparation for research article manuscripts section, and the original article evaluation form for the author) affected the satisfaction of those with knowledge and competency to conduct research with the statistical significance at the level of 0.01. The process of establishing journals had an impact on the satisfaction of those with knowledge and ability to conduct research with statistical significance at the level of .05

Keywords: guidelines, development of management, research journals, international standards

Procedia PDF Downloads 118
16649 A Case Study of Conceptual Framework for Process Performance

Authors: Ljubica Milanović Glavan, Vesna Bosilj Vukšić, Dalia Suša

Abstract:

In order to gain a competitive advantage, many companies are focusing on reorganization of their business processes and implementing process-based management. In this context, assessing process performance is essential because it enables individuals and groups to assess where they stand in comparison to their competitors. In this paper, it is argued that process performance measurement is a necessity for a modern process-oriented company and it should be supported by a holistic process performance measurement system. It seems very unlikely that a universal set of performance indicators can be applied successfully to all business processes. Thus, performance indicators must be process-specific and have to be derived from both the strategic enterprise-wide goals and the process goals. Based on the extensive literature review and interviews conducted in Croatian company a conceptual framework for process performance measurement system was developed. The main objective of such system is to help process managers by providing comprehensive and timely information on the performance of business processes. This information can be used to communicate goals and current performance of a business process directly to the process team, to improve resource allocation and process output regarding quantity and quality, to give early warning signals, to make a diagnosis of the weaknesses of a business process, to decide whether corrective actions are needed and to assess the impact of actions taken.

Keywords: Croatia, key performance indicators, performance measurement, process performance

Procedia PDF Downloads 663
16648 Deep Reinforcement Learning for Advanced Pressure Management in Water Distribution Networks

Authors: Ahmed Negm, George Aggidis, Xiandong Ma

Abstract:

With the diverse nature of urban cities, customer demand patterns, landscape topologies or even seasonal weather trends; managing our water distribution networks (WDNs) has proved a complex task. These unpredictable circumstances manifest as pipe failures, intermittent supply and burst events thus adding to water loss, energy waste and increased carbon emissions. Whilst these events are unavoidable, advanced pressure management has proved an effective tool to control and mitigate them. Henceforth, water utilities have struggled with developing a real-time control method that is resilient when confronting the challenges of water distribution. In this paper we use deep reinforcement learning (DRL) algorithms as a novel pressure control strategy to minimise pressure violations and leakage under both burst and background leakage conditions. Agents based on asynchronous actor critic (A2C) and recurrent proximal policy optimisation (Recurrent PPO) were trained and compared to benchmarked optimisation algorithms (differential evolution, particle swarm optimisation. A2C manages to minimise leakage by 32.48% under burst conditions and 67.17% under background conditions which was the highest performance in the DRL algorithms. A2C and Recurrent PPO performed well in comparison to the benchmarks with higher processing speed and lower computational effort.

Keywords: deep reinforcement learning, pressure management, water distribution networks, leakage management

Procedia PDF Downloads 74
16647 The Effect of Career Decision Self Efficacy on Coping with Career Indecision among Young Adults

Authors: Yuliya Lipshits-Braziler

Abstract:

For many young adults, career decision making is a difficult and complex process that may lead to indecision. Indecision is frequently associated with great psychological distress and low levels of well-being. One important resource for dealing with indecision is career decision self-efficacy (CDSE), which refers to people’s beliefs about their ability to successfully accomplish certain tasks involved in career choice. Drawing from Social Cognitive Theory, it has been hypothesized that CDSE correlates with (a) people’s likelihood to engage in or avoid career decision making tasks, (b) the amount of effort put into the decision making process, (c) the people’s persistence in decision making efforts when faced with difficulties, and (d) the eventual success in arriving at career decisions. Based on these assumptions, the present study examines the associations between the CDSE and 14 strategies for coping with career indecision among young adults. Using the structural equation modeling (SEM), the results showed that CDSE is positively associated with the use of productive coping strategies, such as information-seeking, problem-solving, positive thinking, and self-regulation. In addition, CDSE was negatively associated with nonproductive coping strategies, such as avoidance, isolation, ruminative thinking, and blaming others. Contrary to our expectations, CDSE was not significantly correlated with instrumental help-seeking, while it was negatively correlated with emotional help-seeking. The results of this study can be used to facilitate the development of interventions aiming to reinforce young adults’ career decision making self-efficacy, which may provide them with a basis for overcoming career indecision more effectively.

Keywords: career decision self-efficacy, career indecision, coping strategies, career counseling

Procedia PDF Downloads 248
16646 Particle Size Effect on Shear Strength of Granular Materials in Direct Shear Test

Authors: R. Alias, A. Kasa, M. R. Taha

Abstract:

The effect of particle size on shear strength of granular materials are investigated using direct shear tests. Small direct shear test (60 mm by 60 mm by 24 mm deep) were conducted for particles passing the sieves with opening size of 2.36 mm. Meanwhile, particles passing the standard 20 mm sieves were tested using large direct shear test (300 mm by 300 mm by 200 mm deep). The large direct shear tests and the small direct shear tests carried out using the same shearing rate of 0.09 mm/min and similar normal stresses of 100, 200, and 300 kPa. The results show that the peak and residual shear strength decreases as particle size increases.

Keywords: particle size, shear strength, granular material, direct shear test

Procedia PDF Downloads 476
16645 Comparison of Machine Learning and Deep Learning Algorithms for Automatic Classification of 80 Different Pollen Species

Authors: Endrick Barnacin, Jean-Luc Henry, Jimmy Nagau, Jack Molinie

Abstract:

Palynology is a field of interest in many disciplines due to its multiple applications: chronological dating, climatology, allergy treatment, and honey characterization. Unfortunately, the analysis of a pollen slide is a complicated and time consuming task that requires the intervention of experts in the field, which are becoming increasingly rare due to economic and social conditions. That is why the need for automation of this task is urgent. A lot of studies have investigated the subject using different standard image processing descriptors and sometimes hand-crafted ones.In this work, we make a comparative study between classical feature extraction methods (Shape, GLCM, LBP, and others) and Deep Learning (CNN, Autoencoders, Transfer Learning) to perform a recognition task over 80 regional pollen species. It has been found that the use of Transfer Learning seems to be more precise than the other approaches

Keywords: pollens identification, features extraction, pollens classification, automated palynology

Procedia PDF Downloads 124
16644 Seashore Debris Detection System Using Deep Learning and Histogram of Gradients-Extractor Based Instance Segmentation Model

Authors: Anshika Kankane, Dongshik Kang

Abstract:

Marine debris has a significant influence on coastal environments, damaging biodiversity, and causing loss and damage to marine and ocean sector. A functional cost-effective and automatic approach has been used to look up at this problem. Computer vision combined with a deep learning-based model is being proposed to identify and categorize marine debris of seven kinds on different beach locations of Japan. This research compares state-of-the-art deep learning models with a suggested model architecture that is utilized as a feature extractor for debris categorization. The model is being proposed to detect seven categories of litter using a manually constructed debris dataset, with the help of Mask R-CNN for instance segmentation and a shape matching network called HOGShape, which can then be cleaned on time by clean-up organizations using warning notifications of the system. The manually constructed dataset for this system is created by annotating the images taken by fixed KaKaXi camera using CVAT annotation tool with seven kinds of category labels. A pre-trained HOG feature extractor on LIBSVM is being used along with multiple templates matching on HOG maps of images and HOG maps of templates to improve the predicted masked images obtained via Mask R-CNN training. This system intends to timely alert the cleanup organizations with the warning notifications using live recorded beach debris data. The suggested network results in the improvement of misclassified debris masks of debris objects with different illuminations, shapes, viewpoints and litter with occlusions which have vague visibility.

Keywords: computer vision, debris, deep learning, fixed live camera images, histogram of gradients feature extractor, instance segmentation, manually annotated dataset, multiple template matching

Procedia PDF Downloads 95
16643 Impact of Integrated Signals for Doing Human Activity Recognition Using Deep Learning Models

Authors: Milagros Jaén-Vargas, Javier García Martínez, Karla Miriam Reyes Leiva, María Fernanda Trujillo-Guerrero, Francisco Fernandes, Sérgio Barroso Gonçalves, Miguel Tavares Silva, Daniel Simões Lopes, José Javier Serrano Olmedo

Abstract:

Human Activity Recognition (HAR) is having a growing impact in creating new applications and is responsible for emerging new technologies. Also, the use of wearable sensors is an important key to exploring the human body's behavior when performing activities. Hence, the use of these dispositive is less invasive and the person is more comfortable. In this study, a database that includes three activities is used. The activities were acquired from inertial measurement unit sensors (IMU) and motion capture systems (MOCAP). The main objective is differentiating the performance from four Deep Learning (DL) models: Deep Neural Network (DNN), Convolutional Neural Network (CNN), Recurrent Neural Network (RNN) and hybrid model Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), when considering acceleration, velocity and position and evaluate if integrating the IMU acceleration to obtain velocity and position represent an increment in performance when it works as input to the DL models. Moreover, compared with the same type of data provided by the MOCAP system. Despite the acceleration data is cleaned when integrating, results show a minimal increase in accuracy for the integrated signals.

Keywords: HAR, IMU, MOCAP, acceleration, velocity, position, feature maps

Procedia PDF Downloads 88
16642 Study and Improvement of the Quality of a Production Line

Authors: S. Bouchami, M.N. Lakhoua

Abstract:

The automotive market is a dynamic market that continues to grow. That’s why several companies belonging to this sector adopt a quality improvement approach. Wanting to be competitive and successful in the environment in which they operate, these companies are dedicated to establishing a system of quality management to ensure the achievement of the objective quality, improving the products and process as well as the satisfaction of the customers. In this paper, the management of the quality and the improvement of a production line in an industrial company is presented. In fact, the project is divided into two essential parts: the creation of the technical line documentation and the quality assurance documentation and the resolution of defects at the line, as well as those claimed by the customer. The creation of the documents has required a deep understanding of the manufacturing process. The analysis and problem solving were done through the implementation of PDCA (Plan Do Check Act) and FTA (Fault Tree Analysis). As perspective, in order to better optimize production and improve the efficiency of the production line, a study on the problems associated with the supply of raw materials should be made to solve the problems of stock-outs which cause delays penalizing for the industrial company.

Keywords: quality management, documentary system, Plan Do Check Act (PDCA), fault tree analysis (FTA) method

Procedia PDF Downloads 131
16641 Water Quality Calculation and Management System

Authors: H. M. B. N Jayasinghe

Abstract:

The water is found almost everywhere on Earth. Water resources contain a lot of pollution. Some diseases can be spread through the water to the living beings. So to be clean water it should undergo a number of treatments necessary to make it drinkable. So it is must to have purification technology for the wastewater. So the waste water treatment plants act a major role in these issues. When considering the procedures taken after the water treatment process was always based on manual calculations and recordings. Water purification plants may interact with lots of manual processes. It means the process taking much time consuming. So the final evaluation and chemical, biological treatment process get delayed. So to prevent those types of drawbacks there are some computerized programmable calculation and analytical techniques going to be introduced to the laboratory staff. To solve this problem automated system will be a solution in which guarantees the rational selection. A decision support system is a way to model data and make quality decisions based upon it. It is widely used in the world for the various kind of process automation. Decision support systems that just collect data and organize it effectively are usually called passive models where they do not suggest a specific decision but only reveal information. This web base system is based on global positioning data adding facility with map location. Most worth feature is SMS and E-mail alert service to inform the appropriate person on a critical issue. The technological influence to the system is HTML, MySQL, PHP, and some other web developing technologies. Current issues in the computerized water chemistry analysis are not much deep in progress. For an example the swimming pool water quality calculator. The validity of the system has been verified by test running and comparison with an existing plant data. Automated system will make the life easier in productively and qualitatively.

Keywords: automated system, wastewater, purification technology, map location

Procedia PDF Downloads 240
16640 Development of Digital Twin Concept to Detect Abnormal Changes in Structural Behaviour

Authors: Shady Adib, Vladimir Vinogradov, Peter Gosling

Abstract:

Digital Twin (DT) technology is a new technology that appeared in the early 21st century. The DT is defined as the digital representation of living and non-living physical assets. By connecting the physical and virtual assets, data are transmitted smoothly, allowing the virtual asset to fully represent the physical asset. Although there are lots of studies conducted on the DT concept, there is still limited information about the ability of the DT models for monitoring and detecting unexpected changes in structural behaviour in real time. This is due to the large computational efforts required for the analysis and an excessively large amount of data transferred from sensors. This paper aims to develop the DT concept to be able to detect the abnormal changes in structural behaviour in real time using advanced modelling techniques, deep learning algorithms, and data acquisition systems, taking into consideration model uncertainties. finite element (FE) models were first developed offline to be used with a reduced basis (RB) model order reduction technique for the construction of low-dimensional space to speed the analysis during the online stage. The RB model was validated against experimental test results for the establishment of a DT model of a two-dimensional truss. The established DT model and deep learning algorithms were used to identify the location of damage once it has appeared during the online stage. Finally, the RB model was used again to identify the damage severity. It was found that using the RB model, constructed offline, speeds the FE analysis during the online stage. The constructed RB model showed higher accuracy for predicting the damage severity, while deep learning algorithms were found to be useful for estimating the location of damage with small severity.

Keywords: data acquisition system, deep learning, digital twin, model uncertainties, reduced basis, reduced order model

Procedia PDF Downloads 91
16639 Water-Controlled Fracturing with Fuzzy-Ball Fluid in Tight Gas Reservoirs of Deep Coal Measures in Sulige

Authors: Xiangchun Wang, Lihui Zheng, Maozong Gan, Peng Zhang, Tong Wu, An Chang

Abstract:

The deep coal measure tight gas reservoir in Sulige is usually reformed by fracturing, because the reservoir thickness is small, the water layers can be easily communicated during fracturing, which will lead to water production of gas wells and lower production of gas wells. Therefore, it is necessary to control water during fracturing in deep coal measure tight gas reservoir. Using fuzzy-ball fluid to control water fracturing can not only increase the output but also reduce the water output. The fuzzy-ball fluid was prepared indoors to carry out evaluation experiments. The fuzzy ball fluid was mixed in equal volume with the pre-fluid and formation water to test its compatibility. The core displacement device was used to test the gas and water breaking through the matrix and fractured cores blocked by fuzzy-ball fluid. The breakthrough pressure of the plunger tests its water blocking performance. The experimental results show that there is no precipitation after the fuzzy-ball fluid is mixed with the pad fluid and the formation water, respectively. The breakthrough pressure gradients of gas and water after the fuzzy-ball fluid plugged the cracks were 0.02MPa/cm and 0.04MPa/cm, respectively, and the breakthrough pressure gradients of gas and water after the matrix was plugged were 0.03MPa/cm and 0.2MPa/cm, respectively, which meet the requirements of field operation. Two wells A and B in the Sulige Gas Field were used on site to implement water control fracturing. After the pre-fluid was injected into the two wells, 50m3 of fuzzy-ball fluid was pumped to plug the water. The construction went smoothly. After water control and fracturing, the average daily output in 161 days was increased by 13.71% and 6.99% compared with that of adjacent wells in the same layer. The adjacent wells were bubbled for 3 times and 63 times respectively, while there was no effusion in A and B construction wells. The results show that fuzzy-ball fluid is a water plugging material suitable for water control fracturing in tight gas wells, and its water control mechanism can also provide a new idea for the development of water control fracturing materials.

Keywords: coal seam, deep layer, fracking, fuzzy-ball fluid, reservoir reconstruction

Procedia PDF Downloads 215
16638 Keyframe Extraction Using Face Quality Assessment and Convolution Neural Network

Authors: Rahma Abed, Sahbi Bahroun, Ezzeddine Zagrouba

Abstract:

Due to the huge amount of data in videos, extracting the relevant frames became a necessity and an essential step prior to performing face recognition. In this context, we propose a method for extracting keyframes from videos based on face quality and deep learning for a face recognition task. This method has two steps. We start by generating face quality scores for each face image based on the use of three face feature extractors, including Gabor, LBP, and HOG. The second step consists in training a Deep Convolutional Neural Network in a supervised manner in order to select the frames that have the best face quality. The obtained results show the effectiveness of the proposed method compared to the methods of the state of the art.

Keywords: keyframe extraction, face quality assessment, face in video recognition, convolution neural network

Procedia PDF Downloads 219
16637 Empirical Evaluation of Gradient-Based Training Algorithms for Ordinary Differential Equation Networks

Authors: Martin K. Steiger, Lukas Heisler, Hans-Georg Brachtendorf

Abstract:

Deep neural networks and their variants form the backbone of many AI applications. Based on the so-called residual networks, a continuous formulation of such models as ordinary differential equations (ODEs) has proven advantageous since different techniques may be applied that significantly increase the learning speed and enable controlled trade-offs with the resulting error at the same time. For the evaluation of such models, high-performance numerical differential equation solvers are used, which also provide the gradients required for training. However, whether classical gradient-based methods are even applicable or which one yields the best results has not been discussed yet. This paper aims to redeem this situation by providing empirical results for different applications.

Keywords: deep neural networks, gradient-based learning, image processing, ordinary differential equation networks

Procedia PDF Downloads 150
16636 Fixed Points of Contractive-Like Operators by a Faster Iterative Process

Authors: Safeer Hussain Khan

Abstract:

In this paper, we prove a strong convergence result using a recently introduced iterative process with contractive-like operators. This improves and generalizes corresponding results in the literature in two ways: the iterative process is faster, operators are more general. In the end, we indicate that the results can also be proved with the iterative process with error terms.

Keywords: contractive-like operator, iterative process, fixed point, strong convergence

Procedia PDF Downloads 422
16635 Assessment of Transverse Abdominis Activation during Three Different Exercises in Low Back Pain Patients: Measurement with Real-Time Ultrasonography

Authors: Venus Pagare, Amit Kharat, Dhaval K. Thakkar, Tushar J. Palekar

Abstract:

Introduction: Chronic low back pain (CLBP) is a major public health problem and is the leading musculoskeletal cause of disability. Altered neuromuscular control of core muscles, particulary transverses abdominis (TrA) is thought to be a contributing factor for the development of CLBP. Therefore, various exercises targeting the TrA are commonly incorporated into the rehabilitation. Objectives: To investigate the effects of 3 different core exercises on activation capacity of TrA muscle in individuals with CLBP as compared with healthy controls. Methodology: Thickness of TrA muscle was measured by ultrasound imaging in 30 patients with CLBP and 30 healthy controls. Measurements were taken during 3 different TrA activation exercises i.e Abdominal drawing in maneuver (ADIM), Abdominal drawing in with straight leg raise (ADSLR) and breathe hold at maximum expiration (ME). Thickness of the muscle at rest (at the end of normal tidal expiration) was taken as a baseline measure. Results: There was a significant difference between the healthy subjects and patients with low back pain with regard to the thickness of TrA at rest and thickness during contraction. ADIM produced a significant increase in the thickness of TrA compared to ADSLR and ME (p<0.001). Also, increase in thickness of TrA was more in the control group than patients with low back pain. Conclusion: CLBP patients exhibited atrophy of TrA muscle with delayed activation. Also, of the various core exercises, ADIM can be an effective method for activation of TrA.

Keywords: LBP, CLBP, ADSLR, ADIM

Procedia PDF Downloads 301
16634 Through Seligman’s Lenses: Creating a Culture of Well-Being in Higher-Education

Authors: Neeru Deep, Kimberly McAlister

Abstract:

Mental health issues have been increasing worldwide for many decades, but the COVID-19 pandemic has brought mental health issues into the spotlight. Within higher education, promoting the well-being of students has dramatically increased in focus. The Northwestern State University of Louisiana opened the Center for Positivity, Well-being, and Hope using the action research process of reflecting, planning, acting, and observing. The study’s purpose is two-fold: First, it highlights how to create a collaborative team to reflect, plan, and act to develop a well-being culture in higher education institutions. Second, it investigates the efficacy of the center through Seligman’s lenses. The researchers shared their experience in the first three phases of the action research process and then applied an identical concurrent mixed methods design. A purposive sample evaluated the efficacy of the center through Seligman’s lenses. The researcher administered PERMA-Profiler Measure, the PERMA-Profiler Measure overview, the CoPWH Evaluation I, and the CoPWH Evaluation II questionnaires to collect qualitative and quantitative data. The thematic analysis for qualitative and descriptive statistics for quantitative data concluded that the center creates a well-being culture and promotes well-being in college students. In conclusion, this action research shares the successful implementation of the cyclic process of research in promoting a well-being culture in higher education with the implications for promoting a well-being culture in various educational settings, workplaces, and communities.

Keywords: action research, mixed methods research design, Seligman, well-being.

Procedia PDF Downloads 121
16633 Decision Making Communication in the Process of Technologies Commercialization: Archival Analysis of the Process Content

Authors: Vaida Zemlickiene

Abstract:

Scientists around the world and practitioners are working to identify the factors that influence the results of technology commercialization and to propose the ideal model for the technology commercialization process. In other words, all stakeholders of technology commercialization seek to find a formula or set of rules to succeed in commercializing technologies in order to avoid unproductive investments. In this article, the process of commercialization technology is understood as the process of transforming inventions into marketable products, services, and processes, or the path from the idea of using an invention to a product that incorporates process from 1 to 9 technology readiness level (TRL). There are many publications in the field of management literature, which are aimed at managing the commercialization process. However, there is an apparent lack of research for communication in decision-making in the process of technology commercialization. Works were done in the past, and the last decade's global research analysis led to the unambiguous conclusion that the methodological framework is not mature enough to be of practical use in business. The process of technology commercialization and the decisions made in the process should be explored in-depth. An archival analysis is performed to find insights into decision-making communication in the process of technologies commercialization, to find out the content of technology commercialization process: decision-making stages and participants, to analyze the internal factors of technology commercialization, to perform their critical analysis, to analyze the concept of successful/unsuccessful technology commercialization.

Keywords: the process of technology commercialization, communication in decision-making process, the content of technology commercialization process, successful/unsuccessful technology commercialization

Procedia PDF Downloads 146
16632 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 183
16631 MIMIC: A Multi Input Micro-Influencers Classifier

Authors: Simone Leonardi, Luca Ardito

Abstract:

Micro-influencers are effective elements in the marketing strategies of companies and institutions because of their capability to create an hyper-engaged audience around a specific topic of interest. In recent years, many scientific approaches and commercial tools have handled the task of detecting this type of social media users. These strategies adopt solutions ranging from rule based machine learning models to deep neural networks and graph analysis on text, images, and account information. This work compares the existing solutions and proposes an ensemble method to generalize them with different input data and social media platforms. The deployed solution combines deep learning models on unstructured data with statistical machine learning models on structured data. We retrieve both social media accounts information and multimedia posts on Twitter and Instagram. These data are mapped into feature vectors for an eXtreme Gradient Boosting (XGBoost) classifier. Sixty different topics have been analyzed to build a rule based gold standard dataset and to compare the performances of our approach against baseline classifiers. We prove the effectiveness of our work by comparing the accuracy, precision, recall, and f1 score of our model with different configurations and architectures. We obtained an accuracy of 0.91 with our best performing model.

Keywords: deep learning, gradient boosting, image processing, micro-influencers, NLP, social media

Procedia PDF Downloads 170
16630 Effect of Cryogenic Treatment on Various Mechanical and Metallurgical Properties of Different Material: A Review

Authors: Prashant Dhiman, Viranshu Kumar, Pradeep Joshi

Abstract:

Lot of research is going on to study the effect of cryogenic treatment on materials. Cryogenic treatment is a heat treatment process which is used widely to enhance the mechanical and metallurgical properties of various materials whether the material is ferrous or non ferrous. In almost all ferrous metals, it is found that retained austenite is converted into martensite. Generally deep cryogenic treatment is done using liquid nitrogen having temperature of -195 ℃. The austenite is unstable at this stage and converts into martensite. In non ferrous materials there presents a microcavity and under the action of stress it becomes crack. When this crack propagates, fracture takes place. As the metal contract under low temperature, by doing cryogenic treatment these microcavities will be filled hence increases the soundness of the material. Properties which are enhanced by cryogenic treatment of both ferrous and non ferrous materials are hardness, tensile strength, wear rate, electrical and thermal conductivity, and others. Also there is decrease in residual stress. A large number of manufacturing process (EDM, CNC etc.) are using cryogenic treatment on different tools or workpiece to reduce their wear. In this Review paper the use of cryogenic heat treatment in different manufacturing has been shown along with their advantages.

Keywords: cyrogenic treatment, EDM (Electrical Discharge Machining), CNC (Computer Numeric Control), Mechanical and Metallurgical Properties

Procedia PDF Downloads 424
16629 Clinical Impact of Ultra-Deep Versus Sanger Sequencing Detection of Minority Mutations on the HIV-1 Drug Resistance Genotype Interpretations after Virological Failure

Authors: S. Mohamed, D. Gonzalez, C. Sayada, P. Halfon

Abstract:

Drug resistance mutations are routinely detected using standard Sanger sequencing, which does not detect minor variants with a frequency below 20%. The impact of detecting minor variants generated by ultra-deep sequencing (UDS) on HIV drug-resistance (DR) interpretations has not yet been studied. Fifty HIV-1 patients who experienced virological failure were included in this retrospective study. The HIV-1 UDS protocol allowed the detection and quantification of HIV-1 protease and reverse transcriptase variants related to genotypes A, B, C, E, F, and G. DeepChek®-HIV simplified DR interpretation software was used to compare Sanger sequencing and UDS. The total time required for the UDS protocol was found to be approximately three times longer than Sanger sequencing with equivalent reagent costs. UDS detected all of the mutations found by population sequencing and identified additional resistance variants in all patients. An analysis of DR revealed a total of 643 and 224 clinically relevant mutations by UDS and Sanger sequencing, respectively. Three resistance mutations with > 20% prevalence were detected solely by UDS: A98S (23%), E138A (21%) and V179I (25%). A significant difference in the DR interpretations for 19 antiretroviral drugs was observed between the UDS and Sanger sequencing methods. Y181C and T215Y were the most frequent mutations associated with interpretation differences. A combination of UDS and DeepChek® software for the interpretation of DR results would help clinicians provide suitable treatments. A cut-off of 1% allowed a better characterisation of the viral population by identifying additional resistance mutations and improving the DR interpretation.

Keywords: HIV-1, ultra-deep sequencing, Sanger sequencing, drug resistance

Procedia PDF Downloads 325
16628 Measuring Human Perception and Negative Elements of Public Space Quality Using Deep Learning: A Case Study of Area within the Inner Road of Tianjin City

Authors: Jiaxin Shi, Kaifeng Hao, Qingfan An, Zeng Peng

Abstract:

Due to a lack of data sources and data processing techniques, it has always been difficult to quantify public space quality, which includes urban construction quality and how it is perceived by people, especially in large urban areas. This study proposes a quantitative research method based on the consideration of emotional health and physical health of the built environment. It highlights the low quality of public areas in Tianjin, China, where there are many negative elements. Deep learning technology is then used to measure how effectively people perceive urban areas. First, this work suggests a deep learning model that might simulate how people can perceive the quality of urban construction. Second, we perform semantic segmentation on street images to identify visual elements influencing scene perception. Finally, this study correlated the scene perception score with the proportion of visual elements to determine the surrounding environmental elements that influence scene perception. Using a small-scale labeled Tianjin street view data set based on transfer learning, this study trains five negative spatial discriminant models in order to explore the negative space distribution and quality improvement of urban streets. Then it uses all Tianjin street-level imagery to make predictions and calculate the proportion of negative space. Visualizing the spatial distribution of negative space along the Tianjin Inner Ring Road reveals that the negative elements are mainly found close to the five key districts. The map of Tianjin was combined with the experimental data to perform the visual analysis. Based on the emotional assessment, the distribution of negative materials, and the direction of street guidelines, we suggest guidance content and design strategy points of the negative phenomena in Tianjin street space in the two dimensions of perception and substance. This work demonstrates the utilization of deep learning techniques to understand how people appreciate high-quality urban construction, and it complements both theory and practice in urban planning. It illustrates the connection between human perception and the actual physical public space environment, allowing researchers to make urban interventions.

Keywords: human perception, public space quality, deep learning, negative elements, street images

Procedia PDF Downloads 103
16627 Face Recognition Using Body-Worn Camera: Dataset and Baseline Algorithms

Authors: Ali Almadan, Anoop Krishnan, Ajita Rattani

Abstract:

Facial recognition is a widely adopted technology in surveillance, border control, healthcare, banking services, and lately, in mobile user authentication with Apple introducing “Face ID” moniker with iPhone X. A lot of research has been conducted in the area of face recognition on datasets captured by surveillance cameras, DSLR, and mobile devices. Recently, face recognition technology has also been deployed on body-worn cameras to keep officers safe, enabling situational awareness and providing evidence for trial. However, limited academic research has been conducted on this topic so far, without the availability of any publicly available datasets with a sufficient sample size. This paper aims to advance research in the area of face recognition using body-worn cameras. To this aim, the contribution of this work is two-fold: (1) collection of a dataset consisting of a total of 136,939 facial images of 102 subjects captured using body-worn cameras in in-door and daylight conditions and (2) evaluation of various deep-learning architectures for face identification on the collected dataset. Experimental results suggest a maximum True Positive Rate(TPR) of 99.86% at False Positive Rate(FPR) of 0.000 obtained by SphereFace based deep learning architecture in daylight condition. The collected dataset and the baseline algorithms will promote further research and development. A downloadable link of the dataset and the algorithms is available by contacting the authors.

Keywords: face recognition, body-worn cameras, deep learning, person identification

Procedia PDF Downloads 157
16626 Automating 2D CAD to 3D Model Generation Process: Wall pop-ups

Authors: Mohit Gupta, Chialing Wei, Thomas Czerniawski

Abstract:

In this paper, we have built a neural network that can detect walls on 2D sheets and subsequently create a 3D model in Revit using Dynamo. The training set includes 3500 labeled images, and the detection algorithm used is YOLO. Typically, engineers/designers make concentrated efforts to convert 2D cad drawings to 3D models. This costs a considerable amount of time and human effort. This paper makes a contribution in automating the task of 3D walls modeling. 1. Detecting Walls in 2D cad and generating 3D pop-ups in Revit. 2. Saving designer his/her modeling time in drafting elements like walls from 2D cad to 3D representation. An object detection algorithm YOLO is used for wall detection and localization. The neural network is trained over 3500 labeled images of size 256x256x3. Then, Dynamo is interfaced with the output of the neural network to pop-up 3D walls in Revit. The research uses modern technological tools like deep learning and artificial intelligence to automate the process of generating 3D walls without needing humans to manually model them. Thus, contributes to saving time, human effort, and money.

Keywords: neural networks, Yolo, 2D to 3D transformation, CAD object detection

Procedia PDF Downloads 138
16625 Instance Segmentation of Wildfire Smoke Plumes using Mask-RCNN

Authors: Jamison Duckworth, Shankarachary Ragi

Abstract:

Detection and segmentation of wildfire smoke plumes from remote sensing imagery are being pursued as a solution for early fire detection and response. Smoke plume detection can be automated and made robust by the application of artificial intelligence methods. Specifically, in this study, the deep learning approach Mask Region-based Convolutional Neural Network (RCNN) is being proposed to learn smoke patterns across different spectral bands. This method is proposed to separate the smoke regions from the background and return masks placed over the smoke plumes. Multispectral data was acquired using NASA’s Earthdata and WorldView and services and satellite imagery. Due to the use of multispectral bands along with the three visual bands, we show that Mask R-CNN can be applied to distinguish smoke plumes from clouds and other landscape features that resemble smoke.

Keywords: deep learning, mask-RCNN, smoke plumes, spectral bands

Procedia PDF Downloads 115
16624 Searching the Relationship among Components that Contribute to Interactive Plight and Educational Execution

Authors: Shri Krishna Mishra

Abstract:

In an educational context, technology can prompt interactive plight only when it is used in conjunction with interactive plight methods. This study, therefore, examines the relationships among components that contribute to higher levels of interactive plight and execution, such as interactive Plight methods, technology, intrinsic motivation and deep learning. 526 students participated in this study. With structural equation modelling, the authors test the conceptual model and identify satisfactory model fit. The results indicate that interactive Plight methods, technology and intrinsic motivation have significant relationship with interactive Plight; deep learning mediates the relationships of the other variables with Execution.

Keywords: searching the relationship among components, contribute to interactive plight, educational execution, intrinsic motivation

Procedia PDF Downloads 442
16623 Optimization of the Transfer Molding Process by Implementation of Online Monitoring Techniques for Electronic Packages

Authors: Burcu Kaya, Jan-Martin Kaiser, Karl-Friedrich Becker, Tanja Braun, Klaus-Dieter Lang

Abstract:

Quality of the molded packages is strongly influenced by the process parameters of the transfer molding. To achieve a better package quality and a stable transfer molding process, it is necessary to understand the influence of the process parameters on the package quality. This work aims to comprehend the relationship between the process parameters, and to identify the optimum process parameters for the transfer molding process in order to achieve less voids and wire sweep. To achieve this, a DoE is executed for process optimization and a regression analysis is carried out. A systematic approach is represented to generate models which enable an estimation of the number of voids and wire sweep. Validation experiments are conducted to verify the model and the results are presented.

Keywords: dielectric analysis, electronic packages, epoxy molding compounds, transfer molding process

Procedia PDF Downloads 374
16622 Progress in Combining Image Captioning and Visual Question Answering Tasks

Authors: Prathiksha Kamath, Pratibha Jamkhandi, Prateek Ghanti, Priyanshu Gupta, M. Lakshmi Neelima

Abstract:

Combining Image Captioning and Visual Question Answering (VQA) tasks have emerged as a new and exciting research area. The image captioning task involves generating a textual description that summarizes the content of the image. VQA aims to answer a natural language question about the image. Both these tasks include computer vision and natural language processing (NLP) and require a deep understanding of the content of the image and semantic relationship within the image and the ability to generate a response in natural language. There has been remarkable growth in both these tasks with rapid advancement in deep learning. In this paper, we present a comprehensive review of recent progress in combining image captioning and visual question-answering (VQA) tasks. We first discuss both image captioning and VQA tasks individually and then the various ways in which both these tasks can be integrated. We also analyze the challenges associated with these tasks and ways to overcome them. We finally discuss the various datasets and evaluation metrics used in these tasks. This paper concludes with the need for generating captions based on the context and captions that are able to answer the most likely asked questions about the image so as to aid the VQA task. Overall, this review highlights the significant progress made in combining image captioning and VQA, as well as the ongoing challenges and opportunities for further research in this exciting and rapidly evolving field, which has the potential to improve the performance of real-world applications such as autonomous vehicles, robotics, and image search.

Keywords: image captioning, visual question answering, deep learning, natural language processing

Procedia PDF Downloads 65
16621 Optimal Performance of Plastic Extrusion Process Using Fuzzy Goal Programming

Authors: Abbas Al-Refaie

Abstract:

This study optimized the performance of plastic extrusion process of drip irrigation pipes using fuzzy goal programming. Two main responses were of main interest; roll thickness and hardness. Four main process factors were studied. The L18 array was then used for experimental design. The individual-moving range control charts were used to assess the stability of the process, while the process capability index was used to assess process performance. Confirmation experiments were conducted at the obtained combination of optimal factor setting by fuzzy goal programming. The results revealed that process capability was improved significantly from -1.129 to 0.8148 for roll thickness and from 0.0965 to 0.714 and hardness. Such improvement results in considerable savings in production and quality costs.

Keywords: fuzzy goal programming, extrusion process, process capability, irrigation plastic pipes

Procedia PDF Downloads 257