Search results for: materials processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10046

Search results for: materials processing

9596 Genomic Sequence Representation Learning: An Analysis of K-Mer Vector Embedding Dimensionality

Authors: James Jr. Mashiyane, Risuna Nkolele, Stephanie J. Müller, Gciniwe S. Dlamini, Rebone L. Meraba, Darlington S. Mapiye

Abstract:

When performing language tasks in natural language processing (NLP), the dimensionality of word embeddings is chosen either ad-hoc or is calculated by optimizing the Pairwise Inner Product (PIP) loss. The PIP loss is a metric that measures the dissimilarity between word embeddings, and it is obtained through matrix perturbation theory by utilizing the unitary invariance of word embeddings. Unlike in natural language, in genomics, especially in genome sequence processing, unlike in natural language processing, there is no notion of a “word,” but rather, there are sequence substrings of length k called k-mers. K-mers sizes matter, and they vary depending on the goal of the task at hand. The dimensionality of word embeddings in NLP has been studied using the matrix perturbation theory and the PIP loss. In this paper, the sufficiency and reliability of applying word-embedding algorithms to various genomic sequence datasets are investigated to understand the relationship between the k-mer size and their embedding dimension. This is completed by studying the scaling capability of three embedding algorithms, namely Latent Semantic analysis (LSA), Word2Vec, and Global Vectors (GloVe), with respect to the k-mer size. Utilising the PIP loss as a metric to train embeddings on different datasets, we also show that Word2Vec outperforms LSA and GloVe in accurate computing embeddings as both the k-mer size and vocabulary increase. Finally, the shortcomings of natural language processing embedding algorithms in performing genomic tasks are discussed.

Keywords: word embeddings, k-mer embedding, dimensionality reduction

Procedia PDF Downloads 133
9595 Cost Effective Real-Time Image Processing Based Optical Mark Reader

Authors: Amit Kumar, Himanshu Singal, Arnav Bhavsar

Abstract:

In this modern era of automation, most of the academic exams and competitive exams are Multiple Choice Questions (MCQ). The responses of these MCQ based exams are recorded in the Optical Mark Reader (OMR) sheet. Evaluation of the OMR sheet requires separate specialized machines for scanning and marking. The sheets used by these machines are special and costs more than a normal sheet. Available process is non-economical and dependent on paper thickness, scanning quality, paper orientation, special hardware and customized software. This study tries to tackle the problem of evaluating the OMR sheet without any special hardware and making the whole process economical. We propose an image processing based algorithm which can be used to read and evaluate the scanned OMR sheets with no special hardware required. It will eliminate the use of special OMR sheet. Responses recorded in normal sheet is enough for evaluation. The proposed system takes care of color, brightness, rotation, little imperfections in the OMR sheet images.

Keywords: OMR, image processing, hough circle trans-form, interpolation, detection, binary thresholding

Procedia PDF Downloads 167
9594 Analysis for Shear Spinning of Tubes with Hard-To-Work Materials

Authors: Sukhwinder Singh Jolly

Abstract:

Metal spinning is one such process in which the stresses are localized to a small area and the material is made to flow or move over the mandrel with the help of spinning tool. Spinning of tubular products can be performed by two techniques, forward spinning and backward spinning. Many researchers have studied the process both experimentally and analytically. An effort has been made to apply the process to the spinning of thin wall, highly precision, small bore long tube in hard-to-work materials such as titanium.

Keywords: metal spinning, hard-to-work materials, roller diameter, power consumption

Procedia PDF Downloads 384
9593 Construction Innovation: Support for 3D Printing House

Authors: Andrea Palazzo, Daniel Macek, Veronika Malinova

Abstract:

Contour processing is the new technology challenge for architects and construction companies. The many advantages it promises make it one of the most interesting solutions for construction in terms of automation of building processes. The technology for 3D printing houses offers many application possibilities, from low-cost construction, to being considered by NASA for visionary projects as a good solution for building settlements on other planets. Another very important point is that clients, as architects, will no longer have many limits in design concerning ideas and creativity. The prices for real estate are constantly increasing and the lack of availability of construction materials as well as the speculation that has been created around it in 2021 is bringing prices to such a level that in the future real estate developers risk not being able to find customers for these ultra-expensive homes. Hence, this paper starts with the introduction of 3D printing, which now has the potential to gain an important position in the market, becoming a valid alternative to the classic construction process. This technology is not only beneficial from an economic point of view but it is also a great opportunity to have an impact on the environment by reducing CO2 emissions. Further on in the article we will also understand if, after the COP 26 (2021 United Nations Climate Change Conference), world governments could also push towards building technologies that reduce the waste materials that are needed to be disposed of and at the same time reduce emissions with the contribution of governmental funds. This paper will give us insight on the multiple benefits of 3D printing and emphasise the importance of finding new solutions for materials that can be used by the printer. Therefore, based on the type of material, it will be possible to understand the compatibility with current regulations and how the authorities will be inclined to support this technology. This will help to enable the rise and development of this technology in Europe and in the rest of the world on actual housing projects and not only on prototypes.

Keywords: additive manufacturing, contour crafting, development, new regulation, printing material

Procedia PDF Downloads 191
9592 VIAN-DH: Computational Multimodal Conversation Analysis Software and Infrastructure

Authors: Teodora Vukovic, Christoph Hottiger, Noah Bubenhofer

Abstract:

The development of VIAN-DH aims at bridging two linguistic approaches: conversation analysis/interactional linguistics (IL), so far a dominantly qualitative field, and computational/corpus linguistics and its quantitative and automated methods. Contemporary IL investigates the systematic organization of conversations and interactions composed of speech, gaze, gestures, and body positioning, among others. These highly integrated multimodal behaviour is analysed based on video data aimed at uncovering so called “multimodal gestalts”, patterns of linguistic and embodied conduct that reoccur in specific sequential positions employed for specific purposes. Multimodal analyses (and other disciplines using videos) are so far dependent on time and resource intensive processes of manual transcription of each component from video materials. Automating these tasks requires advanced programming skills, which is often not in the scope of IL. Moreover, the use of different tools makes the integration and analysis of different formats challenging. Consequently, IL research often deals with relatively small samples of annotated data which are suitable for qualitative analysis but not enough for making generalized empirical claims derived quantitatively. VIAN-DH aims to create a workspace where many annotation layers required for the multimodal analysis of videos can be created, processed, and correlated in one platform. VIAN-DH will provide a graphical interface that operates state-of-the-art tools for automating parts of the data processing. The integration of tools that already exist in computational linguistics and computer vision, facilitates data processing for researchers lacking programming skills, speeds up the overall research process, and enables the processing of large amounts of data. The main features to be introduced are automatic speech recognition for the transcription of language, automatic image recognition for extraction of gestures and other visual cues, as well as grammatical annotation for adding morphological and syntactic information to the verbal content. In the ongoing instance of VIAN-DH, we focus on gesture extraction (pointing gestures, in particular), making use of existing models created for sign language and adapting them for this specific purpose. In order to view and search the data, VIAN-DH will provide a unified format and enable the import of the main existing formats of annotated video data and the export to other formats used in the field, while integrating different data source formats in a way that they can be combined in research. VIAN-DH will adapt querying methods from corpus linguistics to enable parallel search of many annotation levels, combining token-level and chronological search for various types of data. VIAN-DH strives to bring crucial and potentially revolutionary innovation to the field of IL, (that can also extend to other fields using video materials). It will allow the processing of large amounts of data automatically and, the implementation of quantitative analyses, combining it with the qualitative approach. It will facilitate the investigation of correlations between linguistic patterns (lexical or grammatical) with conversational aspects (turn-taking or gestures). Users will be able to automatically transcribe and annotate visual, spoken and grammatical information from videos, and to correlate those different levels and perform queries and analyses.

Keywords: multimodal analysis, corpus linguistics, computational linguistics, image recognition, speech recognition

Procedia PDF Downloads 105
9591 Mixotropohic Growth of Chlorella sp. on Raw Food Processing Industrial Wastewater: Effect of COD Tolerance

Authors: Suvidha Gupta, R. A. Pandey, Sanjay Pawar

Abstract:

The effluents from various food processing industries are found with high BOD, COD, suspended solids, nitrate, and phosphate. Mixotrophic growth of microalgae using food processing industrial wastewater as an organic carbon source has emerged as more effective and energy intensive means for the nutrient removal and COD reduction. The present study details the treatment of non-sterilized unfiltered food processing industrial wastewater by microalgae for nutrient removal as well as to determine the tolerance to COD by taking different dilutions of wastewater. In addition, the effect of different inoculum percentages of microalgae on removal efficiency of the nutrients for given dilution has been studied. To see the effect of dilution and COD tolerance, the wastewater having initial COD 5000 mg/L (±5), nitrate 28 mg/L (±10), and phosphate 24 mg/L (±10) was diluted to get COD of 3000 mg/L and 1000 mg/L. The experiments were carried out in 1L conical flask by intermittent aeration with different inoculum percentage i.e. 10%, 20%, and 30% of Chlorella sp. isolated from nearby area of NEERI, Nagpur. The experiments were conducted for 6 days by providing 12:12 light- dark period and determined various parameters such as COD, TOC, NO3-- N, PO4-- P, and total solids on daily basis. Results revealed that, for 10% and 20% inoculum, over 90% COD and TOC reduction was obtained with wastewater containing COD of 3000 mg/L whereas over 80% COD and TOC reduction was obtained with wastewater containing COD of 1000 mg/L. Moreover, microalgae was found to tolerate wastewater containing COD 5000 mg/L and obtained over 60% and 80% reduction in COD and TOC respectively. The obtained results were found similar with 10% and 20% inoculum in all COD dilutions whereas for 30% inoculum over 60% COD and 70% TOC reduction was obtained. In case of nutrient removal, over 70% nitrate removal and 45% phosphate removal was obtained with 20% inoculum in all dilutions. The obtained results indicated that Microalgae assisted nutrient removal gives maximum COD and TOC reduction with 3000 mg/L COD and 20% inoculum. Hence, microalgae assisted wastewater treatment is not only effective for removal of nutrients but also can tolerate high COD up to 5000 mg/L and solid content.

Keywords: Chlorella sp., chemical oxygen demand, food processing industrial wastewater, mixotrophic growth

Procedia PDF Downloads 326
9590 Nanostructured Transition Metal Oxides Doped Graphene for High Performance Solid-State Supercapacitor Electrodes

Authors: G. Nyongombe, Guy L. Kabongo, B. M. Mothudi, M. S. Dhlamini

Abstract:

A series of Transition Metals Oxides (TMOs) doped graphene were synthesized and successfully used as supercapacitor electrode materials. The as-synthesized materials exhibited exceptional electrochemical properties owing to the combined properties of its constituents; high surface area and good conductivity were achieved. Several analytical characterization techniques were employed to investigate the morphology, crystal structure atomic arrangement and elemental chemical state in the materials for which scanning electron microscopy (SEM), X-ray diffraction (XRD) and X-ray photoelectron spectroscopy (XPS) were conducted, respectively. Moreover, the electrochemical properties of the as-synthesized materials were examined by performing cyclic voltammetry (CV), galvanostatic charge-discharge (GCD) and electrochemical impedance spectroscopy (EIS) measurements. Furthermore, the effect of doping concentration on the interlayer distance of the graphene materials and the charge transfer resistance are investigated and correlated to the exceptional current density which was multiplied by a factor of ~80 after TMOs doping in graphene. Finally, the resulting high capacitance obtained confirms the contribution of grapheme exceptional electronic conductivity and large surface area on the electrode materials. Such good-performing electrode materials are highly promising for supercapacitors and other energy storage devices.

Keywords: energy density, graphene, supercapacitors, TMOs

Procedia PDF Downloads 250
9589 Tool Condition Monitoring of Ceramic Inserted Tools in High Speed Machining through Image Processing

Authors: Javier A. Dominguez Caballero, Graeme A. Manson, Matthew B. Marshall

Abstract:

Cutting tools with ceramic inserts are often used in the process of machining many types of superalloy, mainly due to their high strength and thermal resistance. Nevertheless, during the cutting process, the plastic flow wear generated in these inserts enhances and propagates cracks due to high temperature and high mechanical stress. This leads to a very variable failure of the cutting tool. This article explores the relationship between the continuous wear that ceramic SiAlON (solid solutions based on the Si3N4 structure) inserts experience during a high-speed machining process and the evolution of sparks created during the same process. These sparks were analysed through pictures of the cutting process recorded using an SLR camera. Features relating to the intensity and area of the cutting sparks were extracted from the individual pictures using image processing techniques. These features were then related to the ceramic insert’s crater wear area.

Keywords: ceramic cutting tools, high speed machining, image processing, tool condition monitoring, tool wear

Procedia PDF Downloads 291
9588 Techno-Economic Analysis (TEA) of Circular Economy Approach in the Valorisation of Pig Meat Processing Wastes

Authors: Ribeiro A., Vilarinho C., Luisa A., Carvalho J

Abstract:

The pig meat industry generates large volumes of by- and co-products like blood, bones, skin, trimmings, organs, viscera, and skulls, among others, during slaughtering and meat processing and must be treated and disposed of ecologically. The yield of these by-products has been reported to account for about 10% to 15% of the value of the live animal in developed countries, although animal by-products account for about two-thirds of the animal after slaughter. It was selected for further valorization of the principal wastes produced throughout the value chain of pig meat production: Pig Manure, Pig Bones, Fats, Skins, Pig Hair, Wastewater, Wastewater sludges, and other animal subproducts type III. According to the potential valorization options, these wastes will be converted into Biomethane, Fertilizers (phosphorus and digestate), Hydroxyapatite, and protein hydrolysates (Keratin and Collagen). This work includes comprehensive technical and economic analyses (TEA) for each valorization route or applied technology. Metrics such as Net Present Value (NPV), Internal Rate of Return (IRR), and payback periods were used to evaluate economic feasibility. From this analysis, it can be concluded that, for Biogas Production, the scenarios using pig manure, wastewater sludges and mixed grass and leguminous wastes presented a remarkably high economic feasibility. Scenarios showed high economic feasibility with a positive payback period, NPV, and IRR. The optimal scenario combining pig manure with mixed grass and leguminous wastes had a payback period of 1.2 years and produced 427,6269 m³ of biomethane annually. Regarding the Chemical Extraction of Phosphorous and Nitrogen, results proved that the process is economically unviable due to negative cash flows despite high recovery rates. The TEA of Hydrolysis and Extraction of Keratin Hydrolysates indicate that a unit processing and valorizing 10 tons of pig hair per year for the production of keratin hydrolysate has an NPV of 907,940 €, an IRR of 13.07%, and a Payback period of 5.41 years. All of these indicators suggest a highly potential project to explore in the future. On the opposite, the results of Hydrolysis and Extraction of Collagen Hydrolysates showed a process economically unviable with negative cash flows in all scenarios due to the high-fat content in raw materials. In fact, the results from the valorization of 10 tons of pig skin had a negative cash flow of 453 743,88 €. TEA results of Extraction and purification of Hydroxyapatite from Pig Bones with Pyrolysis indicate that unit processing and valorizing 10 tons of pig bones per year for the production of hydroxyapatite has an NPV of 1 274 819,00 €, an IRR of 65.43%, and a Payback period of 1,5 years over a timeline of 10 years with a discount rate of 10%. These valorization routes, circular economy and bio-refinery approach offer significant contributions to sustainable bio-based operations within the agri-food industry. This approach transforms waste into valuable resources, enhancing both environmental and economic outcomes and contributing to a more sustainable and circular bioeconomy.

Keywords: techno-economic analysis (TEA), pig meat processing wastes, circular economy, bio-refinery

Procedia PDF Downloads 4
9587 Reduction of Residual Stress by Variothermal Processing and Validation via Birefringence Measurement Technique on Injection Molded Polycarbonate Samples

Authors: Christoph Lohr, Hanna Wund, Peter Elsner, Kay André Weidenmann

Abstract:

Injection molding is one of the most commonly used techniques in the industrial polymer processing. In the conventional process of injection molding, the liquid polymer is injected into the cavity of the mold, where the polymer directly starts hardening at the cooled walls. To compensate the shrinkage, which is caused predominantly by the immediate cooling, holding pressure is applied. Through that whole process, residual stresses are produced by the temperature difference of the polymer melt and the injection mold and the relocation of the polymer chains, which were oriented by the high process pressures and injection speeds. These residual stresses often weaken or change the structural behavior of the parts or lead to deformation of components. One solution to reduce the residual stresses is the use of variothermal processing. Hereby the mold is heated – i.e. near/over the glass transition temperature of the polymer – the polymer is injected and before opening the mold and ejecting the part the mold is cooled. For the next cycle, the mold gets heated again and the procedure repeats. The rapid heating and cooling of the mold are realized indirectly by convection of heated and cooled liquid (here: water) which is pumped through fluid channels underneath the mold surface. In this paper, the influences of variothermal processing on the residual stresses are analyzed with samples in a larger scale (500 mm x 250 mm x 4 mm). In addition, the influence on functional elements, such as abrupt changes in wall thickness, bosses, and ribs, on the residual stress is examined. Therefore the polycarbonate samples are produced by variothermal and isothermal processing. The melt is injected into a heated mold, which has in our case a temperature varying between 70 °C and 160 °C. After the filling of the cavity, the closed mold is cooled down varying from 70 °C to 100 °C. The pressure and temperature inside the mold are monitored and evaluated with cavity sensors. The residual stresses of the produced samples are illustrated by birefringence where the effect on the refractive index on the polymer under stress is used. The colorful spectrum can be uncovered by placing the sample between a polarized light source and a second polarization filter. To show the achievement and processing effects on the reduction of residual stress the birefringence images of the isothermal and variothermal produced samples are compared and evaluated. In this comparison to the variothermal produced samples have a lower amount of maxima of each color spectrum than the isothermal produced samples, which concludes that the residual stress of the variothermal produced samples is lower.

Keywords: birefringence, injection molding, polycarbonate, residual stress, variothermal processing

Procedia PDF Downloads 281
9586 Understanding the Heart of the Matter: A Pedagogical Framework for Apprehending Successful Second Language Development

Authors: Cinthya Olivares Garita

Abstract:

Untangling language processing in second language development has been either a taken-for-granted and overlooked task for some English language teaching (ELT) instructors or a considerable feat for others. From the most traditional language instruction to the most communicative methodologies, how to assist L2 learners in processing language in the classroom has become a challenging matter in second language teaching. Amidst an ample array of methods, strategies, and techniques to teach a target language, finding a suitable model to lead learners to process, interpret, and negotiate meaning to communicate in a second language has imposed a great responsibility on language teachers; committed teachers are those who are aware of their role in equipping learners with the appropriate tools to communicate in the target language in a 21stcentury society. Unfortunately, one might find some English language teachers convinced that their job is only to lecture students; others are advocates of textbook-based instruction that might hinder second language processing, and just a few might courageously struggle to facilitate second language learning effectively. Grounded on the most representative empirical studies on comprehensible input, processing instruction, and focus on form, this analysis aims to facilitate the understanding of how second language learners process and automatize input and propose a pedagogical framework for the successful development of a second language. In light of this, this paper is structured to tackle noticing and attention and structured input as the heart of processing instruction, comprehensible input as the missing link in second language learning, and form-meaning connections as opposed to traditional grammar approaches to language teaching. The author finishes by suggesting a pedagogical framework involving noticing-attention-comprehensible-input-form (NACIF based on their acronym) to support ELT instructors, teachers, and scholars on the challenging task of facilitating the understanding of effective second language development.

Keywords: second language development, pedagogical framework, noticing, attention, comprehensible input, form

Procedia PDF Downloads 19
9585 The Impact of Acoustic Performance on Neurodiverse Students in K-12 Learning Spaces

Authors: Michael Lekan-Kehinde, Abimbola Asojo, Bonnie Sanborn

Abstract:

Good acoustic performance has been identified as one of the critical Indoor Environmental Quality (IEQ) factors for student learning and development by the National Research Council. Childhood presents the opportunity for children to develop lifelong skills that will support them throughout their adult lives. Acoustic performance of a space has been identified as a factor that can impact language acquisition, concentration, information retention, and general comfort within the environment. Increasingly, students learn by communication between both teachers and fellow students, making speaking and listening crucial. Neurodiversity - while initially coined to describe individuals with autism spectrum disorder (ASD) - widely describes anyone with a different brain process. As the understanding from cognitive and neurosciences increases, the number of people identified as neurodiversity is nearly 30% of the population. This research looks at guidelines and standard for spaces with good acoustical quality and relates it with the experiences of students with autism spectrum disorder (ASD), their parents, teachers, and educators through a mixed methods approach, including selected case studies interviews, and mixed surveys. The information obtained from these sources is used to determine if selected materials, especially properties relating to sound absorption and reverberation reduction, are equally useful in small, medium sized, and large learning spaces and methodologically approaching. The results describe the potential impact of acoustics on Neurodiverse students, considering factors that determine the complexity of sound in relation to the auditory processing capabilities of ASD students. In conclusion, this research extends the knowledge of how materials selection influences the better development of acoustical environments for autism students.

Keywords: acoustics, autism spectrum disorder (ASD), children, education, learning, learning spaces, materials, neurodiversity, sound

Procedia PDF Downloads 102
9584 Studies on Mechanical Behavior of Kevlar/Kenaf/Graphene Reinforced Polymer Based Hybrid Composites

Authors: H. K. Shivanand, Ranjith R. Hombal, Paraveej Shirahatti, Gujjalla Anil Babu, S. ShivaPrakash

Abstract:

When it comes to the selection of materials the knowledge of materials science plays a vital role in selection and enhancements of materials properties. In the world of material science a composite material has the significant role based on its application. The composite materials are those in which two or more components having different physical and chemical properties are combined to create a new enhanced property substance. In this study three different materials (Kenaf, Kevlar and Graphene) been chosen based on their properties and a composite material is developed with help of vacuum bagging process. The fibers (Kenaf and Kevlar) and Resin(vinyl ester) ratio was maintained at 70:30 during the process and 0.5% 1% and 1.5% of Graphene was added during fabrication process. The material was machined to thedimension ofASTM standards(300×300mm and thickness 3mm)with help of water jet cutting machine. The composite materials were tested for Mechanical properties such as Interlaminar shear strength(ILSS) and Flexural strength. It is found that there is significant increase in material properties in the developed composite material.

Keywords: Kevlar, Kenaf, graphene, vacuum bagging process, Interlaminar shear strength test, flexural test

Procedia PDF Downloads 86
9583 Mechanical Properties of Poly(Propylene)-Based Graphene Nanocomposites

Authors: Luiza Melo De Lima, Tito Trindade, Jose M. Oliveira

Abstract:

The development of thermoplastic-based graphene nanocomposites has been of great interest not only to the scientific community but also to different industrial sectors. Due to the possible improvement of performance and weight reduction, thermoplastic nanocomposites are a great promise as a new class of materials. These nanocomposites are of relevance for the automotive industry, namely because the emission limits of CO2 emissions imposed by the European Commission (EC) regulations can be fulfilled without compromising the car’s performance but by reducing its weight. Thermoplastic polymers have some advantages over thermosetting polymers such as higher productivity, lower density, and recyclability. In the automotive industry, for example, poly(propylene) (PP) is a common thermoplastic polymer, which represents more than half of the polymeric raw material used in automotive parts. Graphene-based materials (GBM) are potential nanofillers that can improve the properties of polymer matrices at very low loading. In comparison to other composites, such as fiber-based composites, weight reduction can positively affect their processing and future applications. However, the properties and performance of GBM/polymer nanocomposites depend on the type of GBM and polymer matrix, the degree of dispersion, and especially the type of interactions between the fillers and the polymer matrix. In order to take advantage of the superior mechanical strength of GBM, strong interfacial strength between GBM and the polymer matrix is required for efficient stress transfer from GBM to the polymer. Thus, chemical compatibilizers and physicochemical modifications have been reported as important tools during the processing of these nanocomposites. In this study, PP-based nanocomposites were obtained by a simple melt blending technique, using a Brabender type mixer machine. Graphene nanoplatelets (GnPs) were applied as structural reinforcement. Two compatibilizers were used to improve the interaction between PP matrix and GnPs: PP graft maleic anhydride (PPgMA) and PPgMA modified with tertiary amine alcohol (PPgDM). The samples for tensile and Charpy impact tests were obtained by injection molding. The results suggested the GnPs presence can increase the mechanical strength of the polymer. However, it was verified that the GnPs presence can promote a decrease of impact resistance, turning the nanocomposites more fragile than neat PP. The compatibilizers’ incorporation increases the impact resistance, suggesting that the compatibilizers can enhance the adhesion between PP and GnPs. Compared to neat PP, Young’s modulus of non-compatibilized nanocomposite increase demonstrated that GnPs incorporation can promote a stiffness improvement of the polymer. This trend can be related to the several physical crosslinking points between the PP matrix and the GnPs. Furthermore, the decrease of strain at a yield of PP/GnPs, together with the enhancement of Young’s modulus, confirms that the GnPs incorporation led to an increase in stiffness but to a decrease in toughness. Moreover, the results demonstrated that incorporation of compatibilizers did not affect Young’s modulus and strain at yield results compared to non-compatibilized nanocomposite. The incorporation of these compatibilizers showed an improvement of nanocomposites’ mechanical properties compared both to those the non-compatibilized nanocomposite and to a PP sample used as reference.

Keywords: graphene nanoplatelets, mechanical properties, melt blending processing, poly(propylene)-based nanocomposites

Procedia PDF Downloads 183
9582 Image Processing of Scanning Electron Microscope Micrograph of Ferrite and Pearlite Steel for Recognition of Micro-Constituents

Authors: Subir Gupta, Subhas Ganguly

Abstract:

In this paper, we demonstrate the new area of application of image processing in metallurgical images to develop the more opportunity for structure-property correlation based approaches of alloy design. The present exercise focuses on the development of image processing tools suitable for phrase segmentation, grain boundary detection and recognition of micro-constituents in SEM micrographs of ferrite and pearlite steels. A comprehensive data of micrographs have been experimentally developed encompassing the variation of ferrite and pearlite volume fractions and taking images at different magnification (500X, 1000X, 15000X, 2000X, 3000X and 5000X) under scanning electron microscope. The variation in the volume fraction has been achieved using four different plain carbon steel containing 0.1, 0.22, 0.35 and 0.48 wt% C heat treated under annealing and normalizing treatments. The obtained data pool of micrographs arbitrarily divided into two parts to developing training and testing sets of micrographs. The statistical recognition features for ferrite and pearlite constituents have been developed by learning from training set of micrographs. The obtained features for microstructure pattern recognition are applied to test set of micrographs. The analysis of the result shows that the developed strategy can successfully detect the micro constitutes across the wide range of magnification and variation of volume fractions of the constituents in the structure with an accuracy of about +/- 5%.

Keywords: SEM micrograph, metallurgical image processing, ferrite pearlite steel, microstructure

Procedia PDF Downloads 195
9581 Development of Fake News Model Using Machine Learning through Natural Language Processing

Authors: Sajjad Ahmed, Knut Hinkelmann, Flavio Corradini

Abstract:

Fake news detection research is still in the early stage as this is a relatively new phenomenon in the interest raised by society. Machine learning helps to solve complex problems and to build AI systems nowadays and especially in those cases where we have tacit knowledge or the knowledge that is not known. We used machine learning algorithms and for identification of fake news; we applied three classifiers; Passive Aggressive, Naïve Bayes, and Support Vector Machine. Simple classification is not completely correct in fake news detection because classification methods are not specialized for fake news. With the integration of machine learning and text-based processing, we can detect fake news and build classifiers that can classify the news data. Text classification mainly focuses on extracting various features of text and after that incorporating those features into classification. The big challenge in this area is the lack of an efficient way to differentiate between fake and non-fake due to the unavailability of corpora. We applied three different machine learning classifiers on two publicly available datasets. Experimental analysis based on the existing dataset indicates a very encouraging and improved performance.

Keywords: fake news detection, natural language processing, machine learning, classification techniques.

Procedia PDF Downloads 162
9580 Induction Machine Bearing Failure Detection Using Advanced Signal Processing Methods

Authors: Abdelghani Chahmi

Abstract:

This article examines the detection and localization of faults in electrical systems, particularly those using asynchronous machines. First, the process of failure will be characterized, relevant symptoms will be defined and based on those processes and symptoms, a model of those malfunctions will be obtained. Second, the development of the diagnosis of the machine will be shown. As studies of malfunctions in electrical systems could only rely on a small amount of experimental data, it has been essential to provide ourselves with simulation tools which allowed us to characterize the faulty behavior. Fault detection uses signal processing techniques in known operating phases.

Keywords: induction motor, modeling, bearing damage, airgap eccentricity, torque variation

Procedia PDF Downloads 137
9579 Normalized P-Laplacian: From Stochastic Game to Image Processing

Authors: Abderrahim Elmoataz

Abstract:

More and more contemporary applications involve data in the form of functions defined on irregular and topologically complicated domains (images, meshs, points clouds, networks, etc). Such data are not organized as familiar digital signals and images sampled on regular lattices. However, they can be conveniently represented as graphs where each vertex represents measured data and each edge represents a relationship (connectivity or certain affinities or interaction) between two vertices. Processing and analyzing these types of data is a major challenge for both image and machine learning communities. Hence, it is very important to transfer to graphs and networks many of the mathematical tools which were initially developed on usual Euclidean spaces and proven to be efficient for many inverse problems and applications dealing with usual image and signal domains. Historically, the main tools for the study of graphs or networks come from combinatorial and graph theory. In recent years there has been an increasing interest in the investigation of one of the major mathematical tools for signal and image analysis, which are Partial Differential Equations (PDEs) variational methods on graphs. The normalized p-laplacian operator has been recently introduced to model a stochastic game called tug-of-war-game with noise. Part interest of this class of operators arises from the fact that it includes, as particular case, the infinity Laplacian, the mean curvature operator and the traditionnal Laplacian operators which was extensiveley used to models and to solve problems in image processing. The purpose of this paper is to introduce and to study a new class of normalized p-Laplacian on graphs. The introduction is based on the extension of p-harmonious function introduced in as discrete approximation for both infinity Laplacian and p-Laplacian equations. Finally, we propose to use these operators as a framework for solving many inverse problems in image processing.

Keywords: normalized p-laplacian, image processing, stochastic game, inverse problems

Procedia PDF Downloads 509
9578 A Quantitative Structure-Adsorption Study on Novel and Emerging Adsorbent Materials

Authors: Marc Sader, Michiel Stock, Bernard De Baets

Abstract:

Considering a large amount of adsorption data of adsorbate gases on adsorbent materials in literature, it is interesting to predict such adsorption data without experimentation. A quantitative structure-activity relationship (QSAR) is developed to correlate molecular characteristics of gases and existing knowledge of materials with their respective adsorption properties. The application of Random Forest, a machine learning method, on a set of adsorption isotherms at a wide range of partial pressures and concentrations is studied. The predicted adsorption isotherms are fitted to several adsorption equations to estimate the adsorption properties. To impute the adsorption properties of desired gases on desired materials, leave-one-out cross-validation is employed. Extensive experimental results for a range of settings are reported.

Keywords: adsorption, predictive modeling, QSAR, random forest

Procedia PDF Downloads 223
9577 Sustainability in Space: Implementation of Circular Economy and Material Efficiency Strategies in Space Missions

Authors: Hamda M. Al-Ali

Abstract:

The ultimate aim of space exploration has been centralized around the possibility of life on other planets in the solar system. This aim is driven by the detrimental effects that climate change could potentially have on human survival on Earth in the future. This drives humans to search for feasible solutions to increase environmental and economical sustainability on Earth and to evaluate and explore the ability of human survival on other planets such as Mars. To do that, frequent space missions are required to meet the ambitious human goals. This means that reliable and affordable access to space is required, which could be largely achieved through the use of reusable spacecrafts. Therefore, materials and resources must be used wisely to meet the increasing demand. Space missions are currently extremely expensive to operate. However, reusing materials hence spacecrafts, can potentially reduce overall mission costs as well as the negative impact on both space and Earth environments. This is because reusing materials leads to less waste generated per mission, and therefore fewer landfill sites are required. Reusing materials reduces resource consumption, material production, and the need for processing new and replacement spacecraft and launch vehicle parts. Consequently, this will ease and facilitate human access to outer space as it will reduce the demand for scarce resources, which will boost material efficiency in the space industry. Material efficiency expresses the extent to which resources are consumed in the production cycle and how the waste produced by the industrial process is minimized. The strategies proposed in this paper to boost material efficiency in the space sector are the introduction of key performance indicators that are able to measure material efficiency as well as the introduction of clearly defined policies and legislation that can be easily implemented within the general practices in the space industry. Another strategy to improve material efficiency is by amplifying energy and resource efficiency through reusing materials. The circularity of various spacecraft materials such as Kevlar, steel, and aluminum alloys could be maximized through reusing them directly or after galvanizing them with another layer of material to act as a protective coat. This research paper has an aim to investigate and discuss how to improve material efficiency in space missions considering circular economy concepts so that space and Earth become more economically and environmentally sustainable. The circular economy is a transition from a make-use-waste linear model to a closed-loop socio-economic model, which is regenerative and restorative in nature. The implementation of a circular economy will reduce waste and pollution through maximizing material efficiency, ensuring that businesses can thrive and sustain. Further research into the extent to which reusable launch vehicles reduce space mission costs have been discussed, along with the environmental and economic implications it could have on the space sector and the environment. This has been examined through research and in-depth literature review of published reports, books, scientific articles, and journals. Keywords such as material efficiency, circular economy, reusable launch vehicles and spacecraft materials were used to search for relevant literature.

Keywords: circular economy, key performance indicator, material efficiency, reusable launch vehicles, spacecraft materials

Procedia PDF Downloads 122
9576 General Purpose Graphic Processing Units Based Real Time Video Tracking System

Authors: Mallikarjuna Rao Gundavarapu, Ch. Mallikarjuna Rao, K. Anuradha Bai

Abstract:

Real Time Video Tracking is a challenging task for computing professionals. The performance of video tracking techniques is greatly affected by background detection and elimination process. Local regions of the image frame contain vital information of background and foreground. However, pixel-level processing of local regions consumes a good amount of computational time and memory space by traditional approaches. In our approach we have explored the concurrent computational ability of General Purpose Graphic Processing Units (GPGPU) to address this problem. The Gaussian Mixture Model (GMM) with adaptive weighted kernels is used for detecting the background. The weights of the kernel are influenced by local regions and are updated by inter-frame variations of these corresponding regions. The proposed system has been tested with GPU devices such as GeForce GTX 280, GeForce GTX 280 and Quadro K2000. The results are encouraging with maximum speed up 10X compared to sequential approach.

Keywords: connected components, embrace threads, local weighted kernel, structuring elements

Procedia PDF Downloads 432
9575 Parallel Processing in near Absence of Attention: A Study Using Dual-Task Paradigm

Authors: Aarushi Agarwal, Tara Singh, I.L Singh, Anju Lata Singh, Trayambak Tiwari

Abstract:

Simple discrimination in near absence of attention has been widely observed. Dual-task studies with natural scenes studies have been claimed as being preattentive in nature that facilitated categorization simultaneously with the attentional demanding task. So in this study, multiple images at the periphery are presented, initiating parallel processing in near absence of attention. For the central demanding task rotated letters were presented in both conditions, while in periphery natural and animal images were presented. To understand the breakpoint of ability to perform in near absence of attention one, two and three peripheral images were presented simultaneously with central task and subjects had to respond when all belong to the same category. Individual participant performance did not show a significant difference in both conditions central and peripheral task when the single peripheral image was shown. In case of two images high-level parallel processing could take place with little attentional resources. The eye tracking results supports the evidence as no major saccade was made in a large number of trials. Three image presentations proved to be a breaking point of the capacities to perform outside attentional assistance as participants showed a confused eye gaze pattern which failed to make the natural and animal image discriminations. Thus, we can conclude attention and awareness being independent mechanisms having limited capacities.

Keywords: attention, dual task pardigm, parallel processing, break point, saccade

Procedia PDF Downloads 214
9574 Short-Term Effects of an Open Monitoring Meditation on Cognitive Control and Information Processing

Authors: Sarah Ullrich, Juliane Rolle, Christian Beste, Nicole Wolff

Abstract:

Inhibition and cognitive flexibility are essential parts of executive functions in our daily lives, as they enable the avoidance of unwanted responses or selectively switch between mental processes to generate appropriate behavior. There is growing interest in improving inhibition and response selection through brief mindfulness-based meditations. Arguably, open-monitoring meditation (OMM) improves inhibitory and flexibility performance by optimizing cognitive control and information processing. Yet, the underlying neurophysiological processes have been poorly studied. Using the Simon-Go/Nogo paradigm, the present work examined the effect of a single 15-minute smartphone app-based OMM on inhibitory performance and response selection in meditation novices. We used both behavioral and neurophysiological measures (event-related potentials, ERPs) to investigate which subprocesses of response selection and inhibition are altered after OMM. The study was conducted in a randomized crossover design with N = 32 healthy adults. We thereby investigated Go and Nogo trials in the paradigm. The results show that as little as 15 minutes of OMM can improve response selection and inhibition at behavioral and neurophysiological levels. More specifically, OMM reduces the rate of false alarms, especially during Nogo trials regardless of congruency. It appears that OMM optimizes conflict processing and response inhibition compared to no meditation, also reflected in the ERP N2 and P3 time windows. The results may be explained by the meta control model, which argues in terms of a specific processing mode with increased flexibility and inclusive decision-making under OMM. Importantly, however, the effects of OMM were only evident when there was the prior experience with the task. It is likely that OMM provides more cognitive resources, as the amplitudes of these EKPs decreased. OMM novices seem to induce finer adjustments during conflict processing after familiarization with the task.

Keywords: EEG, inhibition, meditation, Simon Nogo

Procedia PDF Downloads 202
9573 Ranking of Optimal Materials for Building Walls from the Perspective of Cost and Waste of Electricity and Gas Energy Using AHP-TOPSIS 1 Technique: Study Example: Sari City

Authors: Seyedomid Fatemi

Abstract:

The walls of the building, as the main intermediary between the outside and the inside of the building, play an important role in controlling the environmental conditions and ensuring the comfort of the residents, thus reducing the heating and cooling loads. Therefore, the use of suitable materials is considered one of the simplest and most effective ways to reduce the heating and cooling loads of the building, which will also save energy. Therefore, in order to achieve the goal of the research "Ranking of optimal materials for building walls," optimal materials for building walls in a temperate and humid climate (case example: Sari city) from the perspective of embodied energy, waste of electricity and gas energy, cost and reuse been investigated to achieve sustainable architecture. In this regard, using information obtained from Sari Municipality, design components have been presented by experts using the Delphi method. Considering the criteria of experts' opinions (cost and reuse), the amount of embodied energy of the materials, as well as the amount of waste of electricity and gas of different materials of the walls, with the help of the AHP weighting technique and finally with the TOPSIS technique, the best type of materials in the order of 1- 3-D Panel 2-ICF-, 3-Cement block with pumice, 4-Wallcrete block, 5-Clay block, 6-Autoclaved Aerated Concrete (AAC), 7-Foam cement block, 8-Aquapanel and 9-Reinforced concrete wall for use in The walls of the buildings were proposed in Sari city.

Keywords: optimum materials, building walls, moderate and humid climate, sustainable architecture, AHP-TOPSIS technique

Procedia PDF Downloads 72
9572 Use of Polymeric Materials in the Architectural Preservation

Authors: F. Z. Benabid, F. Zouai, A. Douibi, D. Benachour

Abstract:

These Fluorinated polymers and polyacrylics have known a wide use in the field of historical monuments. PVDF provides a great easiness to processing, a good UV resistance and good chemical inertia. Although the quality of physical characteristics of the PMMA and its low price with a respect to PVDF, its deterioration against UV radiations limits its use as protector agent for the stones. On the other hand, PVDF/PMMA blend is a compromise of a great development in the field of architectural restoration, since it is the best method in term of quality and price to make new polymeric materials having enhanced properties. Films of different compositions based on the two polymers within an adequate solvent (DMF) were obtained to perform an exposition to artificial ageing and to the salted fog, a spectroscopic analysis (FTIR and UV) and optical analysis (refractive index). Based on its great interest in the field of building, a variety of standard tests has been elaborated for the first time at the central laboratory of ENAP (Souk-Ahras) in order to evaluate our blend performance. The obtained results have allowed observing the behavior of the different compositions of the blend under various tests. The addition of PVDF to PMMA enhances the properties of this last to know the exhibition to the natural and artificial ageing and to the saline fog. On the other hand, PMMA enhances the optical properties of the blend. Finally, 70/30 composition of the blend is in concordance with results of previous works and it is the adequate proportion for an eventual application.

Keywords: blend, PVDF, PMMA, preservation, historic monuments

Procedia PDF Downloads 304
9571 Increasing Added-Value of Salak Fruit by Freezing Frying to Improve the Welfare of Farmers: Case Study of Sleman Regency, Yogyakarta-Indonesia

Authors: Sucihatiningsih Dian Wisika Prajanti, Himawan Arif Susanto

Abstract:

Fruits are perishable products and have relatively low price, especially at harvest time. Generally, farmers only sell the products shortly after the harvest time without any processing. Farmers also only play role as price takers leading them to have less power to set the price. Sometimes, farmers are manipulated by middlemen, especially during abundant harvest. Therefore, it requires an effort to cultivate fruits and create innovation to make them more durable and have higher economic value. The purpose of this research is how to increase the added- value of fruits that have high economic value. The research involved 60 farmers of Salak fruit as the sample. Then, descriptive analysis was used to analyze the data in this study. The results showed the selling price of Salak fruit is very low. Hence, to increase the added-value of the fruits, fruit processing is carried out by freezing - frying which can cause the fruits last longer. In addition to increase these added-value, the products can be accommodated for further processed without worrying about their crops rotted or unsold.

Keywords: fruits processing, Salak fruit, freezing frying, farmer’s welfare, Sleman, Yogyakarta

Procedia PDF Downloads 346
9570 KCBA, A Method for Feature Extraction of Colonoscopy Images

Authors: Vahid Bayrami Rad

Abstract:

In recent years, the use of artificial intelligence techniques, tools, and methods in processing medical images and health-related applications has been highlighted and a lot of research has been done in this regard. For example, colonoscopy and diagnosis of colon lesions are some cases in which the process of diagnosis of lesions can be improved by using image processing and artificial intelligence algorithms, which help doctors a lot. Due to the lack of accurate measurements and the variety of injuries in colonoscopy images, the process of diagnosing the type of lesions is a little difficult even for expert doctors. Therefore, by using different software and image processing, doctors can be helped to increase the accuracy of their observations and ultimately improve their diagnosis. Also, by using automatic methods, the process of diagnosing the type of disease can be improved. Therefore, in this paper, a deep learning framework called KCBA is proposed to classify colonoscopy lesions which are composed of several methods such as K-means clustering, a bag of features and deep auto-encoder. Finally, according to the experimental results, the proposed method's performance in classifying colonoscopy images is depicted considering the accuracy criterion.

Keywords: colorectal cancer, colonoscopy, region of interest, narrow band imaging, texture analysis, bag of feature

Procedia PDF Downloads 49
9569 A Visual Inspection System for Automotive Sheet Metal Chasis Parts Produced with Cold-Forming Method

Authors: İmren Öztürk Yılmaz, Abdullah Yasin Bilici, Yasin Atalay Candemir

Abstract:

The system consists of 4 main elements: motion system, image acquisition system, image processing software, and control interface. The parts coming out of the production line to enter the image processing system with the conveyor belt at the end of the line. The 3D scanning of the produced part is performed with the laser scanning system integrated into the system entry side. With the 3D scanning method, it is determined at what position and angle the parts enter the system, and according to the data obtained, parameters such as part origin and conveyor speed are calculated with the designed software, and the robot is informed about the position where it will take part. The robot, which receives the information, takes the produced part on the belt conveyor and shows it to high-resolution cameras for quality control. Measurement processes are carried out with a maximum error of 20 microns determined by the experiments.

Keywords: quality control, industry 4.0, image processing, automated fault detection, digital visual inspection

Procedia PDF Downloads 110
9568 Development of the Food Market of the Republic of Kazakhstan in the Field of Milk Processing

Authors: Gulmira Zhakupova, Tamara Tultabayeva, Aknur Muldasheva, Assem Sagandyk

Abstract:

The development of technology and production of products with increased biological value based on the use of natural food raw materials are important tasks in the policy of the food market of the Republic of Kazakhstan. For Kazakhstan, livestock farming, in particular sheep farming, is the most ancient and developed industry and way of life. The history of the Kazakh people is largely connected with this type of agricultural production, with established traditions using dairy products from sheep's milk. Therefore, the development of new technologies from sheep’s milk remains relevant. In addition, one of the most promising areas for the development of food technology for therapeutic and prophylactic purposes is sheep milk products as a source of protein, immunoglobulins, minerals, vitamins, and other biologically active compounds. This article presents the results of research on the study of milk processing technology. The objective of the study is to study the possibilities of processing sheep milk and its role in human nutrition, as well as the results of research to improve the technology of sheep milk products. The studies were carried out on the basis of sanitary and hygienic requirements for dairy products in accordance with the following test methods. To perform microbiological analysis, we used the method for identifying Salmonella bacteria (Horizontal method for identifying, counting, and serotyping Salmonella) in a certain mass or volume of product. Nutritional value is a complex of properties of food products that meet human physiological needs for energy and basic nutrients. The protein mass fraction was determined by the Kjeldahl method. This method is based on the mineralization of a milk sample with concentrated sulfuric acid in the presence of an oxidizing agent, an inert salt - potassium sulfate, and a catalyst - copper sulfate. In this case, the amino groups of the protein are converted into ammonium sulfate dissolved in sulfuric acid. The vitamin composition was determined by HPLC. To determine the content of mineral substances in the studied samples, the method of atomic absorption spectrophotometry was used. The study identified the technological parameters of sheep milk products and determined the prospects for researching sheep milk products. Microbiological studies were used to determine the safety of the study product. According to the results of the microbiological analysis, no deviations from the norm were identified. This means high safety of the products under study. In terms of nutritional value, the resulting products are high in protein. Data on the positive content of amino acids were also obtained. The results obtained will be used in the food industry and will serve as recommendations for manufacturers.

Keywords: dairy, milk processing, nutrition, colostrum

Procedia PDF Downloads 55
9567 Testing the Impact of Formal Interpreting Training on Working Memory Capacity: Evidence from Turkish-English Student-Interpreters

Authors: Elena Antonova Unlu, Cigdem Sagin Simsek

Abstract:

The research presents two studies examining the impact of formal interpreting training (FIT) on Working Memory Capacity (WMC) of student-interpreters. In Study 1, the storage and processing capacities of the working memory (WM) of last-year student-interpreters were compared with those of last-year Foreign Language Education (FLE) students. In Study 2, the impact of FIT on the WMC of student-interpreters was examined via comparing their results on WM tasks at the beginning and the end of their FIT. In both studies, Digit Span Task (DST) and Reading Span Task (RST) were utilized for testing storage and processing capacities of WM. The results of Study 1 revealed that the last-year student-interpreters outperformed the control groups on the RST but not on the DST. The findings of Study 2 were consistent with Study 1 showing that after FIT, the student-interpreters performed better on the RST but not on the DST. Our findings can be considered as evidence supporting the view that FIT has a beneficial effect not only on the interpreting skills of student-interpreters but also on the central executive and processing capacity of their WM.

Keywords: working memory capacity, formal interpreting training, student-interpreters, cross-sectional and longitudinal data

Procedia PDF Downloads 203