Search results for: thermomechanical processing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3754

Search results for: thermomechanical processing

2674 Evaluation of the Suitability of a Microcapsule-Based System for the Manufacturing of Self-Healing Low-Density Polyethylene

Authors: Małgorzata Golonka, Jadwiga Laska

Abstract:

Among self-healing materials, the most unexplored group are thermoplastic polymers. These polymers are used not only to produce packaging with a relatively short life but also to obtain coatings, insulation, casings, or parts of machines and devices. Due to its exceptional resistance to weather conditions, hydrophobicity, sufficient mechanical strength, and ease of extrusion, polyethylene is used in the production of polymer pipelines and as an insulating layer for steel pipelines. Polyethylene or PE coated steel pipelines can be used in difficult conditions such as underground or underwater installations. Both installation and use under such conditions are associated with high stresses and consequently the formation of microdamages in the structure of the material, loss of its integrity and final applicability. The ideal solution would be to include a self-healing system in the polymer material. In the presented study the behavior of resin-coated microcapsules in the extrusion process of low-density polyethylene was examined. Microcapsules are a convenient element of the repair system because they can be filled with appropriate reactive substances to ensure the repair process, but the main problem is their durability under processing conditions. Rapeseed oil, which has a relatively high boiling point of 240⁰C and low volatility, was used as the core material that simulates the reactive agents. The capsule shell, which is a key element responsible for its mechanical strength, was obtained by in situ polymerising urea-formaldehyde, melamine-urea-formaldehyde or melamine-formaldehyde resin on the surface of oil droplets dispersed in water. The strength of the capsules was compared based on the shell material, and in addition, microcapsules with single- and multilayer shells were obtained using different combinations of the chemical composition of the resins. For example, the first layer of appropriate tightness and stiffness was made of melamine-urea-formaldehyde resin, and the second layer was a melamine-formaldehyde reinforcing layer. The size, shape, distribution of capsule diameters and shell thickness were determined using digital optical microscopy and electron microscopy. The efficiency of encapsulation (i.e., the presence of rapeseed oil as the core) and the tightness of the shell were determined by FTIR spectroscopic examination. The mechanical strength and distribution of microcapsules in polyethylene were tested by extruding samples of crushed low-density polyethylene mixed with microcapsules in a ratio of 1 and 2.5% by weight. The extrusion process was carried out in a mini extruder at a temperature of 150⁰C. The capsules obtained had a diameter range of 70-200 µm. FTIR analysis confirmed the presence of rapeseed oil in both single- and multilayer shell microcapsules. Microscopic observations of cross sections of the extrudates confirmed the presence of both intact and cracked microcapsules. However, the melamine-formaldehyde resin shells showed higher processing strength compared to that of the melamine-urea-formaldehyde coating and the urea-formaldehyde coating. Capsules with a urea-formaldehyde shell work very well in resin coating systems and cement composites, i.e., in pressureless processing and moulding conditions. The addition of another layer of melamine-formaldehyde coating to both the melamine-urea-formaldehyde and melamine-formaldehyde resin layers significantly increased the number of microcapsules undamaged during the extrusion process. The properties of multilayer coatings were also determined and compared with each other using computer modelling.

Keywords: self-healing polymers, polyethylene, microcapsules, extrusion

Procedia PDF Downloads 31
2673 Cross Analysis of Gender Discrimination in Print Media of Subcontinent via James Paul Gee Model

Authors: Luqman Shah

Abstract:

The myopic gender discrimination is now a well-documented and recognized fact. However, gender is only one facet of an individual’s multiple identities. The aim of this work is to investigate gender discrimination highlighted in print media in the subcontinent with a specific focus on Pakistan and India. In this study, an approach is adopted by using the James Paul Gee model for the identification of gender discrimination. As a matter of fact, gender discrimination is not consistent in its nature and intensity across global societies and varies as social, geographical, and cultural background change. The World has been changed enormously in every aspect of life, and there are also obvious changes towards gender discrimination, prejudices, and biases, but still, the world has a long way to go to recognize women as equal as men in every sphere of life. The history of the world is full of gender-based incidents and violence. Now the time came that this issue must be seriously addressed and to eradicate this evil, which will lead to harmonize society and consequently heading towards peace and prosperity. The study was carried out by a mixed model research method. The data was extracted from the contents of five Pakistani English newspapers out of a total of 23 daily English newspapers, and likewise, five Indian daily English newspapers out of 52 those were published 2018-2019. Two news stories from each of these newspapers, in total, twenty news stories were taken as sampling for this research. Content and semiotic analysis techniques were used to analyze through James Paul Gee's seven building tasks of language. The resources of renowned e-papers are utilized, and the highlighted cases in Pakistani newspapers of Indian gender-based stories and vice versa are scrutinized as per the requirement of this research paper. For analysis of the written stretches of discourse taken from e-papers and processing of data for the focused problem, James Paul Gee 'Seven Building Tasks of Language' is used. Tabulation of findings is carried to pinpoint the issue with certainty. Findings after processing the data showed that there is a gross human rights violation on the basis of gender discrimination. The print media needs a more realistic representation of what is what not what seems to be. The study recommends the equality and parity of genders.

Keywords: gender discrimination, print media, Paul Gee model, subcontinent

Procedia PDF Downloads 221
2672 Arabic Handwriting Recognition Using Local Approach

Authors: Mohammed Arif, Abdessalam Kifouche

Abstract:

Optical character recognition (OCR) has a main role in the present time. It's capable to solve many serious problems and simplify human activities. The OCR yields to 70's, since many solutions has been proposed, but unfortunately, it was supportive to nothing but Latin languages. This work proposes a system of recognition of an off-line Arabic handwriting. This system is based on a structural segmentation method and uses support vector machines (SVM) in the classification phase. We have presented a state of art of the characters segmentation methods, after that a view of the OCR area, also we will address the normalization problems we went through. After a comparison between the Arabic handwritten characters & the segmentation methods, we had introduced a contribution through a segmentation algorithm.

Keywords: OCR, segmentation, Arabic characters, PAW, post-processing, SVM

Procedia PDF Downloads 74
2671 Polyphenol Stability and Antioxidant Properties of Freeze-Dried Sour Cherry Encapsulates

Authors: Gordana Ćetković, Vesna Tumbas Šaponjac, Jasna Čanadanović-Brunet, Sonja Đilas, Slađana Stajčić, Jelena Vulić, Mirjana Jakišić

Abstract:

Despite the recommended amount of daily intake of fruits, the consumption in modern age remains very low. Therefore there is a need for delivering valuable phytochemicals into the human body through different foods by developing functional food products fortified with natural bioactive compounds from plant sources. Recently, a growing interest rises in exploiting the fruit and vegetable by-products as sources of phytochemicals such as polyphenols, carotenoids, vitamins etc. Cherry contain high amounts of polyphenols, which are known to display a wide range of biological activities like antioxidant, anti-inflammatory, antimicrobial or anti-carcinogenic activities, improvement of vision, induction of apoptosis and neuroprotective effects. Also, cherry pomace, a by-product in juice processing, can also be promising source of phenolic compounds. However, the application of polyphenols as food additives is limited because after extraction these compounds are susceptible to degradation. Microencapsulation is one of the alternative approaches to protect bioactive compounds from degradation during processing and storage. Freeze-drying is one of the most used microencapsulation methods for the protection of thermosensitive and unstable molecules. In this study sour cherry pomace was extracted with food-grade solvent (50% ethanol) to be suitable for application in products for human use. Extracted polyphenols have been concentrated and stabilized on whey (WP) and soy (SP) proteins. Encapsulation efficiency in SP was higher (94.90%), however not significantly (p<0.05) from the one in WP (90.10%). Storage properties of WP and SP encapsulate in terms of total polyphenols, anthocyanins and antioxidant activity was tested for 6 weeks. It was found that the retention of polyphenols after 6 weeks in WP and SP (67.33 and 69.30%, respectively) was similar. The content of anthocyanins has increased in WP (for 47.97%), while their content in SP has very slightly decreased (for 1.45%) after 6-week storage period. In accordance with anthocyanins the decrease in antioxidant activity in WP (87.78%) was higher than in SP (43.02%). According to the results obtained in this study, the technique reported herewith can be used for obtaining quality encapsulates for their further use as functional food additives, and, on the other hand, for fruit waste valorization.

Keywords: cherry pomace, microencapsulation, polyphenols, storage

Procedia PDF Downloads 368
2670 Exploration Study of Civet Coffee: Amino Acids Composition and Cup Quality

Authors: Murna Muzaifa, Dian Hasni, Febriani, Anshar Patria, Amhar Abubakar

Abstract:

Coffee flavour is influenced by many factors such as processing techniques. Civet coffee is known as one of premium coffee due to its unique processing technique and its superior cupping quality. The desirable aroma of coffee is foremost formed during roasting step at a high temperature from precursors that are present in the green bean. Sugars, proteins, acids and trigonelline are the principal flavor precursors compounds in green coffee bean. It is now widely accepted that amino acids act as precursors of the Maillard reaction during which the colour and aroma are formed. To investigate amino acids on civet coffee, concentration of 20 amino acids (L-Isoleucine, L-Valine, L-Proline, L-Phenylalanine, L-Arginine, L-Asparagine, L-Threonine, L-Tryptophan, L-Leucine, L-Serine, L-Glutamine, L-Methionine, L-Histidine, Aspartic acid, L-Tyrosine, L-Lysine, L-Glutamic acid, and L-Cysteine, L-Alanine and Glycine) were determined in green and roasted bean of civet coffee by LCMS analysis. The cup quality of civet coffee performed using professional Q-grader followed SCAA standard method. The measured parameters were fragrance/aroma, flavor, acidity, body, uniformity, clean up, aftertaste, balance, sweetness and overall. The work has been done by collecting samples of civet coffee from six locations in Gayo Higland, Aceh-Indonesia. The results showed that 18 amino acids were detected in green bean of civet coffee (L-Isoleucine, L-Valine, L-Proline, L-Phenylalanine, L-Arginine, L-Asparagine, L-Threonine, L-Tryptophan, L-Leucine, L-Serine, L-Glutamine, L-Methionine, L-Histidine, Aspartic acid, L-Tyrosine, L-Lysine, L-Glutamic acid, and L-Cysteine) and 2 amino acids were not detected (L-Alanine and Glycine). On the other hand, L-Tyrosine and Glycine were not detected in roasted been of civet coffee. Glutamic acid is the amino acid with highest concentration in both green and roasted bean (21,02 mg/g and 24,60 mg/g), followed by L- Valine (19,98 mg/g and 20,22 mg/g) and Aspartic acid (14,93 mg/g and 18,58 mg/g). Civet coffee has a fairly high cupping value (cup quality), ranging from 83.75 to 84.75, categorized as speciality coffee. Moreover, civet coffee noted to have nutty, chocolaty, fishy, herby and watery.

Keywords: amino acids, civet coffee, cupping quality, luwak

Procedia PDF Downloads 189
2669 Flexural Properties of Carbon/Polypropylene Composites: Influence of Matrix Forming Polypropylene in Fiber, Powder, and Film States

Authors: Vijay Goud, Ramasamy Alagirusamy, Apurba Das, Dinesh Kalyanasundaram

Abstract:

Thermoplastic composites render new opportunities as effective processing technology while crafting newer complications into processing. One of the notable challenges is in achieving thorough wettability that is significantly deterred by the high viscosity of the long molecular chains of the thermoplastics. As a result of high viscosity, it is very difficult to impregnate the resin into a tightly interlaced textile structure to fill the voids present in the structure. One potential solution to the above problem, is to pre-deposit resin on the fiber, prior to consolidation. The current study compares DREF spinning, powder coating and film stacking methods of predeposition of resin onto fibers. An investigation into the flexural properties of unidirectional composites (UDC) produced from blending of carbon fiber and polypropylene (PP) matrix in varying forms of fiber, powder and film are reported. Dr. Ernst Fehrer (DREF) yarns or friction spun hybrid yarns were manufactured from PP fibers and carbon tows. The DREF yarns were consolidated to yield unidirectional composites (UDCs) referred to as UDC-D. PP in the form of powder was coated on carbon tows by electrostatic spray coating. The powder-coated towpregs were consolidated to form UDC-P. For the sake of comparison, a third UDC referred as UDC-F was manufactured by the consolidation of PP films stacked between carbon tows. The experiments were designed to yield a matching fiber volume fraction of about 50 % in all the three UDCs. A comparison of mechanical properties of the three composites was studied to understand the efficiency of matrix wetting and impregnation. Approximately 19% and 68% higher flexural strength were obtained for UDC-P than UDC-D and UDC-F respectively. Similarly, 25% and 81% higher modulus were observed in UDC-P than UDC-D and UDC-F respectively. Results from micro-computed tomography, scanning electron microscopy, and short beam tests indicate better impregnation of PP matrix in UDC-P obtained through electrostatic spray coating process and thereby higher flexural strength and modulus.

Keywords: DREF spinning, film stacking, flexural strength, powder coating, thermoplastic composite

Procedia PDF Downloads 223
2668 Arabic Quran Search Tool Based on Ontology

Authors: Mohammad Alqahtani, Eric Atwell

Abstract:

This paper reviews and classifies most of the important types of search techniques that have been applied on the holy Quran. Then, it addresses the limitations in these techniques. Additionally, this paper surveys most existing Quranic ontologies and what are their deficiencies. Finally, it explains a new search tool called: A semantic search tool for Al Quran based on Qur’anic ontologies. This tool will overcome all limitations in the existing Quranic search applications.

Keywords: holy Quran, natural language processing (NLP), semantic search, information retrieval (IR), ontology

Procedia PDF Downloads 572
2667 A Word-to-Vector Formulation for Word Representation

Authors: Sandra Rizkallah, Amir F. Atiya

Abstract:

This work presents a novel word to vector representation that is based on embedding the words into a sphere, whereby the dot product of the corresponding vectors represents the similarity between any two words. Embedding the vectors into a sphere enabled us to take into consideration the antonymity between words, not only the synonymity, because of the suitability to handle the polarity nature of words. For example, a word and its antonym can be represented as a vector and its negative. Moreover, we have managed to extract an adequate vocabulary. The obtained results show that the proposed approach can capture the essence of the language, and can be generalized to estimate a correct similarity of any new pair of words.

Keywords: natural language processing, word to vector, text similarity, text mining

Procedia PDF Downloads 276
2666 Cloud Design for Storing Large Amount of Data

Authors: M. Strémy, P. Závacký, P. Cuninka, M. Juhás

Abstract:

Main goal of this paper is to introduce our design of private cloud for storing large amount of data, especially pictures, and to provide good technological backend for data analysis based on parallel processing and business intelligence. We have tested hypervisors, cloud management tools, storage for storing all data and Hadoop to provide data analysis on unstructured data. Providing high availability, virtual network management, logical separation of projects and also rapid deployment of physical servers to our environment was also needed.

Keywords: cloud, glusterfs, hadoop, juju, kvm, maas, openstack, virtualization

Procedia PDF Downloads 353
2665 Diffusion MRI: Clinical Application in Radiotherapy Planning of Intracranial Pathology

Authors: Pomozova Kseniia, Gorlachev Gennadiy, Chernyaev Aleksandr, Golanov Andrey

Abstract:

In clinical practice, and especially in stereotactic radiosurgery planning, the significance of diffusion-weighted imaging (DWI) is growing. This makes the existence of software capable of quickly processing and reliably visualizing diffusion data, as well as equipped with tools for their analysis in terms of different tasks. We are developing the «MRDiffusionImaging» software on the standard C++ language. The subject part has been moved to separate class libraries and can be used on various platforms. The user interface is Windows WPF (Windows Presentation Foundation), which is a technology for managing Windows applications with access to all components of the .NET 5 or .NET Framework platform ecosystem. One of the important features is the use of a declarative markup language, XAML (eXtensible Application Markup Language), with which you can conveniently create, initialize and set properties of objects with hierarchical relationships. Graphics are generated using the DirectX environment. The MRDiffusionImaging software package has been implemented for processing diffusion magnetic resonance imaging (dMRI), which allows loading and viewing images sorted by series. An algorithm for "masking" dMRI series based on T2-weighted images was developed using a deformable surface model to exclude tissues that are not related to the area of interest from the analysis. An algorithm of distortion correction using deformable image registration based on autocorrelation of local structure has been developed. Maximum voxel dimension was 1,03 ± 0,12 mm. In an elementary brain's volume, the diffusion tensor is geometrically interpreted using an ellipsoid, which is an isosurface of the probability density of a molecule's diffusion. For the first time, non-parametric intensity distributions, neighborhood correlations, and inhomogeneities are combined in one segmentation of white matter (WM), grey matter (GM), and cerebrospinal fluid (CSF) algorithm. A tool for calculating the coefficient of average diffusion and fractional anisotropy has been created, on the basis of which it is possible to build quantitative maps for solving various clinical problems. Functionality has been created that allows clustering and segmenting images to individualize the clinical volume of radiation treatment and further assess the response (Median Dice Score = 0.963 ± 0,137). White matter tracts of the brain were visualized using two algorithms: deterministic (fiber assignment by continuous tracking) and probabilistic using the Hough transform. The proposed algorithms test candidate curves in the voxel, assigning to each one a score computed from the diffusion data, and then selects the curves with the highest scores as the potential anatomical connections. White matter fibers were visualized using a Hough transform tractography algorithm. In the context of functional radiosurgery, it is possible to reduce the irradiation volume of the internal capsule receiving 12 Gy from 0,402 cc to 0,254 cc. The «MRDiffusionImaging» will improve the efficiency and accuracy of diagnostics and stereotactic radiotherapy of intracranial pathology. We develop software with integrated, intuitive support for processing, analysis, and inclusion in the process of radiotherapy planning and evaluating its results.

Keywords: diffusion-weighted imaging, medical imaging, stereotactic radiosurgery, tractography

Procedia PDF Downloads 85
2664 Assessment of Seeding and Weeding Field Robot Performance

Authors: Victor Bloch, Eerikki Kaila, Reetta Palva

Abstract:

Field robots are an important tool for enhancing efficiency and decreasing the climatic impact of food production. There exists a number of commercial field robots; however, since this technology is still new, the robot advantages and limitations, as well as methods for optimal using of robots, are still unclear. In this study, the performance of a commercial field robot for seeding and weeding was assessed. A research 2-ha sugar beet field with 0.5m row width was used for testing, which included robotic sowing of sugar beet and weeding five times during the first two months of the growing. About three and five percent of the field were used as untreated and chemically weeded control areas, respectively. The plant detection was based on the exact plant location without image processing. The robot was equipped with six seeding and weeding tools, including passive between-rows harrow hoes and active hoes cutting inside rows between the plants, and it moved with a maximal speed of 0.9 km/h. The robot's performance was assessed by image processing. The field images were collected by an action camera with a height of 2 m and a resolution 27M pixels installed on the robot and by a drone with a 16M pixel camera flying at 4 m height. To detect plants and weeds, the YOLO model was trained with transfer learning from two available datasets. A preliminary analysis of the entire field showed that in the areas treated by the robot, the weed average density varied across the field from 6.8 to 9.1 weeds/m² (compared with 0.8 in the chemically treated area and 24.3 in the untreated area), the weed average density inside rows was 2.0-2.9 weeds / m (compared with 0 on the chemically treated area), and the emergence rate was 90-95%. The information about the robot's performance has high importance for the application of robotics for field tasks. With the help of the developed method, the performance can be assessed several times during the growth according to the robotic weeding frequency. When it’s used by farmers, they can know the field condition and efficiency of the robotic treatment all over the field. Farmers and researchers could develop optimal strategies for using the robot, such as seeding and weeding timing, robot settings, and plant and field parameters and geometry. The robot producers can have quantitative information from an actual working environment and improve the robots accordingly.

Keywords: agricultural robot, field robot, plant detection, robot performance

Procedia PDF Downloads 87
2663 Sea of Light: A Game 'Based Approach for Evidence-Centered Assessment of Collaborative Problem Solving

Authors: Svenja Pieritz, Jakab Pilaszanovich

Abstract:

Collaborative Problem Solving (CPS) is recognized as being one of the most important skills of the 21st century with having a potential impact on education, job selection, and collaborative systems design. Therefore, CPS has been adopted in several standardized tests, including the Programme for International Student Assessment (PISA) in 2015. A significant challenge of evaluating CPS is the underlying interplay of cognitive and social skills, which requires a more holistic assessment. However, the majority of the existing tests are using a questionnaire-based assessment, which oversimplifies this interplay and undermines ecological validity. Two major difficulties were identified: Firstly, the creation of a controllable, real-time environment allowing natural behaviors and communication between at least two people. Secondly, the development of an appropriate method to collect and synthesize both cognitive and social metrics of collaboration. This paper proposes a more holistic and automated approach to the assessment of CPS. To address these two difficulties, a multiplayer problem-solving game called Sea of Light was developed: An environment allowing students to deploy a variety of measurable collaborative strategies. This controlled environment enables researchers to monitor behavior through the analysis of game actions and chat. The according solution for the statistical model is a combined approach of Natural Language Processing (NLP) and Bayesian network analysis. Social exchanges via the in-game chat are analyzed through NLP and fed into the Bayesian network along with other game actions. This Bayesian network synthesizes evidence to track and update different subdimensions of CPS. Major findings focus on the correlations between the evidences collected through in- game actions, the participants’ chat features and the CPS self- evaluation metrics. These results give an indication of which game mechanics can best describe CPS evaluation. Overall, Sea of Light gives test administrators control over different problem-solving scenarios and difficulties while keeping the student engaged. It enables a more complete assessment based on complex, socio-cognitive information on actions and communication. This tool permits further investigations of the effects of group constellations and personality in collaborative problem-solving.

Keywords: bayesian network, collaborative problem solving, game-based assessment, natural language processing

Procedia PDF Downloads 132
2662 An Assistive Robotic Arm for Defence and Rescue Application

Authors: J. Harrison Kurunathan, R. Jayaparvathy

Abstract:

"Assistive Robotics" is the field that deals with the study of robots that helps in human motion and also empowers human abilities by interfacing the robotic systems to be manipulated by human motion. The proposed model is a robotic arm that works as a haptic interface on the basis on accelerometers and DC motors that will function with respect to the movement of the human muscle. The proposed model would effectively work as a haptic interface that would reduce human effort in the field of defense and rescue. This can be used in very critical conditions like fire accidents to avoid causalities.

Keywords: accelerometers, haptic interface, servo motors, signal processing

Procedia PDF Downloads 398
2661 Study of Chemical Compounds of Garlic

Authors: Bazaraliyeva Aigerim Bakytzhanovna, Turgumbayeva Aknur Amanbekovna

Abstract:

The phytosubstance from garlic was obtained by extraction with liquid carbon dioxide under critical conditions. Methods of processing raw materials are proposed, and the chemical composition of garlic is studied by gas chromatography and mass spectrometry. The garlic extract's composition was determined using gas chromatography (GC) and gas chromatography-mass spectrophotometry (GC-MS). The phytosubstance had 54 constituents. The extract included the following main compounds: Manool (39.56%), Viridifrolol (7%), Podocarpa-1,8,11,13-tetraen-3-one, 14-isopropyl-1,13-dimethoxy- 5,15 percent, (+)-2-Bornanone (4.29%), Thujone (3.49%), Linolic acid ethyl ester (3.41%), and 12-O-Methylcarn.

Keywords: allium sativum, bioactive compounds of garlic, carbon dioxide extraction of garlic, GS-MS method

Procedia PDF Downloads 81
2660 Mechanism of Action of New Sustainable Flame Retardant Additives in Polyamide 6,6

Authors: I. Belyamani, M. K. Hassan, J. U. Otaigbe, W. R. Fielding, K. A. Mauritz, J. S. Wiggins, W. L. Jarrett

Abstract:

We have investigated the flame-retardant efficiency of special new phosphate glass (P-glass) compositions having different glass transition temperatures (Tg) on the processing conditions of polyamide 6,6 (PA6,6) and the final hybrid flame retardancy (FR). We have showed that the low Tg P glass composition (i.e., ILT 1) is a promising flame retardant for PA6,6 at a concentration of up to 15 wt. % compared to intermediate (IIT 3) and high (IHT 1) Tg P glasses. Cone calorimetry data showed that the ILT 1 decreased both the peak heat release rate and the total heat amount released from the PA6,6/ILT 1 hybrids, resulting in an efficient formation of a glassy char layer. These intriguing findings prompted to address several questions concerning the mechanism of action of the different P glasses studied. The general mechanism of action of phosphorous based FR additives occurs during the combustion stage by enhancing the morphology of the char and the thermal shielding effect. However, the present work shows that P glass based FR additives act during melt processing of PA6,6/P glass hybrids. Dynamic mechanical analysis (DMA) revealed that the Tg of PA6,6/ILT 1 was significantly shifted to a lower Tg (~65 oC) and another transition appeared at high temperature (~ 166 oC), thus indicating a strong interaction between PA6,6 and ILT 1. This was supported by a drop in the melting point and crystallinity of the PA6,6/ILT 1 hybrid material as detected by differential scanning calorimetry (DSC). The dielectric spectroscopic investigation of the networks’ molecular level structural variations (i.e. hybrids chain motion, Tg and sub-Tg relaxations) agreed very well with the DMA and DSC findings; it was found that the three different P glass compositions did not show any effect on the PA6,6 sub-Tg relaxations (related to the NH2 and OH chain end groups motions). Nevertheless, contrary to IIT 3 and IHT 1 based hybrids, the PA6,6/ILT 1 hybrid material showed an evidence of splitting the PA6,6 Tg relaxations into two peaks. Finally, the CPMAS 31P-NMR data confirmed the miscibility between ILT 1 and PA6,6 at the molecular level, as a much larger enhancement in cross-polarization for the PA6,6/15%ILT 1 hybrids was observed. It can be concluded that compounding low Tg P-glass (i.e. ILT 1) with PA6,6 facilitates hydrolytic chain scission of the PA6,6 macromolecules through a potential chemical interaction between phosphate and the alpha-Carbon of the amide bonds of the PA6,6, leading to better flame retardant properties.

Keywords: broadband dielectric spectroscopy, composites, flame retardant, polyamide, phosphate glass, sustainable

Procedia PDF Downloads 239
2659 Optimized Processing of Neural Sensory Information with Unwanted Artifacts

Authors: John Lachapelle

Abstract:

Introduction: Neural stimulation is increasingly targeted toward treatment of back pain, PTSD, Parkinson’s disease, and for sensory perception. Sensory recording during stimulation is important in order to examine neural response to stimulation. Most neural amplifiers (headstages) focus on noise efficiency factor (NEF). Conversely, neural headstages need to handle artifacts from several sources including power lines, movement (EMG), and neural stimulation itself. In this work a layered approach to artifact rejection is used to reduce corruption of the neural ENG signal by 60dBv, resulting in recovery of sensory signals in rats and primates that would previously not be possible. Methods: The approach combines analog techniques to reduce and handle unwanted signal amplitudes. The methods include optimized (1) sensory electrode placement, (2) amplifier configuration, and (3) artifact blanking when necessary. The techniques together are like concentric moats protecting a castle; only the wanted neural signal can penetrate. There are two conditions in which the headstage operates: unwanted artifact < 50mV, linear operation, and artifact > 50mV, fast-settle gain reduction signal limiting (covered in more detail in a separate paper). Unwanted Signals at the headstage input: Consider: (a) EMG signals are by nature < 10mV. (b) 60 Hz power line signals may be > 50mV with poor electrode cable conditions; with careful routing much of the signal is common to both reference and active electrode and rejected in the differential amplifier with <50mV remaining. (c) An unwanted (to the neural recorder) stimulation signal is attenuated from stimulation to sensory electrode. The voltage seen at the sensory electrode can be modeled Φ_m=I_o/4πσr. For a 1 mA stimulation signal, with 1 cm spacing between electrodes, the signal is <20mV at the headstage. Headstage ASIC design: The front end ASIC design is designed to produce < 1% THD at 50mV input; 50 times higher than typical headstage ASICs, with no increase in noise floor. This requires careful balance of amplifier stages in the headstage ASIC, as well as consideration of the electrodes effect on noise. The ASIC is designed to allow extremely small signal extraction on low impedance (< 10kohm) electrodes with configuration of the headstage ASIC noise floor to < 700nV/rt-Hz. Smaller high impedance electrodes (> 100kohm) are typically located closer to neural sources and transduce higher amplitude signals (> 10uV); the ASIC low-power mode conserves power with 2uV/rt-Hz noise. Findings: The enhanced neural processing ASIC has been compared with a commercial neural recording amplifier IC. Chronically implanted primates at MGH demonstrated the presence of commercial neural amplifier saturation as a result of large environmental artifacts. The enhanced artifact suppression headstage ASIC, in the same setup, was able to recover and process the wanted neural signal separately from the suppressed unwanted artifacts. Separately, the enhanced artifact suppression headstage ASIC was able to separate sensory neural signals from unwanted artifacts in mouse-implanted peripheral intrafascicular electrodes. Conclusion: Optimizing headstage ASICs allow observation of neural signals in the presence of large artifacts that will be present in real-life implanted applications, and are targeted toward human implantation in the DARPA HAPTIX program.

Keywords: ASIC, biosensors, biomedical signal processing, biomedical sensors

Procedia PDF Downloads 330
2658 Construction of Finite Woven Frames through Bounded Linear Operators

Authors: A. Bhandari, S. Mukherjee

Abstract:

Two frames in a Hilbert space are called woven or weaving if all possible merge combinations between them generate frames of the Hilbert space with uniform frame bounds. Weaving frames are powerful tools in wireless sensor networks which require distributed data processing. Considering the practical applications, this article deals with finite woven frames. We provide methods of constructing finite woven frames, in particular, bounded linear operators are used to construct woven frames from a given frame. Several examples are discussed. We also introduce the notion of woven frame sequences and characterize them through the concepts of gaps and angles between spaces.

Keywords: frames, woven frames, gap, angle

Procedia PDF Downloads 195
2657 Medium-Scale Multi-Juice Extractor for Food Processing

Authors: Flordeliza L. Mercado, Teresito G. Aguinaldo, Helen F. Gavino, Victorino T. Taylan

Abstract:

Most fruits and vegetables are available in large quantities during peak season which are oftentimes marketed at low price and left to rot or fed to farm animals. The lack of efficient storage facilities, and the additional cost and unavailability of small machinery for food processing, results to low price and wastage. Incidentally, processed fresh fruits and vegetables are gaining importance nowadays and health conscious people are also into ‘juicing’. One way to reduce wastage and ensure an all-season availability of crop juices at reasonable costs is to develop equipment for effective extraction of juice. The study was conducted to design, fabricate and evaluate a multi-juice extractor using locally available materials, making it relatively cheaper and affordable for medium-scale enterprises. The study was also conducted to formulate juice blends using extracted juices and calamansi juice at different blending percentage, and evaluate its chemical properties and sensory attributes. Furthermore, the chemical properties of extracted meals were evaluated for future applications. The multi-juice extractor has an overall dimension of 963mm x 300mm x 995mm, a gross weight of 82kg and 5 major components namely; feeding hopper, extracting chamber, juice and meal outlet, transmission assembly, and frame. The machine performance was evaluated based on juice recovery, extraction efficiency, extraction rate, extraction recovery, and extraction loss considering type of crop as apple and carrot with three replications each and was analyzed using T-test. The formulated juice blends were subjected to sensory evaluation and data gathered were analyzed using Analysis of Variance appropriate for Complete Randomized Design. Results showed that the machine’s juice recovery (73.39%), extraction rate (16.40li/hr), and extraction efficiency (88.11%) for apple were significantly higher than for carrot while extraction recovery (99.88%) was higher for apple than for carrot. Extraction loss (0.12%) was lower for apple than for carrot, but was not significantly affected by crop. Based on adding percentage mark-up on extraction cost (Php 2.75/kg), the breakeven weight and payback period for a 35% mark-up is 4,710.69kg and 1.22 years, respectively and for a 50% mark-up, the breakeven weight is 3,492.41kg and the payback period is 0.86 year (10.32 months). Results on the sensory evaluation of juice blends showed that the type of juice significantly influenced all the sensory parameters while the blending percentage including their respective interaction, had no significant effect on all sensory parameters, making the apple-calamansi juice blend more preferred than the carrot-calamansi juice blend in terms of all the sensory parameter. The machine’s performance is higher for apple than for carrot and the cost analysis on the use of the machine revealed that it is financially viable with a payback period of 1.22 years (35% mark-up) and 0.86 year (50% mark-up) for machine cost, generating an income of Php 23,961.60 and Php 34,444.80 per year using 35% and 50% mark-up, respectively. The juice blends were of good qualities based on the values obtained in the chemical analysis and the extracted meal could also be used to produce another product based on the values obtained from proximate analysis.

Keywords: food processing, fruits and vegetables, juice extraction, multi-juice extractor

Procedia PDF Downloads 325
2656 Parallel Version of Reinhard’s Color Transfer Algorithm

Authors: Abhishek Bhardwaj, Manish Kumar Bajpai

Abstract:

An image with its content and schema of colors presents an effective mode of information sharing and processing. By changing its color schema different visions and prospect are discovered by the users. This phenomenon of color transfer is being used by Social media and other channel of entertainment. Reinhard et al’s algorithm was the first one to solve this problem of color transfer. In this paper, we make this algorithm efficient by introducing domain parallelism among different processors. We also comment on the factors that affect the speedup of this problem. In the end by analyzing the experimental data we claim to propose a novel and efficient parallel Reinhard’s algorithm.

Keywords: Reinhard et al’s algorithm, color transferring, parallelism, speedup

Procedia PDF Downloads 614
2655 Particle Size Distribution Estimation of a Mixture of Regular and Irregular Sized Particles Using Acoustic Emissions

Authors: Ejay Nsugbe, Andrew Starr, Ian Jennions, Cristobal Ruiz-Carcel

Abstract:

This works investigates the possibility of using Acoustic Emissions (AE) to estimate the Particle Size Distribution (PSD) of a mixture of particles that comprise of particles of different densities and geometry. The experiments carried out involved the mixture of a set of glass and polyethylene particles that ranged from 150-212 microns and 150-250 microns respectively and an experimental rig that allowed the free fall of a continuous stream of particles on a target plate which the AE sensor was placed. By using a time domain based multiple threshold method, it was observed that the PSD of the particles in the mixture could be estimated.

Keywords: acoustic emissions, particle sizing, process monitoring, signal processing

Procedia PDF Downloads 353
2654 Open-Ended Multi-Modal Relational Reason for Video Question Answering

Authors: Haozheng Luo, Ruiyang Qin

Abstract:

People with visual impairments urgently need assistance, not only on the fundamental tasks such as guiding and retrieving objects but on the advanced like picturing the new environments. More than a guiding dog, they might want such devices that can provide linguistic interaction. Building on this idea, we aim to study the interaction between the robot agent and visually impaired people. In our research, we are going to develop a robot agent that will be able to analyze the test environment and answer the participants’ questions. We also will study the relevant issues regarding the interaction between human beings and the robot agents to figure out which and how the factors will affect the interaction.

Keywords: HRI, video question answering, visual question answering, natural language processing

Procedia PDF Downloads 217
2653 Diversification of Indonesian Terasi Shrimp (Acetes indicus) Powder as Alternative and Sustainable Food for the Double Burden of Malnutrition

Authors: Galuh Asri Bestari, Hajar Shofiyya

Abstract:

Double burden of malnutrition (DBM) has been a global problem in these last decades occurs in both developed and developing countries. Overweight in adults and stunting among preschool children have dramatically increased and become the main problems of malnutrition that should be solved immediately since they are directly related with the health status and productivity. Reformulation of food product by using the local sea resources called terasi shrimp (Acetes indicus) has a potential possibility in facing the DBM. A study was carried out in Indonesia to determine the acceptability of terasi shrimp powder through sensory evaluation. Terasi shrimps were processed into powder form through sun drying and pounding methods. The powder form was directly added in food as alternative seasonings and tested among stunted and normal preschool children. Meanwhile, a further processing method is given to the shrimp powder tested in overweight and normal-weighed adults. The shrimp powder was mixed with sago flour and formed into balls, then steamed for 15-20 minutes, and finally served as alternative snacks. Based on the sensory evaluation, the shrimp powder has a good acceptance in taste (54%), shape (60%), and color properties (63%), while the shrimp balls has a good acceptance in size (65%), shape (50%), color (48%), taste (40%), and texture (36%). Terasi shrimp powder can be stored for a month in room temperature. In addition, carried out chemical analysis revealed that terasi shrimp (Acetes indicus) has higher percentage of protein, calcium, and iron than other animal sources, but conversely contains zero sodium and very low percentage of fat. Terasi shrimp’s shell also contains a substance called chitosan which acts by forming gels in the intestinal tract to entrap lipids, thus interfering with their absorption. After going through some processing methods, the shrimp powder and balls did not show any significant changes in their nutrient contents. So that, terasi shrimp powder is good to be consumed not only by overweight adults, but also by children to support their optimum growth. Intervention of terasi shrimp powder should be implemented step by step from national up to global governance program to face the DBM.

Keywords: Acetes indicus, alternative food, double burden of malnutrition, sensory evaluation

Procedia PDF Downloads 306
2652 Medicompills Architecture: A Mathematical Precise Tool to Reduce the Risk of Diagnosis Errors on Precise Medicine

Authors: Adriana Haulica

Abstract:

Powered by Machine Learning, Precise medicine is tailored by now to use genetic and molecular profiling, with the aim of optimizing the therapeutic benefits for cohorts of patients. As the majority of Machine Language algorithms come from heuristics, the outputs have contextual validity. This is not very restrictive in the sense that medicine itself is not an exact science. Meanwhile, the progress made in Molecular Biology, Bioinformatics, Computational Biology, and Precise Medicine, correlated with the huge amount of human biology data and the increase in computational power, opens new healthcare challenges. A more accurate diagnosis is needed along with real-time treatments by processing as much as possible from the available information. The purpose of this paper is to present a deeper vision for the future of Artificial Intelligence in Precise medicine. In fact, actual Machine Learning algorithms use standard mathematical knowledge, mostly Euclidian metrics and standard computation rules. The loss of information arising from the classical methods prevents obtaining 100% evidence on the diagnosis process. To overcome these problems, we introduce MEDICOMPILLS, a new architectural concept tool of information processing in Precise medicine that delivers diagnosis and therapy advice. This tool processes poly-field digital resources: global knowledge related to biomedicine in a direct or indirect manner but also technical databases, Natural Language Processing algorithms, and strong class optimization functions. As the name suggests, the heart of this tool is a compiler. The approach is completely new, tailored for omics and clinical data. Firstly, the intrinsic biological intuition is different from the well-known “a needle in a haystack” approach usually used when Machine Learning algorithms have to process differential genomic or molecular data to find biomarkers. Also, even if the input is seized from various types of data, the working engine inside the MEDICOMPILLS does not search for patterns as an integrative tool. This approach deciphers the biological meaning of input data up to the metabolic and physiologic mechanisms, based on a compiler with grammars issued from bio-algebra-inspired mathematics. It translates input data into bio-semantic units with the help of contextual information iteratively until Bio-Logical operations can be performed on the base of the “common denominator “rule. The rigorousness of MEDICOMPILLS comes from the structure of the contextual information on functions, built to be analogous to mathematical “proofs”. The major impact of this architecture is expressed by the high accuracy of the diagnosis. Detected as a multiple conditions diagnostic, constituted by some main diseases along with unhealthy biological states, this format is highly suitable for therapy proposal and disease prevention. The use of MEDICOMPILLS architecture is highly beneficial for the healthcare industry. The expectation is to generate a strategic trend in Precise medicine, making medicine more like an exact science and reducing the considerable risk of errors in diagnostics and therapies. The tool can be used by pharmaceutical laboratories for the discovery of new cures. It will also contribute to better design of clinical trials and speed them up.

Keywords: bio-semantic units, multiple conditions diagnosis, NLP, omics

Procedia PDF Downloads 70
2651 Product Separation of Green Processes and Catalyst Recycling of a Homogeneous Polyoxometalate Catalyst Using Nanofiltration Membranes

Authors: Dorothea Voß, Tobias Esser, Michael Huber, Jakob Albert

Abstract:

The growing world population and the associated increase in demand for energy and consumer goods, as well as increasing waste production, requires the development of sustainable processes. In addition, the increasing environmental awareness of our society is a driving force for the requirement that processes must be as resource and energy efficient as possible. In this context, the use of polyoxometalate catalysts (POMs) has emerged as a promising approach for the development of green processes. POMs are bifunctional polynuclear metal-oxo-anion cluster characterized by a strong Brønsted acidity, a high proton mobility combined with fast multi-electron transfer and tunable redox potential. In addition, POMs are soluble in many commonly known solvents and exhibit resistance to hydrolytic and oxidative degradation. Due to their structure and excellent physicochemical properties, POMs are efficient acid and oxidation catalysts that have attracted much attention in recent years. Oxidation processes with molecular oxygen are worth mentioning here. However, the fact that the POM catalysts are homogeneous poses a challenge for downstream processing of product solutions and recycling of the catalysts. In this regard, nanofiltration membranes have gained increasing interest in recent years, particularly due to their relative sustainability advantage over other technologies and their unique properties such as increased selectivity towards multivalent ions. In order to establish an efficient downstream process for the highly selective separation of homogeneous POM catalysts from aqueous solutions using nanofiltration membranes, a laboratory-scale membrane system was designed and constructed. By varying various process parameters, a sensitivity analysis was performed on a model system to develop an optimized method for the recovery of POM catalysts. From this, process-relevant key figures such as the rejection of various system components were derived. These results form the basis for further experiments on other systems to test the transferability to serval separation tasks with different POMs and products, as well as for recycling experiments of the catalysts in processes on laboratory scale.

Keywords: downstream processing, nanofiltration, polyoxometalates, homogeneous catalysis, green chemistry

Procedia PDF Downloads 89
2650 Analysis of Histogram Asymmetry for Waste Recognition

Authors: Janusz Bobulski, Kamila Pasternak

Abstract:

Despite many years of effort and research, the problem of waste management is still current. So far, no fully effective waste management system has been developed. Many programs and projects improve statistics on the percentage of waste recycled every year. In these efforts, it is worth using modern Computer Vision techniques supported by artificial intelligence. In the article, we present a method of identifying plastic waste based on the asymmetry analysis of the histogram of the image containing the waste. The method is simple but effective (94%), which allows it to be implemented on devices with low computing power, in particular on microcomputers. Such de-vices will be used both at home and in waste sorting plants.

Keywords: waste management, environmental protection, image processing, computer vision

Procedia PDF Downloads 120
2649 Replication of Meaningful Gesture Study for N400 Detection Using a Commercial Brain-Computer Interface

Authors: Thomas Ousterhout

Abstract:

In an effort to test the ability of a commercial grade EEG headset to effectively measure the N400 ERP, a replication study was conducted to see if similar results could be produced as that which used a medical grade EEG. Pictures of meaningful and meaningless hand postures were borrowed from the original author and subjects were required to perform a semantic discrimination task. The N400 was detected indicating semantic processing of the meaningfulness of the hand postures. The results corroborate those of the original author and support the use of some commercial grade EEG headsets for non-critical research applications.

Keywords: EEG, ERP, N400, semantics, congruency, gestures, emotiv

Procedia PDF Downloads 264
2648 Assessment and Forecasting of the Impact of Negative Environmental Factors on Public Health

Authors: Nurlan Smagulov, Aiman Konkabayeva, Akerke Sadykova, Arailym Serik

Abstract:

Introduction. Adverse environmental factors do not immediately lead to pathological changes in the body. They can exert the growth of pre-pathology characterized by shifts in physiological, biochemical, immunological and other indicators of the body state. These disorders are unstable, reversible and indicative of body reactions. There is an opportunity to objectively judge the internal structure of the adaptive body reactions at the level of individual organs and systems. In order to obtain a stable response of the body to the chronic effects of unfavorable environmental factors of low intensity (compared to production environment factors), a time called the «lag time» is needed. The obtained results without considering this factor distort reality and, for the most part, cannot be a reliable statement of the main conclusions in any work. A technique is needed to reduce methodological errors and combine mathematical logic using statistical methods and a medical point of view, which ultimately will affect the obtained results and avoid a false correlation. Objective. Development of a methodology for assessing and predicting the environmental factors impact on the population health considering the «lag time.» Methods. Research objects: environmental and population morbidity indicators. The database on the environmental state was compiled from the monthly newsletters of Kazhydromet. Data on population morbidity were obtained from regional statistical yearbooks. When processing static data, a time interval (lag) was determined for each «argument-function» pair. That is the required interval, after which the harmful factor effect (argument) will fully manifest itself in the indicators of the organism's state (function). The lag value was determined by cross-correlation functions of arguments (environmental indicators) with functions (morbidity). Correlation coefficients (r) and their reliability (t), Fisher's criterion (F) and the influence share (R2) of the main factor (argument) per indicator (function) were calculated as a percentage. Results. The ecological situation of an industrially developed region has an impact on health indicators, but it has some nuances. Fundamentally opposite results were obtained in the mathematical data processing, considering the «lag time». Namely, an expressed correlation was revealed after two databases (ecology-morbidity) shifted. For example, the lag period was 4 years for dust concentration, general morbidity, and 3 years – for childhood morbidity. These periods accounted for the maximum values of the correlation coefficients and the largest percentage of the influencing factor. Similar results were observed in relation to the concentration of soot, dioxide, etc. The comprehensive statistical processing using multiple correlation-regression variance analysis confirms the correctness of the above statement. This method provided the integrated approach to predicting the degree of pollution of the main environmental components to identify the most dangerous combinations of concentrations of leading negative environmental factors. Conclusion. The method of assessing the «environment-public health» system (considering the «lag time») is qualitatively different from the traditional (without considering the «lag time»). The results significantly differ and are more amenable to a logical explanation of the obtained dependencies. The method allows presenting the quantitative and qualitative dependence in a different way within the «environment-public health» system.

Keywords: ecology, morbidity, population, lag time

Procedia PDF Downloads 82
2647 Carbon Capture: Growth and Development of Membranes in Gas Sequestration

Authors: Sreevalli Bokka

Abstract:

Various technologies are emerging to capture or reduce carbon intensity from a gas stream, such as industrial effluent air and atmosphere. Of these technologies, filter membranes are emerging as a key player in carbon sequestering. The key advantages of these membranes are their high surface area and porosity. Fabricating a filter membrane that has high selectivity for carbon sequestration is challenging as material properties and processing parameters affect the membrane properties. In this study, the growth of the filter membranes and the critical material properties that impact carbon sequestration are presented.

Keywords: membranes, filtration, separations, polymers, carbon capture

Procedia PDF Downloads 69
2646 Impulsivity Leads to Compromise Effect

Authors: Sana Maidullah, Ankita Sharma

Abstract:

The present study takes naturalistic decision-making approach to examine the role of personality in information processing in consumer decision making. In the technological era, most of the information comes in form of HTML or similar language via the internet; processing of this situation could be ambiguous, laborious and painful. The present study explores the role of impulsivity in creating an extreme effect on consumer decision making. Specifically, the study explores the role of impulsivity in extreme effect, i.e., extremeness avoidance (compromise effect) and extremeness seeking; the role of demographic variables, i.e. age and gender, in the relation between impulsivity and extreme effect. The study was conducted with the help of a questionnaire and two experiments. The experiment was designed in the form of two shopping websites with two product types: Hotel choice and Mobile choice. Both experimental interfaces were created with the Xampp software, the frontend of interfaces was HTML CSS JAVASCRIPT and backend was PHP MySQL. The mobile experiment was designed to measure the extreme effect and hotel experiment was designed to measure extreme effect with alignability of attributes. To observe the possibilities of the combined effect of individual difference and context effects, the manipulation of price, a number of alignable attributes and number of the non-alignable attributes is done. The study was conducted on 100 undergraduate and post-graduate engineering students within the age range of 18-35. The familiarity and level of use of internet and shopping website were assessed and controlled in the analysis. The analysis was done by using a t-test, ANOVA and regression analysis. The results indicated that the impulsivity leads to compromise effect and at the same time it also increases the relationship between alignability of attribute among choices and the compromise effect. The demographic variables were found to play a significant role in the relationship. The subcomponents of impulsivity were significantly influencing compromise effect, but the cognitive impulsivity was significant for women, and motor impulsivity was significant for males only. The impulsivity was significantly positively predicted by age, though there were no significant gender differences in impulsivity. The results clearly indicate the importance of individual factors in decision making. The present study, with precise and direct results, provides a significant suggestion for market analyst and business providers.

Keywords: impulsivity, extreme effect, personality, alignability, consumer decision making

Procedia PDF Downloads 189
2645 Research of Concentratibility of Low Quality Bauxite Raw Materials

Authors: Nadezhda Nikolaeva, Tatyana Alexandrova, Alexandr Alexandrov

Abstract:

Processing of high-silicon bauxite on the base of the traditional clinkering method is related to high power consumption and capital investments, which makes production of alumina from those ores non-competitive in terms of basic economic showings. For these reasons, development of technological solutions enabling to process bauxites with various chemical and mineralogical structures efficiently with low level of thermal power consumption is important. Flow sheet of the studies on washability of ores from the Timanskoe and the Severo-Onezhskoe deposits is on the base of the flotation method.

Keywords: low-quality bauxite, resource-saving technology, optimization, aluminum, conditioning of composition, separation characteristics

Procedia PDF Downloads 290