Search results for: decomposition process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15216

Search results for: decomposition process

15156 Effect of Resveratrol and Ascorbic Acid on the Stability of Alfa-Tocopherol in Whey Protein Isolate Stabilized O/W Emulsions

Authors: Lei Wang, Yingzhou Ni, Amr M. Bakry, Hao Cheng, Li Liang

Abstract:

Food proteins have been widely used as carrier materials because of their multiple functional properties. In this study, alfa-tocopherol was encapsulated in the oil phase of an oil-in-water emulsion stabilized with whey protein isolate (WPI). The influence of WPI concentration and resveratrol or ascorbic acid on the decomposition of alfa-tocopherol in the emulsion during storage is discussed. Decomposition decreased as WPI concentrations increased. Decomposition was delayed at ascorbic acid/WPI molar ratios lower than 5 but was promoted at higher ratios. Resveratrol partitioned into the oil-water interface by binding to WPI and its cis-isomer is believed to have contributed most of the protective effect of this polyphenol. These results suggest the possibility of using the emulsifying and ligand-binging properties of WPI to produce carriers for simultaneous encapsulation of alfa-tocopherol and resveratrol in a single emulsion system.

Keywords: stability, alfa-tocopherol, resveratrol, whey protein isolate

Procedia PDF Downloads 487
15155 The Wage Differential between Migrant and Native Workers in Australia: Decomposition Approach

Authors: Sabrina Tabassum

Abstract:

Using Census Data for Housing and Population of Australia 2001, 2006, 2011, and 2016, this paper shows the existence of wage differences between natives and immigrants in Australia. Addressing the heterogeneous nature of immigrants, this study group the immigrants in three broad categories- migrants from English speaking countries and migrants from India and China. Migrants from English speaking countries and India earn more than the natives per week, whereas migrants from China earn far less than the natives per week. Oaxaca decomposition suggests that major part of this differential is unexplained. Using the occupational segregation concept and Brown decomposition, this study indicates that migrants from India and China would have been earned more than the natives if they had the same occupation distribution as natives due to their individual characteristics. Within occupation, wage differences are more prominent than inter-occupation wage differences for immigrants from China and India.

Keywords: Australia, labour, migration, wage

Procedia PDF Downloads 99
15154 The Influence of Oil Price Fluctuations on Macroeconomics Variables of the Kingdom of Saudi Arabia

Authors: Khalid Mujaljal, Hassan Alhajhoj

Abstract:

This paper empirically investigates the influence of oil price fluctuations on the key macroeconomic variables of the Kingdom of Saudi Arabia using unrestricted VAR methodology. Two analytical tools- Granger-causality and variance decomposition are used. The Granger-causality test reveals that almost all specifications of oil price shocks significantly Granger-cause GDP and demonstrates evidence of causality between oil price changes and money supply (M3) and consumer price index percent (CPIPC) in the case of positive oil price shocks. Surprisingly, almost all specifications of oil price shocks do not Granger-cause government expenditure. The outcomes from variance decomposition analysis suggest that positive oil shocks contribute about 25 percent in causing inflation in the country. Also, contribution of symmetric linear oil price shocks and asymmetric positive oil price shocks is significant and persistent with 25 percent explaining variation in world consumer price index till end of the period.

Keywords: Granger causality, oil prices changes, Saudi Arabian economy, variance decomposition

Procedia PDF Downloads 295
15153 Functional Decomposition Based Effort Estimation Model for Software-Intensive Systems

Authors: Nermin Sökmen

Abstract:

An effort estimation model is needed for software-intensive projects that consist of hardware, embedded software or some combination of the two, as well as high level software solutions. This paper first focuses on functional decomposition techniques to measure functional complexity of a computer system and investigates its impact on system development effort. Later, it examines effects of technical difficulty and design team capability factors in order to construct the best effort estimation model. With using traditional regression analysis technique, the study develops a system development effort estimation model which takes functional complexity, technical difficulty and design team capability factors as input parameters. Finally, the assumptions of the model are tested.

Keywords: functional complexity, functional decomposition, development effort, technical difficulty, design team capability, regression analysis

Procedia PDF Downloads 258
15152 Atomic Decomposition Audio Data Compression and Denoising Using Sparse Dictionary Feature Learning

Authors: T. Bryan , V. Kepuska, I. Kostnaic

Abstract:

A method of data compression and denoising is introduced that is based on atomic decomposition of audio data using “basis vectors” that are learned from the audio data itself. The basis vectors are shown to have higher data compression and better signal-to-noise enhancement than the Gabor and gammatone “seed atoms” that were used to generate them. The basis vectors are the input weights of a Sparse AutoEncoder (SAE) that is trained using “envelope samples” of windowed segments of the audio data. The envelope samples are extracted from the audio data by performing atomic decomposition with Gabor or gammatone seed atoms. This process identifies segments of audio data that are locally coherent with the seed atoms. Envelope samples are extracted by identifying locally coherent audio data segments with Gabor or gammatone seed atoms, found by matching pursuit. The envelope samples are formed by taking the kronecker products of the atomic envelopes with the locally coherent data segments. Oracle signal-to-noise ratio (SNR) verses data compression curves are generated for the seed atoms as well as the basis vectors learned from Gabor and gammatone seed atoms. SNR data compression curves are generated for speech signals as well as early American music recordings. The basis vectors are shown to have higher denoising capability for data compression rates ranging from 90% to 99.84% for speech as well as music. Envelope samples are displayed as images by folding the time series into column vectors. This display method is used to compare of the output of the SAE with the envelope samples that produced them. The basis vectors are also displayed as images. Sparsity is shown to play an important role in producing the highest denoising basis vectors.

Keywords: sparse dictionary learning, autoencoder, sparse autoencoder, basis vectors, atomic decomposition, envelope sampling, envelope samples, Gabor, gammatone, matching pursuit

Procedia PDF Downloads 223
15151 Evolution of Microstructure through Phase Separation via Spinodal Decomposition in Spinel Ferrite Thin Films

Authors: Nipa Debnath, Harinarayan Das, Takahiko Kawaguchi, Naonori Sakamoto, Kazuo Shinozaki, Hisao Suzuki, Naoki Wakiya

Abstract:

Nowadays spinel ferrite magnetic thin films have drawn considerable attention due to their interesting magnetic and electrical properties with enhanced chemical and thermal stability. Spinel ferrite magnetic films can be implemented in magnetic data storage, sensors, and spin filters or microwave devices. It is well established that the structural, magnetic and transport properties of the magnetic thin films are dependent on microstructure. Spinodal decomposition (SD) is a phase separation process, whereby a material system is spontaneously separated into two phases with distinct compositions. The periodic microstructure is the characteristic feature of SD. Thus, SD can be exploited to control the microstructure at the nanoscale level. In bulk spinel ferrites having general formula, MₓFe₃₋ₓ O₄ (M= Co, Mn, Ni, Zn), phase separation via SD has been reported only for cobalt ferrite (CFO); however, long time post-annealing is required to occur the spinodal decomposition. We have found that SD occurs in CoF thin film without using any post-deposition annealing process if we apply magnetic field during thin film growth. Dynamic Aurora pulsed laser deposition (PLD) is a specially designed PLD system through which in-situ magnetic field (up to 2000 G) can be applied during thin film growth. The in-situ magnetic field suppresses the recombination of ions in the plume. In addition, the peak’s intensity of the ions in the spectra of the plume also increases when magnetic field is applied to the plume. As a result, ions with high kinetic energy strike into the substrate. Thus, ion-impingement occurred under magnetic field during thin film growth. The driving force of SD is the ion-impingement towards the substrates that is induced by in-situ magnetic field. In this study, we report about the occurrence of phase separation through SD and evolution of microstructure after phase separation in spinel ferrite thin films. The surface morphology of the phase separated films show checkerboard like domain structure. The cross-sectional microstructure of the phase separated films reveal columnar type phase separation. Herein, the decomposition wave propagates in lateral direction which has been confirmed from the lateral composition modulations in spinodally decomposed films. Large magnetic anisotropy has been found in spinodally decomposed nickel ferrite (NFO) thin films. This approach approves that magnetic field is also an important thermodynamic parameter to induce phase separation by the enhancement of up-hill diffusion in thin films. This thin film deposition technique could be a more efficient alternative for the fabrication of self-organized phase separated thin films and employed in controlling of the microstructure at nanoscale level.

Keywords: Dynamic Aurora PLD, magnetic anisotropy, spinodal decomposition, spinel ferrite thin film

Procedia PDF Downloads 337
15150 On Cold Roll Bonding of Polymeric Films

Authors: Nikhil Padhye

Abstract:

Recently a new phenomenon for bonding of polymeric films in solid-state, at ambient temperatures well below the glass transition temperature of the polymer, has been reported. This is achieved by bulk plastic compression of polymeric films held in contact. Here we analyze the process of cold-rolling of polymeric films via finite element simulations and illustrate a flexible and modular experimental rolling-apparatus that can achieve bonding of polymeric films through cold-rolling. Firstly, the classical theory of rolling a rigid-plastic thin-strip is utilized to estimate various deformation fields such as strain-rates, velocities, loads etc. in rolling the polymeric films at the specified feed-rates and desired levels of thickness-reduction(s). Predicted magnitudes of slow strain-rates, particularly at ambient temperatures during rolling, and moderate levels of plastic deformation (at which Bauschinger effect can be neglected for the particular class of polymeric materials studied here), greatly simplifies the task of material modeling and allows us to deploy a computationally efficient, yet accurate, finite deformation rate-independent elastic-plastic material behavior model (with inclusion of isotropic-hardening) for analyzing the rolling of these polymeric films. The interfacial behavior between the roller and polymer surfaces is modeled using Coulombic friction; consistent with the rate-independent behavior. The finite deformation elastic-plastic material behavior based on (i) the additive decomposition of stretching tensor (D = De + Dp, i.e. a hypoelastic formulation) with incrementally objective time integration and, (ii) multiplicative decomposition of deformation gradient (F = FeFp) into elastic and plastic parts, are programmed and carried out for cold-rolling within ABAQUS Explicit. Predictions from both the formulations, i.e., hypoelastic and multiplicative decomposition, exhibit a close match. We find that no specialized hyperlastic/visco-plastic model is required to describe the behavior of the blend of polymeric films, under the conditions described here, thereby speeding up the computation process .

Keywords: Polymer Plasticity, Bonding, Deformation Induced Mobility, Rolling

Procedia PDF Downloads 151
15149 Torrefaction of Biomass Pellets: Modeling of the Process in a Fixed Bed Reactor

Authors: Ekaterina Artiukhina, Panagiotis Grammelis

Abstract:

Torrefaction of biomass pellets is considered as a useful pretreatment technology in order to convert them into a high quality solid biofuel that is more suitable for pyrolysis, gasification, combustion and co-firing applications. In the course of torrefaction the temperature varies across the pellet, and therefore chemical reactions proceed unevenly within the pellet. However, the uniformity of the thermal distribution along the pellet is generally assumed. The torrefaction process of a single cylindrical pellet is modeled here, accounting for heat transfer coupled with chemical kinetics. The drying sub-model was also introduced. The non-stationary process of wood pellet decomposition is described by the system of non-linear partial differential equations over the temperature and mass. The model captures well the main features of the experimental data.

Keywords: torrefaction, biomass pellets, model, heat, mass transfer

Procedia PDF Downloads 446
15148 Finding Bicluster on Gene Expression Data of Lymphoma Based on Singular Value Decomposition and Hierarchical Clustering

Authors: Alhadi Bustaman, Soeganda Formalidin, Titin Siswantining

Abstract:

DNA microarray technology is used to analyze thousand gene expression data simultaneously and a very important task for drug development and test, function annotation, and cancer diagnosis. Various clustering methods have been used for analyzing gene expression data. However, when analyzing very large and heterogeneous collections of gene expression data, conventional clustering methods often cannot produce a satisfactory solution. Biclustering algorithm has been used as an alternative approach to identifying structures from gene expression data. In this paper, we introduce a transform technique based on singular value decomposition to identify normalized matrix of gene expression data followed by Mixed-Clustering algorithm and the Lift algorithm, inspired in the node-deletion and node-addition phases proposed by Cheng and Church based on Agglomerative Hierarchical Clustering (AHC). Experimental study on standard datasets demonstrated the effectiveness of the algorithm in gene expression data.

Keywords: agglomerative hierarchical clustering (AHC), biclustering, gene expression data, lymphoma, singular value decomposition (SVD)

Procedia PDF Downloads 247
15147 Toward Subtle Change Detection and Quantification in Magnetic Resonance Neuroimaging

Authors: Mohammad Esmaeilpour

Abstract:

One of the important open problems in the field of medical image processing is detection and quantification of small changes. In this poster, we try to investigate that, how the algebraic decomposition techniques can be used for semiautomatically detecting and quantifying subtle changes in Magnetic Resonance (MR) neuroimaging volumes. We mostly focus on the low-rank values of the matrices achieved from decomposing MR image pairs during a period of time. Besides, a skillful neuroradiologist will help the algorithm to distinguish between noises and small changes.

Keywords: magnetic resonance neuroimaging, subtle change detection and quantification, algebraic decomposition, basis functions

Procedia PDF Downloads 440
15146 A Multi Sensor Monochrome Video Fusion Using Image Quality Assessment

Authors: M. Prema Kumar, P. Rajesh Kumar

Abstract:

The increasing interest in image fusion (combining images of two or more modalities such as infrared and visible light radiation) has led to a need for accurate and reliable image assessment methods. This paper gives a novel approach of merging the information content from several videos taken from the same scene in order to rack up a combined video that contains the finest information coming from different source videos. This process is known as video fusion which helps in providing superior quality (The term quality, connote measurement on the particular application.) image than the source images. In this technique different sensors (whose redundant information can be reduced) are used for various cameras that are imperative for capturing the required images and also help in reducing. In this paper Image fusion technique based on multi-resolution singular value decomposition (MSVD) has been used. The image fusion by MSVD is almost similar to that of wavelets. The idea behind MSVD is to replace the FIR filters in wavelet transform with singular value decomposition (SVD). It is computationally very simple and is well suited for real time applications like in remote sensing and in astronomy.

Keywords: multi sensor image fusion, MSVD, image processing, monochrome video

Procedia PDF Downloads 542
15145 Fast and Accurate Model to Detect Ictal Waveforms in Electroencephalogram Signals

Authors: Piyush Swami, Bijaya Ketan Panigrahi, Sneh Anand, Manvir Bhatia, Tapan Gandhi

Abstract:

Visual inspection of electroencephalogram (EEG) signals to detect epileptic signals is very challenging and time-consuming task even for any expert neurophysiologist. This problem is most challenging in under-developed and developing countries due to shortage of skilled neurophysiologists. In the past, notable research efforts have gone in trying to automate the seizure detection process. However, due to high false alarm detections and complexity of the models developed so far, have vastly delimited their practical implementation. In this paper, we present a novel scheme for epileptic seizure detection using empirical mode decomposition technique. The intrinsic mode functions obtained were then used to calculate the standard deviations. This was followed by probability density based classifier to discriminate between non-ictal and ictal patterns in EEG signals. The model presented here demonstrated very high classification rates ( > 97%) without compromising the statistical performance. The computation timings for each testing phase were also very low ( < 0.029 s) which makes this model ideal for practical applications.

Keywords: electroencephalogram (EEG), epilepsy, ictal patterns, empirical mode decomposition

Procedia PDF Downloads 373
15144 Application of Regularized Low-Rank Matrix Factorization in Personalized Targeting

Authors: Kourosh Modarresi

Abstract:

The Netflix problem has brought the topic of “Recommendation Systems” into the mainstream of computer science, mathematics, and statistics. Though much progress has been made, the available algorithms do not obtain satisfactory results. The success of these algorithms is rarely above 5%. This work is based on the belief that the main challenge is to come up with “scalable personalization” models. This paper uses an adaptive regularization of inverse singular value decomposition (SVD) that applies adaptive penalization on the singular vectors. The results show far better matching for recommender systems when compared to the ones from the state of the art models in the industry.

Keywords: convex optimization, LASSO, regression, recommender systems, singular value decomposition, low rank approximation

Procedia PDF Downloads 417
15143 Atom Probe Study of Early Stage of Precipitation on Binary Al-Li, Al-Cu Alloys and Ternary Al-Li-Cu Alloys

Authors: Muna Khushaim

Abstract:

Aluminum-based alloys play a key role in modern engineering, especially in the aerospace industry. Introduction of solute atoms such as Li and Cu is the main approach to improve the strength in age-hardenable Al alloys via the precipitation hardening phenomenon. Knowledge of the decomposition process of the microstructure during the precipitation reaction is particularly important for future technical developments. The objective of this study is to investigate the nano-scale chemical composition in the Al-Cu, Al-Li and Al-Li-Cu during the early stage of the precipitation sequence and to describe whether this compositional difference correlates with variations in the observed precipitation kinetics. Comparing the random binomial frequency distribution and the experimental frequency distribution of concentrations in atom probe tomography data was used to investigate the early stage of decomposition in the different binary and ternary alloys which were experienced different heat treatments. The results show that an Al-1.7 at.% Cu alloy requires a long ageing time of approximately 8 h at 160 °C to allow the diffusion of Cu atoms into Al matrix. For the Al-8.2 at.% Li alloy, a combination of both the natural ageing condition (48 h at room temperature) and a short artificial ageing condition (5 min at 160 °C) induces increasing on the number density of the Li clusters and hence increase number of precipitated δ' particles. Applying this combination of natural ageing and short artificial ageing conditions onto the ternary Al-4 at.% Li-1.7 at.% Cu alloy induces the formation of a Cu-rich phase. Increasing the Li content in the ternary alloy up to 8 at.% and increasing the ageing time to 30 min resulted in the precipitation processes ending with δ' particles. Thus, the results contribute to the understanding of Al-alloy design.

Keywords: aluminum alloy, atom probe tomography, early stage, decomposition

Procedia PDF Downloads 317
15142 Catalytic Decomposition of Formic Acid into H₂/CO₂ Gas: A Distinct Approach

Authors: Ayman Hijazi, Witold Kwapinski, J. J. Leahy

Abstract:

Finding a sustainable alternative energy to fossil fuel is an urgent need as various environmental challenges in the world arise. Therefore, formic acid (FA) decomposition has been an attractive field that lies at the center of the biomass platform, comprising a potential pool of hydrogen energy that stands as a distinct energy vector. Liquid FA features considerable volumetric energy density of 6.4 MJ/L and a specific energy density of 5.3 MJ/Kg that qualifies it in the prime seat as an energy source for transportation infrastructure. Additionally, the increasing research interest in FA decomposition is driven by the need for in-situ H₂ production, which plays a key role in the hydrogenation reactions of biomass into higher-value components. It is reported elsewhere in the literature that catalytic decomposition of FA is usually performed in poorly designed setups using simple glassware under magnetic stirring, thus demanding further energy investment to retain the used catalyst. Our work suggests an approach that integrates designing a distinct catalyst featuring magnetic properties with a robust setup that minimizes experimental & measurement discrepancies. One of the most prominent active species for the dehydrogenation/hydrogenation of biomass compounds is palladium. Accordingly, we investigate the potential of engrafting palladium metal onto functionalized magnetic nanoparticles as a heterogeneous catalyst to favor the production of CO-free H₂ gas from FA. Using an ordinary magnet to collect the spent catalyst renders core-shell magnetic nanoparticles as the backbone of the process. Catalytic experiments were performed in a jacketed batch reactor equipped with an overhead stirrer under an inert medium. Through a distinct approach, FA is charged into the reactor via a high-pressure positive displacement pump at steady-state conditions. The produced gas (H₂+CO₂) was measured by connecting the gas outlet to a measuring system based on the amount of the displaced water. The uniqueness of this work lies in designing a very responsive catalyst, pumping a consistent amount of FA into a sealed reactor running at steady-state mild temperatures, and continuous gas measurement, along with collecting the used catalyst without the need for centrifugation. Catalyst characterization using TEM, XRD, SEM, and CHN elemental analyzer provided us with details of catalyst preparation and facilitated new venues to alter the nanostructure of the catalyst framework. Consequently, the introduction of amine groups has led to appreciable improvements in terms of dispersion of the doped metals and eventually attaining nearly complete conversion (100%) of FA after 7 hours. The relative importance of the process parameters such as temperature (35-85°C), stirring speed (150-450rpm), catalyst loading (50-200mgr.), and Pd doping ratio (0.75-1.80wt.%) on gas yield was assessed by a Taguchi design-of-experiment based model. Experimental results showed that operating at a lower temperature range (35-50°C) yielded more gas, while the catalyst loading and Pd doping wt.% were found to be the most significant factors with P-values 0.026 & 0.031, respectively.

Keywords: formic acid decomposition, green catalysis, hydrogen, mesoporous silica, process optimization, nanoparticles

Procedia PDF Downloads 11
15141 Magnetohydrodynamics (MHD) Boundary Layer Flow Past A Stretching Plate with Heat Transfer and Viscous Dissipation

Authors: Jiya Mohammed, Tsadu Shuaib, Yusuf Abdulhakeem

Abstract:

The research work focuses on the cases of MHD boundary layer flow past a stretching plate with heat transfer and viscous dissipation. The non-linear of momentum and energy equation are transform into ordinary differential equation by using similarity transformation, the resulting equation are solved using Adomian Decomposition Method (ADM). An attempt has been made to show the potentials and wide range application of the Adomian decomposition method in the comparison with the previous one in solving heat transfer problems. The Pade approximates value (η= 11[11, 11]) is use on the difficulty at infinity. The results are compared by numerical technique method. A vivid conclusion can be drawn from the results that ADM provides highly precise numerical solution for non-linear differential equations. The result where accurate especially for η ≤ 4, a general equating terms of Eckert number (Ec), Prandtl number (Pr) and magnetic parameter ( ) is derived which was used to investigate velocity and temperature profiles in boundary layer.

Keywords: MHD, Adomian decomposition, boundary layer, viscous dissipation

Procedia PDF Downloads 517
15140 Industry Openness, Human Capital and Wage Inequality: Evidence from Chinese Manufacturing Firms

Authors: Qiong Huang, Satish Chand

Abstract:

This paper uses a primary data from 670 Chinese manufacturing firms, together with the newly introduced regressionbased inequality decomposition method, to study the effect of openness on wage inequality. We find that openness leads to a positive industry wage premium, but its contribution to firm-level wage inequality is relatively small, only 4.69%. The major contributor to wage inequality is human capital, which could explain 14.3% of wage inequality across sample firms.  

Keywords: openness, human capital, wage inequality, decomposition, China

Procedia PDF Downloads 394
15139 Evaluation Methods for Question Decomposition Formalism

Authors: Aviv Yaniv, Ron Ben Arosh, Nadav Gasner, Michael Konviser, Arbel Yaniv

Abstract:

This paper introduces two methods for the evaluation of Question Decomposition Meaning Representation (QDMR) as predicted by sequence-to-sequence model and COPYNET parser for natural language questions processing, motivated by the fact that previous evaluation metrics used for this task do not take into account some characteristics of the representation, such as partial ordering structure. To this end, several heuristics to extract such partial dependencies are formulated, followed by the hereby proposed evaluation methods denoted as Proportional Graph Matcher (PGM) and Conversion to Normal String Representation (Nor-Str), designed to better capture the accuracy level of QDMR predictions. Experiments are conducted to demonstrate the efficacy of the proposed evaluation methods and show the added value suggested by one of them- the Nor-Str, for better distinguishing between high and low-quality QDMR when predicted by models such as COPYNET. This work represents an important step forward in the development of better evaluation methods for QDMR predictions, which will be critical for improving the accuracy and reliability of natural language question-answering systems.

Keywords: NLP, question answering, question decomposition meaning representation, QDMR evaluation metrics

Procedia PDF Downloads 34
15138 Classification Systems of Peat Soils Based on Their Geotechnical, Physical and Chemical Properties

Authors: Mohammad Saberian, Reza Porhoseini, Mohammad Ali Rahgozar

Abstract:

Peat is a partially carbonized vegetable tissue which is formed in wet conditions by decomposition of various plants, mosses and animal remains. This restricted definition, including only materials which are entirely of vegetative origin, conflicts with several established soil classification systems. Peat soils are usually defined as soils having more than 75 percent organic matter. Due to this composition, the structure of peat soil is highly different from the mineral soils such as silt, clay and sand. Peat has high compressibility, high moisture content, low shear strength and low bearing capacity, so it is considered to be in the category of problematic. Since this kind of soil is generally found in many countries and various zones, except for desert and polar zones, recognizing this soil is inevitably significant. The objective of this paper is to review the classification of peats based on various properties of peat soils such as organic contents, water content, color, odor, and decomposition, scholars offer various classification systems which Von Post classification system is one of the most well-known and efficient system.

Keywords: peat soil, degree of decomposition, organic content, water content, Von Post classification

Procedia PDF Downloads 560
15137 Co-Liquefaction of Cellulosic Biomass and Waste Plastics

Authors: Katsumi Hirano, Yusuke Kakuta, Koji Yoshida, Shozo Itagaki, Masahiko Kajioka, Toshihiko Okada

Abstract:

A conversion technology of cellulosic biomass and waste plastics to liquid fuel at low pressure and low temperature has been investigated. This study aims at the production of the liquefied fuel (CPLF) of substituting diesel oil by mixing cellulosic biomass and waste plastics in the presence of solvent. Co-liquefaction of cellulosic biomass (Japan cedar) and polypropylene (PP) using wood tar or mineral oil as solvent at 673K with an autoclave was carried out. It was confirmed that the co-liquefaction gave CPLF in a high yield among the cases of wood or of polypropylene Which was ascribed the acceleration of decomposition of plastics by radicals derived from the decomposition of wood. The co-liquefaction was also conducted by a small twin screw extruder. It was found that CPLF was obtained in the co-liquefaction, And the acceleration of decomposition of plastics in the presence of cellulosic biomass. The engine test of CPLF showed that the engine performances, Compression ignition and combustion characteristics were almost similar to those of diesel fuel at any mixing ratio of CPLF and any load, Therefore, CPLF could be practically used as alternative fuel for diesel engines.

Keywords: Cellulosic Biomass, Co-liquefaction, Solvent, Waste Plastics

Procedia PDF Downloads 338
15136 Identification of Nonlinear Systems Structured by Hammerstein-Wiener Model

Authors: A. Brouri, F. Giri, A. Mkhida, A. Elkarkri, M. L. Chhibat

Abstract:

Standard Hammerstein-Wiener models consist of a linear subsystem sandwiched by two memoryless nonlinearities. Presently, the linear subsystem is allowed to be parametric or not, continuous- or discrete-time. The input and output nonlinearities are polynomial and may be noninvertible. A two-stage identification method is developed such the parameters of all nonlinear elements are estimated first using the Kozen-Landau polynomial decomposition algorithm. The obtained estimates are then based upon in the identification of the linear subsystem, making use of suitable pre-ad post-compensators.

Keywords: nonlinear system identification, Hammerstein-Wiener systems, frequency identification, polynomial decomposition

Procedia PDF Downloads 475
15135 Application of Local Mean Decomposition for Rolling Bearing Fault Diagnosis Based On Vibration Signals

Authors: Toufik Bensana, Slimane Mekhilef, Kamel Tadjine

Abstract:

Vibration analysis has been frequently applied in the condition monitoring and fault diagnosis of rolling element bearings. Unfortunately, the vibration signals collected from a faulty bearing are generally non stationary, nonlinear and with strong noise interference, so it is essential to obtain the fault features correctly. In this paper, a novel numerical analysis method based on local mean decomposition (LMD) is proposed. LMD decompose the signal into a series of product functions (PFs), each of which is the product of an envelope signal and a purely frequency modulated FM signal. The envelope of a PF is the instantaneous amplitude (IA) and the derivative of the unwrapped phase of a purely flat frequency demodulated (FM) signal is the IF. After that the fault characteristic frequency of the roller bearing can be extracted by performing spectrum analysis to the instantaneous amplitude of PF component containing dominant fault information. The results show the effectiveness of the proposed technique in fault detection and diagnosis of rolling element bearing.

Keywords: fault diagnosis, condition monitoring, local mean decomposition, rolling element bearing, vibration analysis

Procedia PDF Downloads 359
15134 System Identification of Timber Masonry Walls Using Shaking Table Test

Authors: Timir Baran Roy, Luis Guerreiro, Ashutosh Bagchi

Abstract:

Dynamic study is important in order to design, repair and rehabilitation of structures. It has played an important role in the behavior characterization of structures; such as bridges, dams, high-rise buildings etc. There had been a substantial development in this area over the last few decades, especially in the field of dynamic identification techniques of structural systems. Frequency Domain Decomposition (FDD) and Time Domain Decomposition are most commonly used methods to identify modal parameters; such as natural frequency, modal damping, and mode shape. The focus of the present research is to study the dynamic characteristics of typical timber masonry walls commonly used in Portugal. For that purpose, a multi-storey structural prototypes of such walls have been tested on a seismic shake table at the National Laboratory for Civil Engineering, Portugal (LNEC). Signal processing has been performed of the output response, which is collected from the shaking table experiment of the prototype using accelerometers. In the present work signal processing of the output response, based on the input response has been done in two ways: FDD and Stochastic Subspace Identification (SSI). In order to estimate the values of the modal parameters, algorithms for FDD are formulated, and parametric functions for the SSI are computed. Finally, estimated values from both the methods are compared to measure the accuracy of both the techniques.

Keywords: frequency domain decomposition (fdd), modal parameters, signal processing, stochastic subspace identification (ssi), time domain decomposition

Procedia PDF Downloads 237
15133 Study of Inhibition of the End Effect Based on AR Model Predict of Combined Data Extension and Window Function

Authors: Pan Hongxia, Wang Zhenhua

Abstract:

In this paper, the EMD decomposition in the process of endpoint effect adopted data based on AR model to predict the continuation and window function method of combining the two effective inhibition. Proven by simulation of the simulation signal obtained the ideal effect, then, apply this method to the gearbox test data is also achieved good effect in the process, for the analysis of the subsequent data processing to improve the calculation accuracy. In the end, under various working conditions for the gearbox fault diagnosis laid a good foundation.

Keywords: gearbox, fault diagnosis, ar model, end effect

Procedia PDF Downloads 336
15132 Advanced Exergetic Analysis: Decomposition Method Applied to a Membrane-Based Hard Coal Oxyfuel Power Plant

Authors: Renzo Castillo, George Tsatsaronis

Abstract:

High-temperature ceramic membranes for air separation represents an important option to reduce the significant efficiency drops incurred in state-of-the-art cryogenic air separation for high tonnage oxygen production required in oxyfuel power stations. This study is focused on the thermodynamic analysis of two power plant model designs: the state-of-the-art supercritical 600ᵒC hard coal plant (reference power plant Nordrhein-Westfalen) and the membrane-based oxyfuel concept implemented in this reference plant. In the latter case, the oxygen is separated through a mixed-conducting hollow fiber perovskite membrane unit in the three-end operation mode, which has been simulated under vacuum conditions on the permeate side and at high-pressure conditions on the feed side. The thermodynamic performance of each plant concept is assessed by conventional exergetic analysis, which determines location, magnitude and sources of efficiency losses, and advanced exergetic analysis, where endogenous/exogenous and avoidable/unavoidable parts of exergy destruction are calculated at the component and full process level. These calculations identify thermodynamic interdependencies among components and reveal the real potential for efficiency improvements. The endogenous and exogenous exergy destruction portions are calculated by the decomposition method, a recently developed straightforward methodology, which is suitable for complex power stations with a large number of process components. Lastly, an improvement priority ranking for relevant components, as well as suggested changes in process layouts are presented for both power stations.

Keywords: exergy, carbon capture and storage, ceramic membranes, perovskite, oxyfuel combustion

Procedia PDF Downloads 160
15131 Effect of Tissue Preservation Chemicals on Decomposition in Different Soil Types

Authors: Onyekachi Ogbonnaya Iroanya, Taiye Abdullahi Gegele, Frank Tochukwu Egwuatu

Abstract:

Introduction: Forensic taphonomy is a multifaceted area that incorporates decomposition, chemical and biological cadaver exposure in post-mortem event chronology and reconstruction to predict the Post Mortem Interval (PMI). The aim of this study was to evaluate the integrity of DNA extracted from the remains of embalmed decomposed Sus domesticus tissues buried in different soil types. Method: A total of 12 limbs of Sus domesticus weighing between 0.7-1.4 kg were used. Each of the samples across the groups was treated with 10% formaldehyde, absolute methanol and 50% Pine oil for 24 hours before burial except the control samples, which were buried immediately. All samples were buried in shallow simulated Clay, Sandy and Loamy soil graves for 12 months. The DNA for each sample was extracted and quantified with Nanodrop Spectrophotometer (6305 JENWAY spectrometers). The rate of decomposition was examined through the modified qualitative decomposition analysis. Extracted DNA was amplified through PCR and bands visualized via gel electrophoresis. A biochemical enzyme assay was done for each burial grave soil. Result: The limbs in all burial groups had lost weight over the burial period. There was a significant increase in the soil urease level in the samples preserved in formaldehyde across the 3 soil type groups (p≤0.01). Also, the control grave soils recorded significantly higher alkaline phosphatase, dehydrogenase and calcium carbonate values compared to experimental grave soils (p≤0.01). The experimental samples showed a significant decrease in DNA concentration and purity when compared to the control groups (p≤0.01). Obtained findings of the soil biochemical analysis showed the embalming treatment altered the relationship between organic matter decomposition and soil biochemical properties as observed in the fluctuations that were recorded in the soil biochemical parameters. The PCR amplified DNA showed no bands on the gel electrophoresis plates. Conclusion: In criminal investigations, factors such as burial grave soil, grave soil biochemical properties, antemortem exposure to embalming chemicals should be considered in post-mortem interval (PMI) determination.

Keywords: forensic taphonomy, post-mortem interval (PMI), embalmment, decomposition, grave soil

Procedia PDF Downloads 134
15130 Iron Catalyst for Decomposition of Methane: Influence of Al/Si Ratio Support

Authors: A. S. Al-Fatesh, A. A. Ibrahim, A. M. AlSharekh, F. S. Alqahtani, S. O. Kasim, A. H. Fakeeha

Abstract:

Hydrogen is the expected future fuel since it produces energy without any pollution. It can be used as a fuel directly or through the fuel cell. It is also used in chemical and petrochemical industry as reducing agent or in hydrogenation processes. It is produced by different methods such as reforming of hydrocarbon, electrolytic method and methane decomposition. The objective of the present paper is to study the decomposition of methane reaction at 700°C and 800°C. The catalysts were prepared via impregnation method using 20%Fe and different proportions of combined alumina and silica support using the following ratios [100%, 90%, 80%, and 0% Al₂O₃/SiO₂]. The prepared catalysts were calcined and activated at 600 OC and 500 OC respectively. The reaction was carried out in fixed bed reactor at atmospheric pressure using 0.3g of catalyst and feed gas ratio of 1.5/1 CH₄/N₂ with a total flow rate 25 mL/min. Catalyst characterizations (TPR, TGA, BET, XRD, etc.) have been employed to study the behavior of catalysts before and after the reaction. Moreover, a brief description of the weight loss and the CH₄ conversions versus time on stream relating the different support ratios over 20%Fe/Al₂O₃/SiO₂ catalysts has been added as well. The results of TGA analysis provided higher weights losses for catalysts operated at 700°C than 800°C. For the 90% Al₂O₃/SiO₂, the activity decreases with the time on stream using 800°C reaction temperature from 73.9% initial CH₄ conversion to 46.3% for a period of 300min, whereas the activity for the same catalyst increases from 47.1% to 64.8% when 700°C reaction temperature is employed. Likewise, for 80% Al₂O₃/SiO₂ the trend of activity is similar to that of 90% Al₂O₃/SiO₂ but with a different rate of activity variation. It can be inferred from the activity results that the ratio of Al₂O₃ to SiO₂ is crucial and it is directly proportional with the activity. Whenever the Al/Si ratio decreases the activity declines. Indeed, the CH₄ conversion of 100% SiO₂ support was less than 5%.

Keywords: Al₂O₃, SiO₂, CH₄ decomposition, hydrogen, iron

Procedia PDF Downloads 147
15129 Determination of Activation Energy for Thermal Decomposition of Selected Soft Tissues Components

Authors: M. Ekiert, T. Uhl, A. Mlyniec

Abstract:

Tendons are the biological soft tissue structures composed of collagen, proteoglycan, glycoproteins, water and cells of extracellular matrix (ECM). Tendons, which primary function is to transfer force generated by the muscles to the bones causing joints movement, are exposed to many micro and macro damages. In fact, tendons and ligaments trauma are one of the most numerous injuries of human musculoskeletal system, causing for many people (particularly for athletes and physically active people), recurring disorders, chronic pain or even inability of movement. The number of tendons reconstruction and transplantation procedures is increasing every year. Therefore, studies on soft tissues storage conditions (influencing i.e. tissue aging) seem to be an extremely important issue. In this study, an atomic-scale investigation on the kinetics of decomposition of two selected tendon components – collagen type I (which forms a 60-85% of a tendon dry mass) and elastin protein (which combine with ECM creates elastic fibers of connective tissues) is presented. A molecular model of collagen and elastin was developed based on crystal structure of triple-helical collagen-like 1QSU peptide and P15502 human elastin protein, respectively. Each model employed 4 linear strands collagen/elastin strands per unit cell, distributed in 2x2 matrix arrangement, placed in simulation box filled with water molecules. A decomposition phenomena was simulated with molecular dynamics (MD) method using ReaxFF force field and periodic boundary conditions. A set of NVT-MD runs was performed for 1000K temperature range in order to obtained temperature-depended rate of production of decomposition by-products. Based on calculated reaction rates activation energies and pre-exponential factors, required to formulate Arrhenius equations describing kinetics of decomposition of tested soft tissue components, were calculated. Moreover, by adjusting a model developed for collagen, system scalability and correct implementation of the periodic boundary conditions were evaluated. An obtained results provide a deeper insight into decomposition of selected tendon components. A developed methodology may also be easily transferred to other connective tissue elements and therefore might be used for further studies on soft tissues aging.

Keywords: decomposition, molecular dynamics, soft tissue, tendons

Procedia PDF Downloads 181
15128 Using the Combination of Food Waste and Animal Waste as a Reliable Energy Source in Rural Guatemala

Authors: Jina Lee

Abstract:

Methane gas is a common byproduct in any process of rot and degradation of organic matter. This gas, when decomposition occurs, is emitted directly into the atmosphere. Methane is the simplest alkane hydrocarbon that exists. Its chemical formula is CH₄. This means that there are four atoms of hydrogen and one of carbon, which is linked by covalent bonds. Methane is found in nature in the form of gas at normal temperatures and pressures. In addition, it is colorless and odorless, despite being produced by the rot of plants. It is a non-toxic gas, and the only real danger is that of burns if it were to ignite. There are several ways to generate methane gas in homes, and the amount of methane gas generated by the decomposition of organic matter varies depending on the type of matter in question. An experiment was designed to measure the efficiency, such as a relationship between the amount of raw material and the amount of gas generated, of three different mixtures of organic matter: 1. food remains of home; 2. animal waste (excrement) 3. equal parts mixing of food debris and animal waste. The results allowed us to conclude which of the three mixtures is the one that grants the highest efficiency in methane gas generation and which would be the most suitable for methane gas generation systems for homes in order to occupy less space generating an equal amount of gas.

Keywords: alternative energy source, energy conversion, methane gas conversion system, waste management

Procedia PDF Downloads 143
15127 Comparison of Irradiance Decomposition and Energy Production Methods in a Solar Photovoltaic System

Authors: Tisciane Perpetuo e Oliveira, Dante Inga Narvaez, Marcelo Gradella Villalva

Abstract:

Installations of solar photovoltaic systems have increased considerably in the last decade. Therefore, it has been noticed that monitoring of meteorological data (solar irradiance, air temperature, wind velocity, etc.) is important to predict the potential of a given geographical area in solar energy production. In this sense, the present work compares two computational tools that are capable of estimating the energy generation of a photovoltaic system through correlation analyzes of solar radiation data: PVsyst software and an algorithm based on the PVlib package implemented in MATLAB. In order to achieve the objective, it was necessary to obtain solar radiation data (measured and from a solarimetric database), analyze the decomposition of global solar irradiance in direct normal and horizontal diffuse components, as well as analyze the modeling of the devices of a photovoltaic system (solar modules and inverters) for energy production calculations. Simulated results were compared with experimental data in order to evaluate the performance of the studied methods. Errors in estimation of energy production were less than 30% for the MATLAB algorithm and less than 20% for the PVsyst software.

Keywords: energy production, meteorological data, irradiance decomposition, solar photovoltaic system

Procedia PDF Downloads 104