Search results for: discrete Fourrier transform
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2060

Search results for: discrete Fourrier transform

1640 Rapid Discrimination of Porcine and Tilapia Fish Gelatin by Fourier Transform Infrared- Attenuated Total Reflection Combined with 2 Dimensional Infrared Correlation Analysis

Authors: Norhidayu Muhamad Zain

Abstract:

Gelatin, a purified protein derived mostly from porcine and bovine sources, is used widely in food manufacturing, pharmaceutical, and cosmetic industries. However, the presence of any porcine-related products are strictly forbidden for Muslim and Jewish consumption. Therefore, analytical methods offering reliable results to differentiate the sources of gelatin are needed. The aim of this study was to differentiate the sources of gelatin (porcine and tilapia fish) using Fourier transform infrared- attenuated total reflection (FTIR-ATR) combined with two dimensional infrared (2DIR) correlation analysis. Porcine gelatin (PG) and tilapia fish gelatin (FG) samples were diluted in distilled water at concentrations ranged from 4-20% (w/v). The samples were then analysed using FTIR-ATR and 2DIR correlation software. The results showed a significant difference in the pattern map of synchronous spectra at the region of 1000 cm⁻¹ to 1100 cm⁻¹ between PG and FG samples. The auto peak at 1080 cm⁻¹ that attributed to C-O functional group was observed at high intensity in PG samples compared to FG samples. Meanwhile, two auto peaks (1080 cm⁻¹ and 1030 cm⁻¹) at lower intensity were identified in FG samples. In addition, using 2D correlation analysis, the original broad water OH bands in 1D IR spectra can be effectively differentiated into six auto peaks located at 3630, 3340, 3230, 3065, 2950 and 2885 cm⁻¹ for PG samples and five auto peaks at 3630, 3330, 3230, 3060 and 2940 cm⁻¹ for FG samples. Based on the rule proposed by Noda, the sequence of the spectral changes in PG samples is as following: NH₃⁺ amino acid > CH₂ and CH₃ aliphatic > OH stretch > carboxylic acid OH stretch > NH in secondary amide > NH in primary amide. In contrast, the sequence was totally in the opposite direction for FG samples and thus both samples provide different 2D correlation spectra ranged from 2800 cm-1 to 3700 cm⁻¹. This method may provide a rapid determination of gelatin source for application in food, pharmaceutical, and cosmetic products.

Keywords: 2 dimensional infrared (2DIR) correlation analysis, Fourier transform infrared- attenuated total reflection (FTIR-ATR), porcine gelatin, tilapia fish gelatin

Procedia PDF Downloads 227
1639 Rock-Bed Thermocline Storage: A Numerical Analysis of Granular Bed Behavior and Interaction with Storage Tank

Authors: Nahia H. Sassine, Frédéric-Victor Donzé, Arnaud Bruch, Barthélemy Harthong

Abstract:

Thermal Energy Storage (TES) systems are central elements of various types of power plants operated using renewable energy sources. Packed bed TES can be considered as a cost–effective solution in concentrated solar power plants (CSP). Such a device is made up of a tank filled with a granular bed through which heat-transfer fluid circulates. However, in such devices, the tank might be subjected to catastrophic failure induced by a mechanical phenomenon known as thermal ratcheting. Thermal stresses are accumulated during cycles of loading and unloading until the failure happens. For instance, when rocks are used as storage material, the tank wall expands more than the solid medium during charge process, a gap is created between the rocks and tank walls and the filler material settles down to fill it. During discharge, the tank contracts against the bed, resulting in thermal stresses that may exceed the wall tank yield stress and generate plastic deformation. This phenomenon is repeated over the cycles and the tank will be slowly ratcheted outward until it fails. This paper aims at studying the evolution of tank wall stresses over granular bed thermal cycles, taking into account both thermal and mechanical loads, with a numerical model based on the discrete element method (DEM). Simulations were performed to study two different thermal configurations: (i) the tank is heated homogeneously along its height or (ii) with a vertical gradient of temperature. Then, the resulting loading stresses applied on the tank are compared as well the response of the internal granular material. Besides the study of the influence of different thermal configurations on the storage tank response, other parameters are varied, such as the internal angle of friction of the granular material, the dispersion of particles diameters as well as the tank’s dimensions. Then, their influences on the kinematics of the granular bed submitted to thermal cycles are highlighted.

Keywords: discrete element method (DEM), thermal cycles, thermal energy storage, thermocline

Procedia PDF Downloads 390
1638 Existence of Positive Solutions for Second-Order Difference Equation with Discrete Boundary Value Problem

Authors: Thanin Sitthiwirattham, Jiraporn Reunsumrit

Abstract:

We study the existence of positive solutions to the three points difference summation boundary value problem. We show the existence of at least one positive solution if f is either superlinear or sublinear by applying the fixed point theorem due to Krasnoselskii in cones.

Keywords: positive solution, boundary value problem, fixed point theorem, cone

Procedia PDF Downloads 425
1637 Fault Detection and Diagnosis of Broken Bar Problem in Induction Motors Base Wavelet Analysis and EMD Method: Case Study of Mobarakeh Steel Company in Iran

Authors: M. Ahmadi, M. Kafil, H. Ebrahimi

Abstract:

Nowadays, induction motors have a significant role in industries. Condition monitoring (CM) of this equipment has gained a remarkable importance during recent years due to huge production losses, substantial imposed costs and increases in vulnerability, risk, and uncertainty levels. Motor current signature analysis (MCSA) is one of the most important techniques in CM. This method can be used for rotor broken bars detection. Signal processing methods such as Fast Fourier transformation (FFT), Wavelet transformation and Empirical Mode Decomposition (EMD) are used for analyzing MCSA output data. In this study, these signal processing methods are used for broken bar problem detection of Mobarakeh steel company induction motors. Based on wavelet transformation method, an index for fault detection, CF, is introduced which is the variation of maximum to the mean of wavelet transformation coefficients. We find that, in the broken bar condition, the amount of CF factor is greater than the healthy condition. Based on EMD method, the energy of intrinsic mode functions (IMF) is calculated and finds that when motor bars become broken the energy of IMFs increases.

Keywords: broken bar, condition monitoring, diagnostics, empirical mode decomposition, fourier transform, wavelet transform

Procedia PDF Downloads 137
1636 Consumer Preferences when Buying Second Hand Luxury Items

Authors: K. A. Schuck, J. K. Perret, A. Mehn, K. Rommel

Abstract:

Consumers increasingly consider sustainability aspects in their consumption behavior. Although, few fashion brands are already active in the second-hand luxury market with their own online platforms. Separating between base and high-end luxury brands, two online discrete choice experiments determine the drivers behind consumers’ willingness-to-pay for platform characteristics like the type of ownership, giving brands the opportunity to elicit a financial scope they can operate within.

Keywords: choice experiment, luxury, preferences, second-hand, platform, online

Procedia PDF Downloads 112
1635 Consumer Utility Analysis of Halal Certification on Beef Using Discrete Choice Experiment: A Case Study in the Netherlands

Authors: Rosa Amalia Safitri, Ine van der Fels-Klerx, Henk Hogeveen

Abstract:

Halal is a dietary law observed by people following Islamic faith. It is considered as a type of credence food quality which cannot be easily assured by consumers even upon and after consumption. Therefore, Halal certification takes place as a practical tool for the consumers to make an informed choice particularly in a non-Muslim majority country, including the Netherlands. Discrete choice experiment (DCE) was employed in this study for its ability to assess the importance of attributes attached to Halal beef in the Dutch market and to investigate consumer utilities. Furthermore, willingness to pay (WTP) for the desired Halal certification was estimated. Four most relevant attributes were selected, i.e., the slaughter method, traceability information, place of purchase, and Halal certification. Price was incorporated as an attribute to allow estimation of willingness to pay for Halal certification. There were 242 Muslim respondents who regularly consumed Halal beef completed the survey, from Dutch (53%) and non-Dutch consumers living in the Netherlands (47%). The vast majority of the respondents (95%) were within the age of 18-45 years old, with the largest group being student (43%) followed by employee (30%) and housewife (12%). Majority of the respondents (76%) had disposable monthly income less than € 2,500, while the rest earned more than € 2,500. The respondents assessed themselves of having good knowledge of the studied attributes, except for traceability information with 62% of the respondents considered themselves not knowledgeable. The findings indicated that slaughter method was valued as the most important attribute, followed by Halal certificate, place of purchase, price, and traceability information. This order of importance varied across sociodemographic variables, except for the slaughter method. Both Dutch and non-Dutch subgroups valued Halal certification as the third most important attributes. However, non-Dutch respondents valued it with higher importance (0,20) than their Dutch counterparts (0,16). For non-Dutch, the price was more important than Halal certification. The ideal product preferred by the consumers indicated the product serving the highest utilities for consumers, and characterized by beef obtained without pre-slaughtering stunning, with traceability info, available at Halal store, certified by an official certifier, and sold at 2.75 € per 500 gr. In general, an official Halal certifier was mostly preferred. However, consumers were not willing to pay for premium for any type of Halal certifiers, indicated by negative WTP of -0.73 €, -0.93 €, and -1,03€ for small, official, and international certifiers, respectively. This finding indicated that consumers tend to lose their utility when confronted with price. WTP estimates differ across socio-demographic variables with male and non-Dutch respondents had the lowest WTP. The unfamiliarity to traceability information might cause respondents to perceive it as the least important attribute. In the context of Halal certified meat, adding traceability information into meat packaging can serve two functions, first consumers can justify for themselves whether the processes comply with Halal requirements, for example, the use of pre-slaughtering stunning, and secondly to assure its safety. Therefore, integrating traceability info into meat packaging can help to make informed decision for both Halal status and food safety.

Keywords: consumer utilities, discrete choice experiments, Halal certification, willingness to pay

Procedia PDF Downloads 111
1634 Reallocation of Bed Capacity in a Hospital Combining Discrete Event Simulation and Integer Linear Programming

Authors: Muhammed Ordu, Eren Demir, Chris Tofallis

Abstract:

The number of inpatient admissions in the UK has been significantly increasing over the past decade. These increases cause bed occupancy rates to exceed the target level (85%) set by the Department of Health in England. Therefore, hospital service managers are struggling to better manage key resource such as beds. On the other hand, this severe demand pressure might lead to confusion in wards. For example, patients can be admitted to the ward of another inpatient specialty due to lack of resources (i.e., bed). This study aims to develop a simulation-optimization model to reallocate the available number of beds in a mid-sized hospital in the UK. A hospital simulation model was developed to capture the stochastic behaviours of the hospital by taking into account the accident and emergency department, all outpatient and inpatient services, and the interactions between each other. A couple of outputs of the simulation model (e.g., average length of stay and revenue) were generated as inputs to be used in the optimization model. An integer linear programming was developed under a number of constraints (financial, demand, target level of bed occupancy rate and staffing level) with the aims of maximizing number of admitted patients. In addition, a sensitivity analysis was carried out by taking into account unexpected increases on inpatient demand over the next 12 months. As a result, the major findings of the approach proposed in this study optimally reallocate the available number of beds for each inpatient speciality and reveal that 74 beds are idle. In addition, the findings of the study indicate that the hospital wards will be able to cope with 14% demand increase at most in the projected year. In conclusion, this paper sheds a new light on how best to reallocate beds in order to cope with current and future demand for healthcare services.

Keywords: bed occupancy rate, bed reallocation, discrete event simulation, inpatient admissions, integer linear programming, projected usage

Procedia PDF Downloads 128
1633 Contribution to the Study of Automatic Epileptiform Pattern Recognition in Long Term EEG Signals

Authors: Christine F. Boos, Fernando M. Azevedo

Abstract:

Electroencephalogram (EEG) is a record of the electrical activity of the brain that has many applications, such as monitoring alertness, coma and brain death; locating damaged areas of the brain after head injury, stroke and tumor; monitoring anesthesia depth; researching physiology and sleep disorders; researching epilepsy and localizing the seizure focus. Epilepsy is a chronic condition, or a group of diseases of high prevalence, still poorly explained by science and whose diagnosis is still predominantly clinical. The EEG recording is considered an important test for epilepsy investigation and its visual analysis is very often applied for clinical confirmation of epilepsy diagnosis. Moreover, this EEG analysis can also be used to help define the types of epileptic syndrome, determine epileptiform zone, assist in the planning of drug treatment and provide additional information about the feasibility of surgical intervention. In the context of diagnosis confirmation the analysis is made using long term EEG recordings with at least 24 hours long and acquired by a minimum of 24 electrodes in which the neurophysiologists perform a thorough visual evaluation of EEG screens in search of specific electrographic patterns called epileptiform discharges. Considering that the EEG screens usually display 10 seconds of the recording, the neurophysiologist has to evaluate 360 screens per hour of EEG or a minimum of 8,640 screens per long term EEG recording. Analyzing thousands of EEG screens in search patterns that have a maximum duration of 200 ms is a very time consuming, complex and exhaustive task. Because of this, over the years several studies have proposed automated methodologies that could facilitate the neurophysiologists’ task of identifying epileptiform discharges and a large number of methodologies used neural networks for the pattern classification. One of the differences between all of these methodologies is the type of input stimuli presented to the networks, i.e., how the EEG signal is introduced in the network. Five types of input stimuli have been commonly found in literature: raw EEG signal, morphological descriptors (i.e. parameters related to the signal’s morphology), Fast Fourier Transform (FFT) spectrum, Short-Time Fourier Transform (STFT) spectrograms and Wavelet Transform features. This study evaluates the application of these five types of input stimuli and compares the classification results of neural networks that were implemented using each of these inputs. The performance of using raw signal varied between 43 and 84% efficiency. The results of FFT spectrum and STFT spectrograms were quite similar with average efficiency being 73 and 77%, respectively. The efficiency of Wavelet Transform features varied between 57 and 81% while the descriptors presented efficiency values between 62 and 93%. After simulations we could observe that the best results were achieved when either morphological descriptors or Wavelet features were used as input stimuli.

Keywords: Artificial neural network, electroencephalogram signal, pattern recognition, signal processing

Procedia PDF Downloads 512
1632 Green Synthesis of Magnetic, Silica Nanocomposite and Its Adsorptive Performance against Organochlorine Pesticides

Authors: Waleed A. El-Said, Dina M. Fouad, Mohamed H. Aly, Mohamed A. El-Gahami

Abstract:

Green synthesis of nanomaterials has received increasing attention as an eco-friendly technology in materials science. Here, we have used two types of extractions from green tea leaf (i.e. total extraction and tannin extraction) as reducing agents for a rapid, simple and one step synthesis method of mesoporous silica nanoparticles (MSNPs)/iron oxide (Fe3O4) nanocomposite based on deposition of Fe3O4 onto MSNPs. MSNPs/Fe3O4 nanocomposite were characterized by X-ray diffraction, Fourier transform infrared spectroscopy, scanning electron microscopy, energy dispersive X-ray, vibrating sample magnetometer, N2 adsorption, and high-resolution transmission electron microscopy. The average mesoporous silica particle diameter was found to be around 30 nm with high surface area (818 m2/gm). MSNPs/Fe3O4 nanocomposite was used for removing lindane pesticide (an environmental hazard material) from aqueous solutions. Fourier transform infrared, UV-vis, High-performance liquid chromatography and gas chromatography techniques were used to confirm the high ability of MSNPs/Fe3O4 nanocomposite for sensing and capture of lindane molecules with high sorption capacity (more than 89%) that could develop a new eco-friendly strategy for detection and removing of pesticide and as a promising material for water treatment application.

Keywords: green synthesis, mesoporous silica, magnetic iron oxide NPs, adsorption Lindane

Procedia PDF Downloads 419
1631 Approximation Property Pass to Free Product

Authors: Kankeyanathan Kannan

Abstract:

On approximation properties of group C* algebras is everywhere; it is powerful, important, backbone of countless breakthroughs. For a discrete group G, let A(G) denote its Fourier algebra, and let M₀A(G) denote the space of completely bounded Fourier multipliers on G. An approximate identity on G is a sequence (Φn) of finitely supported functions such that (Φn) uniformly converge to constant function 1 In this paper we prove that approximation property pass to free product.

Keywords: approximation property, weakly amenable, strong invariant approximation property, invariant approximation property

Procedia PDF Downloads 661
1630 A Double Ended AC Series Arc Fault Location Algorithm Based on Currents Estimation and a Fault Map Trace Generation

Authors: Edwin Calderon-Mendoza, Patrick Schweitzer, Serge Weber

Abstract:

Series arc faults appear frequently and unpredictably in low voltage distribution systems. Many methods have been developed to detect this type of faults and commercial protection systems such AFCI (arc fault circuit interrupter) have been used successfully in electrical networks to prevent damage and catastrophic incidents like fires. However, these devices do not allow series arc faults to be located on the line in operating mode. This paper presents a location algorithm for series arc fault in a low-voltage indoor power line in an AC 230 V-50Hz home network. The method is validated through simulations using the MATLAB software. The fault location method uses electrical parameters (resistance, inductance, capacitance, and conductance) of a 49 m indoor power line. The mathematical model of a series arc fault is based on the analysis of the V-I characteristics of the arc and consists basically of two antiparallel diodes and DC voltage sources. In a first step, the arc fault model is inserted at some different positions across the line which is modeled using lumped parameters. At both ends of the line, currents and voltages are recorded for each arc fault generation at different distances. In the second step, a fault map trace is created by using signature coefficients obtained from Kirchhoff equations which allow a virtual decoupling of the line’s mutual capacitance. Each signature coefficient obtained from the subtraction of estimated currents is calculated taking into account the Discrete Fast Fourier Transform of currents and voltages and also the fault distance value. These parameters are then substituted into Kirchhoff equations. In a third step, the same procedure described previously to calculate signature coefficients is employed but this time by considering hypothetical fault distances where the fault can appear. In this step the fault distance is unknown. The iterative calculus from Kirchhoff equations considering stepped variations of the fault distance entails the obtaining of a curve with a linear trend. Finally, the fault distance location is estimated at the intersection of two curves obtained in steps 2 and 3. The series arc fault model is validated by comparing current registered from simulation with real recorded currents. The model of the complete circuit is obtained for a 49m line with a resistive load. Also, 11 different arc fault positions are considered for the map trace generation. By carrying out the complete simulation, the performance of the method and the perspectives of the work will be presented.

Keywords: indoor power line, fault location, fault map trace, series arc fault

Procedia PDF Downloads 121
1629 Globalization as Instrument for Multi-National Corporation in Transforming Asian’s Perspective towards Clean Water Consumption

Authors: Atanta Gian

Abstract:

It is inevitable that globalization has succeeded in transforming the world today. The influence of globalization has emerged in almost every aspect of life nowadays, especially in shaping the perception of the people. It can be seen on how easy for people are affected by the information surrounding them. Due to globalization, the flow of information has become more rapid along with the development of technology. People tend to believe in information that they actually get by themselves, if there is information where most of the people believe it is true, then this information could be categorized as factual and relevant. Therefore if people gain information on what is best for them in terms of daily consumption, then this information could transform their perspective, and it becomes a consideration in selecting their needs for daily consumption. By looking at this trend, the author sees that globalization could be used by Multi-National Corporation (MNC) to enhance the promotion of their products. This is applied by shaping the perspectives of the world regarding what is the best for them. Multi-National Corporation which has better technology in terms of the development of their external promotion could utilize this opportunity to affect people’s perspectives into what they want. In this paper, the author would like to elaborate how globalization is applied by MNC to shape people’s perspective regarding what is the best for them. The author would utilize a case study to analyze on how MNC could transform the perspectives of Asian people regarding the necessary of having a better quality drinking water, which in this case, MNC has shaped the perspective of Asian people in choosing their product by promoting the bottled water as the best choice for them. In the end of this paper, author would come to a conclusion that MNCs are able to shape the world’s perspective regarding the needs of their products which is supported by the globalization that is happening now.

Keywords: consumption, globalisation, influence, information technology, multi-national corporations

Procedia PDF Downloads 197
1628 Impact of Zeolite NaY Synthesized from Kaolin on the Properties of Pyrolytic Oil Derived from Used Tire

Authors: Julius Ilawe Osayi, Peter Osifo

Abstract:

Solid waste disposal, such as used tires is a global challenge as well as energy crisis due to rising energy demand amidst price uncertainty and depleting fossil fuel reserves. Therefore, the effectiveness of pyrolysis as a disposal method that can transform used tires into liquid fuel and other end-products has made the process attractive to researchers. Although used tires have been converted to liquid fuel using pyrolysis, there is the need to improve on the liquid fuel properties. Hence, this paper reports the investigation of zeolite NaY synthesized from kaolin, a locally abundant soil material in the Benin metropolis as a suitable catalyst and its effect on the properties of pyrolytic oil produced from used tires. The pyrolysis process was conducted for a range of 1 to 10 wt.% of catalyst concentration to used tire at a temperature of 600 oC, a heating rate of 15oC/min and particle size of 6mm. Although no significant increase in pyrolytic oil yield was observed compared to the previously investigated non-catalytic pyrolysis of a used tire. However, the Fourier transform infrared (FTIR), Nuclear Magnetic Resonance (NMR); and Gas chromatography-mass spectrometry (GC-MS) characterization results revealed the pyrolytic oil to possess an improved physicochemical and fuel properties alongside valuable industrial chemical species. This confirms the possibility of transforming kaolin into a catalyst suitable for improved fuel properties of the liquid fraction obtainable from thermal cracking of hydrocarbon materials.

Keywords: catalytic pyrolysis, fossil fuel, kaolin, pyrolytic oil, used tyres, Zeolite NaY

Procedia PDF Downloads 159
1627 Computer-Aided Detection of Simultaneous Abdominal Organ CT Images by Iterative Watershed Transform

Authors: Belgherbi Aicha, Hadjidj Ismahen, Bessaid Abdelhafid

Abstract:

Interpretation of medical images benefits from anatomical and physiological priors to optimize computer-aided diagnosis applications. Segmentation of liver, spleen and kidneys is regarded as a major primary step in the computer-aided diagnosis of abdominal organ diseases. In this paper, a semi-automated method for medical image data is presented for the abdominal organ segmentation data using mathematical morphology. Our proposed method is based on hierarchical segmentation and watershed algorithm. In our approach, a powerful technique has been designed to suppress over-segmentation based on mosaic image and on the computation of the watershed transform. Our algorithm is currency in two parts. In the first, we seek to improve the quality of the gradient-mosaic image. In this step, we propose a method for improving the gradient-mosaic image by applying the anisotropic diffusion filter followed by the morphological filters. Thereafter, we proceed to the hierarchical segmentation of the liver, spleen and kidney. To validate the segmentation technique proposed, we have tested it on several images. Our segmentation approach is evaluated by comparing our results with the manual segmentation performed by an expert. The experimental results are described in the last part of this work.

Keywords: anisotropic diffusion filter, CT images, morphological filter, mosaic image, simultaneous organ segmentation, the watershed algorithm

Procedia PDF Downloads 422
1626 Sand Production Modelled with Darcy Fluid Flow Using Discrete Element Method

Authors: M. N. Nwodo, Y. P. Cheng, N. H. Minh

Abstract:

In the process of recovering oil in weak sandstone formations, the strength of sandstones around the wellbore is weakened due to the increase of effective stress/load from the completion activities around the cavity. The weakened and de-bonded sandstone may be eroded away by the produced fluid, which is termed sand production. It is one of the major trending subjects in the petroleum industry because of its significant negative impacts, as well as some observed positive impacts. For efficient sand management therefore, there has been need for a reliable study tool to understand the mechanism of sanding. One method of studying sand production is the use of the widely recognized Discrete Element Method (DEM), Particle Flow Code (PFC3D) which represents sands as granular individual elements bonded together at contact points. However, there is limited knowledge of the particle-scale behavior of the weak sandstone, and the parameters that affect sanding. This paper aims to investigate the reliability of using PFC3D and a simple Darcy flow in understanding the sand production behavior of a weak sandstone. An isotropic tri-axial test on a weak oil sandstone sample was first simulated at a confining stress of 1MPa to calibrate and validate the parallel bond models of PFC3D using a 10m height and 10m diameter solid cylindrical model. The effect of the confining stress on the number of bonds failure was studied using this cylindrical model. With the calibrated data and sample material properties obtained from the tri-axial test, simulations without and with fluid flow were carried out to check on the effect of Darcy flow on bonds failure using the same model geometry. The fluid flow network comprised of every four particles connected with tetrahedral flow pipes with a central pore or flow domain. Parametric studies included the effects of confining stress, and fluid pressure; as well as validating flow rate – permeability relationship to verify Darcy’s fluid flow law. The effect of model size scaling on sanding was also investigated using 4m height, 2m diameter model. The parallel bond model successfully calibrated the sample’s strength of 4.4MPa, showing a sharp peak strength before strain-softening, similar to the behavior of real cemented sandstones. There seems to be an exponential increasing relationship for the bigger model, but a curvilinear shape for the smaller model. The presence of the Darcy flow induced tensile forces and increased the number of broken bonds. For the parametric studies, flow rate has a linear relationship with permeability at constant pressure head. The higher the fluid flow pressure, the higher the number of broken bonds/sanding. The DEM PFC3D is a promising tool to studying the micromechanical behavior of cemented sandstones.

Keywords: discrete element method, fluid flow, parametric study, sand production/bonds failure

Procedia PDF Downloads 304
1625 Denoising Transient Electromagnetic Data

Authors: Lingerew Nebere Kassie, Ping-Yu Chang, Hsin-Hua Huang, , Chaw-Son Chen

Abstract:

Transient electromagnetic (TEM) data plays a crucial role in hydrogeological and environmental applications, providing valuable insights into geological structures and resistivity variations. However, the presence of noise often hinders the interpretation and reliability of these data. Our study addresses this issue by utilizing a FASTSNAP system for the TEM survey, which operates at different modes (low, medium, and high) with continuous adjustments to discretization, gain, and current. We employ a denoising approach that processes the raw data obtained from each acquisition mode to improve signal quality and enhance data reliability. We use a signal-averaging technique for each mode, increasing the signal-to-noise ratio. Additionally, we utilize wavelet transform to suppress noise further while preserving the integrity of the underlying signals. This approach significantly improves the data quality, notably suppressing severe noise at late times. The resulting denoised data exhibits a substantially improved signal-to-noise ratio, leading to increased accuracy in parameter estimation. By effectively denoising TEM data, our study contributes to a more reliable interpretation and analysis of underground structures. Moreover, the proposed denoising approach can be seamlessly integrated into existing ground-based TEM data processing workflows, facilitating the extraction of meaningful information from noisy measurements and enhancing the overall quality and reliability of the acquired data.

Keywords: data quality, signal averaging, transient electromagnetic, wavelet transform

Procedia PDF Downloads 73
1624 Constructions of Linear and Robust Codes Based on Wavelet Decompositions

Authors: Alla Levina, Sergey Taranov

Abstract:

The classical approach to the providing noise immunity and integrity of information that process in computing devices and communication channels is to use linear codes. Linear codes have fast and efficient algorithms of encoding and decoding information, but this codes concentrate their detect and correct abilities in certain error configurations. To protect against any configuration of errors at predetermined probability can robust codes. This is accomplished by the use of perfect nonlinear and almost perfect nonlinear functions to calculate the code redundancy. The paper presents the error-correcting coding scheme using biorthogonal wavelet transform. Wavelet transform applied in various fields of science. Some of the wavelet applications are cleaning of signal from noise, data compression, spectral analysis of the signal components. The article suggests methods for constructing linear codes based on wavelet decomposition. For developed constructions we build generator and check matrix that contain the scaling function coefficients of wavelet. Based on linear wavelet codes we develop robust codes that provide uniform protection against all errors. In article we propose two constructions of robust code. The first class of robust code is based on multiplicative inverse in finite field. In the second robust code construction the redundancy part is a cube of information part. Also, this paper investigates the characteristics of proposed robust and linear codes.

Keywords: robust code, linear code, wavelet decomposition, scaling function, error masking probability

Procedia PDF Downloads 476
1623 Reconstruction of Binary Matrices Satisfying Neighborhood Constraints by Simulated Annealing

Authors: Divyesh Patel, Tanuja Srivastava

Abstract:

This paper considers the NP-hard problem of reconstructing binary matrices satisfying exactly-1-4-adjacency constraint from its row and column projections. This problem is formulated into a maximization problem. The objective function gives a measure of adjacency constraint for the binary matrices. The maximization problem is solved by the simulated annealing algorithm and experimental results are presented.

Keywords: discrete tomography, exactly-1-4-adjacency, simulated annealing, binary matrices

Procedia PDF Downloads 391
1622 Traditional Industries Innovation and Brand Value Analysis in Taiwan: Case Study of a Certain Plastic Company

Authors: Ju Shan Lin

Abstract:

The challenges for traditional industries in Taiwan the past few years are the changes of overall domestic and foreign industry structure, the entrepreneurs not only need to keep on improving their profession skills but also continuously research and develop new products. It is also necessary for the all traditional industries to keep updating the business strategy, let the enterprises continue to progress, and won't be easily replaced by the other industries. The traditional industry in Taiwan attach great importance to the field of enterprises upgrading and innovation in recent years, by the enterprise innovation and transformation can enhance the overall business situation also enable them to obtain more additional profits than in the past. Except the original industry structure's need to transform and upgrade, the brand's business and marketing strategy are also essential. This study will take a certain plastic company as case analysis, for the brand promotion of traditional industries, brand values and business innovation model for further exploration. It will also be mentioned that the other traditional industries cases which were already achieved success on the enterprise's upgrading and innovation, at the same time, the difficulties which they faced with and the way they overcome will be explored as well. This study will use the case study method combined with expert interviews to discuss and analyze this certain plastic company's current business situation, the existing products and the possible trends in the future. Looking forward to providing an innovative business model that will enable this plastic company to upgrade its corporate image and the brand could transform successfully.

Keywords: brand marketing strategy, enterprise upgrade, industrial transformation, traditional industry

Procedia PDF Downloads 226
1621 Tracking of Linarin from the Ethyl Acetate Fraction of Melinjo (Gnetum gnemon L.) Seeds Using Preparative High Performance Liquid Chromatography

Authors: Asep Sukohar, Ramadhan Triyandi, Muhammad Iqbal, Sahidin, Suharyani

Abstract:

Introduction: Resveratrol is a class of bioactive chemicals found in melinjo, which has a wide range of biological actions. The purpose of this study is to determine the linarin content of the melinjo fraksi by using preparative-high-performance liquid chromatography (prep-HPLC). Method: Extraction used the soxhletation method with 96% ethanol solvent. Fractionation used ethyl acetate and ethanol in a ratio of 1:1. Tracing of linarin compound used prep-HPLC with a mobile phase ratio of distilled water: methanol (55: 45, v/v). The presence of linarin was detected using a wavelength of 215 nm. Fourier Transform Infrared (FTIR) was used to identify the functional groups of compound. Result: The retention time required to elute the ethyl acetate fraction was 2.601 minutes. Compound separation identification using Fourier Transform Infrared Spectroscopy - Quest Attenuated Total Reflectance (FTIR - QATR) has a similarity value range with standards from 0 to 1000. The elution results of the ethyl acetate fraction have similar values with the standard compounds linarin (668), resveratrol (578), and catechin (455). Conclusion: Tracing for active compound in the ethyl acetate fraction of Gnetum Gnemon L. using prep-HPLC showed a strong suspicion of the presence of linarin compound.

Keywords: Gnetum gnemon L., linarin, prep-HPLC, fraction ethyl acetate

Procedia PDF Downloads 80
1620 The Mechanism Study of Degradative Solvent Extraction of Biomass by Liquid Membrane-Fourier Transform Infrared Spectroscopy

Authors: W. Ketren, J. Wannapeera, Z. Heishun, A. Ryuichi, K. Toshiteru, M. Kouichi, O. Hideaki

Abstract:

Degradative solvent extraction is the method developed for biomass upgrading by dewatering and fractionation of biomass under the mild condition. However, the conversion mechanism of the degradative solvent extraction method has not been fully understood so far. The rice straw was treated in 1-methylnaphthalene (1-MN) at a different solvent-treatment temperature varied from 250 to 350 oC with the residence time for 60 min. The liquid membrane-Fourier Transform Infrared Spectroscopy (FTIR) technique is applied to study the processing mechanism in-depth without separation of the solvent. It has been found that the strength of the oxygen-hydrogen stretching  (3600-3100 cm-1) decreased slightly with increasing temperature in the range of 300-350 oC. The decrease of the hydroxyl group in the solvent soluble suggested dehydration reaction taking place between 300 and 350 oC. FTIR spectra in the carbonyl stretching region (1800-1600 cm-1) revealed the presence of esters groups, carboxylic acid and ketonic groups in the solvent-soluble of biomass. The carboxylic acid increased in the range of 200 to 250 oC and then decreased. The prevailing of aromatic groups showed that the aromatization took place during extraction at above 250 oC. From 300 to 350 oC, the carbonyl functional groups in the solvent-soluble noticeably decreased. The removal of the carboxylic acid and the decrease of esters into the form of carbon dioxide indicated that the decarboxylation reaction occurred during the extraction process.

Keywords: biomass waste, degradative solvent extraction, mechanism, upgrading

Procedia PDF Downloads 272
1619 Investigation of Contact Pressure Distribution at Expanded Polystyrene Geofoam Interfaces Using Tactile Sensors

Authors: Chen Liu, Dawit Negussey

Abstract:

EPS (Expanded Polystyrene) geofoam as light-weight material in geotechnical applications are made of pre-expanded resin beads that form fused cellular micro-structures. The strength and deformation properties of geofoam blocks are determined by unconfined compression of small test samples between rigid loading plates. Applied loads are presumed to be supported uniformly over the entire mating end areas. Predictions of field performance on the basis of such laboratory tests widely over-estimate actual post-construction settlements and exaggerate predictions of long-term creep deformations. This investigation examined the development of contact pressures at a large number of discrete points at low and large strain levels for different densities of geofoam. Development of pressure patterns for fine and coarse interface material textures as well as for molding skin and hot wire cut geofoam surfaces were examined. The lab testing showed that I-Scan tactile sensors are useful for detailed observation of contact pressures at a large number of discrete points simultaneously. At low strain level (1%), the lower density EPS block presents low variations in localized stress distribution compared to higher density EPS. At high strain level (10%), the dense geofoam reached the sensor cut-off limit. The imprint and pressure patterns for different interface textures can be distinguished with tactile sensing. The pressure sensing system can be used in many fields with real-time pressure detection. The research findings provide a better understanding of EPS geofoam behavior for improvement of design methods and performance prediction of critical infrastructures, which will be anticipated to guide future improvements in design and rapid construction of critical transportation infrastructures with geofoam in geotechnical applications.

Keywords: geofoam, pressure distribution, tactile pressure sensors, interface

Procedia PDF Downloads 158
1618 Voice Liveness Detection Using Kolmogorov Arnold Networks

Authors: Arth J. Shah, Madhu R. Kamble

Abstract:

Voice biometric liveness detection is customized to certify an authentication process of the voice data presented is genuine and not a recording or synthetic voice. With the rise of deepfakes and other equivalently sophisticated spoofing generation techniques, it’s becoming challenging to ensure that the person on the other end is a live speaker or not. Voice Liveness Detection (VLD) system is a group of security measures which detect and prevent voice spoofing attacks. Motivated by the recent development of the Kolmogorov-Arnold Network (KAN) based on the Kolmogorov-Arnold theorem, we proposed KAN for the VLD task. To date, multilayer perceptron (MLP) based classifiers have been used for the classification tasks. We aim to capture not only the compositional structure of the model but also to optimize the values of univariate functions. This study explains the mathematical as well as experimental analysis of KAN for VLD tasks, thereby opening a new perspective for scientists to work on speech and signal processing-based tasks. This study emerges as a combination of traditional signal processing tasks and new deep learning models, which further proved to be a better combination for VLD tasks. The experiments are performed on the POCO and ASVSpoof 2017 V2 database. We used Constant Q-transform, Mel, and short-time Fourier transform (STFT) based front-end features and used CNN, BiLSTM, and KAN as back-end classifiers. The best accuracy is 91.26 % on the POCO database using STFT features with the KAN classifier. In the ASVSpoof 2017 V2 database, the lowest EER we obtained was 26.42 %, using CQT features and KAN as a classifier.

Keywords: Kolmogorov Arnold networks, multilayer perceptron, pop noise, voice liveness detection

Procedia PDF Downloads 14
1617 Soil Macronutrients Sensing for Precision Agriculture Purpose Using Fourier Transform Infrared Spectroscopy

Authors: Hossein Navid, Maryam Adeli Khadem, Shahin Oustan, Mahmoud Zareie

Abstract:

Among the nutrients needed by the plants, three elements containing nitrate, phosphorus and potassium are more important. The objective of this research was measuring these nutrient amounts in soil using Fourier transform infrared spectroscopy in range of 400- 4000 cm-1. Soil samples for different soil types (sandy, clay and loam) were collected from different areas of East Azerbaijan. Three types of fertilizers in conventional farming (urea, triple superphosphate, potassium sulphate) were used for soil treatment. Each specimen was divided into two categories: The first group was used in the laboratory (direct measurement) to extract nitrate, phosphorus and potassium uptake by colorimetric method of Olsen and ammonium acetate. The second group was used to measure drug absorption spectrometry. In spectrometry, the small amount of soil samples mixed with KBr and was taken in a small pill form. For the tests, the pills were put in the center of infrared spectrometer and graphs were obtained. Analysis of data was done using MINITAB and PLSR software. The data obtained from spectrometry method were compared with amount of soil nutrients obtained from direct drug absorption using EXCEL software. There were good fitting between these two data series. For nitrate, phosphorus and potassium R2 was 79.5%, 92.0% and 81.9%, respectively. Also, results showed that the range of MIR (mid-infrared) is appropriate for determine the amount of soil nitrate and potassium and can be used in future research to obtain detailed maps of land in agricultural use.

Keywords: nitrate, phosphorus, potassium, soil nutrients, spectroscopy

Procedia PDF Downloads 382
1616 Realization of Hybrid Beams Inertial Amplifier

Authors: Somya Ranjan Patro, Abhigna Bhatt, Arnab Banerjee

Abstract:

Inertial amplifier has recently gained increasing attention as a new mechanism for vibration control of structures. Currently, theoretical investigations are undertaken by researchers to reveal its fundamentals and to understand its underline principles in altering the structural response of structures against dynamic loadings. This paper investigates experimental and analytical studies on the dynamic characteristics of hybrid beam inertial amplifier (HBIA). The analytical formulation of the HBIA has been derived by implementing the spectral element method and rigid body dynamics. This formulation gives the relation between dynamic force and the response of the structure in the frequency domain. Further, for validation of the proposed HBIA, the experiments have been performed. The experimental setup consists of a 3D printed HBIA of polylactic acid (PLA) material screwed at the base plate of the shaker system. Two numbers of accelerometers are used to study the response, one at the base plate of the shaker second one placed at the top of the inertial amplifier. A force transducer is also placed in between the base plate and the inertial amplifier to calculate the total amount of load transferred from the base plate to the inertial amplifier. The obtained time domain response from the accelerometers have been converted into the frequency domain using the Fast Fourier Transform (FFT) algorithm. The experimental transmittance values are successfully validated with the analytical results, providing us essential confidence in our proposed methodology.

Keywords: inertial amplifier, fast fourier transform, natural frequencies, polylactic acid, transmittance, vibration absorbers

Procedia PDF Downloads 83
1615 Fault Analysis of Induction Machine Using Finite Element Method (FEM)

Authors: Wiem Zaabi, Yemna Bensalem, Hafedh Trabelsi

Abstract:

The paper presents a finite element (FE) based efficient analysis procedure for induction machine (IM). The FE formulation approaches are proposed to achieve this goal: the magnetostatic and the non-linear transient time stepped formulations. The study based on finite element models offers much more information on the phenomena characterizing the operation of electrical machines than the classical analytical models. This explains the increase of the interest for the finite element investigations in electrical machines. Based on finite element models, this paper studies the influence of the stator and the rotor faults on the behavior of the IM. In this work, a simple dynamic model for an IM with inter-turn winding fault and a broken bar fault is presented. This fault model is used to study the IM under various fault conditions and severity. The simulation results are conducted to validate the fault model for different levels of fault severity. The comparison of the results obtained by simulation tests allowed verifying the precision of the proposed FEM model. This paper presents a technical method based on Fast Fourier Transform (FFT) analysis of stator current and electromagnetic torque to detect the faults of broken rotor bar. The technique used and the obtained results show clearly the possibility of extracting signatures to detect and locate faults.

Keywords: Finite element Method (FEM), Induction motor (IM), short-circuit fault, broken rotor bar, Fast Fourier Transform (FFT) analysis

Procedia PDF Downloads 282
1614 Hindi Speech Synthesis by Concatenation of Recognized Hand Written Devnagri Script Using Support Vector Machines Classifier

Authors: Saurabh Farkya, Govinda Surampudi

Abstract:

Optical Character Recognition is one of the current major research areas. This paper is focussed on recognition of Devanagari script and its sound generation. This Paper consists of two parts. First, Optical Character Recognition of Devnagari handwritten Script. Second, speech synthesis of the recognized text. This paper shows an implementation of support vector machines for the purpose of Devnagari Script recognition. The Support Vector Machines was trained with Multi Domain features; Transform Domain and Spatial Domain or Structural Domain feature. Transform Domain includes the wavelet feature of the character. Structural Domain consists of Distance Profile feature and Gradient feature. The Segmentation of the text document has been done in 3 levels-Line Segmentation, Word Segmentation, and Character Segmentation. The pre-processing of the characters has been done with the help of various Morphological operations-Otsu's Algorithm, Erosion, Dilation, Filtration and Thinning techniques. The Algorithm was tested on the self-prepared database, a collection of various handwriting. Further, Unicode was used to convert recognized Devnagari text into understandable computer document. The document so obtained is an array of codes which was used to generate digitized text and to synthesize Hindi speech. Phonemes from the self-prepared database were used to generate the speech of the scanned document using concatenation technique.

Keywords: Character Recognition (OCR), Text to Speech (TTS), Support Vector Machines (SVM), Library of Support Vector Machines (LIBSVM)

Procedia PDF Downloads 475
1613 Synchronization of a Perturbed Satellite Attitude Motion

Authors: Sadaoui Djaouida

Abstract:

In this paper, the predictive control method is proposed to control the synchronization of two perturbed satellites attitude motion. Based on delayed feedback control of continuous-time systems combines with the prediction-based method of discrete-time systems, this approach only needs a single controller to realize synchronization, which has considerable significance in reducing the cost and complexity for controller implementation.

Keywords: predictive control, synchronization, satellite attitude, control engineering

Procedia PDF Downloads 540
1612 Embedded System of Signal Processing on FPGA: Underwater Application Architecture

Authors: Abdelkader Elhanaoui, Mhamed Hadji, Rachid Skouri, Said Agounad

Abstract:

The purpose of this paper is to study the phenomenon of acoustic scattering by using a new method. The signal processing (Fast Fourier Transform FFT Inverse Fast Fourier Transform iFFT and BESSEL functions) is widely applied to obtain information with high precision accuracy. Signal processing has a wider implementation in general-purpose pro-cessors. Our interest was focused on the use of FPGAs (Field-Programmable Gate Ar-rays) in order to minimize the computational complexity in single processor architecture, then be accelerated on FPGA and meet real-time and energy efficiency requirements. Gen-eral-purpose processors are not efficient for signal processing. We implemented the acous-tic backscattered signal processing model on the Altera DE-SOC board and compared it to Odroid xu4. By comparison, the computing latency of Odroid xu4 and FPGA is 60 sec-onds and 3 seconds, respectively. The detailed SoC FPGA-based system has shown that acoustic spectra are performed up to 20 times faster than the Odroid xu4 implementation. FPGA-based system of processing algorithms is realized with an absolute error of about 10⁻³. This study underlines the increasing importance of embedded systems in underwater acoustics, especially in non-destructive testing. It is possible to obtain information related to the detection and characterization of submerged cells. So we have achieved good exper-imental results in real-time and energy efficiency.

Keywords: DE1 FPGA, acoustic scattering, form function, signal processing, non-destructive testing

Procedia PDF Downloads 62
1611 Existence and Uniqueness of Solutions to Singular Higher Order Two-Point BVPs on Time Scales

Authors: Zhenjie Liu

Abstract:

This paper investigates the existence and uniqueness of solutions for singular higher order boundary value problems on time scales by using mixed monotone method. The theorems obtained are very general. For the different time scale, the problem may be the corresponding continuous or discrete boundary value problem.

Keywords: mixed monotone operator, boundary value problem, time scale, green's function, positive solution, singularity

Procedia PDF Downloads 243