Search results for: noise reduction techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11641

Search results for: noise reduction techniques

11071 Thermal Reduction of Perfect Well Identified Hexagonal Graphene Oxide Nano-Sheets for Super-Capacitor Applications

Authors: A. N. Fouda

Abstract:

A novel well identified hexagonal graphene oxide (GO) nano-sheets were synthesized using modified Hummer method. Low temperature thermal reduction at 350°C in air ambient was performed. After thermal reduction, typical few layers of thermal reduced GO (TRGO) with dimension of few hundreds nanometers were observed using high resolution transmission electron microscopy (HRTEM). GO has a lot of structure models due to variation of the preparation process. Determining the atomic structure of GO is essential for a better understanding of its fundamental properties and for realization of the future technological applications. Structural characterization was identified by x-ray diffraction (XRD), Fourier transform infra-red spectroscopy (FTIR) measurements. A comparison between exper- imental and theoretical IR spectrum were done to confirm the match between experimentally and theoretically proposed GO structure. Partial overlap of the experimental IR spectrum with the theoretical IR was confirmed. The electrochemical properties of TRGO nano-sheets as electrode materials for supercapacitors were investigated by cyclic voltammetry and electrochemical impedance spectroscopy (EIS) measurements. An enhancement in supercapacitance after reduction was confirmed and the area of the CV curve for the TRGO electrode is larger than those for the GO electrode indicating higher specific capacitance which is promising in super-capacitor applications

Keywords: hexagonal graphene oxide, thermal reduction, cyclic voltammetry

Procedia PDF Downloads 481
11070 Methodology to Achieve Non-Cooperative Target Identification Using High Resolution Range Profiles

Authors: Olga Hernán-Vega, Patricia López-Rodríguez, David Escot-Bocanegra, Raúl Fernández-Recio, Ignacio Bravo

Abstract:

Non-Cooperative Target Identification has become a key research domain in the Defense industry since it provides the ability to recognize targets at long distance and under any weather condition. High Resolution Range Profiles, one-dimensional radar images where the reflectivity of a target is projected onto the radar line of sight, are widely used for identification of flying targets. According to that, to face this problem, an approach to Non-Cooperative Target Identification based on the exploitation of Singular Value Decomposition to a matrix of range profiles is presented. Target Identification based on one-dimensional radar images compares a collection of profiles of a given target, namely test set, with the profiles included in a pre-loaded database, namely training set. The classification is improved by using Singular Value Decomposition since it allows to model each aircraft as a subspace and to accomplish recognition in a transformed domain where the main features are easier to extract hence, reducing unwanted information such as noise. Singular Value Decomposition permits to define a signal subspace which contain the highest percentage of the energy, and a noise subspace which will be discarded. This way, only the valuable information of each target is used in the recognition process. The identification algorithm is based on finding the target that minimizes the angle between subspaces and takes place in a transformed domain. Two metrics, F1 and F2, based on Singular Value Decomposition are accomplished in the identification process. In the case of F2, the angle is weighted, since the top vectors set the importance in the contribution to the formation of a target signal, on the contrary F1 simply shows the evolution of the unweighted angle. In order to have a wide database or radar signatures and evaluate the performance, range profiles are obtained through numerical simulation of seven civil aircraft at defined trajectories taken from an actual measurement. Taking into account the nature of the datasets, the main drawback of using simulated profiles instead of actual measured profiles is that the former implies an ideal identification scenario, since measured profiles suffer from noise, clutter and other unwanted information and simulated profiles don't. In this case, the test and training samples have similar nature and usually a similar high signal-to-noise ratio, so as to assess the feasibility of the approach, the addition of noise has been considered before the creation of the test set. The identification results applying the unweighted and weighted metrics are analysed for demonstrating which algorithm provides the best robustness against noise in an actual possible scenario. So as to confirm the validity of the methodology, identification experiments of profiles coming from electromagnetic simulations are conducted, revealing promising results. Considering the dissimilarities between the test and training sets when noise is added, the recognition performance has been improved when weighting is applied. Future experiments with larger sets are expected to be conducted with the aim of finally using actual profiles as test sets in a real hostile situation.

Keywords: HRRP, NCTI, simulated/synthetic database, SVD

Procedia PDF Downloads 337
11069 A New Sign Subband Adaptive Filter Based on Dynamic Selection of Subbands

Authors: Mohammad Shams Esfand Abadi, Mehrdad Zalaghi, Reza ebrahimpour

Abstract:

In this paper, we propose a sign adaptive filter algorithm with the ability of dynamic selection of subband filters which leads to low computational complexity compared with conventional sign subband adaptive filter (SSAF) algorithm. Dynamic selection criterion is based on largest reduction of the mean square deviation at each adaption. We demonstrate that this simple proposed algorithm has the same performance of the conventional SSAF and somewhat faster than it. In the presence of impulsive interferences robustness of the simple proposed algorithm as well as the conventional SSAF and outperform the conventional normalized subband adaptive filter (NSAF) algorithm. Therefore, it is preferred for environments under impulsive interferences. Simulation results are presented to verify these above considerations very well have been achieved.

Keywords: acoustic echo cancellation (AEC), normalized subband adaptive filter (NSAF), dynamic selection subband adaptive filter (DS-NSAF), sign subband adaptive filter (SSAF), impulsive noise, robust filtering

Procedia PDF Downloads 576
11068 Variants of Mathematical Induction as Strong Proof Techniques in Theory of Computing

Authors: Ahmed Tarek, Ahmed Alveed

Abstract:

In the theory of computing, there are a wide variety of direct and indirect proof techniques. However, mathematical induction (MI) stands out to be one of the most powerful proof techniques for proving hypotheses, theorems, and new results. There are variations of mathematical induction-based proof techniques, which are broadly classified into three categories, such as structural induction (SI), weak induction (WI), and strong induction (SI). In this expository paper, several different variants of the mathematical induction techniques are explored, and the specific scenarios are discussed where a specific induction technique stands out to be more advantageous as compared to other induction strategies. Also, the essential difference among the variants of mathematical induction are explored. The points of separation among mathematical induction, recursion, and logical deduction are precisely analyzed, and the relationship among variations of recurrence relations, and mathematical induction are being explored. In this context, the application of recurrence relations, and mathematical inductions are considered together in a single framework for codewords over a given alphabet.

Keywords: alphabet, codeword, deduction, mathematical, induction, recurrence relation, strong induction, structural induction, weak induction

Procedia PDF Downloads 150
11067 3-D Strain Imaging of Nanostructures Synthesized via CVD

Authors: Sohini Manna, Jong Woo Kim, Oleg Shpyrko, Eric E. Fullerton

Abstract:

CVD techniques have emerged as a promising approach in the formation of a broad range of nanostructured materials. The realization of many practical applications will require efficient and economical synthesis techniques that preferably avoid the need for templates or costly single-crystal substrates and also afford process adaptability. Towards this end, we have developed a single-step route for the reduction-type synthesis of nanostructured Ni materials using a thermal CVD method. By tuning the CVD growth parameters, we can synthesize morphologically dissimilar nanostructures including single-crystal cubes and Au nanostructures which form atop untreated amorphous SiO2||Si substrates. An understanding of the new properties that emerge in these nanostructures materials and their relationship to function will lead to for a broad range of magnetostrictive devices as well as other catalysis, fuel cell, sensor, and battery applications based on high-surface-area transition-metal nanostructures. We use coherent X-ray diffraction imaging technique to obtain 3-D image and strain maps of individual nanocrystals. Coherent x-ray diffractive imaging (CXDI) is a technique that provides the overall shape of a nanostructure and the lattice distortion based on the combination of highly brilliant coherent x-ray sources and phase retrieval algorithm. We observe a fine interplay of reduction of surface energy vs internal stress, which plays an important role in the morphology of nano-crystals. The strain distribution is influenced by the metal-substrate interface and metal-air interface, which arise due to differences in their thermal expansion. We find the lattice strain at the surface of the octahedral gold nanocrystal agrees well with the predictions of the Young-Laplace equation quantitatively, but exhibits a discrepancy near the nanocrystal-substrate interface resulting from the interface. The strain in the bottom side of the Ni nanocube, which is contacted on the substrate surface is compressive. This is caused by dissimilar thermal expansion coefficients between Ni nanocube and Si substrate. Research at UCSD support by NSF DMR Award # 1411335.

Keywords: CVD, nanostructures, strain, CXRD

Procedia PDF Downloads 377
11066 Changes in Student Definition of De-Escalation in Professional Peace Officer Education

Authors: Pat Nelson

Abstract:

Since the release of the 21st century policing report in the United States, the techniques of de-escalation have received a lot of attention and focus in political systems, policy changes, and the media. The challenge in professional peace officer education is that there is a vast range of defining de-escalation and understanding the various techniques involved, many of which are based on popular media. This research surveyed professional peace officer education university students on their definition of de-escalation and the techniques associated with de-escalation before specific communications coursework was completed. The students were then surveyed after the communication coursework was completed to determine the changes in defining and understanding de-escalation techniques. This research has found that clearly defining de-escalation and emphasizing the broad range of techniques available enhances the students’ understanding and application of proper de-escalation. This research demonstrates the need for professional peace officer education to move students from media concepts of law enforcement to theoretical concepts.

Keywords: criminal justice education, communication theory, de-escalation, peace officer communication

Procedia PDF Downloads 146
11065 Liver Lesion Extraction with Fuzzy Thresholding in Contrast Enhanced Ultrasound Images

Authors: Abder-Rahman Ali, Adélaïde Albouy-Kissi, Manuel Grand-Brochier, Viviane Ladan-Marcus, Christine Hoeffl, Claude Marcus, Antoine Vacavant, Jean-Yves Boire

Abstract:

In this paper, we present a new segmentation approach for focal liver lesions in contrast enhanced ultrasound imaging. This approach, based on a two-cluster Fuzzy C-Means methodology, considers type-II fuzzy sets to handle uncertainty due to the image modality (presence of speckle noise, low contrast, etc.), and to calculate the optimum inter-cluster threshold. Fine boundaries are detected by a local recursive merging of ambiguous pixels. The method has been tested on a representative database. Compared to both Otsu and type-I Fuzzy C-Means techniques, the proposed method significantly reduces the segmentation errors.

Keywords: defuzzification, fuzzy clustering, image segmentation, type-II fuzzy sets

Procedia PDF Downloads 464
11064 Efficacy of Comprehensive Diabetic Care Program with the Reduction of HbA1c in Overweight Type II Diabetes Mellitus Patients: A Retrospective Study

Authors: Rohit Sane, Pravin Ghadigaonkar, Purvi Ahuja, Suvarna Tirmare, Archana Kelhe, Kranti Shinde, Rahul Mandole

Abstract:

To evaluate the efficacy of Comprehensive Diabetic Care Program with the reduction of HbA1c in overweight Diabetes Mellitus Type II patients retrospectively. Methods: Retrospective study was carried out on 34 overweight type II diabetic patients (Mean Age = 54.58 ±11.38 yrs). A total of 34 patients were enrolled after screening of 68 patients (HbA1c 7-10%). The patients were on concomitant drugs namely insulin (11.76%), DPP-4 inhibitor (17.64%), Biguanide (55.88%), Sulfonylurea (52.94%), thiazolidinedione (11.76%), other medications (20.58%) and no allopathic medications (14.70%). The patients were given Comprehensive Diabetic Care Program consisting of panchkarma procedures namely snehana (external oleation), swedana (passive heat therapy) and basti (enema), which was completed in 15 sittings. During the therapy and next 90 days, the patients followed low carbohydrate and moderate protein & fat diet. The primary endpoint of this study was the evaluation of reduction in HbA1c at the end of the follow-up after 90 days. Results: Thirty-four overweight type II diabetic patients (mean age: 54.58[±11.38], HbA1c[7-10%], 67.64% male and 32.35% female) were enrolled in the study. A significant reduction was observed in HbA1c levels (14.30%, p<0.05) at the end of the 90 days follow-up as compared to baseline. Also, BMI was reduced by 5.87%. There was reduction in the usage of the concomitant drugs namely insulin (2.94%), DPP-4 inhibitor (2.94%), Biguanide (32.35%), Sulfonylurea (35.29%), thiazolidinedione (5.88%), other medications(17.64%) and no allopathic medications (32.35%). Conclusion: The results of the study highlight not only in the reduction of HbA1c, but also in BMI and drug tapering of the CDC program in the overweight type II diabetic patients with HbA1c (7-10%).

Keywords: HbA1c, low carb diet, Panchakarma therapy, Type II Diabetes

Procedia PDF Downloads 260
11063 Novel Algorithm for Restoration of Retina Images

Authors: P. Subbuthai, S. Muruganand

Abstract:

Diabetic Retinopathy is one of the complicated diseases and it is caused by the changes in the blood vessels of the retina. Extraction of retina image through Fundus camera sometimes produced poor contrast and noises. Because of this noise, detection of blood vessels in the retina is very complicated. So preprocessing is needed, in this paper, a novel algorithm is implemented to remove the noisy pixel in the retina image. The proposed algorithm is Extended Median Filter and it is applied to the green channel of the retina because green channel vessels are brighter than the background. Proposed extended median filter is compared with the existing standard median filter by performance metrics such as PSNR, MSE and RMSE. Experimental results show that the proposed Extended Median Filter algorithm gives a better result than the existing standard median filter in terms of noise suppression and detail preservation.

Keywords: fundus retina image, diabetic retinopathy, median filter, microaneurysms, exudates

Procedia PDF Downloads 325
11062 Fold and Thrust Belts Seismic Imaging and Interpretation

Authors: Sunjay

Abstract:

Plate tectonics is of very great significance as it represents the spatial relationships of volcanic rock suites at plate margins, the distribution in space and time of the conditions of different metamorphic facies, the scheme of deformation in mountain belts, or orogens, and the association of different types of economic deposit. Orogenic belts are characterized by extensive thrust faulting, movements along large strike-slip fault zones, and extensional deformation that occur deep within continental interiors. Within oceanic areas there also are regions of crustal extension and accretion in the backarc basins that are located on the landward sides of many destructive plate margins.Collisional orogens develop where a continent or island arc collides with a continental margin as a result of subduction. collisional and noncollisional orogens can be explained by differences in the strength and rheology of the continental lithosphere and by processes that influence these properties during orogenesis.Seismic Imaging Difficulties-In triangle zones, several factors reduce the effectiveness of seismic methods. The topography in the central part of the triangle zone is usually rugged and is associated with near-surface velocity inversions which degrade the quality of the seismic image. These characteristics lead to low signal-to-noise ratio, inadequate penetration of energy through overburden, poor geophone coupling with the surface and wave scattering. Depth Seismic Imaging Techniques-Seismic processing relates to the process of altering the seismic data to suppress noise, enhancing the desired signal (higher signal-to-noise ratio) and migrating seismic events to their appropriate location in space and depth. Processing steps generally include analysis of velocities, static corrections, moveout corrections, stacking and migration. Exploration seismology Bow-tie effect -Shadow Zones-areas with no reflections (dead areas). These are called shadow zones and are common in the vicinity of faults and other discontinuous areas in the subsurface. Shadow zones result when energy from a reflector is focused on receivers that produce other traces. As a result, reflectors are not shown in their true positions. Subsurface Discontinuities-Diffractions occur at discontinuities in the subsurface such as faults and velocity discontinuities (as at “bright spot” terminations). Bow-tie effect caused by the two deep-seated synclines. Seismic imaging of thrust faults and structural damage-deepwater thrust belts, Imaging deformation in submarine thrust belts using seismic attributes,Imaging thrust and fault zones using 3D seismic image processing techniques, Balanced structural cross sections seismic interpretation pitfalls checking, The seismic pitfalls can originate due to any or all of the limitations of data acquisition, processing, interpretation of the subsurface geology,Pitfalls and limitations in seismic attribute interpretation of tectonic features, Seismic attributes are routinely used to accelerate and quantify the interpretation of tectonic features in 3D seismic data. Coherence (or variance) cubes delineate the edges of megablocks and faulted strata, curvature delineates folds and flexures, while spectral components delineate lateral changes in thickness and lithology. Carbon capture and geological storage leakage surveillance because fault behave as a seal or a conduit for hydrocarbon transportation to a trap,etc.

Keywords: tectonics, seismic imaging, fold and thrust belts, seismic interpretation

Procedia PDF Downloads 52
11061 Control of Listeria monocytogenes ATCC7644 in Fresh Tomato and Carrot with Zinc Oxide Nanoparticles

Authors: Oluwatosin A. Ijabadeniyi, Faith Semwayo

Abstract:

Preference for consumption of fresh and minimally processed fruits and vegetables continues to be on the upward trend however food-borne outbreaks related to them have also been on the increase. In this study the effect of zinc oxide nanoparticles on controlling Listeria monocytogenes ATCC 7644 in tomatoes and carrots during storage was investigated. Nutrient broth was inoculated with Listeria monocytogenes ATCC 7644 and thereafter inoculated with 0.3mg/ml nano-zinc oxide solution and 1.2mg/ml nano-zinc oxide solution and 200ppm chlorine was used as a control. Whole tomatoes and carrots were also inoculated with Listeria monocytogenes ATCC 7644 after which they were dipped into zinc oxide nanoparticle solutions and chlorine solutions. 1.2 mg/ml had a 2.40 log reduction; 0.3mg/ml nano-zinc oxide solution had a log reduction of 2.15 in the broth solution. There was however a 4.89 log and 4.46 reduction by 200 ppm chlorine in tomato and carrot respectively. Control with 0.3 mg/ml zinc oxide nanoparticles resulted in a log reduction of 5.19 in tomato and 3.66 in carrots. 1.2 mg/ml nanozinc oxide solution resulted in a 5.53 log reduction in tomato and a 4.44 log reduction in carrots. A combination of 50ppm Chlorine and 0.3 mg/ml nanozinc oxide was also used and resulted in log reductions of 5.76 and 4.84 respectively in tomatoes and carrots. Treatments were more effective in tomatoes than in carrots and the combination of 50ppm Chlorine and 0.3 mg/ml ZnO resulted in the highest log reductions in both vegetables. Statistical analysis however showed that there was no significant difference between treatments with Chlorine and nanoparticle solutions. This study therefore indicates that zinc oxide nanoparticles have the potential for use as a control agent in the fresh produce industry.

Keywords: Listeria monocytogenes, nanoparticles, tomato, carrot

Procedia PDF Downloads 487
11060 Computational and Experimental Determination of Acoustic Impedance of Internal Combustion Engine Exhaust

Authors: A. O. Glazkov, A. S. Krylova, G. G. Nadareishvili, A. S. Terenchenko, S. I. Yudin

Abstract:

The topic of the presented materials concerns the design of the exhaust system for a certain internal combustion engine. The exhaust system can be divided into two parts. The first is the engine exhaust manifold, turbocharger, and catalytic converters, which are called “hot part.” The second part is the gas exhaust system, which contains elements exclusively for reducing exhaust noise (mufflers, resonators), the accepted designation of which is the "cold part." The design of the exhaust system from the point of view of acoustics, that is, reducing the exhaust noise to a predetermined level, consists of working on the second part. Modern computer technology and software make it possible to design "cold part" with high accuracy in a given frequency range but with the condition of accurately specifying the input parameters, namely, the amplitude spectrum of the input noise and the acoustic impedance of the noise source in the form of an engine with a "hot part". Getting this data is a difficult problem: high temperatures, high exhaust gas velocities (turbulent flows), and high sound pressure levels (non-linearity mode) do not allow the calculated results to be applied with sufficient accuracy. The aim of this work is to obtain the most reliable acoustic output parameters of an engine with a "hot part" based on a complex of computational and experimental studies. The presented methodology includes several parts. The first part is a finite element simulation of the "cold part" of the exhaust system (taking into account the acoustic impedance of radiation of outlet pipe into open space) with the result in the form of the input impedance of "cold part". The second part is a finite element simulation of the "hot part" of the exhaust system (taking into account acoustic characteristics of catalytic units and geometry of turbocharger) with the result in the form of the input impedance of the "hot part". The next third part of the technique consists of the mathematical processing of the results according to the proposed formula for the convergence of the mathematical series of summation of multiple reflections of the acoustic signal "cold part" - "hot part". This is followed by conducting a set of tests on an engine stand with two high-temperature pressure sensors measuring pulsations in the nozzle between "hot part" and "cold part" of the exhaust system and subsequent processing of test results according to a well-known technique in order to separate the "incident" and "reflected" waves. The final stage consists of the mathematical processing of all calculated and experimental data to obtain a result in the form of a spectrum of the amplitude of the engine noise and its acoustic impedance.

Keywords: acoustic impedance, engine exhaust system, FEM model, test stand

Procedia PDF Downloads 38
11059 Green Aviation System: The Way Forward for Better Environment

Authors: Ramana Reddy, Vijay Kothari

Abstract:

Aircraft provide a fast, reliable mode of transport with no comparable alternative for long distance travel. Throughout the years, technology improvements have been made to aircraft and engines to make them more fuel efficient. Air traffic continues to grow around the world and needs more aircrafts to accommodate such rapid growth. This has direct consequences on two of the most important environmental factors i.e. emissions and noise. Aviation contributes about 2% of global greenhouse gas emissions. Aviation emits a number of pollutants that alter the chemical composition of the atmosphere, changing its radiative balance and hence influencing the climate. In order to reduce or if possible eliminate potential harm to the environment and also make air travel efficient and economical, an environmentally beneficial concept called “Green Aviation System” is required. This is a structured frame work with elements like innovative technologies/tools in engineering design, manufacturing, airport and fleet operations.

Keywords: air traffic, environment, emissions, noise, green aviation system

Procedia PDF Downloads 438
11058 A Review on Medical Image Registration Techniques

Authors: Shadrack Mambo, Karim Djouani, Yskandar Hamam, Barend van Wyk, Patrick Siarry

Abstract:

This paper discusses the current trends in medical image registration techniques and addresses the need to provide a solid theoretical foundation for research endeavours. Methodological analysis and synthesis of quality literature was done, providing a platform for developing a good foundation for research study in this field which is crucial in understanding the existing levels of knowledge. Research on medical image registration techniques assists clinical and medical practitioners in diagnosis of tumours and lesion in anatomical organs, thereby enhancing fast and accurate curative treatment of patients. Literature review aims to provide a solid theoretical foundation for research endeavours in image registration techniques. Developing a solid foundation for a research study is possible through a methodological analysis and synthesis of existing contributions. Out of these considerations, the aim of this paper is to enhance the scientific community’s understanding of the current status of research in medical image registration techniques and also communicate to them, the contribution of this research in the field of image processing. The gaps identified in current techniques can be closed by use of artificial neural networks that form learning systems designed to minimise error function. The paper also suggests several areas of future research in the image registration.

Keywords: image registration techniques, medical images, neural networks, optimisaztion, transformation

Procedia PDF Downloads 161
11057 Recreational Nitrous Oxide Use: Increasing Risks and Harms

Authors: Julaine Allan, Jacqui Cameron, Helen Simpson, Kenny Kor

Abstract:

The pleasurable and intoxicating effects of psychoactive substances result in widespread use. However, deaths and injuries from psychoactive substance use, particularly among young people, are a global public health problem. Understanding the benefits and problems associated with different drugs is an important part of creating contextually and physiologically relevant harm reduction strategies. Nitrous oxide use is increasing. A systematic review sought information for harm reduction strategies. The aim of this study was to systematically collate and synthesize the disparate body of research on recreational nitrous oxide use to inform harm reduction approaches tailored for young people. A mixed-methods systematic review combined quantitative data such as prevalence and incidence statistics as well as interpretive data on the experience of N₂O use. Thirty-four studies were included in the final analysis. There was minimal information available to inform policy, health care, or individuals using N₂O. The cultural, contextual, and personal reasons for N₂O use are largely unexplored.

Keywords: substance misuse, nitrous oxide, harms, harm reduction, systematic review

Procedia PDF Downloads 76
11056 Wasteless Solid-Phase Method for Conversion of Iron Ores Contaminated with Silicon and Phosphorus Compounds

Authors: А. V. Panko, Е. V. Ablets, I. G. Kovzun, М. А. Ilyashov

Abstract:

Based upon generalized analysis of modern know-how in the sphere of processing, concentration and purification of iron-ore raw materials (IORM), in particular, the most widespread ferrioxide-silicate materials (FOSM), containing impurities of phosphorus and other elements compounds, noted special role of nano technological initiatives in improvement of such processes. Considered ideas of role of nano particles in processes of FOSM carbonization with subsequent direct reduction of ferric oxides contained in them to metal phase, as well as in processes of alkali treatment and separation of powered iron from phosphorus compounds. Using the obtained results the wasteless solid-phase processing, concentration and purification of IORM and FOSM from compounds of phosphorus, silicon and other impurities excelling known methods of direct iron reduction from iron ores and metallurgical slimes.

Keywords: iron ores, solid-phase reduction, nanoparticles in reduction and purification of iron from silicon and phosphorus, wasteless method of ores processing

Procedia PDF Downloads 467
11055 Laboratory Model Tests on Encased Group Columns

Authors: Kausar Ali

Abstract:

There are several ground treatment techniques which may meet the twin objectives of increasing the bearing capacity with simultaneous reduction of settlements, but the use of stone columns is one of the most suited techniques for flexible structures such as embankments, oil storage tanks etc. that can tolerate some settlement and used worldwide. However, when the stone columns in very soft soils are loaded; stone columns undergo excessive settlement due to low lateral confinement provided by the soft soil, leading to the failure of the structure. The poor performance of stone columns under these conditions can be improved by encasing the columns with a suitable geosynthetic. In this study, the effect of reinforcement on bearing capacity of composite soil has been investigated by conducting laboratory model tests on floating and end bearing long stone columns with l/d ratio of 12. The columns were reinforced by providing geosynthetic encasement over varying column length (upper 25%, 50%, 75%, and 100% column length). In this study, a group of columns has been used instead of single column, because in the field, columns used for the purpose always remain in groups. The tests indicate that the encasement over the full column length gives higher failure stress as compared to the encasement over the partial column length for both floating and end bearing long columns. The performance of end-bearing columns was found much better than the floating columns.

Keywords: geosynthetic, ground improvement, soft clay, stone column

Procedia PDF Downloads 409
11054 Prediction of a Nanostructure Called Porphyrin-Like Buckyball, Using Density Functional Theory and Investigating Electro Catalytic Reduction of Co₂ to Co by Cobalt– Porphyrin-Like Buckyball

Authors: Mohammad Asadpour, Maryam Sadeghi, Mahmoud Jafari

Abstract:

The transformation of carbon dioxide into fuels and commodity chemicals is considered one of the most attractive methods to meet energy demands and reduce atmospheric CO₂ levels. Cobalt complexes have previously shown high faradaic efficiency in the reduction of CO₂ to CO. In this study, a nanostructure, referred to as a porphyrin-like buckyball, is simulated and analyzed for its electrical properties. The investigation aims to understand the unique characteristics of this material and its potential applications in electronic devices. Through computational simulations and analysis, the electrocatalytic reduction of CO₂ to CO by Cobalt-porphyrin-like buckyball is explored. The findings of this study offer valuable insights into the electrocatalytic properties of this predicted structure, paving the way for further research and development in the field of nanotechnology.

Keywords: porphyrin-like buckyball, DFT, nanomaterials, CO₂ to CO

Procedia PDF Downloads 14
11053 Convolutional Neural Network Based on Random Kernels for Analyzing Visual Imagery

Authors: Ja-Keoung Koo, Kensuke Nakamura, Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Byung-Woo Hong

Abstract:

The machine learning techniques based on a convolutional neural network (CNN) have been actively developed and successfully applied to a variety of image analysis tasks including reconstruction, noise reduction, resolution enhancement, segmentation, motion estimation, object recognition. The classical visual information processing that ranges from low level tasks to high level ones has been widely developed in the deep learning framework. It is generally considered as a challenging problem to derive visual interpretation from high dimensional imagery data. A CNN is a class of feed-forward artificial neural network that usually consists of deep layers the connections of which are established by a series of non-linear operations. The CNN architecture is known to be shift invariant due to its shared weights and translation invariance characteristics. However, it is often computationally intractable to optimize the network in particular with a large number of convolution layers due to a large number of unknowns to be optimized with respect to the training set that is generally required to be large enough to effectively generalize the model under consideration. It is also necessary to limit the size of convolution kernels due to the computational expense despite of the recent development of effective parallel processing machinery, which leads to the use of the constantly small size of the convolution kernels throughout the deep CNN architecture. However, it is often desired to consider different scales in the analysis of visual features at different layers in the network. Thus, we propose a CNN model where different sizes of the convolution kernels are applied at each layer based on the random projection. We apply random filters with varying sizes and associate the filter responses with scalar weights that correspond to the standard deviation of the random filters. We are allowed to use large number of random filters with the cost of one scalar unknown for each filter. The computational cost in the back-propagation procedure does not increase with the larger size of the filters even though the additional computational cost is required in the computation of convolution in the feed-forward procedure. The use of random kernels with varying sizes allows to effectively analyze image features at multiple scales leading to a better generalization. The robustness and effectiveness of the proposed CNN based on random kernels are demonstrated by numerical experiments where the quantitative comparison of the well-known CNN architectures and our models that simply replace the convolution kernels with the random filters is performed. The experimental results indicate that our model achieves better performance with less number of unknown weights. The proposed algorithm has a high potential in the application of a variety of visual tasks based on the CNN framework. Acknowledgement—This work was supported by the MISP (Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by IITP, and NRF-2014R1A2A1A11051941, NRF2017R1A2B4006023.

Keywords: deep learning, convolutional neural network, random kernel, random projection, dimensionality reduction, object recognition

Procedia PDF Downloads 266
11052 Authentication of Physical Objects with Dot-Based 2D Code

Authors: Michał Glet, Kamil Kaczyński

Abstract:

Counterfeit goods and documents are a global problem, which needs more and more sophisticated methods of resolving it. Existing techniques using watermarking or embedding symbols on objects are not suitable for all use cases. To address those special needs, we created complete system allowing authentication of paper documents and physical objects with flat surface. Objects are marked using orientation independent and resistant to camera noise 2D graphic codes, named DotAuth. Based on the identifier stored in 2D code, the system is able to perform basic authentication and allows to conduct more sophisticated analysis methods, e.g., relying on augmented reality and physical properties of the object. In this paper, we present the complete architecture, algorithms and applications of the proposed system. Results of the features comparison of the proposed solution and other products are presented as well, pointing to the existence of many advantages that increase usability and efficiency in the means of protecting physical objects.

Keywords: anti-forgery, authentication, paper documents, security

Procedia PDF Downloads 116
11051 E-Learning Approaches Based on Artificial Intelligence Techniques: A Survey

Authors: Nabila Daly, Hamdi Ellouzi, Hela Ltifi

Abstract:

In last year’s, several recent researches’ that focus on e-learning approaches having as goal to improve pedagogy and student’s academy level assessment. E-learning-related works have become an important research file nowadays due to several problems that make it impossible for students join classrooms, especially in last year’s. Among those problems, we note the current epidemic problems in the word case of Covid-19. For those reasons, several e-learning-related works based on Artificial Intelligence techniques are proposed to improve distant education targets. In the current paper, we will present a short survey of the most relevant e-learning based on Artificial Intelligence techniques giving birth to newly developed e-learning tools that rely on new technologies.

Keywords: artificial intelligence techniques, decision, e-learning, support system, survey

Procedia PDF Downloads 203
11050 Analyzing Tools and Techniques for Classification In Educational Data Mining: A Survey

Authors: D. I. George Amalarethinam, A. Emima

Abstract:

Educational Data Mining (EDM) is one of the newest topics to emerge in recent years, and it is concerned with developing methods for analyzing various types of data gathered from the educational circle. EDM methods and techniques with machine learning algorithms are used to extract meaningful and usable information from huge databases. For scientists and researchers, realistic applications of Machine Learning in the EDM sectors offer new frontiers and present new problems. One of the most important research areas in EDM is predicting student success. The prediction algorithms and techniques must be developed to forecast students' performance, which aids the tutor, institution to boost the level of student’s performance. This paper examines various classification techniques in prediction methods and data mining tools used in EDM.

Keywords: classification technique, data mining, EDM methods, prediction methods

Procedia PDF Downloads 105
11049 Numerical Analysis of Flow in the Gap between a Simplified Tractor-Trailer Model and Cross Vortex Trap Device

Authors: Terrance Charles, Zhiyin Yang, Yiling Lu

Abstract:

Heavy trucks are aerodynamically inefficient due to their un-streamlined body shapes, leading to more than of 60% engine power being required to overcome the aerodynamics drag at 60 m/hr. There are many aerodynamics drag reduction devices developed and this paper presents a study on a drag reduction device called Cross Vortex Trap Device (CVTD) deployed in the gap between the tractor and the trailer of a simplified tractor-trailer model. Numerical simulations have been carried out at Reynolds number 0.51×106 based on inlet flow velocity and height of the trailer using the Reynolds-Averaged Navier-Stokes (RANS) approach. Three different configurations of CVTD have been studied, ranging from single to three slabs, equally spaced on the front face of the trailer. Flow field around three different configurations of trap device have been analysed and presented. The results show that a maximum of 12.25% drag reduction can be achieved when a triple vortex trap device is used. Detailed flow field analysis along with pressure contours are presented to elucidate the drag reduction mechanisms of CVTD and why the triple vortex trap configuration produces the maximum drag reduction among the three configurations tested.

Keywords: aerodynamic drag, cross vortex trap device, truck, Reynolds-Averaged Navier-Stokes, RANS

Procedia PDF Downloads 120
11048 Aeroacoustics Investigations of Unsteady 3D Airfoil for Different Angle Using Computational Fluid Dynamics Software

Authors: Haydar Kepekçi, Baha Zafer, Hasan Rıza Güven

Abstract:

Noise disturbance is one of the major factors considered in the fast development of aircraft technology. This paper reviews the flow field, which is examined on the 2D NACA0015 and 3D NACA0012 blade profile using SST k-ω turbulence model to compute the unsteady flow field. We inserted the time-dependent flow area variables in Ffowcs-Williams and Hawkings (FW-H) equations as an input and Sound Pressure Level (SPL) values will be computed for different angles of attack (AoA) from the microphone which is positioned in the computational domain to investigate effect of augmentation of unsteady 2D and 3D airfoil region noise level. The computed results will be compared with experimental data which are available in the open literature. As results; one of the calculated Cp is slightly lower than the experimental value. This difference could be due to the higher Reynolds number of the experimental data. The ANSYS Fluent software was used in this study. Fluent includes well-validated physical modeling capabilities to deliver fast, accurate results across the widest range of CFD and multiphysics applications. This paper includes a study which is on external flow over an airfoil. The case of 2D NACA0015 has approximately 7 million elements and solves compressible fluid flow with heat transfer using the SST turbulence model. The other case of 3D NACA0012 has approximately 3 million elements.

Keywords: 3D blade profile, noise disturbance, aeroacoustics, Ffowcs-Williams and Hawkings (FW-H) equations, k-ω-SST turbulence model

Procedia PDF Downloads 191
11047 Learning outside the Box by Using Memory Techniques Skill: Case Study in Indonesia Memory Sports Council

Authors: Muhammad Fajar Suardi, Fathimatufzzahra, Dela Isnaini Sendra

Abstract:

Learning is an activity that has been used to do, especially for a student or academics. But a handful of people have not been using and maximizing their brains work and some also do not know a good brain work time in capturing the lessons, so that knowledge is absorbed is also less than the maximum. Indonesia Memory Sports Council (IMSC) is an institution which is engaged in the performance of the brain and the development of effective learning methods by using several techniques that can be used in considering the lessons and knowledge to grasp well, including: loci method, substitution method, and chain method. This study aims to determine the techniques and benefits of using the method given in learning and memorization by applying memory techniques taught by Indonesia Memory Sports Council (IMSC) to students and the difference if not using this method. This research uses quantitative research with survey method addressed to students of Indonesian Memory Sports Council (IMSC). The results of this study indicate that learn, understand and remember the lesson using the techniques of memory which is taught in Indonesia Memory Sport Council is very effective and faster to absorb the lesson than learning without using the techniques of memory, and this affects the academic achievement of students in each educational institution.

Keywords: chain method, Indonesia memory sports council, loci method, substitution method

Procedia PDF Downloads 272
11046 ICanny: CNN Modulation Recognition Algorithm

Authors: Jingpeng Gao, Xinrui Mao, Zhibin Deng

Abstract:

Aiming at the low recognition rate on the composite signal modulation in low signal to noise ratio (SNR), this paper proposes a modulation recognition algorithm based on ICanny-CNN. Firstly, the radar signal is transformed into the time-frequency image by Choi-Williams Distribution (CWD). Secondly, we propose an image processing algorithm using the Guided Filter and the threshold selection method, which is combined with the hole filling and the mask operation. Finally, the shallow convolutional neural network (CNN) is combined with the idea of the depth-wise convolution (Dw Conv) and the point-wise convolution (Pw Conv). The proposed CNN is designed to complete image classification and realize modulation recognition of radar signal. The simulation results show that the proposed algorithm can reach 90.83% at 0dB and 71.52% at -8dB. Therefore, the proposed algorithm has a good classification and anti-noise performance in radar signal modulation recognition and other fields.

Keywords: modulation recognition, image processing, composite signal, improved Canny algorithm

Procedia PDF Downloads 175
11045 Geophysical Methods and Machine Learning Algorithms for Stuck Pipe Prediction and Avoidance

Authors: Ammar Alali, Mahmoud Abughaban

Abstract:

Cost reduction and drilling optimization is the goal of many drilling operators. Historically, stuck pipe incidents were a major segment of non-productive time (NPT) associated costs. Traditionally, stuck pipe problems are part of the operations and solved post-sticking. However, the real key to savings and success is in predicting the stuck pipe incidents and avoiding the conditions leading to its occurrences. Previous attempts in stuck-pipe predictions have neglected the local geology of the problem. The proposed predictive tool utilizes geophysical data processing techniques and Machine Learning (ML) algorithms to predict drilling activities events in real-time using surface drilling data with minimum computational power. The method combines two types of analysis: (1) real-time prediction, and (2) cause analysis. Real-time prediction aggregates the input data, including historical drilling surface data, geological formation tops, and petrophysical data, from wells within the same field. The input data are then flattened per the geological formation and stacked per stuck-pipe incidents. The algorithm uses two physical methods (stacking and flattening) to filter any noise in the signature and create a robust pre-determined pilot that adheres to the local geology. Once the drilling operation starts, the Wellsite Information Transfer Standard Markup Language (WITSML) live surface data are fed into a matrix and aggregated in a similar frequency as the pre-determined signature. Then, the matrix is correlated with the pre-determined stuck-pipe signature for this field, in real-time. The correlation used is a machine learning Correlation-based Feature Selection (CFS) algorithm, which selects relevant features from the class and identifying redundant features. The correlation output is interpreted as a probability curve of stuck pipe incidents prediction in real-time. Once this probability passes a fixed-threshold defined by the user, the other component, cause analysis, alerts the user of the expected incident based on set pre-determined signatures. A set of recommendations will be provided to reduce the associated risk. The validation process involved feeding of historical drilling data as live-stream, mimicking actual drilling conditions, of an onshore oil field. Pre-determined signatures were created for three problematic geological formations in this field prior. Three wells were processed as case studies, and the stuck-pipe incidents were predicted successfully, with an accuracy of 76%. This accuracy of detection could have resulted in around 50% reduction in NPT, equivalent to 9% cost saving in comparison with offset wells. The prediction of stuck pipe problem requires a method to capture geological, geophysical and drilling data, and recognize the indicators of this issue at a field and geological formation level. This paper illustrates the efficiency and the robustness of the proposed cross-disciplinary approach in its ability to produce such signatures and predicting this NPT event.

Keywords: drilling optimization, hazard prediction, machine learning, stuck pipe

Procedia PDF Downloads 204
11044 High-Efficiency Comparator for Low-Power Application

Authors: M. Yousefi, N. Nasirzadeh

Abstract:

In this paper, dynamic comparator structure employing two methods for power consumption reduction with applications in low-power high-speed analog-to-digital converters have been presented. The proposed comparator has low consumption thanks to power reduction methods. They have the ability for offset adjustment. The comparator consumes 14.3 μW at 100 MHz which is equal to 11.8 fJ. The comparator has been designed and simulated in 180 nm CMOS. Layouts occupy 210 μm2.

Keywords: efficiency, comparator, power, low

Procedia PDF Downloads 338
11043 An Open-Label Phase I Clinical Study: Safety, Tolerability and Pharmacodynamics of Mutant Collagenase Injection in Adults for Localized Fat Reduction

Authors: Yong Cang

Abstract:

RJV001 is a subcutaneous injection containing mutated recombinant Collagenase H (ColH), leading to disruption of collagen matrix in adipose tissue and programmed cell death of adipocytes. Here we reported our clinical investigation of the safety, tolerance and pharmacodynamics of localized RJV001 injection into healthy human abdominal fat tissues (NCT04821648, Arizona Research Center). Investigate the safety, tolerance and clinical pharmacodynamics of subcutaneous RJV001 in humans. In the dose-escalating study, 18 subjects completed the study, 100% female, 78% white, with a mean age of 42[±9.9]. All three tested dose (0.05, 0.075 and 0.15 mg/injection), up to 30 injections, were safe and well-tolerated. Bruising and tenderness to palpation, mild to moderate, were the most frequent local skin reactions but nearly all resolved within 30 days. Additionally, physician-monitored ultrasound measurement showed that a reduction in abdominal fat tissue thickness was consistently observed in Cohort C (0.075, 0.15 mg/injection, 30injections), with a mean reduction of 7.37 [± 2.020] mm. Based on this clinical study, RJV001 has been advanced to phase II clinical studies. In the dose-escalating study, subcutaneously administered RJV001 was safe and well-tolerated in healthy adults up to 0.15 mg/injection, 30 injections. Fat reduction and adipocytolysis were observed by ultrasound measurements and histological analysis for exploratory purposes.

Keywords: fat reduction, mutant collagenase, clinical trial, subcutaneous injection

Procedia PDF Downloads 89
11042 A High Performance Piano Note Recognition Scheme via Precise Onset Detection and Segmented Short-Time Fourier Transform

Authors: Sonali Banrjee, Swarup Kumar Mitra, Aritra Acharyya

Abstract:

A piano note recognition method has been proposed by the authors in this paper. The authors have used a comprehensive method for onset detection of each note present in a piano piece followed by segmented short-time Fourier transform (STFT) for the identification of piano notes. The performance evaluation of the proposed method has been carried out in different harsh noisy environments by adding different levels of additive white Gaussian noise (AWGN) having different signal-to-noise ratio (SNR) in the original signal and evaluating the note detection error rate (NDER) of different piano pieces consisting of different number of notes at different SNR levels. The NDER is found to be remained within 15% for all piano pieces under consideration when the SNR is kept above 8 dB.

Keywords: AWGN, onset detection, piano note, STFT

Procedia PDF Downloads 150