Search results for: discrete cosine transform
1752 An Integrated Approach to the Carbonate Reservoir Modeling: Case Study of the Eastern Siberia Field
Authors: Yana Snegireva
Abstract:
Carbonate reservoirs are known for their heterogeneity, resulting from various geological processes such as diagenesis and fracturing. These complexities may cause great challenges in understanding fluid flow behavior and predicting the production performance of naturally fractured reservoirs. The investigation of carbonate reservoirs is crucial, as many petroleum reservoirs are naturally fractured, which can be difficult due to the complexity of their fracture networks. This can lead to geological uncertainties, which are important for global petroleum reserves. The problem outlines the key challenges in carbonate reservoir modeling, including the accurate representation of fractures and their connectivity, as well as capturing the impact of fractures on fluid flow and production. Traditional reservoir modeling techniques often oversimplify fracture networks, leading to inaccurate predictions. Therefore, there is a need for a modern approach that can capture the complexities of carbonate reservoirs and provide reliable predictions for effective reservoir management and production optimization. The modern approach to carbonate reservoir modeling involves the utilization of the hybrid fracture modeling approach, including the discrete fracture network (DFN) method and implicit fracture network, which offer enhanced accuracy and reliability in characterizing complex fracture systems within these reservoirs. This study focuses on the application of the hybrid method in the Nepsko-Botuobinskaya anticline of the Eastern Siberia field, aiming to prove the appropriateness of this method in these geological conditions. The DFN method is adopted to model the fracture network within the carbonate reservoir. This method considers fractures as discrete entities, capturing their geometry, orientation, and connectivity. But the method has significant disadvantages since the number of fractures in the field can be very high. Due to limitations in the amount of main memory, it is very difficult to represent these fractures explicitly. By integrating data from image logs (formation micro imager), core data, and fracture density logs, a discrete fracture network (DFN) model can be constructed to represent fracture characteristics for hydraulically relevant fractures. The results obtained from the DFN modeling approaches provide valuable insights into the East Siberia field's carbonate reservoir behavior. The DFN model accurately captures the fracture system, allowing for a better understanding of fluid flow pathways, connectivity, and potential production zones. The analysis of simulation results enables the identification of zones of increased fracturing and optimization opportunities for reservoir development with the potential application of enhanced oil recovery techniques, which were considered in further simulations on the dual porosity and dual permeability models. This approach considers fractures as separate, interconnected flow paths within the reservoir matrix, allowing for the characterization of dual-porosity media. The case study of the East Siberia field demonstrates the effectiveness of the hybrid model method in accurately representing fracture systems and predicting reservoir behavior. The findings from this study contribute to improved reservoir management and production optimization in carbonate reservoirs with the use of enhanced and improved oil recovery methods.Keywords: carbonate reservoir, discrete fracture network, fracture modeling, dual porosity, enhanced oil recovery, implicit fracture model, hybrid fracture model
Procedia PDF Downloads 761751 Statically Fused Unbiased Converted Measurements Kalman Filter
Authors: Zhengkun Guo, Yanbin Li, Wenqing Wang, Bo Zou
Abstract:
The statically fused converted position and doppler measurements Kalman filter (SF-CMKF) with additive debiased measurement conversion has been previously presented to combine the resulting states of converted position measurements Kalman filter (CPMKF) and converted doppler measurement Kalman filter (CDMKF) to yield the final state estimates under minimum mean squared error (MMSE) criterion. However, the exact compensation for the bias in the polar-to-cartesian and spherical-to-cartesian conversion are multiplicative and depend on the statistics of the cosine of the angle measurement errors. As a result, the consistency and performance of the SF-CMKF may be suboptimal in large-angle error situations. In this paper, the multiplicative unbiased position and Doppler measurement conversion for 2D (polar-to-cartesian) tracking are derived, and the SF-CMKF is improved to use those conversions. Monte Carlo simulations are presented to demonstrate the statistical consistency of the multiplicative unbiased conversion and the superior performance of the modified SF-CMKF (SF-UCMKF).Keywords: measurement conversion, Doppler, Kalman filter, estimation, tracking
Procedia PDF Downloads 2091750 A Speeded up Robust Scale-Invariant Feature Transform Currency Recognition Algorithm
Authors: Daliyah S. Aljutaili, Redna A. Almutlaq, Suha A. Alharbi, Dina M. Ibrahim
Abstract:
All currencies around the world look very different from each other. For instance, the size, color, and pattern of the paper are different. With the development of modern banking services, automatic methods for paper currency recognition become important in many applications like vending machines. One of the currency recognition architecture’s phases is Feature detection and description. There are many algorithms that are used for this phase, but they still have some disadvantages. This paper proposes a feature detection algorithm, which merges the advantages given in the current SIFT and SURF algorithms, which we call, Speeded up Robust Scale-Invariant Feature Transform (SR-SIFT) algorithm. Our proposed SR-SIFT algorithm overcomes the problems of both the SIFT and SURF algorithms. The proposed algorithm aims to speed up the SIFT feature detection algorithm and keep it robust. Simulation results demonstrate that the proposed SR-SIFT algorithm decreases the average response time, especially in small and minimum number of best key points, increases the distribution of the number of best key points on the surface of the currency. Furthermore, the proposed algorithm increases the accuracy of the true best point distribution inside the currency edge than the other two algorithms.Keywords: currency recognition, feature detection and description, SIFT algorithm, SURF algorithm, speeded up and robust features
Procedia PDF Downloads 2351749 Classification on Statistical Distributions of a Complex N-Body System
Authors: David C. Ni
Abstract:
Contemporary models for N-body systems are based on temporal, two-body, and mass point representation of Newtonian mechanics. Other mainstream models include 2D and 3D Ising models based on local neighborhood the lattice structures. In Quantum mechanics, the theories of collective modes are for superconductivity and for the long-range quantum entanglement. However, these models are still mainly for the specific phenomena with a set of designated parameters. We are therefore motivated to develop a new construction directly from the complex-variable N-body systems based on the extended Blaschke functions (EBF), which represent a non-temporal and nonlinear extension of Lorentz transformation on the complex plane – the normalized momentum spaces. A point on the complex plane represents a normalized state of particle momentums observed from a reference frame in the theory of special relativity. There are only two key parameters, normalized momentum and nonlinearity for modelling. An algorithm similar to Jenkins-Traub method is adopted for solving EBF iteratively. Through iteration, the solution sets show a form of σ + i [-t, t], where σ and t are the real numbers, and the [-t, t] shows various distributions, such as 1-peak, 2-peak, and 3-peak etc. distributions and some of them are analog to the canonical distributions. The results of the numerical analysis demonstrate continuum-to-discreteness transitions, evolutional invariance of distributions, phase transitions with conjugate symmetry, etc., which manifest the construction as a potential candidate for the unification of statistics. We hereby classify the observed distributions on the finite convergent domains. Continuous and discrete distributions both exist and are predictable for given partitions in different regions of parameter-pair. We further compare these distributions with canonical distributions and address the impacts on the existing applications.Keywords: blaschke, lorentz transformation, complex variables, continuous, discrete, canonical, classification
Procedia PDF Downloads 3111748 Hardware Implementation of Local Binary Pattern Based Two-Bit Transform Motion Estimation
Authors: Seda Yavuz, Anıl Çelebi, Aysun Taşyapı Çelebi, Oğuzhan Urhan
Abstract:
Nowadays, demand for using real-time video transmission capable devices is ever-increasing. So, high resolution videos have made efficient video compression techniques an essential component for capturing and transmitting video data. Motion estimation has a critical role in encoding raw video. Hence, various motion estimation methods are introduced to efficiently compress the video. Low bit‑depth representation based motion estimation methods facilitate computation of matching criteria and thus, provide small hardware footprint. In this paper, a hardware implementation of a two-bit transformation based low-complexity motion estimation method using local binary pattern approach is proposed. Image frames are represented in two-bit depth instead of full-depth by making use of the local binary pattern as a binarization approach and the binarization part of the hardware architecture is explained in detail. Experimental results demonstrate the difference between the proposed hardware architecture and the architectures of well-known low-complexity motion estimation methods in terms of important aspects such as resource utilization, energy and power consumption.Keywords: binarization, hardware architecture, local binary pattern, motion estimation, two-bit transform
Procedia PDF Downloads 3141747 Spatial Interpolation of Intermediate Soil Properties to Enhance Geotechnical Surveying for Foundation Design
Authors: Yelbek B. Utepov, Assel T. Mukhamejanova, Aliya K. Aldungarova, Aida G. Nazarova, Sabit A. Karaulov, Nurgul T. Alibekova, Aigul K. Kozhas, Dias Kazhimkanuly, Akmaral K. Tleubayeva
Abstract:
This research focuses on enhancing geotechnical surveying for foundation design through the spatial interpolation of intermediate soil properties. Traditional geotechnical practices rely on discrete data from borehole drilling, soil sampling, and laboratory analyses, often neglecting the continuous nature of soil properties and disregarding values in intermediate locations. This study challenges these omissions by emphasizing interpolation techniques such as Kriging, Inverse Distance Weighting, and Spline interpolation to capture the nuanced spatial variations in soil properties. The methodology is applied to geotechnical survey data from two construction sites in Astana, Kazakhstan, revealing continuous representations of Young's Modulus, Cohesion, and Friction Angle. The spatial heatmaps generated through interpolation offered valuable insights into the subsurface environment, highlighting heterogeneity and aiding in more informed foundation design decisions for considered cites. Moreover, intriguing patterns of heterogeneity, as well as visual clusters and transitions between soil classes, were explored within seemingly uniform layers. The study bridges the gap between discrete borehole samples and the continuous subsurface, contributing to the evolution of geotechnical engineering practices. The proposed approach, utilizing open-source software geographic information systems, provides a practical tool for visualizing soil characteristics and may pave the way for future advancements in geotechnical surveying and foundation design.Keywords: soil mechanical properties, spatial interpolation, inverse distance weighting, heatmaps
Procedia PDF Downloads 861746 Solution to Riemann Hypothesis Critical Strip Zone Using Non-Linear Complex Variable Functions
Authors: Manojkumar Sabanayagam
Abstract:
The Riemann hypothesis is an unsolved millennium problem and the search for a solution to the Riemann hypothesis is to study the pattern of prime number distribution. The scope of this paper is to identify the solution for the critical strip and the critical line axis, which has the non-trivial zero solutions using complex plane functions. The Riemann graphical plot is constructed using a linear complex variable function (X+iY) and is applicable only when X>1. But the investigation shows that complex variable behavior has two zones. The first zone is the transformation zone, where the definition of the complex plane should be a non-linear variable which is the critical strip zone in the graph (X=0 to 1). The second zone is the transformed zone (X>1) defined using linear variables conventionally. This paper deals with the Non-linear function in the transformation zone derived using cosine and sinusoidal time lag w.r.t imaginary number ‘i’. The alternate complex variable (Cosθ+i Sinθ) is used to understand the variables in the critical strip zone. It is concluded that the non-trivial zeros present in the Real part 0.5 are because the linear function is not the correct approach in the critical strip. This paper provides the solution to Reimann's hypothesis.Keywords: Reimann hypothesis, critical strip, complex plane, transformation zone
Procedia PDF Downloads 2081745 Design of a Real Time Heart Sounds Recognition System
Authors: Omer Abdalla Ishag, Magdi Baker Amien
Abstract:
Physicians used the stethoscope for listening patient heart sounds in order to make a diagnosis. However, the determination of heart conditions by acoustic stethoscope is a difficult task so it requires special training of medical staff. This study developed an accurate model for analyzing the phonocardiograph signal based on PC and DSP processor. The system has been realized into two phases; offline and real time phase. In offline phase, 30 cases of heart sounds files were collected from medical students and doctor's world website. For experimental phase (real time), an electronic stethoscope has been designed, implemented and recorded signals from 30 volunteers, 17 were normal cases and 13 were various pathologies cases, these acquired 30 signals were preprocessed using an adaptive filter to remove lung sounds. The background noise has been removed from both offline and real data, using wavelet transform, then graphical and statistics features vector elements were extracted, finally a look-up table was used for classification heart sounds cases. The obtained results of the implemented system showed accuracy of 90%, 80% and sensitivity of 87.5%, 82.4% for offline data, and real data respectively. The whole system has been designed on TMS320VC5509a DSP Platform.Keywords: code composer studio, heart sounds, phonocardiograph, wavelet transform
Procedia PDF Downloads 4481744 Decision Support System for Hospital Selection in Emergency Medical Services: A Discrete Event Simulation Approach
Authors: D. Tedesco, G. Feletti, P. Trucco
Abstract:
The present study aims to develop a Decision Support System (DSS) to support the operational decision of the Emergency Medical Service (EMS) regarding the assignment of medical emergency requests to Emergency Departments (ED). In the literature, this problem is also known as “hospital selection” and concerns the definition of policies for the selection of the ED to which patients who require further treatment are transported by ambulance. The employed research methodology consists of the first phase of revision of the technical-scientific literature concerning DSSs to support the EMS management and, in particular, the hospital selection decision. From the literature analysis, it emerged that current studies are mainly focused on the EMS phases related to the ambulance service and consider a process that ends when the ambulance is available after completing a request. Therefore, all the ED-related issues are excluded and considered as part of a separate process. Indeed, the most studied hospital selection policy turned out to be proximity, thus allowing to minimize the transport time and release the ambulance in the shortest possible time. The purpose of the present study consists in developing an optimization model for assigning medical emergency requests to the EDs, considering information relating to the subsequent phases of the process, such as the case-mix, the expected service throughput times, and the operational capacity of different EDs in hospitals. To this end, a Discrete Event Simulation (DES) model was created to evaluate different hospital selection policies. Therefore, the next steps of the research consisted of the development of a general simulation architecture, its implementation in the AnyLogic software and its validation on a realistic dataset. The hospital selection policy that produced the best results was the minimization of the Time To Provider (TTP), considered as the time from the beginning of the ambulance journey to the ED at the beginning of the clinical evaluation by the doctor. Finally, two approaches were further compared: a static approach, which is based on a retrospective estimate of the TTP, and a dynamic approach, which is based on a predictive estimate of the TTP determined with a constantly updated Winters model. Findings reveal that considering the minimization of TTP as a hospital selection policy raises several benefits. It allows to significantly reduce service throughput times in the ED with a minimum increase in travel time. Furthermore, an immediate view of the saturation state of the ED is produced and the case-mix present in the ED structures (i.e., the different triage codes) is considered, as different severity codes correspond to different service throughput times. Besides, the use of a predictive approach is certainly more reliable in terms of TTP estimation than a retrospective approach but entails a more difficult application. These considerations can support decision-makers in introducing different hospital selection policies to enhance EMSs performance.Keywords: discrete event simulation, emergency medical services, forecast model, hospital selection
Procedia PDF Downloads 921743 Jensen's Inequality and M-Convex Functions
Authors: Yamin Sayyari
Abstract:
In this paper, we generalized the Jensen's inequality for m-convex functions and also we present a correction of Jensen's inequality which is a better than the generalization of this inequality for m-convex functions. Finally, we have found new lower and new upper bounds for Jensen's discrete inequality.Keywords: Jensen's inequality, m-convex function, Convex function, Inequality
Procedia PDF Downloads 1461742 Ischemic Stroke Detection in Computed Tomography Examinations
Authors: Allan F. F. Alves, Fernando A. Bacchim Neto, Guilherme Giacomini, Marcela de Oliveira, Ana L. M. Pavan, Maria E. D. Rosa, Diana R. Pina
Abstract:
Stroke is a worldwide concern, only in Brazil it accounts for 10% of all registered deaths. There are 2 stroke types, ischemic (87%) and hemorrhagic (13%). Early diagnosis is essential to avoid irreversible cerebral damage. Non-enhanced computed tomography (NECT) is one of the main diagnostic techniques used due to its wide availability and rapid diagnosis. Detection depends on the size and severity of lesions and the time spent between the first symptoms and examination. The Alberta Stroke Program Early CT Score (ASPECTS) is a subjective method that increases the detection rate. The aim of this work was to implement an image segmentation system to enhance ischemic stroke and to quantify the area of ischemic and hemorrhagic stroke lesions in CT scans. We evaluated 10 patients with NECT examinations diagnosed with ischemic stroke. Analyzes were performed in two axial slices, one at the level of the thalamus and basal ganglion and one adjacent to the top edge of the ganglionic structures with window width between 80 and 100 Hounsfield Units. We used different image processing techniques such as morphological filters, discrete wavelet transform and Fuzzy C-means clustering. Subjective analyzes were performed by a neuroradiologist according to the ASPECTS scale to quantify ischemic areas in the middle cerebral artery region. These subjective analysis results were compared with objective analyzes performed by the computational algorithm. Preliminary results indicate that the morphological filters actually improve the ischemic areas for subjective evaluations. The comparison in area of the ischemic region contoured by the neuroradiologist and the defined area by computational algorithm showed no deviations greater than 12% in any of the 10 examination tests. Although there is a tendency that the areas contoured by the neuroradiologist are smaller than those obtained by the algorithm. These results show the importance of a computer aided diagnosis software to assist neuroradiology decisions, especially in critical situations as the choice of treatment for ischemic stroke.Keywords: ischemic stroke, image processing, CT scans, Fuzzy C-means
Procedia PDF Downloads 3691741 Abdominal Organ Segmentation in CT Images Based On Watershed Transform and Mosaic Image
Authors: Belgherbi Aicha, Hadjidj Ismahen, Bessaid Abdelhafid
Abstract:
Accurate Liver, spleen and kidneys segmentation in abdominal CT images is one of the most important steps for computer aided abdominal organs pathology diagnosis. In this paper, we have proposed a new semi-automatic algorithm for Liver, spleen and kidneys area extraction in abdominal CT images. Our proposed method is based on hierarchical segmentation and watershed algorithm. In our approach, a powerful technique has been designed to suppress over-segmentation based on mosaic image and on the computation of the watershed transform. The algorithm is currency in two parts. In the first, we seek to improve the quality of the gradient-mosaic image. In this step, we propose a method for improving the gradient-mosaic image by applying the anisotropic diffusion filter followed by the morphological filters. Thereafter we proceed to the hierarchical segmentation of the liver, spleen and kidney. To validate the segmentation technique proposed, we have tested it on several images. Our segmentation approach is evaluated by comparing our results with the manual segmentation performed by an expert. The experimental results are described in the last part of this work.Keywords: anisotropic diffusion filter, CT images, morphological filter, mosaic image, multi-abdominal organ segmentation, mosaic image, the watershed algorithm
Procedia PDF Downloads 4991740 Analytical Authentication of Butter Using Fourier Transform Infrared Spectroscopy Coupled with Chemometrics
Authors: M. Bodner, M. Scampicchio
Abstract:
Fourier Transform Infrared (FT-IR) spectroscopy coupled with chemometrics was used to distinguish between butter samples and non-butter samples. Further, quantification of the content of margarine in adulterated butter samples was investigated. Fingerprinting region (1400-800 cm–1) was used to develop unsupervised pattern recognition (Principal Component Analysis, PCA), supervised modeling (Soft Independent Modelling by Class Analogy, SIMCA), classification (Partial Least Squares Discriminant Analysis, PLS-DA) and regression (Partial Least Squares Regression, PLS-R) models. PCA of the fingerprinting region shows a clustering of the two sample types. All samples were classified in their rightful class by SIMCA approach; however, nine adulterated samples (between 1% and 30% w/w of margarine) were classified as belonging both at the butter class and at the non-butter one. In the two-class PLS-DA model’s (R2 = 0.73, RMSEP, Root Mean Square Error of Prediction = 0.26% w/w) sensitivity was 71.4% and Positive Predictive Value (PPV) 100%. Its threshold was calculated at 7% w/w of margarine in adulterated butter samples. Finally, PLS-R model (R2 = 0.84, RMSEP = 16.54%) was developed. PLS-DA was a suitable classification tool and PLS-R a proper quantification approach. Results demonstrate that FT-IR spectroscopy combined with PLS-R can be used as a rapid, simple and safe method to identify pure butter samples from adulterated ones and to determine the grade of adulteration of margarine in butter samples.Keywords: adulterated butter, margarine, PCA, PLS-DA, PLS-R, SIMCA
Procedia PDF Downloads 1471739 Ethical Utility of Artificial Intelligence in Education
Authors: Layan Kateb, Jawaher Ragban, Rawia Alamri, Rewaa Alhazmi
Abstract:
This paper evaluates user satisfaction with the AI functionality of Coconote, an innovative app designed to transform raw input -documents, voice recordings, and video recordings- into structured study materials such as flashcards, study guides, and transcripts. Leveraging AI-driven natural language processing and summarization algorithms, Coconote aims to streamline the study process by providing personalized, accessible, and accurate learning tools. To measure user satisfaction, we conducted a survey among 20 participants, including students, educators, and professionals, who tested the app’s core features over a two-week period. The survey assessed criteria such as ease of use, accuracy of generated materials, and perceived value in enhancing productivity. Results indicate high satisfaction levels with the app’s ability to produce coherent and concise outputs, with 83% of respondents reporting improved study efficiency. However, feedback highlighted areas for improvement, including occasional inaccuracies in transcription and customization limitations in flashcard formatting. This study demonstrates the potential of AI to transform learning methods and identifies key areas for refinement in user-centric design. Coconote represents a step forward in leveraging AI to enhance educational technologyKeywords: tools for students, artificial intelligence in education, AI study tool, ethical AI tools
Procedia PDF Downloads 01738 Analytical Model of Locomotion of a Thin-Film Piezoelectric 2D Soft Robot Including Gravity Effects
Authors: Zhiwu Zheng, Prakhar Kumar, Sigurd Wagner, Naveen Verma, James C. Sturm
Abstract:
Soft robots have drawn great interest recently due to a rich range of possible shapes and motions they can take on to address new applications, compared to traditional rigid robots. Large-area electronics (LAE) provides a unique platform for creating soft robots by leveraging thin-film technology to enable the integration of a large number of actuators, sensors, and control circuits on flexible sheets. However, the rich shapes and motions possible, especially when interacting with complex environments, pose significant challenges to forming well-generalized and robust models necessary for robot design and control. In this work, we describe an analytical model for predicting the shape and locomotion of a flexible (steel-foil-based) piezoelectric-actuated 2D robot based on Euler-Bernoulli beam theory. It is nominally (unpowered) lying flat on the ground, and when powered, its shape is controlled by an array of piezoelectric thin-film actuators. Key features of the models are its ability to incorporate the significant effects of gravity on the shape and to precisely predict the spatial distribution of friction against the contacting surfaces, necessary for determining inchworm-type motion. We verified the model by developing a distributed discrete element representation of a continuous piezoelectric actuator and by comparing its analytical predictions to discrete-element robot simulations using PyBullet. Without gravity, predicting the shape of a sheet with a linear array of piezoelectric actuators at arbitrary voltages is straightforward. However, gravity significantly distorts the shape of the sheet, causing some segments to flatten against the ground. Our work includes the following contributions: (i) A self-consistent approach was developed to exactly determine which parts of the soft robot are lifted off the ground, and the exact shape of these sections, for an arbitrary array of piezoelectric voltages and configurations. (ii) Inchworm-type motion relies on controlling the relative friction with the ground surface in different sections of the robot. By adding torque-balance to our model and analyzing shear forces, the model can then determine the exact spatial distribution of the vertical force that the ground is exerting on the soft robot. Through this, the spatial distribution of friction forces between ground and robot can be determined. (iii) By combining this spatial friction distribution with the shape of the soft robot, in the function of time as piezoelectric actuator voltages are changed, the inchworm-type locomotion of the robot can be determined. As a practical example, we calculated the performance of a 5-actuator system on a 50-µm thick steel foil. Piezoelectric properties of commercially available thin-film piezoelectric actuators were assumed. The model predicted inchworm motion of up to 200 µm per step. For independent verification, we also modelled the system using PyBullet, a discrete-element robot simulator. To model a continuous thin-film piezoelectric actuator, we broke each actuator into multiple segments, each of which consisted of two rigid arms with appropriate mass connected with a 'motor' whose torque was set by the applied actuator voltage. Excellent agreement between our analytical model and the discrete-element simulator was shown for both for the full deformation shape and motion of the robot.Keywords: analytical modeling, piezoelectric actuators, soft robot locomotion, thin-film technology
Procedia PDF Downloads 1811737 Numerical Simulation of Unsteady Natural Convective Nanofluid Flow within a Trapezoidal Enclosure Using Meshfree Method
Authors: S. Nandal, R. Bhargava
Abstract:
The paper contains a numerical study of the unsteady magneto-hydrodynamic natural convection flow of nanofluids within a symmetrical wavy walled trapezoidal enclosure. The length and height of enclosure are both considered equal to L. Two-phase nanofluid model is employed. The governing equations of nanofluid flow along with boundary conditions are non-dimensionalized and are solved using one of Meshfree technique (EFGM method). Meshfree numerical technique does not require a predefined mesh for discretization purpose. The bottom wavy wall of the enclosure is defined using a cosine function. Element free Galerkin method (EFGM) does not require the domain. The effects of various parameters namely time t, amplitude of bottom wavy wall a, Brownian motion parameter Nb and thermophoresis parameter Nt is examined on rate of heat and mass transfer to get a visualization of cooling and heating effects. Such problems have important applications in heat exchangers or solar collectors, as wavy walled enclosures enhance heat transfer in comparison to flat walled enclosures.Keywords: heat transfer, meshfree methods, nanofluid, trapezoidal enclosure
Procedia PDF Downloads 1581736 Biosynthesis of Selenium Oxide Nanoparticles by Streptomyces bikiniensis and Its Cytotoxicity as Antitumor Agents against Hepatocellular and Breast Cells Carcinoma
Authors: Maged Syed Ahamd, Manal Mohamed Yasser, Essam Sholkamy
Abstract:
In this paper, we reported that selenium (Se) nanoparticles were firstly biosynthesized with a simple and eco-friendly biological method. Their shape, size, FTIR (Fourier Transform Infrared spectroscopy), UV–vis spectra, TEM (Transmission Electron Microscopy) images and EDS (Energy Dispersive Spectroscopy) pattern have been analyzed. TEM analyses of the samples obtained at different stages indicated that the formation of these Se nanostructures was governed by an incubation time (12- 24- 48 hours). The Se nanoparticles were initially generated and then would transform into crystal seeds for the subsequent growth of nanowires; however obtaining stable Se nanowire with a diameter of about 15-100 nm. EDS shows that Se nanoparticles are entirely pure. The IR spectra showed the peaks at 550 cm-1, 1635 cm-1, 1994 cm-1 and 3430 cm-1 correspond to the presence of Se-O bending and stretching vibrations. The concentrations of Se-NPs (0, 1, 2, 5 µg/ml) did not give significantly effect on both two cell lines while the highest concentrations (10- 100 µg/ml gave significantly effects on them. The lethal dose (ID50%) of Se-NPs on Hep2 G and MCF-7 cells was obtained at 75.96 and 61.86 µg/ml, respectively. Results showed that Se nanoparticles as anticancer agent against MCF-7 cells were more effective than Hep2 G cells. Our results suggest that Se-NPs may be a candidate for further evaluation as a chemotherapeutic agent for breast and liver cancers.Keywords: selenium nanoparticle, Streptomyces bikiniensis, nanowires, chemotherapeutic agent
Procedia PDF Downloads 4451735 Preparation and Characterization of Bioplastic from Sorghum Husks
Authors: Hannatu Abubakar Sani, Abubakar Umar Birnin Yauri, Aliyu Muhammad, Mujahid Salau, Aminu Musa, Hadiza Adamu Kwazo
Abstract:
The increase in the global population and advances in technology have made plastic materials to have wide applications in every aspect of life. However, the non-biodegradability of these petrochemical-based materials and their increasing accumulation in the environment has been a threat to the planet and has been a source of environmental concerns and hence, the driving force in the search for ‘green’ alternatives for which agricultural waste remains the front liner. Sorghum husk, an agricultural waste with potentials as a raw material in the production of bioplastic, was used in this research to prepare bioplastic using sulphuric acid-catalyzed acetylation process. The prepared bioplastic was characterized by X-ray diffraction and Fourier transform infrared spectroscopy (FTIR), and the structure of the prepared bioplastic was confirmed. The Fourier transform infrared spectroscopy (FTIR) spectra of the product displayed the presence of OH, C-H, C=O, and C-O absorption peaks. The bioplastic obtained is biodegradable and is affected by acid, salt, and alkali to a lesser extent. Other tests like solubility and swelling studies were carried out to ensure the commercial properties of these bioplastic materials. Therefore, this revealed that new bioplastics with better environmental and sustainable properties could be produced from agricultural waste, which may have applications in many industries.Keywords: agricultural waste, bioplastic, characterization, Sorghum Husk
Procedia PDF Downloads 1601734 Continuous and Discontinuos Modeling of Wellbore Instability in Anisotropic Rocks
Authors: C. Deangeli, P. Obentaku Obenebot, O. Omwanghe
Abstract:
The study focuses on the analysis of wellbore instability in rock masses affected by weakness planes. The occurrence of failure in such a type of rocks can occur in the rock matrix and/ or along the weakness planes, in relation to the mud weight gradient. In this case the simple Kirsch solution coupled with a failure criterion cannot supply a suitable scenario for borehole instabilities. Two different numerical approaches have been used in order to investigate the onset of local failure at the wall of a borehole. For each type of approach the influence of the inclination of weakness planes has been investigates, by considering joint sets at 0°, 35° and 90° to the horizontal. The first set of models have been carried out with FLAC 2D (Fast Lagrangian Analysis of Continua) by considering the rock material as a continuous medium, with a Mohr Coulomb criterion for the rock matrix and using the ubiquitous joint model for accounting for the presence of the weakness planes. In this model yield may occur in either the solid or along the weak plane, or both, depending on the stress state, the orientation of the weak plane and the material properties of the solid and weak plane. The second set of models have been performed with PFC2D (Particle Flow code). This code is based on the Discrete Element Method and considers the rock material as an assembly of grains bonded by cement-like materials, and pore spaces. The presence of weakness planes is simulated by the degradation of the bonds between grains along given directions. In general the results of the two approaches are in agreement. However the discrete approach seems to capture more complex phenomena related to local failure in the form of grain detachment at wall of the borehole. In fact the presence of weakness planes in the discontinuous medium leads to local instability along the weak planes also in conditions not predicted from the continuous solution. In general slip failure locations and directions do not follow the conventional wellbore breakout direction but depend upon the internal friction angle and the orientation of the bedding planes. When weakness plane is at 0° and 90° the behaviour are similar to that of a continuous rock material, but borehole instability is more severe when weakness planes are inclined at an angle between 0° and 90° to the horizontal. In conclusion, the results of the numerical simulations show that the prediction of local failure at the wall of the wellbore cannot disregard the presence of weakness planes and consequently the higher mud weight required for stability for any specific inclination of the joints. Despite the discrete approach can simulate smaller areas because of the large number of particles required for the generation of the rock material, however it seems to investigate more correctly the occurrence of failure at the miscroscale and eventually the propagation of the failed zone to a large portion of rock around the wellbore.Keywords: continuous- discontinuous, numerical modelling, weakness planes wellbore, FLAC 2D
Procedia PDF Downloads 5011733 Visualization of Wave Propagation in Monocoupled System with Effective Negative Stiffness, Effective Negative Mass, and Inertial Amplifier
Authors: Abhigna Bhatt, Arnab Banerjee
Abstract:
A periodic system with only a single coupling degree of freedom is called a monocoupled system. Monocoupled systems with mechanisms like mass in the mass system generates effective negative mass, mass connected with rigid links generates inertial amplification, and spring-mass connected with a rigid link generateseffective negative stiffness. In this paper, the representative unit cell is introduced, considering all three mechanisms combined. Further, the dynamic stiffness matrix of the unit cell is constructed, and the dispersion relation is obtained by applying the Bloch theorem. The frequency response function is also calculated for the finite length of periodic unit cells. Moreover, the input displacement signal is given to the finite length of periodic structure and using inverse Fourier transform to visualize the wave propagation in the time domain. This visualization explains the sudden attenuation in metamaterial due to energy dissipation by an embedded resonator at the resonance frequency. The visualization created for wave propagation is found necessary to understand the insights of physics behind the attenuation characteristics of the system.Keywords: mono coupled system, negative effective mass, negative effective stiffness, inertial amplifier, fourier transform
Procedia PDF Downloads 1271732 Quality Evaluation of Backfill Grout in Tunnel Boring Machine Tail Void Using Impact-Echo (IE): Short-Time Fourier Transform (STFT) Numerical Analysis
Authors: Ju-Young Choi, Ki-Il Song, Kyoung-Yul Kim
Abstract:
During Tunnel Boring Machine (TBM) tunnel excavation, backfill grout should be injected after the installation of segment lining to ensure the stability of the tunnel and to minimize ground deformation. If grouting is not sufficient to fill the gap between the segments and rock mass, hydraulic pressures occur in the void, which can negatively influence the stability of the tunnel. Recently the tendency to use TBM tunnelling method to replace the drill and blast(NATM) method is increasing. However, there are only a few studies of evaluation of backfill grout. This study evaluates the TBM tunnel backfill state using Impact-Echo(IE). 3-layers, segment-grout-rock mass, are simulated by FLAC 2D, FDM-based software. The signals obtained from numerical analysis and IE test are analyzed by Short-Time Fourier Transform(STFT) in time domain, frequency domain, and time-frequency domain. The result of this study can be used to evaluate the quality of backfill grouting in tail void.Keywords: tunnel boring machine, backfill grout, impact-echo method, time-frequency domain analysis, finite difference method
Procedia PDF Downloads 2671731 Enhancing Patch Time Series Transformer with Wavelet Transform for Improved Stock Prediction
Authors: Cheng-yu Hsieh, Bo Zhang, Ahmed Hambaba
Abstract:
Stock market prediction has long been an area of interest for both expert analysts and investors, driven by its complexity and the noisy, volatile conditions it operates under. This research examines the efficacy of combining the Patch Time Series Transformer (PatchTST) with wavelet transforms, specifically focusing on Haar and Daubechies wavelets, in forecasting the adjusted closing price of the S&P 500 index for the following day. By comparing the performance of the augmented PatchTST models with traditional predictive models such as Recurrent Neural Networks (RNNs), Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and Transformers, this study highlights significant enhancements in prediction accuracy. The integration of the Daubechies wavelet with PatchTST notably excels, surpassing other configurations and conventional models in terms of Mean Absolute Error (MAE) and Mean Squared Error (MSE). The success of the PatchTST model paired with Daubechies wavelet is attributed to its superior capability in extracting detailed signal information and eliminating irrelevant noise, thus proving to be an effective approach for financial time series forecasting.Keywords: deep learning, financial forecasting, stock market prediction, patch time series transformer, wavelet transform
Procedia PDF Downloads 551730 Predicting Potential Protein Therapeutic Candidates from the Gut Microbiome
Authors: Prasanna Ramachandran, Kareem Graham, Helena Kiefel, Sunit Jain, Todd DeSantis
Abstract:
Microbes that reside inside the mammalian GI tract, commonly referred to as the gut microbiome, have been shown to have therapeutic effects in animal models of disease. We hypothesize that specific proteins produced by these microbes are responsible for this activity and may be used directly as therapeutics. To speed up the discovery of these key proteins from the big-data metagenomics, we have applied machine learning techniques. Using amino acid sequences of known epitopes and their corresponding binding partners, protein interaction descriptors (PID) were calculated, making a positive interaction set. A negative interaction dataset was calculated using sequences of proteins known not to interact with these same binding partners. Using Random Forest and positive and negative PID, a machine learning model was trained and used to predict interacting versus non-interacting proteins. Furthermore, the continuous variable, cosine similarity in the interaction descriptors was used to rank bacterial therapeutic candidates. Laboratory binding assays were conducted to test the candidates for their potential as therapeutics. Results from binding assays reveal the accuracy of the machine learning prediction and are subsequently used to further improve the model.Keywords: protein-interactions, machine-learning, metagenomics, microbiome
Procedia PDF Downloads 3761729 Influence of Flight Design on Discharging Profiles of Granular Material in Rotary Dryer
Authors: I. Benhsine, M. Hellou, F. Lominé, Y. Roques
Abstract:
During the manufacture of fertilizer, it is necessary to add water for granulation purposes. The water content is then removed or reduced using rotary dryers. They are commonly used to dry wet granular materials and they are usually fitted with lifting flights. The transport of granular materials occurs when particles cascade from the lifting flights and fall into the air stream. Each cascade consists of a lifting and a falling cycle. Lifting flights are thus of great importance for the transport of granular materials along the dryer. They also enhance the contact between solid particles and the air stream. Optimization of the drying process needs an understanding of the behavior of granular materials inside a rotary dryer. Different approaches exist to study the movement of granular materials inside the dryer. Most common of them are based on empirical formulations or on study the movement of the bulk material. In the present work, we are interested in the behavior of each particle in the cross section of the dryer using Discrete Element Method (DEM) to understand. In this paper, we focus on studying the hold-up, the cascade patterns, the falling time and the falling length of the particles leaving the flights. We will be using two segment flights. Three different profiles are used: a straight flight (180° between both segments), an angled flight (with an angle of 150°), and a right-angled flight (90°). The profile of the flight affects significantly the movement of the particles in the dryer. Changing the flight angle changes the flight capacity which leads to different discharging profile of the flight, thus affecting the hold-up in the flight. When the angle of the flight is reduced, the range of the discharge angle increases leading to a more uniformed cascade pattern in time. The falling length and the falling time of the particles also increase up to a maximum value then they start decreasing. Moreover, the results show an increase in the falling length and the falling time up to 70% and 50%, respectively, when using a right-angled flight instead of a straight one.Keywords: discrete element method, granular materials, lifting flight, rotary dryer
Procedia PDF Downloads 3281728 Optoelectronic Hardware Architecture for Recurrent Learning Algorithm in Image Processing
Authors: Abdullah Bal, Sevdenur Bal
Abstract:
This paper purposes a new type of hardware application for training of cellular neural networks (CNN) using optical joint transform correlation (JTC) architecture for image feature extraction. CNNs require much more computation during the training stage compare to test process. Since optoelectronic hardware applications offer possibility of parallel high speed processing capability for 2D data processing applications, CNN training algorithm can be realized using Fourier optics technique. JTC employs lens and CCD cameras with laser beam that realize 2D matrix multiplication and summation in the light speed. Therefore, in the each iteration of training, JTC carries more computation burden inherently and the rest of mathematical computation realized digitally. The bipolar data is encoded by phase and summation of correlation operations is realized using multi-object input joint images. Overlapping properties of JTC are then utilized for summation of two cross-correlations which provide less computation possibility for training stage. Phase-only JTC does not require data rearrangement, electronic pre-calculation and strict system alignment. The proposed system can be incorporated simultaneously with various optical image processing or optical pattern recognition techniques just in the same optical system.Keywords: CNN training, image processing, joint transform correlation, optoelectronic hardware
Procedia PDF Downloads 5071727 Sidelobe Free Inverse Synthetic Aperture Radar Imaging of Non Cooperative Moving Targets Using WiFi
Authors: Jiamin Huang, Shuliang Gui, Zengshan Tian, Fei Yan, Xiaodong Wu
Abstract:
In recent years, with the rapid development of radio frequency technology, the differences between radar sensing and wireless communication in terms of receiving and sending channels, signal processing, data management and control are gradually shrinking. There has been a trend of integrated communication radar sensing. However, most of the existing radar imaging technologies based on communication signals are combined with synthetic aperture radar (SAR) imaging, which does not conform to the practical application case of the integration of communication and radar. Therefore, in this paper proposes a high-precision imaging method using communication signals based on the imaging mechanism of inverse synthetic aperture radar (ISAR) imaging. This method makes full use of the structural characteristics of the orthogonal frequency division multiplexing (OFDM) signal, so the sidelobe effect in distance compression is removed and combines radon transform and Fractional Fourier Transform (FrFT) parameter estimation methods to achieve ISAR imaging of non-cooperative targets. The simulation experiment and measured results verify the feasibility and effectiveness of the method, and prove its broad application prospects in the field of intelligent transportation.Keywords: integration of communication and radar, OFDM, radon, FrFT, ISAR
Procedia PDF Downloads 1261726 Fourier Transform and Machine Learning Techniques for Fault Detection and Diagnosis of Induction Motors
Authors: Duc V. Nguyen
Abstract:
Induction motors are widely used in different industry areas and can experience various kinds of faults in stators and rotors. In general, fault detection and diagnosis techniques for induction motors can be supervised by measuring quantities such as noise, vibration, and temperature. The installation of mechanical sensors in order to assess the health conditions of a machine is typically only done for expensive or load-critical machines, where the high cost of a continuous monitoring system can be Justified. Nevertheless, induced current monitoring can be implemented inexpensively on machines with arbitrary sizes by using current transformers. In this regard, effective and low-cost fault detection techniques can be implemented, hence reducing the maintenance and downtime costs of motors. This work proposes a method for fault detection and diagnosis of induction motors, which combines classical fast Fourier transform and modern/advanced machine learning techniques. The proposed method is validated on real-world data and achieves a precision of 99.7% for fault detection and 100% for fault classification with minimal expert knowledge requirement. In addition, this approach allows users to be able to optimize/balance risks and maintenance costs to achieve the highest benet based on their requirements. These are the key requirements of a robust prognostics and health management system.Keywords: fault detection, FFT, induction motor, predictive maintenance
Procedia PDF Downloads 1731725 Low-Density Polyethylene Film Biodegradation Potential by Fungal Species From Thailand
Authors: Patcharee Pripdeevech, Sarunpron Khruengsai
Abstract:
Thirty fungi were tested for their degradation ability on low-density polyethylene (LDPE) plastic film. Biodegradation of all fungi was screened in mineral salt medium broth containing LDPE film as the sole carbon source for 30 days. Diaporthe italiana, Thyrostroma jaczewskii, Colletotrichum fructicola, and Stagonosporopsis citrulli were able to colonize and cover the surface of LDPE film in media. The degradation test result was compared to those obtained from Aspergillus niger. LDPE films cocultured with D. italiana, T. jaczewskii, C. fructicola, S. citrulli, A. niger, and control showed weight loss of 43.90%, 46.34%, 48.78%, 45.12%, 28.78%, and 10.85%, respectively. The tensile strength of degraded LDPE films cocultured with D. italiana, T. jaczewskii, C. fructicola, S. citrulli, A. niger, and control also reduced significantly by 1.56 MPa, 1.78 MPa, 0.43 MPa, 1.86 MPa, 3.34 MPa, and 9.98 MPa, respectively. Analysis of LDPE films by Fourier transform infrared spectroscopy and scanning electron microscopy confirmed the biodegradation by the presence of morphological changes such as cracks, scions, and holes on the surface of the film. These fungi have the ability to break down and consume the LDPE film, especially C. fructicola. These findings showed the potential of fungi in Thailand that play an important role in LDPE film degradation.Keywords: plastic biodegradation, LDPE film, fungi, Fourier transform infrared, scanning electron microscopy
Procedia PDF Downloads 1341724 Numerical Solution of Transient Natural Convection in Vertical Heated Rectangular Channel between Two Vertical Parallel MTR-Type Fuel Plates
Authors: Djalal Hamed
Abstract:
The aim of this paper is to perform, by mean of the finite volume method, a numerical solution of the transient natural convection in a narrow rectangular channel between two vertical parallel Material Testing Reactor (MTR)-type fuel plates, imposed under a heat flux with a cosine shape to determine the margin of the nuclear core power at which the natural convection cooling mode can ensure a safe core cooling, where the cladding temperature should not reach a specific safety limits (90 °C). For this purpose, a computer program is developed to determine the principal parameters related to the nuclear core safety, such as the temperature distribution in the fuel plate and in the coolant (light water) as a function of the reactor core power. Throughout the obtained results, we noticed that the core power should not reach 400 kW, to ensure a safe passive residual heat removing from the nuclear core by the upward natural convection cooling mode.Keywords: buoyancy force, friction force, finite volume method, transient natural convection
Procedia PDF Downloads 1971723 Curvature Based-Methods for Automatic Coarse and Fine Registration in Dimensional Metrology
Authors: Rindra Rantoson, Hichem Nouira, Nabil Anwer, Charyar Mehdi-Souzani
Abstract:
Multiple measurements by means of various data acquisition systems are generally required to measure the shape of freeform workpieces for accuracy, reliability and holisticity. The obtained data are aligned and fused into a common coordinate system within a registration technique involving coarse and fine registrations. Standardized iterative methods have been established for fine registration such as Iterative Closest Points (ICP) and its variants. For coarse registration, no conventional method has been adopted yet despite a significant number of techniques which have been developed in the literature to supply an automatic rough matching between data sets. Two main issues are addressed in this paper: the coarse registration and the fine registration. For coarse registration, two novel automated methods based on the exploitation of discrete curvatures are presented: an enhanced Hough Transformation (HT) and an improved Ransac Transformation. The use of curvature features in both methods aims to reduce computational cost. For fine registration, a new variant of ICP method is proposed in order to reduce registration error using curvature parameters. A specific distance considering the curvature similarity has been combined with Euclidean distance to define the distance criterion used for correspondences searching. Additionally, the objective function has been improved by combining the point-to-point (P-P) minimization and the point-to-plane (P-Pl) minimization with automatic weights. These ones are determined from the preliminary calculated curvature features at each point of the workpiece surface. The algorithms are applied on simulated and real data performed by a computer tomography (CT) system. The obtained results reveal the benefit of the proposed novel curvature-based registration methods.Keywords: discrete curvature, RANSAC transformation, hough transformation, coarse registration, ICP variant, point-to-point and point-to-plane minimization combination, computer tomography
Procedia PDF Downloads 424