Search results for: lumped mass parameter
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5430

Search results for: lumped mass parameter

3630 Electrical Tortuosity across Electrokinetically Remediated Soils

Authors: Waddah S. Abdullah, Khaled F. Al-Omari

Abstract:

Electrokinetic remediation is one of the most influential and effective methods to decontaminate contaminated soils. Electroosmosis and electromigration are the processes of electrochemical extraction of contaminants from soils. The driving force that causes removing contaminants from soils (electroosmosis process or electromigration process) is voltage gradient. Therefore, the electric field distribution throughout the soil domain is extremely important to investigate and to determine the factors that help to establish a uniform electric field distribution in order to make the clean-up process work properly and efficiently. In this study, small-sized passive electrodes (made of graphite) were placed at predetermined locations within the soil specimen, and the voltage drop between these passive electrodes was measured in order to observe the electrical distribution throughout the tested soil specimens. The electrokinetic test was conducted on two types of soils; a sandy soil and a clayey soil. The electrical distribution throughout the soil domain was conducted with different tests properties; and the electrical field distribution was observed in three-dimensional pattern in order to establish the electrical distribution within the soil domain. The effects of density, applied voltages, and degree of saturation on the electrical distribution within the remediated soil were investigated. The distribution of the moisture content, concentration of the sodium ions, and the concentration of the calcium ions were determined and established in three-dimensional scheme. The study has shown that the electrical conductivity within soil domain depends on the moisture content and concentration of electrolytes present in the pore fluid. The distribution of the electrical field in the saturated soil was found not be affected by its density. The study has also shown that high voltage gradient leads to non-uniform electric field distribution within the electroremediated soil. Very importantly, it was found that even when the electric field distribution is uniform globally (i.e. between the passive electrodes), local non-uniformity could be established within the remediated soil mass. Cracks or air gaps formed due to temperature rise (because of electric flow in low conductivity regions) promotes electrical tortuosity. Thus, fracturing or cracking formed in the remediated soil mass causes disconnection of electric current and hence, no removal of contaminant occur within these areas.

Keywords: contaminant removal, electrical tortuousity, electromigration, electroosmosis, voltage distribution

Procedia PDF Downloads 421
3629 Population Size Estimation Based on the GPD

Authors: O. Anan, D. Böhning, A. Maruotti

Abstract:

The purpose of the study is to estimate the elusive target population size under a truncated count model that accounts for heterogeneity. The purposed estimator is based on the generalized Poisson distribution (GPD), which extends the Poisson distribution by adding a dispersion parameter. Thus, it becomes an useful model for capture-recapture data where concurrent events are not homogeneous. In addition, it can account for over-dispersion and under-dispersion. The ratios of neighboring frequency counts are used as a tool for investigating the validity of whether generalized Poisson or Poisson distribution. Since capture-recapture approaches do not provide the zero counts, the estimated parameters can be achieved by modifying the EM-algorithm technique for the zero-truncated generalized Poisson distribution. The properties and the comparative performance of proposed estimator were investigated through simulation studies. Furthermore, some empirical examples are represented insights on the behavior of the estimators.

Keywords: capture, recapture methods, ratio plot, heterogeneous population, zero-truncated count

Procedia PDF Downloads 435
3628 Applications of Artificial Intelligence (AI) in Cardiac imaging

Authors: Angelis P. Barlampas

Abstract:

The purpose of this study is to inform the reader, about the various applications of artificial intelligence (AI), in cardiac imaging. AI grows fast and its role is crucial in medical specialties, which use large amounts of digital data, that are very difficult or even impossible to be managed by human beings and especially doctors.Artificial intelligence (AI) refers to the ability of computers to mimic human cognitive function, performing tasks such as learning, problem-solving, and autonomous decision making based on digital data. Whereas AI describes the concept of using computers to mimic human cognitive tasks, machine learning (ML) describes the category of algorithms that enable most current applications described as AI. Some of the current applications of AI in cardiac imaging are the follows: Ultrasound: Automated segmentation of cardiac chambers across five common views and consequently quantify chamber volumes/mass, ascertain ejection fraction and determine longitudinal strain through speckle tracking. Determine the severity of mitral regurgitation (accuracy > 99% for every degree of severity). Identify myocardial infarction. Distinguish between Athlete’s heart and hypertrophic cardiomyopathy, as well as restrictive cardiomyopathy and constrictive pericarditis. Predict all-cause mortality. CT Reduce radiation doses. Calculate the calcium score. Diagnose coronary artery disease (CAD). Predict all-cause 5-year mortality. Predict major cardiovascular events in patients with suspected CAD. MRI Segment of cardiac structures and infarct tissue. Calculate cardiac mass and function parameters. Distinguish between patients with myocardial infarction and control subjects. It could potentially reduce costs since it would preclude the need for gadolinium-enhanced CMR. Predict 4-year survival in patients with pulmonary hypertension. Nuclear Imaging Classify normal and abnormal myocardium in CAD. Detect locations with abnormal myocardium. Predict cardiac death. ML was comparable to or better than two experienced readers in predicting the need for revascularization. AI emerge as a helpful tool in cardiac imaging and for the doctors who can not manage the overall increasing demand, in examinations such as ultrasound, computed tomography, MRI, or nuclear imaging studies.

Keywords: artificial intelligence, cardiac imaging, ultrasound, MRI, CT, nuclear medicine

Procedia PDF Downloads 78
3627 Optimal Relaxation Parameters for Obtaining Efficient Iterative Methods for the Solution of Electromagnetic Scattering Problems

Authors: Nadaniela Egidi, Pierluigi Maponi

Abstract:

The approximate solution of a time-harmonic electromagnetic scattering problem for inhomogeneous media is required in several application contexts, and its two-dimensional formulation is a Fredholm integral equation of the second kind. This integral equation provides a formulation for the direct scattering problem, but it has to be solved several times also in the numerical solution of the corresponding inverse scattering problem. The discretization of this Fredholm equation produces large and dense linear systems that are usually solved by iterative methods. In order to improve the efficiency of these iterative methods, we use the Symmetric SOR preconditioning, and we propose an algorithm for the evaluation of the associated relaxation parameter. We show the efficiency of the proposed algorithm by several numerical experiments, where we use two Krylov subspace methods, i.e., Bi-CGSTAB and GMRES.

Keywords: Fredholm integral equation, iterative method, preconditioning, scattering problem

Procedia PDF Downloads 104
3626 Mapping Tunnelling Parameters for Global Optimization in Big Data via Dye Laser Simulation

Authors: Sahil Imtiyaz

Abstract:

One of the biggest challenges has emerged from the ever-expanding, dynamic, and instantaneously changing space-Big Data; and to find a data point and inherit wisdom to this space is a hard task. In this paper, we reduce the space of big data in Hamiltonian formalism that is in concordance with Ising Model. For this formulation, we simulate the system using dye laser in FORTRAN and analyse the dynamics of the data point in energy well of rhodium atom. After mapping the photon intensity and pulse width with energy and potential we concluded that as we increase the energy there is also increase in probability of tunnelling up to some point and then it starts decreasing and then shows a randomizing behaviour. It is due to decoherence with the environment and hence there is a loss of ‘quantumness’. This interprets the efficiency parameter and the extent of quantum evolution. The results are strongly encouraging in favour of the use of ‘Topological Property’ as a source of information instead of the qubit.

Keywords: big data, optimization, quantum evolution, hamiltonian, dye laser, fermionic computations

Procedia PDF Downloads 194
3625 Perception and Implementation of Machine Translation Applications by the Iranian English Translators

Authors: Abdul Amir Hazbavi

Abstract:

The present study is an attempt to provide a relatively comprehensive preview of the Iranian English translators’ perception on Machine Translation. Furthermore, the study tries to shed light on the status of implementation of Machine Translation among the Iranian English Translators. To reach the aforementioned objectives, the Localization Industry Standards Association’s questioner for measuring perceptions with regard to the adoption of a technology innovation was adapted and used to investigate three parameter among the participants of the study, namely familiarity with Machine Translation, general perception on Machine Translation and implementation of Machine Translation systems in translation tasks. The participants of the study were 224 last-year undergraduate Iranian students of English translation at 10 universities across the country. The study revealed a very low level of adoption and a very high level of willingness to get familiar with and learn about Machine Translation, as well as a positive perception of and attitude toward Machine Translation by the Iranian English translators.

Keywords: translation technology, machine translation, perception, implementation

Procedia PDF Downloads 523
3624 Analysis of Delamination in Drilling of Composite Materials

Authors: Navid Zarif Karimi, Hossein Heidary, Giangiacomo Minak, Mehdi Ahmadi

Abstract:

In this paper analytical model based on the mechanics of oblique cutting, linear elastic fracture mechanics (LEFM) and bending plate theory has been presented to determine the critical feed rate causing delamination in drilling of composite materials. Most of the models in this area used LEFM and bending plate theory; hence, they can only determine the critical thrust force which is an incorporable parameter. In this model by adding cutting oblique mechanics to previous models, critical feed rate has been determined. Also instead of simplification in loading condition, actual thrust force induced by chisel edge and cutting lips on composite plate is modeled.

Keywords: composite material, delamination, drilling, thrust force

Procedia PDF Downloads 515
3623 Automatic Calibration of Agent-Based Models Using Deep Neural Networks

Authors: Sima Najafzadehkhoei, George Vega Yon

Abstract:

This paper presents an approach for calibrating Agent-Based Models (ABMs) efficiently, utilizing Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks. These machine learning techniques are applied to Susceptible-Infected-Recovered (SIR) models, which are a core framework in the study of epidemiology. Our method replicates parameter values from observed trajectory curves, enhancing the accuracy of predictions when compared to traditional calibration techniques. Through the use of simulated data, we train the models to predict epidemiological parameters more accurately. Two primary approaches were explored: one where the number of susceptible, infected, and recovered individuals is fully known, and another using only the number of infected individuals. Our method shows promise for application in other ABMs where calibration is computationally intensive and expensive.

Keywords: ABM, calibration, CNN, LSTM, epidemiology

Procedia PDF Downloads 24
3622 Numerical Simulation of a Solar Photovoltaic Panel Cooled by a Forced Air System

Authors: Djamila Nebbali, Rezki Nebbali, Ahmed Ouibrahim

Abstract:

This study focuses on the cooling of a photovoltaic panel (PV). Indeed, the cooling improves the conversion capacity of this one and maintains, under extreme conditions of air temperature, the panel temperature at an appreciable level which avoids the altering. To do this, a fan provides forced circulation of air. Because the fan is supplied by the panel, it is necessary to determine the optimum operating point that unites efficiency of the PV with the consumption of the fan. For this matter, numerical simulations are performed at varying mass flow rates of air, under two extreme air temperatures (50°C, 25°C) and a fixed solar radiation (1000 W.m2) in a case of no wind.

Keywords: energy conversion, efficiency, balance energy, solar cell

Procedia PDF Downloads 416
3621 Structural Analysis and Modelling in an Evolving Iron Ore Operation

Authors: Sameh Shahin, Nannang Arrys

Abstract:

Optimizing pit slope stability and reducing strip ratio of a mining operation are two key tasks in geotechnical engineering. With a growing demand for minerals and an increasing cost associated with extraction, companies are constantly re-evaluating the viability of mineral deposits and challenging their geological understanding. Within Rio Tinto Iron Ore, the Structural Geology (SG) team investigate and collect critical data, such as point based orientations, mapping and geological inferences from adjacent pits to re-model deposits where previous interpretations have failed to account for structurally controlled slope failures. Utilizing innovative data collection methods and data-driven investigation, SG aims to address the root causes of slope instability. Committing to a resource grid drill campaign as the primary source of data collection will often bias data collection to a specific orientation and significantly reduce the capability to identify and qualify complexity. Consequently, these limitations make it difficult to construct a realistic and coherent structural model that identifies adverse structural domains. Without the consideration of complexity and the capability of capturing these structural domains, mining operations run the risk of inadequately designed slopes that may fail and potentially harm people. Regional structural trends have been considered in conjunction with surface and in-pit mapping data to model multi-batter fold structures that were absent from previous iterations of the structural model. The risk is evident in newly identified dip-slope and rock-mass controlled sectors of the geotechnical design rather than a ubiquitous dip-slope sector across the pit. The reward is two-fold: 1) providing sectors of rock-mass controlled design in previously interpreted structurally controlled domains and 2) the opportunity to optimize the slope angle for mineral recovery and reduced strip ratio. Furthermore, a resulting high confidence model with structures and geometries that can account for historic slope instabilities in structurally controlled domains where design assumptions failed.

Keywords: structural geology, geotechnical design, optimization, slope stability, risk mitigation

Procedia PDF Downloads 46
3620 EEG Signal Processing Methods to Differentiate Mental States

Authors: Sun H. Hwang, Young E. Lee, Yunhan Ga, Gilwon Yoon

Abstract:

EEG is a very complex signal with noises and other bio-potential interferences. EOG is the most distinct interfering signal when EEG signals are measured and analyzed. It is very important how to process raw EEG signals in order to obtain useful information. In this study, the EEG signal processing techniques such as EOG filtering and outlier removal were examined to minimize unwanted EOG signals and other noises. The two different mental states of resting and focusing were examined through EEG analysis. A focused state was induced by letting subjects to watch a red dot on the white screen. EEG data for 32 healthy subjects were measured. EEG data after 60-Hz notch filtering were processed by a commercially available EOG filtering and our presented algorithm based on the removal of outliers. The ratio of beta wave to theta wave was used as a parameter for determining the degree of focusing. The results show that our algorithm was more appropriate than the existing EOG filtering.

Keywords: EEG, focus, mental state, outlier, signal processing

Procedia PDF Downloads 283
3619 Advancing Communication Theory in the Age of Digital Technology: Bridging the Gap Between Traditional Models and Emerging Platforms

Authors: Sidique Fofanah

Abstract:

This paper explores the intersection of traditional communication theories and modern digital technologies, analyzing how established models adapt to contemporary communication platforms. It examines the evolving nature of interpersonal, group, and mass communication within digital environments, emphasizing the role of social media, AI-driven communication tools, and virtual reality in reshaping communication paradigms. The paper also discusses the implications for future research and practice in communication studies, proposing an integrated framework that accommodates both classical and emerging theories.

Keywords: communication, traditional models, emerging platforms, digital media

Procedia PDF Downloads 25
3618 Reliability Based Optimal Design of Laterally Loaded Pile with Limited Residual Strain Energy Capacity

Authors: M. Movahedi Rad

Abstract:

In this study, a general approach to the reliability based limit analysis of laterally loaded piles is presented. In engineering practice, the uncertainties play a very important role. The aim of this study is to evaluate the lateral load capacity of free head and fixed-head long pile when the plastic limit analysis is considered. In addition to the plastic limit analysis to control the plastic behaviour of the structure, uncertain bound on the complementary strain energy of the residual forces is also applied. This bound has a significant effect for the load parameter. The solution to reliability-based problems is obtained by a computer program which is governed by the reliability index calculation.

Keywords: reliability, laterally loaded pile, residual strain energy, probability, limit analysis

Procedia PDF Downloads 349
3617 Effect of Fault Depth on Near-Fault Peak Ground Velocity

Authors: Yanyan Yu, Haiping Ding, Pengjun Chen, Yiou Sun

Abstract:

Fault depth is an important parameter to be determined in ground motion simulation, and peak ground velocity (PGV) demonstrates good application prospect. Using numerical simulation method, the variations of distribution and peak value of near-fault PGV with different fault depth were studied in detail, and the reason of some phenomena were discussed. The simulation results show that the distribution characteristics of PGV of fault-parallel (FP) component and fault-normal (FN) component are distinctly different; the value of PGV FN component is much larger than that of FP component. With the increase of fault depth, the distribution region of the FN component strong PGV moves forward along the rupture direction, while the strong PGV zone of FP component becomes gradually far away from the fault trace along the direction perpendicular to the strike. However, no matter FN component or FP component, the strong PGV distribution area and its value are both quickly reduced with increased fault depth. The results above suggest that the fault depth have significant effect on both FN component and FP component of near-fault PGV.

Keywords: fault depth, near-fault, PGV, numerical simulation

Procedia PDF Downloads 346
3616 Biomarkers, A Reliable Tool for Delineating Spill Trajectory

Authors: Okpor Victor, Selegha Abrakasa

Abstract:

Oil (Petroleum) spill occur frequently and in this era of a higher degree of awareness, it is pertinent that the trajectory of the spill is properly defined, to make certain of the area of impact by the spill. In this study, biomarkers that are known as the custodians of paleo information in oils are suggested to be used as reliable tools for defining the pathway of a spill. Samples were collected as tills alongside the GPS coordinates of the sample points suspected to have been impacted by a spill. Oils in the samples were extracted and analyzed as whole oil using GC–MS. Some biomarker parametric ratios were derived, and the ratio showed consistency of values along the sample trail from sample 1 to sample 20. The consistency of the values indicates that the oils at each sample point are the same hence the same value. This method can be used to validate the trajectory/pathway of a spill and also to define or establish a suspected pathway for a spill. The Oleanane/C30Hopane ratio showed good consistency and was suggested as a reliable parameter for establishing the trajectory of an oil spill.

Keywords: spill, biomarkers, trajectory, pathway

Procedia PDF Downloads 65
3615 Minimizing Fresh and Wastewater Using Water Pinch Technique in Petrochemical Industries

Authors: Wasif Mughees, Malik Al-Ahmad, Muhammad Naeem

Abstract:

This research involves the design and analysis of pinch-based water/wastewater networks to minimize water utility in the petrochemical and petroleum industries. A study has been done on Tehran Oil Refinery to analyze feasibilities of regeneration, reuse and recycling of water network. COD is considered as a single key contaminant. Amount of freshwater was reduced about 149m3/h (43.8%) regarding COD. Re-design (or retrofitting) of water allocation in the networks was undertaken. The results were analyzed through graphical method and mathematical programming technique which clearly demonstrated that amount of required water would be determined by mass transfer of COD.

Keywords: minimization, water pinch, water management, pollution prevention

Procedia PDF Downloads 448
3614 Thermal Characterisation of Multi-Coated Lightweight Brake Rotors for Passenger Cars

Authors: Ankit Khurana

Abstract:

The sufficient heat storage capacity or ability to dissipate heat is the most decisive parameter to have an effective and efficient functioning of Friction-based Brake Disc systems. The primary aim of the research was to analyse the effect of multiple coatings on lightweight disk rotors surface which not only alleviates the mass of vehicle & also, augments heat transfer. This research is projected to aid the automobile fraternity with an enunciated view over the thermal aspects in a braking system. The results of the project indicate that with the advent of modern coating technologies a brake system’s thermal curtailments can be removed and together with forced convection, heat transfer processes can see a drastic improvement leading to increased lifetime of the brake rotor. Other advantages of modifying the surface of a lightweight rotor substrate will be to reduce the overall weight of the vehicle, decrease the risk of thermal brake failure (brake fade and fluid vaporization), longer component life, as well as lower noise and vibration characteristics. A mathematical model was constructed in MATLAB which encompassing the various thermal characteristics of the proposed coatings and substrate materials required to approximate the heat flux values in a free and forced convection environment; resembling to a real-time braking phenomenon which could easily be modelled into a full cum scaled version of the alloy brake rotor part in ABAQUS. The finite element of a brake rotor was modelled in a constrained environment such that the nodal temperature between the contact surfaces of the coatings and substrate (Wrought Aluminum alloy) resemble an amalgamated solid brake rotor element. The initial results obtained were for a Plasma Electrolytic Oxidized (PEO) substrate wherein the Aluminum alloy gets a hard ceramic oxide layer grown on its transitional phase. The rotor was modelled and then evaluated in real-time for a constant ‘g’ braking event (based upon the mathematical heat flux input and convective surroundings), which reflected the necessity to deposit a conducting coat (sacrificial) above the PEO layer in order to inhibit thermal degradation of the barrier coating prematurely. Taguchi study was then used to bring out certain critical factors which may influence the maximum operating temperature of a multi-coated brake disc by simulating brake tests: a) an Alpine descent lasting 50 seconds; b) an Autobahn stop lasting 3.53 seconds; c) a Six–high speed repeated stop in accordance to FMVSS 135 lasting 46.25 seconds. Thermal Barrier coating thickness and Vane heat transfer coefficient were the two most influential factors and owing to their design and manufacturing constraints a final optimized model was obtained which survived the 6-high speed stop test as per the FMVSS -135 specifications. The simulation data highlighted the merits for preferring Wrought Aluminum alloy 7068 over Grey Cast Iron and Aluminum Metal Matrix Composite in coherence with the multiple coating depositions.

Keywords: lightweight brakes, surface modification, simulated braking, PEO, aluminum

Procedia PDF Downloads 408
3613 Multi-Agent Coverage Control with Bounded Gain Forgetting Composite Adaptive Controller

Authors: Mert Turanli, Hakan Temeltas

Abstract:

In this paper, we present an adaptive controller for decentralized coordination problem of multiple non-holonomic agents. The performance of the presented Multi-Agent Bounded Gain Forgetting (BGF) Composite Adaptive controller is compared against the tracking error criterion with a Feedback Linearization controller. By using the method, the sensor nodes move and reconfigure themselves in a coordinated way in response to a sensed environment. The multi-agent coordination is achieved through Centroidal Voronoi Tessellations and Coverage Control. Also, a consensus protocol is used for synchronization of the parameter vectors. The two controllers are given with their Lyapunov stability analysis and their stability is verified with simulation results. The simulations are carried out in MATLAB and ROS environments. Better performance is obtained with BGF Adaptive Controller.

Keywords: adaptive control, centroidal voronoi tessellations, composite adaptation, coordination, multi robots

Procedia PDF Downloads 348
3612 Using Single Decision Tree to Assess the Impact of Cutting Conditions on Vibration

Authors: S. Ghorbani, N. I. Polushin

Abstract:

Vibration during machining process is crucial since it affects cutting tool, machine, and workpiece leading to a tool wear, tool breakage, and an unacceptable surface roughness. This paper applies a nonparametric statistical method, single decision tree (SDT), to identify factors affecting on vibration in machining process. Workpiece material (AISI 1045 Steel, AA2024 Aluminum alloy, A48-class30 Gray Cast Iron), cutting tool (conventional, cutting tool with holes in toolholder, cutting tool filled up with epoxy-granite), tool overhang (41-65 mm), spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev) and depth of cut (0.05-0.15 mm) were used as input variables, while vibration was the output parameter. It is concluded that workpiece material is the most important parameters for natural frequency followed by cutting tool and overhang.

Keywords: cutting condition, vibration, natural frequency, decision tree, CART algorithm

Procedia PDF Downloads 336
3611 Analysis of the Result for the Accelerated Life Cycle Test of the Motor for Washing Machine by Using Acceleration Factor

Authors: Youn-Sung Kim, Jin-Ho Jo, Mi-Sung Kim, Jae-Kun Lee

Abstract:

Accelerated life cycle test is applied to various products or components in order to reduce the time of life cycle test in industry. It must be considered for many test conditions according to the product characteristics for the test and the selection of acceleration parameter is especially very important. We have carried out the general life cycle test and the accelerated life cycle test by applying the acceleration factor (AF) considering the characteristics of brushless DC (BLDC) motor for washing machine. The final purpose of this study is to verify the validity by analyzing the results of the general life cycle test and the accelerated life cycle test. It will make it possible to reduce the life test time through the reasonable accelerated life cycle test.

Keywords: accelerated life cycle test, reliability test, motor for washing machine, brushless dc motor test

Procedia PDF Downloads 611
3610 Polyclonal IgG glycosylation in Patients with Pediatric Appendicitis

Authors: Dalma Dojcsák, Csaba Váradi, Flóra Farkas, Tamás Farkas, János Papp, Béla Viskolcz

Abstract:

Background: Appendicitis is a common acute inflammatory condition in both children and adults, but current laboratory markers such as C-reactive protein (CRP), white blood cell count (WBC), absolute neutrophil count (ANC), and red blood cell count (RNC) lack specificity in detecting appendicitis-related inflammation. N-glycosylation, an asparagine-linked glycosylation process, plays a vital role in cellular interactions, angiogenesis, immune response, and effector functions. Altered N-glycosylation impacts tumor growth and both acute and chronic inflammatory processes. IgG, the second most abundant glycoprotein in serum, shows altered glycosylation patterns during inflammation, suggesting that IgG glycan modifications may serve as potential biomarkers for appendicitis. Specifically, increased levels of agalactosylated IgG glycans are a known feature of various inflammatory conditions, potentially including appendicitis. Identifying pediatric appendicitis remains challenging due to the absence of specific biomarkers, which makes diagnosis reliant on clinical symptoms, imaging such as ultrasound, and nonspecific lab indicators (e.g., CRP, WBC, ANC). In this study, we analyzed the IgG derived N-glycome in pediatric patients with appendicitis compared with healthy controls. Methodology: The N-glycome was analyzed by high-performance liquid-chromatography combined with mass spectrometry. IgG was isolated from serum samples by Protein G column. The IgG derived glycans were released by enzymatic deglycosylation and fluorescent tags were attached to each glycan moiety, which made necessitates the sample clean-up for further reliable quantitation. Overall, 38 controls and 40 serum samples diagnosed with pediatric appendicitis were analyzed by HILIC-MS methods. Multivariate statistical tests were performed with area percentage under the peak data derived from the integrated peaks, which were obtained from the chromatograms. Conclusions: Our results represented the altered N-glycome of IgG in pediatric appendicitis is similar with other observations. The glycosylation pattern reported so far for IgG is characterized by decreased galactosylation and sialylation, and an increase in fucosylation.

Keywords: N-glycosylation, liquid chromatography, mass spectrometry, inflammation, appendicitis, immunoglobulin G

Procedia PDF Downloads 9
3609 Association of Sleep Duration and Insomnia with Body Mass Index Among Brazilian Adults

Authors: Giovana Longo-Silva, Risia Cristina Egito de Menezes, Renan Serenini, Márcia de Oliveira Lima, Júlia Souza de Melo, Larissa de Lima Soares

Abstract:

Introduction: Sleep duration and quality have been increasingly recognized as important factors affecting overall health and well-being, including their potential impact on body weight and composition. Previous research has shown inconsistent results regarding the association between sleep patterns and body mass index (BMI), particularly among diverse populations such as Brazilian adults. Understanding these relationships is crucial for developing targeted interventions to address obesity and related health issues. Objective: This study aimed to investigate the association between sleep duration, insomnia, and BMI among Brazilian adults using data from a large national survey focused on chronic nutrition and sleep habits. Materials and Methods: The study included 2050 participants from a population-based virtual survey. BMI was calculated using self-reported weight and height measurements. Participants also reported usual bedtime and wake time on weekdays and weekends and whether they experienced symptoms of insomnia. The average sleep duration across the entire week was calculated as follows: [(5×sleep duration on weekdays) + (2×sleep duration on weekends)]/7. Linear regression analyses were conducted to assess the association between sleep duration, insomnia, and BMI, adjusting for potential confounding factors, including age, sex, marital status, physical exercise duration, and diet quality. Results: After adjusting for confounding variables, the study found that BMI decreased by 0.19 kg/m² for each additional hour of sleep duration (95% CI = -0.37, -0.02; P = 0.03). Conversely, individuals with insomnia had a higher BMI, with an increase of 0.75 kg/m² (95% CI = 0.28, 1.22; P = 0.002) compared to those without insomnia. Conclusions: The findings suggest a significant association between sleep duration, insomnia, and BMI among Brazilian adults. Longer sleep duration was associated with lower BMI, while insomnia was associated with higher BMI. These results underscore the importance of considering sleep patterns in strategies aimed at preventing and managing obesity in this population. Further research is needed to explore the underlying mechanisms and potential interventions targeting sleep-related factors to promote healthier body weight outcomes.

Keywords: sleep, obesity, chronobiology, nutrition

Procedia PDF Downloads 44
3608 Colonialism and Modernism in Architecture, the Case of a Blank Page Opportunity in Casablanka

Authors: Nezha Alaoui

Abstract:

The early 1950s French colonial context in Morocco provided an opportunity for architects to question the modernist established order by building dwellings for the local population. The dwellings were originally designed to encourage Muslims to adopt an urban lifestyle based on local customs. However, the inhabitants transformed their dwelling into a hybrid habitation. This paper aims to prove the relevance of the design process in accordance with the local colonial context by analyzing the dwellers' appropriation process and the modification of their habitat.

Keywords: colonial heritage, appropriation process, islamic spatial habit, housing experiment, modernist mass housing

Procedia PDF Downloads 128
3607 Cognitive Models of Future in Political Texts

Authors: Solopova Olga

Abstract:

The present paper briefly recalls theoretical preconditions for investigating cognitive-discursive models of future in political discourse. The author reviews theories and methods used for strengthening a future focus in this discourse working out two main tools – a model of future and a metaphorical scenario. The paper examines the implications of metaphorical analogies for modeling future in mass media. It argues that metaphor is not merely a rhetorical ornament in the political discourse of media regulation but a conceptual model that legislates and regulates our understanding of future.

Keywords: cognitive approach, future research, political discourse, model, scenario, metaphor

Procedia PDF Downloads 394
3606 Use of Artificial Intelligence Based Models to Estimate the Use of a Spectral Band in Cognitive Radio

Authors: Danilo López, Edwin Rivas, Fernando Pedraza

Abstract:

Currently, one of the major challenges in wireless networks is the optimal use of radio spectrum, which is managed inefficiently. One of the solutions to existing problem converges in the use of Cognitive Radio (CR), as an essential parameter so that the use of the available licensed spectrum is possible (by secondary users), well above the usage values that are currently detected; thus allowing the opportunistic use of the channel in the absence of primary users (PU). This article presents the results found when estimating or predicting the future use of a spectral transmission band (from the perspective of the PU) for a chaotic type channel arrival behavior. The time series prediction method (which the PU represents) used is ANFIS (Adaptive Neuro Fuzzy Inference System). The results obtained were compared to those delivered by the RNA (Artificial Neural Network) algorithm. The results show better performance in the characterization (modeling and prediction) with the ANFIS methodology.

Keywords: ANFIS, cognitive radio, prediction primary user, RNA

Procedia PDF Downloads 420
3605 Model Order Reduction for Frequency Response and Effect of Order of Method for Matching Condition

Authors: Aref Ghafouri, Mohammad javad Mollakazemi, Farhad Asadi

Abstract:

In this paper, model order reduction method is used for approximation in linear and nonlinearity aspects in some experimental data. This method can be used for obtaining offline reduced model for approximation of experimental data and can produce and follow the data and order of system and also it can match to experimental data in some frequency ratios. In this study, the method is compared in different experimental data and influence of choosing of order of the model reduction for obtaining the best and sufficient matching condition for following the data is investigated in format of imaginary and reality part of the frequency response curve and finally the effect and important parameter of number of order reduction in nonlinear experimental data is explained further.

Keywords: frequency response, order of model reduction, frequency matching condition, nonlinear experimental data

Procedia PDF Downloads 402
3604 Automatic Seizure Detection Using Weighted Permutation Entropy and Support Vector Machine

Authors: Noha Seddik, Sherine Youssef, Mohamed Kholeif

Abstract:

The automated epileptic seizure detection research field has emerged in the recent years; this involves analyzing the Electroencephalogram (EEG) signals instead of the traditional visual inspection performed by expert neurologists. In this study, a Support Vector Machine (SVM) that uses Weighted Permutation Entropy (WPE) as the input feature is proposed for classifying normal and seizure EEG records. WPE is a modified statistical parameter of the permutation entropy (PE) that measures the complexity and irregularity of a time series. It incorporates both the mapped ordinal pattern of the time series and the information contained in the amplitude of its sample points. The proposed system utilizes the fact that entropy based measures for the EEG segments during epileptic seizure are lower than in normal EEG.

Keywords: electroencephalogram (EEG), epileptic seizure detection, weighted permutation entropy (WPE), support vector machine (SVM)

Procedia PDF Downloads 371
3603 Calculation of Stress Intensity Factors in Rotating Disks Containing 3D Semi-Elliptical Cracks

Authors: Mahdi Fakoor, Seyed Mohammad Navid Ghoreishi

Abstract:

Initiation and propagation of cracks may cause catastrophic failures in rotating disks, and hence determination of fracture parameter in rotating disks under the different working condition is very important issue. In this paper, a comprehensive study of stress intensity factors in rotating disks containing 3D semi-elliptical cracks under the different working condition is investigated. In this regard, after verification of modeling and analytical procedure, the effects of mechanical properties, rotational velocity, and orientation of cracks on Stress Intensity Factors (SIF) in rotating disks under centrifugal loading are investigated. Also, the effects of using composite patch in reduction of SIF in rotating disks are studied. By that way, the effects of patching design variables like mechanical properties, thickness, and ply angle are investigated individually.

Keywords: stress intensity factor, semi-elliptical crack, rotating disk, finite element analysis (FEA)

Procedia PDF Downloads 364
3602 Reliability Analysis of Geometric Performance of Onboard Satellite Sensors: A Study on Location Accuracy

Authors: Ch. Sridevi, A. Chalapathi Rao, P. Srinivasulu

Abstract:

The location accuracy of data products is a critical parameter in assessing the geometric performance of satellite sensors. This study focuses on reliability analysis of onboard sensors to evaluate their performance in terms of location accuracy performance over time. The analysis utilizes field failure data and employs the weibull distribution to determine the reliability and in turn to understand the improvements or degradations over a period of time. The analysis begins by scrutinizing the location accuracy error which is the root mean square (RMS) error of differences between ground control point coordinates observed on the product and the map and identifying the failure data with reference to time. A significant challenge in this study is to thoroughly analyze the possibility of an infant mortality phase in the data. To address this, the Weibull distribution is utilized to determine if the data exhibits an infant stage or if it has transitioned into the operational phase. The shape parameter beta plays a crucial role in identifying this stage. Additionally, determining the exact start of the operational phase and the end of the infant stage poses another challenge as it is crucial to eliminate residual infant mortality or wear-out from the model, as it can significantly increase the total failure rate. To address this, an approach utilizing the well-established statistical Laplace test is applied to infer the behavior of sensors and to accurately ascertain the duration of different phases in the lifetime and the time required for stabilization. This approach also helps in understanding if the bathtub curve model, which accounts for the different phases in the lifetime of a product, is appropriate for the data and whether the thresholds for the infant period and wear-out phase are accurately estimated by validating the data in individual phases with Weibull distribution curve fitting analysis. Once the operational phase is determined, reliability is assessed using Weibull analysis. This analysis not only provides insights into the reliability of individual sensors with regards to location accuracy over the required period of time, but also establishes a model that can be applied to automate similar analyses for various sensors and parameters using field failure data. Furthermore, the identification of the best-performing sensor through this analysis serves as a benchmark for future missions and designs, ensuring continuous improvement in sensor performance and reliability. Overall, this study provides a methodology to accurately determine the duration of different phases in the life data of individual sensors. It enables an assessment of the time required for stabilization and provides insights into the reliability during the operational phase and the commencement of the wear-out phase. By employing this methodology, designers can make informed decisions regarding sensor performance with regards to location accuracy, contributing to enhanced accuracy in satellite-based applications.

Keywords: bathtub curve, geometric performance, Laplace test, location accuracy, reliability analysis, Weibull analysis

Procedia PDF Downloads 65
3601 Improvement of Reaction Technology of Decalin Halogenation

Authors: Dmitriy Yu. Korulkin, Ravshan M. Nuraliev, Raissa A. Muzychkina

Abstract:

In this research paper, we investigated the main regularities of a radical bromination reaction of decalin. We studied the temperature effect, durations of reaction, frequency rate of process, ratio of initial components, type and number of the initiator on decalin bromination degree. We found specified optimum conditions of synthesis of a perbromodecalin by the method of a decalin bromination. We developed the technological flowchart of receiving a perbromodecalin and the mass balance of process on the first and the subsequent loadings of components. The results of the research of antibacterial and antifungal activity of synthesized bromoderivatives have been represented.

Keywords: decalin, optimum technology, perbromodecalin, radical bromination

Procedia PDF Downloads 225