Search results for: unknown input observer
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3139

Search results for: unknown input observer

1819 Comparative Evaluation of Ultrasound Guided Internal Jugular Vein Cannulation Using Measured Guided Needle and Conventional Size Needle for Success and Complication of Cannulation

Authors: Devendra Gupta, Vikash Arya, Prabhat K. Singh

Abstract:

Background: Ultrasound guidance could be beneficial in placing central venous catheters by improving the success rate, reducing the number of needle passes, and decreasing complications. Central venous cannulation set has a single puncture needle of a fixed length of 6.4 cm. However, the average distance of midpoint of IJV to the skin is around 1 cm to 2 cm. The long length needle has tendency to go in depth more than required and this is very common during learning period of any individual. Therefore, we devised a long needle with a guard which can be adjusted according to the required length. Methods: After approval from the institute ethics committee and patient’s written informed consent, a prospective, randomized, single-blinded controlled study was conducted. Adult patient aged of both sexes with ASA grade 1-2 undergoing surgery requiring internal jugular venous (IJV) access was included. After intubation, the head was rotated to the contralateral side at 30 degree head rotation on the position of the right IJV. The transducer probe a 6.5 to 13-MHz linear transducer (Sonosite, USA) had been placed at the apex of triangle with minimal pressure to avoid IJV compression. The distance from skin to midpoint of the right IJV and skin to anterior wall of Common Carotid Artery (CCA) had been done using B-mode duplex sonography with a 6.5 to 13-MHz linear transducer. Depending upon the results of randomization 420 patients had been divided into two groups of equal numbers (n=210). Group 1. USG guided right sided IJV cannulation was done with conventional (6.4 cm) needle; and Group 2. USG guided right sided IJV cannulation was done with conventional (6.4 cm) needle with guard fixed to a required length (length between skin and midpoint of IJV) by an experienced anesthesiologist. Independent observer has noted the number of attempts and occurrence of complications (CCA puncture, pneumothorax or adjacent tissue damage). Results: Demographic data were similar in both the group. The groups were comparable when considered for relationship of IJV to CCA. There was no significant difference between groups as regard to distance of midpoint of IJV to the skin (p<0.05). IJV cannulation was successfully done in single attempts in 180 (85.7%), in two attempts in 27 (12.9%) and three attempts in 3 (1.4%) in group I, whereas in single attempt in 207 (98.6%) and second attempts in 3 (1.4%) in group II (p <0.000). Incidence of carotid artery puncture was significantly more in group I (7.1%) compared to group II (0%) (p<0.000). Incidence of adjacent tissue puncture was significantly more in group I (8.6%) compared to group II (0%) (p<0.000). Conclusion: Therefore IJV catheterization using guard over the needle at predefined length with the help of real-time ultrasound results in better success rates and lower immediate complications.

Keywords: ultrasound guided, internal jugular vein cannulation, measured guided needle, common carotid artery puncture

Procedia PDF Downloads 207
1818 Development of Risk Management System for Urban Railroad Underground Structures and Surrounding Ground

Authors: Y. K. Park, B. K. Kim, J. W. Lee, S. J. Lee

Abstract:

To assess the risk of the underground structures and surrounding ground, we collect basic data by the engineering method of measurement, exploration and surveys and, derive the risk through proper analysis and each assessment for urban railroad underground structures and surrounding ground including station inflow. Basic data are obtained by the fiber-optic sensors, MEMS sensors, water quantity/quality sensors, tunnel scanner, ground penetrating radar, light weight deflectometer, and are evaluated if they are more than the proper value or not. Based on these data, we analyze the risk level of urban railroad underground structures and surrounding ground. And we develop the risk management system to manage efficiently these data and to support a convenient interface environment at input/output of data.

Keywords: urban railroad, underground structures, ground subsidence, station inflow, risk

Procedia PDF Downloads 329
1817 Tip60’s Novel RNA-Binding Function Modulates Alternative Splicing of Pre-mRNA Targets Implicated in Alzheimer’s Disease

Authors: Felice Elefant, Akanksha Bhatnaghar, Keegan Krick, Elizabeth Heller

Abstract:

Context: The severity of Alzheimer’s Disease (AD) progression involves an interplay of genetics, age, and environmental factors orchestrated by histone acetyltransferase (HAT) mediated neuroepigenetic mechanisms. While disruption of Tip60 HAT action in neural gene control is implicated in AD, alternative mechanisms underlying Tip60 function remain unexplored. Altered RNA splicing has recently been highlighted as a widespread hallmark in the AD transcriptome that is implicated in the disease. Research Aim: The aim of this study was to identify a novel RNA binding/splicing function for Tip60 in human hippocampus and impaired in brains from AD fly models and AD patients. Methodology/Analysis: The authors used RNA immunoprecipitation using RNA isolated from 200 pooled wild type Drosophila brains for each of the 3 biological replicates. To identify Tip60’s RNA targets, they performed genome sequencing (DNB-SequencingTM technology, BGI genomics) on 3 replicates for Input RNA and RNA IPs by Tip60. Findings: The authors' transcriptomic analysis of RNA bound to Tip60 by Tip60-RNA immunoprecipitation (RIP) revealed Tip60 RNA targets enriched for critical neuronal processes implicated in AD. Remarkably, 79% of Tip60’s RNA targets overlap with its chromatin gene targets, supporting a model by which Tip60 orchestrates bi-level transcriptional regulation at both the chromatin and RNA level, a function unprecedented for any HAT to date. Since RNA splicing occurs co-transcriptionally and splicing defects are implicated in AD, the authors investigated whether Tip60-RNA targeting modulates splicing decisions and if this function is altered in AD. Replicate multivariate analysis of transcript splicing (rMATS) analysis of RNA-Seq data sets from wild-type and AD fly brains revealed a multitude of mammalian-like AS defects. Strikingly, over half of these altered RNAs were bonafide Tip60-RNA targets enriched for in the AD-gene curated database, with some AS alterations prevented against by increasing Tip60 in fly brain. Importantly, human orthologs of several Tip60-modulated spliced genes in Drosophila are well characterized aberrantly spliced genes in human AD brains, implicating disruption of Tip60’s splicing function in AD pathogenesis. Theoretical Importance: The authors' findings support a novel RNA interaction and splicing regulatory function for Tip60 that may underlie AS impairments that hallmark AD etiology. Data Collection: The authors collected data from RNA immunoprecipitation experiments using RNA isolated from 200 pooled wild type Drosophila brains for each of the 3 biological replicates. They also performed genome sequencing (DNBSequencingTM technology, BGI genomics) on 3 replicates for Input RNA and RNA IPs by Tip60. Questions: The question addressed by this study was whether Tip60 has a novel RNA binding/splicing function in human hippocampus and whether this function is impaired in brains from AD fly models and AD patients. Conclusions: The authors' findings support a novel RNA interaction and splicing regulatory function for Tip60 that may underlie AS impairments that hallmark AD etiology.

Keywords: Alzheimer's disease, cognition, aging, neuroepigenetics

Procedia PDF Downloads 68
1816 Method of Estimating Absolute Entropy of Municipal Solid Waste

Authors: Francis Chinweuba Eboh, Peter Ahlström, Tobias Richards

Abstract:

Entropy, as an outcome of the second law of thermodynamics, measures the level of irreversibility associated with any process. The identification and reduction of irreversibility in the energy conversion process helps to improve the efficiency of the system. The entropy of pure substances known as absolute entropy is determined at an absolute reference point and is useful in the thermodynamic analysis of chemical reactions; however, municipal solid waste (MSW) is a structurally complicated material with unknown absolute entropy. In this work, an empirical model to calculate the absolute entropy of MSW based on the content of carbon, hydrogen, oxygen, nitrogen, sulphur, and chlorine on a dry ash free basis (daf) is presented. The proposed model was derived from 117 relevant organic substances which represent the main constituents in MSW with known standard entropies using statistical analysis. The substances were divided into different waste fractions; namely, food, wood/paper, textiles/rubber and plastics waste and the standard entropies of each waste fraction and for the complete mixture were calculated. The correlation of the standard entropy of the complete waste mixture derived was found to be somsw= 0.0101C + 0.0630H + 0.0106O + 0.0108N + 0.0155S + 0.0084Cl (kJ.K-1.kg) and the present correlation can be used for estimating the absolute entropy of MSW by using the elemental compositions of the fuel within the range of 10.3%  C 95.1%, 0.0%  H  14.3%, 0.0%  O  71.1%, 0.0  N  66.7%, 0.0%  S  42.1%, 0.0%  Cl  89.7%. The model is also applicable for the efficient modelling of a combustion system in a waste-to-energy plant.

Keywords: absolute entropy, irreversibility, municipal solid waste, waste-to-energy

Procedia PDF Downloads 301
1815 Investigation of Cascade Loop Heat Pipes

Authors: Nandy Putra, Atrialdipa Duanovsah, Kristofer Haliansyah

Abstract:

The aim of this research is to design a LHP with low thermal resistance and low condenser temperature. A Self-designed cascade LHP was tested by using biomaterial, sintered copper powder, and aluminum screen mesh as the wick. Using pure water as the working fluid for the first level of the LHP and 96% alcohol as the working fluid for the second level of LHP, the experiments were run with 10W, 20W, and 30W heat input. Experimental result shows that the usage of biomaterial as wick could reduce more temperature at evaporator than by using sintered copper powder and screen mesh up to 22.63% and 37.41% respectively. The lowest thermal resistance occurred during the usage of biomaterial as wick of heat pipe, which is 2.06 oC/W. The usage of cascade system could be applied to LHP to reduce the temperature at condenser and reduced thermal resistance up to 17.6%.

Keywords: biomaterial, cascade loop heat pipe, screen mesh, sintered Cu

Procedia PDF Downloads 255
1814 Comprehensive Evaluation of Thermal Environment and Its Countermeasures: A Case Study of Beijing

Authors: Yike Lamu, Jieyu Tang, Jialin Wu, Jianyun Huang

Abstract:

With the development of economy and science and technology, the urban heat island effect becomes more and more serious. Taking Beijing city as an example, this paper divides the value of each influence index of heat island intensity and establishes a mathematical model – neural network system based on the fuzzy comprehensive evaluation index of heat island effect. After data preprocessing, the algorithm of weight of each factor affecting heat island effect is generated, and the data of sex indexes affecting heat island intensity of Shenyang City and Shanghai City, Beijing, and Hangzhou City are input, and the result is automatically output by the neural network system. It is of practical significance to show the intensity of heat island effect by visual method, which is simple, intuitive and can be dynamically monitored.

Keywords: heat island effect, neural network, comprehensive evaluation, visualization

Procedia PDF Downloads 129
1813 Fuzzy Logic Control for Flexible Joint Manipulator: An Experimental Implementation

Authors: Sophia Fry, Mahir Irtiza, Alexa Hoffman, Yousef Sardahi

Abstract:

This study presents an intelligent control algorithm for a flexible robotic arm. Fuzzy control is used to control the motion of the arm to maintain the arm tip at the desired position while reducing vibration and increasing the system speed of response. The Fuzzy controller (FC) is based on adding the tip angular position to the arm deflection angle and using their sum as a feedback signal to the control algorithm. This reduces the complexity of the FC in terms of the input variables, number of membership functions, fuzzy rules, and control structure. Also, the design of the fuzzy controller is model-free and uses only our knowledge about the system. To show the efficacy of the FC, the control algorithm is implemented on the flexible joint manipulator (FJM) developed by Quanser. The results show that the proposed control method is effective in terms of response time, overshoot, and vibration amplitude.

Keywords: fuzzy logic control, model-free control, flexible joint manipulators, nonlinear control

Procedia PDF Downloads 97
1812 The Processing of Context-Dependent and Context-Independent Scalar Implicatures

Authors: Liu Jia’nan

Abstract:

The default accounts hold the view that there exists a kind of scalar implicature which can be processed without context and own a psychological privilege over other scalar implicatures which depend on context. In contrast, the Relevance Theorist regards context as a must because all the scalar implicatures have to meet the need of relevance in discourse. However, in Katsos, the experimental results showed: Although quantitatively the adults rejected under-informative utterance with lexical scales (context-independent) and the ad hoc scales (context-dependent) at almost the same rate, adults still regarded the violation of utterance with lexical scales much more severe than with ad hoc scales. Neither default account nor Relevance Theory can fully explain this result. Thus, there are two questionable points to this result: (1) Is it possible that the strange discrepancy is due to other factors instead of the generation of scalar implicature? (2) Are the ad hoc scales truly formed under the possible influence from mental context? Do the participants generate scalar implicatures with ad hoc scales instead of just comparing semantic difference among target objects in the under- informative utterance? In my Experiment 1, the question (1) will be answered by repetition of Experiment 1 by Katsos. Test materials will be showed by PowerPoint in the form of pictures, and each procedure will be done under the guidance of a tester in a quiet room. Our Experiment 2 is intended to answer question (2). The test material of picture will be transformed into the literal words in DMDX and the target sentence will be showed word-by-word to participants in the soundproof room in our lab. Reading time of target parts, i.e. words containing scalar implicatures, will be recorded. We presume that in the group with lexical scale, standardized pragmatically mental context would help generate scalar implicature once the scalar word occurs, which will make the participants hope the upcoming words to be informative. Thus if the new input after scalar word is under-informative, more time will be cost for the extra semantic processing. However, in the group with ad hoc scale, scalar implicature may hardly be generated without the support from fixed mental context of scale. Thus, whether the new input is informative or not does not matter at all, and the reading time of target parts will be the same in informative and under-informative utterances. People’s mind may be a dynamic system, in which lots of factors would co-occur. If Katsos’ experimental result is reliable, will it shed light on the interplay of default accounts and context factors in scalar implicature processing? We might be able to assume, based on our experiments, that one single dominant processing paradigm may not be plausible. Furthermore, in the processing of scalar implicature, the semantic interpretation and the pragmatic interpretation may be made in a dynamic interplay in the mind. As to the lexical scale, the pragmatic reading may prevail over the semantic reading because of its greater exposure in daily language use, which may also lead the possible default or standardized paradigm override the role of context. However, those objects in ad hoc scale are not usually treated as scalar membership in mental context, and thus lexical-semantic association of the objects may prevent their pragmatic reading from generating scalar implicature. Only when the sufficient contextual factors are highlighted, can the pragmatic reading get privilege and generate scalar implicature.

Keywords: scalar implicature, ad hoc scale, dynamic interplay, default account, Mandarin Chinese processing

Procedia PDF Downloads 311
1811 Using Machine Learning to Monitor the Condition of the Cutting Edge during Milling Hardened Steel

Authors: Pawel Twardowski, Maciej Tabaszewski, Jakub Czyżycki

Abstract:

The main goal of the work was to use machine learning to predict cutting-edge wear. The research was carried out while milling hardened steel with sintered carbide cutters at various cutting speeds. During the tests, cutting-edge wear was measured, and vibration acceleration signals were also measured. Appropriate measures were determined from the vibration signals and served as input data in the machine-learning process. Two approaches were used in this work. The first one involved a two-state classification of the cutting edge - suitable and unfit for further work. In the second approach, prediction of the cutting-edge state based on vibration signals was used. The obtained research results show that the appropriate use of machine learning algorithms gives excellent results related to monitoring cutting edge during the process.

Keywords: milling of hardened steel, tool wear, vibrations, machine learning

Procedia PDF Downloads 46
1810 Electrophysiological Correlates of Statistical Learning in Children with and without Developmental Language Disorder

Authors: Ana Paula Soares, Alexandrina Lages, Helena Oliveira, Francisco-Javier Gutiérrez-Domínguez, Marisa Lousada

Abstract:

From an early age, exposure to a spoken language allows us to implicitly capture the structure underlying the succession of the speech sounds in that language and to segment it into meaningful units (words). Statistical learning (SL), i.e., the ability to pick up patterns in the sensory environment even without intention or consciousness of doing it, is thus assumed to play a central role in the acquisition of the rule-governed aspects of language and possibly to lie behind the language difficulties exhibited by children with development language disorder (DLD). The research conducted so far has, however, led to inconsistent results, which might stem from the behavioral tasks used to test SL. In a classic SL experiment, participants are first exposed to a continuous stream (e.g., syllables) in which, unbeknownst to the participants, stimuli are grouped into triplets that always appear together in the stream (e.g., ‘tokibu’, ‘tipolu’), with no pauses between each other (e.g., ‘tokibutipolugopilatokibu’) and without any information regarding the task or the stimuli. Following exposure, SL is assessed by asking participants to discriminate between triplets previously presented (‘tokibu’) from new sequences never presented together during exposure (‘kipopi’), i.e., to perform a two-alternative-forced-choice (2-AFC) task. Despite the widespread use of the 2-AFC to test SL, it has come under increasing criticism as it is an offline post-learning task that only assesses the result of the learning that had occurred during the previous exposure phase and that might be affected by other factors beyond the computation of regularities embedded in the input, typically the likelihood two syllables occurring together, a statistic known as transitional probability (TP). One solution to overcome these limitations is to assess SL as exposure to the stream unfolds using online techniques such as event-related potentials (ERP) that is highly sensitive to the time-course of the learning in the brain. Here we collected ERPs to examine the neurofunctional correlates of SL in preschool children with DLD, and chronological-age typical language development (TLD) controls who were exposed to an auditory stream in which eight three-syllable nonsense words, four of which presenting high-TPs and the other four low-TPs, to further analyze whether the ability of DLD and TLD children to extract-word-like units from the steam was modulated by words’ predictability. Moreover, to ascertain if the previous knowledge of the to-be-learned-regularities affected the neural responses to high- and low-TP words, children performed the auditory SL task, firstly, under implicit, and, subsequently, under explicit conditions. Although behavioral evidence of SL was not obtained in either group, the neural responses elicited during the exposure phases of the SL tasks differentiated children with DLD from children with TLD. Specifically, the results indicated that only children from the TDL group showed neural evidence of SL, particularly in the SL task performed under explicit conditions, firstly, for the low-TP, and, subsequently, for the high-TP ‘words’. Taken together, these findings support the view that children with DLD showed deficits in the extraction of the regularities embedded in the auditory input which might underlie the language difficulties.

Keywords: development language disorder, statistical learning, transitional probabilities, word segmentation

Procedia PDF Downloads 185
1809 Numerical Investigation of the Influence on Buckling Behaviour Due to Different Launching Bearings

Authors: Nadine Maier, Martin Mensinger, Enea Tallushi

Abstract:

In general, today, two types of launching bearings are used in the construction of large steel and steel concrete composite bridges. These are sliding rockers and systems with hydraulic bearings. The advantages and disadvantages of the respective systems are under discussion. During incremental launching, the center of the webs of the superstructure is not perfectly in line with the center of the launching bearings due to unavoidable tolerances, which may have an influence on the buckling behavior of the web plates. These imperfections are not considered in the current design against plate buckling, according to DIN EN 1993-1-5. It is therefore investigated whether the design rules have to take into account any eccentricities which occur during incremental launching and also if this depends on the respective launching bearing. Therefore, at the Technical University Munich, large-scale buckling tests were carried out on longitudinally stiffened plates under biaxial stresses with the two different types of launching bearings and eccentric load introduction. Based on the experimental results, a numerical model was validated. Currently, we are evaluating different parameters for both types of launching bearings, such as load introduction length, load eccentricity, the distance between longitudinal stiffeners, the position of the rotation point of the spherical bearing, which are used within the hydraulic bearings, web, and flange thickness and imperfections. The imperfection depends on the geometry of the buckling field and whether local or global buckling occurs. This and also the size of the meshing is taken into account in the numerical calculations of the parametric study. As a geometric imperfection, the scaled first buckling mode is applied. A bilinear material curve is used so that a GMNIA analysis is performed to determine the load capacity. Stresses and displacements are evaluated in different directions, and specific stress ratios are determined at the critical points of the plate at the time of the converging load step. To evaluate the load introduction of the transverse load, the transverse stress concentration is plotted on a defined longitudinal section on the web. In the same way, the rotation of the flange is evaluated in order to show the influence of the different degrees of freedom of the launching bearings under eccentric load introduction and to be able to make an assessment for the case, which is relevant in practice. The input and the output are automatized and depend on the given parameters. Thus we are able to adapt our model to different geometric dimensions and load conditions. The programming is done with the help of APDL and a Python code. This allows us to evaluate and compare more parameters faster. Input and output errors are also avoided. It is, therefore, possible to evaluate a large spectrum of parameters in a short time, which allows a practical evaluation of different parameters for buckling behavior. This paper presents the results of the tests as well as the validation and parameterization of the numerical model and shows the first influences on the buckling behavior under eccentric and multi-axial load introduction.

Keywords: buckling behavior, eccentric load introduction, incremental launching, large scale buckling tests, multi axial stress states, parametric numerical modelling

Procedia PDF Downloads 106
1808 Simple Procedure for Probability Calculation of Tensile Crack Occurring in Rigid Pavement: A Case Study

Authors: Aleš Florian, Lenka Ševelová, Jaroslav Žák

Abstract:

Formation of tensile cracks in concrete slabs of rigid pavement can be (among others) the initiation point of the other, more serious failures which can ultimately lead to complete degradation of the concrete slab and thus the whole pavement. Two measures can be used for reliability assessment of this phenomenon - the probability of failure and/or the reliability index. Different methods can be used for their calculation. The simple ones are called moment methods and simulation techniques. Two methods - FOSM Method and Simple Random Sampling Method - are verified and their comparison is performed. The influence of information about the probability distribution and the statistical parameters of input variables as well as of the limit state function on the calculated reliability index and failure probability are studied in three points on the lower surface of concrete slabs of the older type of rigid pavement formerly used in the Czech Republic.

Keywords: failure, pavement, probability, reliability index, simulation, tensile crack

Procedia PDF Downloads 541
1807 Geometric and Algebraic Properties of the Eigenvalues of Monotone Matrices

Authors: Brando Vagenende, Marie-Anne Guerry

Abstract:

For stochastic matrices of any order, the geometric description of the convex set of eigenvalues is completely known. The purpose of this study is to investigate the subset of the monotone matrices. This type of matrix appears in contexts such as intergenerational occupational mobility, equal-input modeling, and credit ratings-based systems. Monotone matrices are stochastic matrices in which each row stochastically dominates the previous row. The monotonicity property of a stochastic matrix can be expressed by a nonnegative lower-order matrix with the same eigenvalues as the original monotone matrix (except for the eigenvalue 1). Specifically, the aim of this research is to focus on the properties of eigenvalues of monotone matrices. For those matrices up to order 3, there already exists a complete description of the convex set of eigenvalues. For monotone matrices of order at least 4, this study gives, through simulations, more insight into the geometric description of their eigenvalues. Furthermore, this research treats in a geometric and algebraic way the properties of eigenvalues of monotone matrices of order at least 4.

Keywords: eigenvalues of matrices, finite Markov chains, monotone matrices, nonnegative matrices, stochastic matrices

Procedia PDF Downloads 71
1806 Annual Water Level Simulation Using Support Vector Machine

Authors: Maryam Khalilzadeh Poshtegal, Seyed Ahmad Mirbagheri, Mojtaba Noury

Abstract:

In this paper, by application of the input yearly data of rainfall, temperature and flow to the Urmia Lake, the simulation of water level fluctuation were applied by means of three models. According to the climate change investigation the fluctuation of lakes water level are of high interest. This study investigate data-driven models, support vector machines (SVM), SVM method which is a new regression procedure in water resources are applied to the yearly level data of Lake Urmia that is the biggest and the hyper saline lake in Iran. The evaluated lake levels are found to be in good correlation with the observed values. The results of SVM simulation show better accuracy and implementation. The mean square errors, mean absolute relative errors and determination coefficient statistics are used as comparison criteria.

Keywords: simulation, water level fluctuation, urmia lake, support vector machine

Procedia PDF Downloads 358
1805 Hydrodynamic Analysis with Heat Transfer in Solid Gas Fluidized Bed Reactor for Solar Thermal Applications

Authors: Sam Rasoulzadeh, Atefeh Mousavi

Abstract:

Fluidized bed reactors are known as highly exothermic and endothermic according to uniformity in temperature as a safe and effective mean for catalytic reactors. In these reactors, a wide range of catalyst particles can be used and by using a continuous operation proceed to produce in succession. Providing optimal conditions for the operation of these types of reactors will prevent the exorbitant costs necessary to carry out laboratory work. In this regard, a hydrodynamic analysis was carried out with heat transfer in the solid-gas fluidized bed reactor for solar thermal applications. The results showed that in the fluid flow the input of the reactor has a lower temperature than the outlet, and when the fluid is passing from the reactor, the heat transfer happens between cylinder and solar panel and fluid. It increases the fluid temperature in the outlet pump and also the kinetic energy of the fluid has been raised in the outlet areas.

Keywords: heat transfer, solar reactor, fluidized bed reactor, CFD, computational fluid dynamics

Procedia PDF Downloads 168
1804 Crosssampler: A Digital Convolution Cross Synthesis Instrument

Authors: Jimmy Eadie

Abstract:

Convolutional Cross Synthesis (CCS) has emerged as a powerful technique for blending input signals to create hybrid sounds. It has significantly expanded the horizons of digital signal processing, enabling artists to explore audio effects. However, the conventional applications of CCS primarily revolve around reverberation and room simulation rather than being utilized as a creative synthesis method. In this paper, we present the design of a digital instrument called CrossSampler that harnesses a parametric approach to convolution cross-synthesis, which involves using adjustable parameters to control the blending of audio signals through convolution. These parameters allow for customization of the resulting sound, offering greater creative control and flexibility. It enables users to shape the output by manipulating factors such as duration, intensity, and spectral characteristics. This approach facilitates experimentation and exploration in sound design and opens new sonic possibilities.

Keywords: convolution, synthesis, sampling, virtual instrument

Procedia PDF Downloads 56
1803 Dairy Products on the Algerian Market: Proportion of Imitation and Degree of Processing

Authors: Bentayeb-Ait Lounis Saïda, Cheref Zahia, Cherifi Thizi, Ri Kahina Bahmed, Kahina Hallali Yasmine Abdellaoui, Kenza Adli

Abstract:

Algeria is the leading consumer of dairy products in North Africa. This is a fact. However, the nutritional quality of the latter remains unknown. The aim of this study is to characterise the dairy products available on the Algerian market in order to assess whether they constitute a healthy and safe choice. To do this, it collected data on the labelling of 390 dairy products, including cheese, yoghurt, UHT milk and milk drinks, infant formula and dairy creams. We assessed their degree of processing according to the NOVA classification, as well as the proportion of imitation products. The study was carried out between March 2020 and August 2023. The results show that 88% are ultra-processed; 84% for 'cheese', 92% for dairy creams, 92% for 'yoghurt', 100% for infant formula, 92% for margarines and 36% for UHT milk/dairy drinks. As for imitation/analogue dairy products, the study revealed the following proportions: 100% for infant formula, 78% for butter/margarine, 18% for UHT milk/milk-based drinks, 54% for cheese, 2% for camembert and 75% for dairy cream. The harmful effects of consuming ultra-processed products on long-term health are increasingly documented in dozens of publications. The findings of this study sound the alarm about the health risks to which Algerian consumers are exposed. Various scientific, economic and industrial bodies need to be involved in order to safeguard consumer health in both the short and long term. Food awareness and education campaigns should be organised.

Keywords: dairy, UPF, NOVA, yoghurt, cheese

Procedia PDF Downloads 24
1802 Impact of Cellular Recharge Technology 8 (CRT8) Supplement Mix on Cellular Senescence and Autophagy in 2D Cellular and 3D Organoid Models

Authors: Vincent Gauthier, Om Rangacharya, Jordan Lemorande, Senthil Sundaram, James Green, Ghada Alsaleh

Abstract:

Introduction: Cellular senescence and autophagy are critical processes in aging and age-related diseases. This study examines the effects of Cellular Recharge Technology 8 (CRT8), a proprietary blend of certain supplements on both two-dimensional Human dermal fibroblasts (HDF) and three-dimensional organoid models. Some of the individual ingredients of CRT8 are known to individually have an effect on cellular senescence and autophagy, but their overall effect in the proprietary blend is unknown. The focus of this study is on the impact of CRT8 on cellular senescence and autophagy. Methods: Human dermal fibroblasts (HDF) were treated with CRT8 in 2D cellular studies, assessing effects on p16 (senescence marker) Mean Fluorescence Intensity (MFI), and LC3 (autophagy marker) MFI. The same CRT8 mix was applied to 3D organoid models, composed of HDF and endothelial cells, with results measured after 6 weeks of treatment, focusing on autophagy and senescence markers. Results: In 2D cellular models, CRT8 significantly reduced p16 MFI levels by 40%, indicating a potential decrease in cellular senescence, and increased LC3 MFI by more than 2-fold, demonstrating enhanced autophagic activity. Similarly, in the 3D organoid models, CRT8 significantly reduced p16 MFI levels by 63% and increased autophagic flux by 1.62-fold, further reinforcing its potential to reduce senescence and promote autophagy. Conclusion: CRT8 effectively reduces p16 levels and enhances autophagy in both 2D and 3D models, highlighting its potential as a powerful anti-aging intervention. CRT8 may be valuable in promoting cellular health and longevity.

Keywords: cellular senescence, autophagy, aging, supplements

Procedia PDF Downloads 7
1801 Targeting NLRP3 Inflammasome Activation: A New Mechanism Underlying the Protective Effects of Nafamostat Against Acute Pancreatitis

Authors: Jiandong Ren, Lijun Zhao, Peng Chen

Abstract:

Nafamostat (NA), a synthetic broad-spectrum serine protease inhibitor, has been routinely employed for the treatment of acute pancreatitis (AP) and other inflammatory-associated diseases in some East Asia countries. Although the potent inhibitory activity against inflammation-related proteases such as thrombin, trypsin, kallikrein, plasmin, coagulation factors and complement factors is generally considered to be responsible for the anti-inflammatory effects of NA, precise target and molecular mechanism underlying the anti-inflammatory activity in the treatment of AP remain largely unknown yet. As an intracellular inflammatory signaling platform, the NOD-like receptor protein 3 (NLRP3) inflammasome is recently identified to be involved in the development of AP. In present study, we have revealed that NA alleviated pancreatic injury in a caerulein-induced AP model by inhibiting the NLRP3 inflammasome activation in pancreas. Mechanistically, NA interacted with HDAC6, a cytoplasmic deacetylase implicated in the NLRP3 inflammasome pathway, and efficiently abrogated the function of HDAC6. This property enabled NA to influence HDAC6 dependent NF-κB transcriptional activity and thus block NF-κB-driven transcriptional priming of NLRP3 inflammasome. Moreover, NA exerted the potential to interfere HDAC6-mediated intracellular transport of NLRP3, thereby leading to the failure of NLRP3 inflammasome activation. Our current work has provided valuable insight into the molecular mechanism underlying the immunomodulatory effect of NA in treatment of AP, highlighting its promising application in prevention of NLRP3 inflammasome-associated inflammatory pathological damage.

Keywords: acute pancreatitis, HDAC6, nafamostat, NLRP3 inflammasome

Procedia PDF Downloads 60
1800 Neural Synchronization - The Brain’s Transfer of Sensory Data

Authors: David Edgar

Abstract:

To understand how the brain’s subconscious and conscious functions, we must conquer the physics of Unity, which leads to duality’s algorithm. Where the subconscious (bottom-up) and conscious (top-down) processes function together to produce and consume intelligence, we use terms like ‘time is relative,’ but we really do understand the meaning. In the brain, there are different processes and, therefore, different observers. These different processes experience time at different rates. A sensory system such as the eyes cycles measurement around 33 milliseconds, the conscious process of the frontal lobe cycles at 300 milliseconds, and the subconscious process of the thalamus cycle at 5 milliseconds. Three different observers experience time differently. To bridge observers, the thalamus, which is the fastest of the processes, maintains a synchronous state and entangles the different components of the brain’s physical process. The entanglements form a synchronous cohesion between the brain components allowing them to share the same state and execute in the same measurement cycle. The thalamus uses the shared state to control the firing sequence of the brain’s linear subconscious process. Sharing state also allows the brain to cheat on the amount of sensory data that must be exchanged between components. Only unpredictable motion is transferred through the synchronous state because predictable motion already exists in the shared framework. The brain’s synchronous subconscious process is entirely based on energy conservation, where prediction regulates energy usage. So, the eyes every 33 milliseconds dump their sensory data into the thalamus every day. The thalamus is going to perform a motion measurement to identify the unpredictable motion in the sensory data. Here is the trick. The thalamus conducts its measurement based on the original observation time of the sensory system (33 ms), not its own process time (5 ms). This creates a data payload of synchronous motion that preserves the original sensory observation. Basically, a frozen moment in time (Flat 4D). The single moment in time can then be processed through the single state maintained by the synchronous process. Other processes, such as consciousness (300 ms), can interface with the synchronous state to generate awareness of that moment. Now, synchronous data traveling through a separate faster synchronous process creates a theoretical time tunnel where observation time is tunneled through the synchronous process and is reproduced on the other side in the original time-relativity. The synchronous process eliminates time dilation by simply removing itself from the equation so that its own process time does not alter the experience. To the original observer, the measurement appears to be instantaneous, but in the thalamus, a linear subconscious process generating sensory perception and thought production is being executed. It is all just occurring in the time available because other observation times are slower than thalamic measurement time. For life to exist in the physical universe requires a linear measurement process, it just hides by operating at a faster time relativity. What’s interesting is time dilation is not the problem; it’s the solution. Einstein said there was no universal time.

Keywords: neural synchronization, natural intelligence, 99.95% IoT data transmission savings, artificial subconscious intelligence (ASI)

Procedia PDF Downloads 119
1799 The Effect of Environmental Consciousness on Firm Performance

Authors: Hossein Emari, Hossein Vazifehdoust, Hashem Nikoo Maram

Abstract:

This study aims to develop an original framework of Environmental Consciousness (EC) to explore the positive effect of environmental consciousness on financial performance through the partial mediator - green intellectual capital. A questionnaire survey on the environmental consciousness, intellectual capital, and financial performance of Iran’s manufacturing firms was conducted, and 324 samples were analyzed. This study utilizes structural equation modeling to explore the direct and indirect influences of EC on financial performance. Research results reveal that environmental consciousness had an indirect impact on financial performance through investment in green intellectual capital. It was thus known that green intellectual capital is a mediator of the relationship between environmental consciousness and financial performance. This paper may serve as a reference for firms mapping out future environmental policies and provide an input of various perspectives and arguments into the discipline of green management.

Keywords: environmental consciousness, social responsibility, green intellectual capital, financial performance

Procedia PDF Downloads 481
1798 Application of Neural Networks to Predict Changing the Diameters of Bubbles in Pool Boiling Distilled Water

Authors: V. Nikkhah Rashidabad, M. Manteghian, M. Masoumi, S. Mousavian, D. Ashouri

Abstract:

In this research, the capability of neural networks in modeling and learning complicated and nonlinear relations has been used to develop a model for the prediction of changes in the diameter of bubbles in pool boiling distilled water. The input parameters used in the development of this network include element temperature, heat flux, and retention time of bubbles. The test data obtained from the experiment of the pool boiling of distilled water, and the measurement of the bubbles form on the cylindrical element. The model was developed based on training algorithm, which is typologically of back-propagation type. Considering the correlation coefficient obtained from this model is 0.9633. This shows that this model can be trusted for the simulation and modeling of the size of bubble and thermal transfer of boiling.

Keywords: bubble diameter, heat flux, neural network, training algorithm

Procedia PDF Downloads 439
1797 Exploring the Strategy to Identify Seed-Specific Acyl-Hydrolases from Arabidopsis thaliana by Activity-Based Protein Profiling

Authors: M. Latha, Achintya K. Dolui, P. Vijayaraj

Abstract:

Vegetable oils mainly triacylglycerol (TAG) are an essential nutrient in the human diet as well as one of the major global commodity. There is a pressing need to enhance the yield of oil production to meet the world’s growing demand. Oil content is controlled by the balance between synthesis and breakdown in the cells. Several studies have established to increase the oil content by the overexpression of oil biosynthetic enzymes. Interestingly the significant oil accumulation was observed with impaired TAG hydrolysis. Unfortunately, the structural, as well as the biochemical properties of the lipase enzymes, is widely unknown, and so far, no candidate gene was identified in seeds except sugar-dependent1 (SDP1). Evidence has shown that SDP1directly responsible for initiation of oil breakdown in the seeds during germination. The present study is the identification of seed-specific acyl-hydrolases by activity based proteome profiling (ABPP) using Arabidopsis thaliana as a model system. The ABPP reveals that around 8 to 10 proteins having the serine hydrolase domain and are expressed during germination of Arabidopsis seed. The N-term sequencing, as well as LC-MS/MS analysis, was performed for the differentially expressed protein during germination. The coding region of the identified proteins was cloned, and lipases activity was assessed with purified recombinant protein. The enzyme assay was performed against various lipid substrates, and we have observed the acylhydrolase activity towards lysophosphatidylcholine and monoacylglycerol. Further, the functional characteristic of the identified protein will reveal the physiological significance the enzyme in oil accumulation.

Keywords: lipase, lipids, vegetable oil, triacylglycerol

Procedia PDF Downloads 174
1796 Trust and Conflict Resolution: Relationship Building for Learning

Authors: Jeff Dickie

Abstract:

This research paper combined grounded coding and research questions with the objective to investigate conflict resolution in the classroom. The students’ answers concerning teaching were coded according to phrasal meanings which revealed concepts. These concept codes then became input data into theoretical frameworks. The investigation indicated two conflicts: whether the information was valid and whether to make the study effort which was discussed as perceptions of teacher’s competence in helping to learn. The relevant factors in helping to learn were predominately emotional. These factors were important in the negotiation process to develop relationships. Information validity seemed to be the motivator to begin and participate effectively with the learning process. In effect, confidence in the learning negotiation process with the focus towards relationship building with the subject matter seemed to be the motivator to make the study effort.

Keywords: coding, confidence, competence, conflict resolution, risk, trust, relationship building

Procedia PDF Downloads 424
1795 A Design of Beam-Steerable Antenna Array for Use in Future Mobile Handsets

Authors: Naser Ojaroudi Parchin, Atta Ullah, Haleh Jahanbakhsh Basherlou, Raed A. Abd-Alhameed, Peter S. Excell

Abstract:

A design of beam-steerable antenna array for the future cellular communication (5G) is presented. The proposed design contains eight elements of compact end-fire antennas arranged on the top edge of smartphone printed circuit board (PCB). Configuration of the antenna element consists of the conductive patterns on the top and bottom copper foil layers and a substrate layer with a via-hole. The simulated results including input-impedance and also fundamental radiation properties have been presented and discussed. The impedance bandwidth (S11 ≤ -10 dB) of the antenna spans from 17.5 to 21 GHz (more than 3 GHz bandwidth) with a resonance at 19 GHz. The antenna exhibits end-fire (directional) radiation beams with wide-angle scanning property and could be used for the future 5G beam-forming. Furthermore, the characteristics of the array design in the vicinity of user-hand are studied.

Keywords: beam-steering, end-fire radiation mode, mobile-phone antenna, phased array

Procedia PDF Downloads 148
1794 Seismic Soil-Pile Interaction Considering Nonlinear Soil Column Behavior in Saturated and Dry Soil Conditions

Authors: Mohammad Moeini, Mehrdad Ghyabi, Kiarash Mohtasham Dolatshahi

Abstract:

This paper investigates seismic soil-pile interaction using the Beam on Nonlinear Winkler Foundation (BNWF) approach. Three soil types are considered to cover all the possible responses, as well as nonlinear site response analysis using finite element method in OpenSees platform. Excitations at each elevation that are output of the site response analysis are used as the input excitation to the soil pile system implementing multi-support excitation method. Spectral intensities of acceleration show that the extent of the response in sand is more severe than that of clay, in addition, increasing the PGA of ground strong motion will affect the sandy soil more, in comparison with clayey medium, which is an indicator of the sensitivity of soil-pile systems in sandy soil.

Keywords: BNWF method, multi-support excitation, nonlinear site response analysis, seismic soil-pile interaction

Procedia PDF Downloads 381
1793 Spatial Rank-Based High-Dimensional Monitoring through Random Projection

Authors: Chen Zhang, Nan Chen

Abstract:

High-dimensional process monitoring becomes increasingly important in many application domains, where usually the process distribution is unknown and much more complicated than the normal distribution, and the between-stream correlation can not be neglected. However, since the process dimension is generally much bigger than the reference sample size, most traditional nonparametric multivariate control charts fail in high-dimensional cases due to the curse of dimensionality. Furthermore, when the process goes out of control, the influenced variables are quite sparse compared with the whole dimension, which increases the detection difficulty. Targeting at these issues, this paper proposes a new nonparametric monitoring scheme for high-dimensional processes. This scheme first projects the high-dimensional process into several subprocesses using random projections for dimension reduction. Then, for every subprocess with the dimension much smaller than the reference sample size, a local nonparametric control chart is constructed based on the spatial rank test to detect changes in this subprocess. Finally, the results of all the local charts are fused together for decision. Furthermore, after an out-of-control (OC) alarm is triggered, a diagnostic framework is proposed. using the square-root LASSO. Numerical studies demonstrate that the chart has satisfactory detection power for sparse OC changes and robust performance for non-normally distributed data, The diagnostic framework is also effective to identify truly changed variables. Finally, a real-data example is presented to demonstrate the application of the proposed method.

Keywords: random projection, high-dimensional process control, spatial rank, sequential change detection

Procedia PDF Downloads 294
1792 Towards a Methodology for the Assessment of Neighbourhood Design for Happiness

Authors: Tina Pujara

Abstract:

Urban and regional research in the new emerging inter-disciplinary field of happiness is seemingly limited. However, it is progressively being recognized that there is enormous potential for social and behavioral scientists to add a spatial dimension to it. In fact, the happiness of communities can be notably influenced by the design and maintenance of the neighborhoods they inhabit. The probable key reasons being that places can facilitate human social connections and relationships. While it is increasingly being acknowledged that some neighborhood designs appear better suited for social connectedness than others, the plausible reasons for places to deter these characteristics and perhaps their influence on happiness are outwardly unknown. In addition, an explicit step wise methodology to assess neighborhood designs for happiness (of their communities) is not known to exist. This paper is an attempt towards developing such a methodological framework. The paper presents the development of a methodological framework for assessing neighborhood designs for happiness, with a particular focus on the outdoor shared spaces in neighborhoods. The developed methodological framework of investigation follows a mixed method approach and draws upon four different sources of information. The framework proposes an empirical examination of the contribution of neighborhood factors, particularly outdoor shared spaces, to individual happiness. One of the main tools proposed for this empirical examination is Jan Gehl’s Public Space Public Life (PSPL) Survey. The developed framework, as presented in the paper, is a contribution towards the development of a consolidated methodology for assessing neighborhood designs for happiness, which can further serve as a unique tool to inform urban designers, architects and other decision makers.

Keywords: happiness, methodology, neighbourhood design, outdoor shared spaces

Procedia PDF Downloads 159
1791 Control of a Stewart Platform for Minimizing Impact Energy in Simulating Spacecraft Docking Operations

Authors: Leonardo Herrera, Shield B. Lin, Stephen J. Montgomery-Smith, Ziraguen O. Williams

Abstract:

Three control algorithms: Proportional-Integral-Derivative, Linear-Quadratic-Gaussian, and Linear-Quadratic-Gaussian with the shift, were applied to the computer simulation of a one-directional dynamic model of a Stewart Platform. The goal was to compare the dynamic system responses under the three control algorithms and to minimize the impact energy when simulating spacecraft docking operations. Equations were derived for the control algorithms and the input and output of the feedback control system. Using MATLAB, Simulink diagrams were created to represent the three control schemes. A switch selector was used for the convenience of changing among different controllers. The simulation demonstrated the controller using the algorithm of Linear-Quadratic-Gaussian with the shift resulting in the lowest impact energy.

Keywords: controller, Stewart platform, docking operation, spacecraft

Procedia PDF Downloads 39
1790 Measuring Environmental Efficiency of Energy in OPEC Countries

Authors: Bahram Fathi, Seyedhossein Sajadifar, Naser Khiabani

Abstract:

Data envelopment analysis (DEA) has recently gained popularity in energy efficiency analysis. A common feature of the previously proposed DEA models for measuring energy efficiency performance is that they treat energy consumption as an input within a production framework without considering undesirable outputs. However, energy use results in the generation of undesirable outputs as byproducts of producing desirable outputs. Within a joint production framework of both desirable and undesirable outputs, this paper presents several DEA-type linear programming models for measuring energy efficiency performance. In addition to considering undesirable outputs, our models treat different energy sources as different inputs so that changes in energy mix could be accounted for in evaluating energy efficiency. The proposed models are applied to measure the energy efficiency performances of 12 OPEC countries and the results obtained are presented.

Keywords: energy efficiency, undesirable outputs, data envelopment analysis

Procedia PDF Downloads 730