Search results for: matrix geometric method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20462

Search results for: matrix geometric method

19802 An Improved Multiple Scattering Reflectance Model Based on Specular V-Cavity

Authors: Hongbin Yang, Mingxue Liao, Changwen Zheng, Mengyao Kong, Chaohui Liu

Abstract:

Microfacet-based reflection models are widely used to model light reflections for rough surfaces. Microfacet models have become the standard surface material building block for describing specular components with varying roughness; and yet, while they possess many desirable properties as well as produce convincing results, their design ignores important sources of scattering, which can cause a significant loss of energy. Specifically, they only simulate the single scattering on the microfacets and ignore the subsequent interactions. As the roughness increases, the interaction will become more and more important. So a multiple-scattering microfacet model based on specular V-cavity is presented for this important open problem. However, it spends much unnecessary rendering time because of setting the same number of scatterings for different roughness surfaces. In this paper, we design a geometric attenuation term G to compute the BRDF (Bidirectional reflection distribution function) of multiple scattering of rough surfaces. Moreover, we consider determining the number of scattering by deterministic heuristics for different roughness surfaces. As a result, our model produces a similar appearance of the objects with the state of the art model with significantly improved rendering efficiency. Finally, we derive a multiple scattering BRDF based on the original microfacet framework.

Keywords: bidirectional reflection distribution function, BRDF, geometric attenuation term, multiple scattering, V-cavity model

Procedia PDF Downloads 106
19801 Signal Processing Techniques for Adaptive Beamforming with Robustness

Authors: Ju-Hong Lee, Ching-Wei Liao

Abstract:

Adaptive beamforming using antenna array of sensors is useful in the process of adaptively detecting and preserving the presence of the desired signal while suppressing the interference and the background noise. For conventional adaptive array beamforming, we require a prior information of either the impinging direction or the waveform of the desired signal to adapt the weights. The adaptive weights of an antenna array beamformer under a steered-beam constraint are calculated by minimizing the output power of the beamformer subject to the constraint that forces the beamformer to make a constant response in the steering direction. Hence, the performance of the beamformer is very sensitive to the accuracy of the steering operation. In the literature, it is well known that the performance of an adaptive beamformer will be deteriorated by any steering angle error encountered in many practical applications, e.g., the wireless communication systems with massive antennas deployed at the base station and user equipment. Hence, developing effective signal processing techniques to deal with the problem due to steering angle error for array beamforming systems has become an important research work. In this paper, we present an effective signal processing technique for constructing an adaptive beamformer against the steering angle error. The proposed array beamformer adaptively estimates the actual direction of the desired signal by using the presumed steering vector and the received array data snapshots. Based on the presumed steering vector and a preset angle range for steering mismatch tolerance, we first create a matrix related to the direction vector of signal sources. Two projection matrices are generated from the matrix. The projection matrix associated with the desired signal information and the received array data are utilized to iteratively estimate the actual direction vector of the desired signal. The estimated direction vector of the desired signal is then used for appropriately finding the quiescent weight vector. The other projection matrix is set to be the signal blocking matrix required for performing adaptive beamforming. Accordingly, the proposed beamformer consists of adaptive quiescent weights and partially adaptive weights. Several computer simulation examples are provided for evaluating and comparing the proposed technique with the existing robust techniques.

Keywords: adaptive beamforming, robustness, signal blocking, steering angle error

Procedia PDF Downloads 114
19800 Nanofibrous Ion Exchangers

Authors: Jaromír Marek, Jakub Wiener, Yan Wang

Abstract:

The main goal of this study was to find simple and industrially applicable production of ion exchangers based on nanofibrous polystyrene matrix and characterization of prepared material. Starting polystyrene nanofibers were sulfonated and crosslinked under appropriate conditions at the same time by sulfuric acid. Strongly acidic cation exchanger was obtained in such a way. The polymer matrix was made from polystyrene nanofibers prepared by Nanospider technology. Various types postpolymerization reactions and other methods of crosslinking were studied. Greatly different behavior between nano and microsize materials was observed. The final nanofibrous material was characterized and compared to common granular ion exchangers and available microfibrous ion exchangers. The sorption properties of nanofibrous ion exchangers were compared with the granular ion exchangers. For nanofibrous ion exchangers of comparable ion exchange capacity was observed considerably faster adsorption kinetics.

Keywords: electrospinning, ion exchangers, nanofibers, polystyrene

Procedia PDF Downloads 245
19799 Combline Cavity Bandpass Filter Design and Implementation Using EM Simulation Tool

Authors: Taha Ahmed Özbey, Sedat Nazlıbilek, Alparslan Çağrı Yapıcı

Abstract:

Combline cavity filters have gained significant attention in recent years due to their exceptional narrowband characteristics, high unloaded Q, remarkable out-of-band rejection, and versatile post-manufacturing tuning capabilities. These filters play a vital role in various wireless communication systems, radar applications, and other advanced technologies where stringent frequency selectivity and superior performance are required. This paper represents combined cavity filter design and implementation by coupling matrix synthesis. Limited filter length, 50 dB out-of-band rejection, and agile design were aimed. To do so, CAD tools and intuitive methods were used.

Keywords: cavity, band pass filter, cavity combline filter, coupling matrix synthesis

Procedia PDF Downloads 63
19798 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem

Authors: Newroz Nooralddin Abdulrazaq, Thuraya Mahmood Qaradaghi

Abstract:

The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.

Keywords: McEliece cryptosystem, Goppa code, separable, irreducible

Procedia PDF Downloads 258
19797 Determination of Marbofloxacin in Pig Plasma Using LC-MS/MS and Its Application to the Pharmacokinetic Studies

Authors: Jeong Woo Kang, MiYoung Baek, Ki-Suk Kim, Kwang-Jick Lee, ByungJae So

Abstract:

Introduction: A fast, easy and sensitive detection method was developed and validated by liquid chromatography tandem mass spectrometry for the determination of marbofloxacin in pig plasma which was further applied to study the pharmacokinetics of marbofloxacin. Materials and Methods: The plasma sample (500 μL) was mixed with 1.5 ml of 0.1% formic acid in MeCN to precipitate plasma proteins. After shaking for 20 min, The mixture was centrifuged at 5,000 × g for 30 min. It was dried under a nitrogen flow at 50℃. 500 μL aliquot of the sample was injected into the LC-MS/MS system. Chromatographic analysis was carried out mobile phase gradient consisting 0.1% formic acid in D.W. (A) and 0.1% formic acid in MeCN (B) with C18 reverse phase column. Mass spectrometry was performed using the positive ion mode and the selected ion monitoring (MRM). Results and Conclusions: The method validation was performed in the sample matrix. Good linearities (R2>0.999) were observed and the quantified average recoveries of marbofloxacin were 87 - 92% at level of 10 ng g-1 -100 ng g-1. The percent of coefficient of variation (CV) for the described method was less than 10 % over the range of concentrations studied. The limits of detection (LOD) and quantification (LOQ) were 2 and 5 ng g-1, respectively. This method has also been applied successfully to pharmacokinetic analysis of marbofloxacin after intravenous (IV), intramuscular (IM) and oral administration (PO). The mean peak plasma concentration (Cmax) was 2,597 ng g-1at 0.25 h, 2,587 ng g-1at 0.44 h and 2,355 ng g-1at 1.58 h for IV, IM and PO, respectively. The area under the plasma concentration-time curve (AUC0–t) was 24.8, 29.0 and 25.2 h μg/mL for IV, IM and PO, respectively. The elimination half-life (T1/2) was 8.6, 13.1 and 9.5 for IV, IM and PO, respectively. Bioavailability (F) of the marbofloxacin in pig was 117 and 101 % for IM and PO, respectively. Based on these result, marbofloxacin does not have any obstacles as therapeutics to develop the oral formulations such as tablets and capsules.

Keywords: marbofloxacin, LC-MS/MS, pharmacokinetics, chromatographic

Procedia PDF Downloads 532
19796 Classification of Barley Varieties by Artificial Neural Networks

Authors: Alper Taner, Yesim Benal Oztekin, Huseyin Duran

Abstract:

In this study, an Artificial Neural Network (ANN) was developed in order to classify barley varieties. For this purpose, physical properties of barley varieties were determined and ANN techniques were used. The physical properties of 8 barley varieties grown in Turkey, namely thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain, were determined and it was found that these properties were statistically significant with respect to varieties. As ANN model, three models, N-l, N-2 and N-3 were constructed. The performances of these models were compared. It was determined that the best-fit model was N-1. In the N-1 model, the structure of the model was designed to be 11 input layers, 2 hidden layers and 1 output layer. Thousand kernel weight, geometric mean diameter, sphericity, kernel volume, surface area, bulk density, true density, porosity and colour parameters of grain were used as input parameter; and varieties as output parameter. R2, Root Mean Square Error and Mean Error for the N-l model were found as 99.99%, 0.00074 and 0.009%, respectively. All results obtained by the N-l model were observed to have been quite consistent with real data. By this model, it would be possible to construct automation systems for classification and cleaning in flourmills.

Keywords: physical properties, artificial neural networks, barley, classification

Procedia PDF Downloads 167
19795 Evaluation of Methodologies for Measuring Harmonics and Inter-Harmonics in Photovoltaic Facilities

Authors: Anésio de Leles Ferreira Filho, Wesley Rodrigues de Oliveira, Jéssica Santoro Gonçalves, Jorge Andrés Cormane Angarita

Abstract:

The increase in electric power demand in face of environmental issues has intensified the participation of renewable energy sources such as photovoltaics, in the energy matrix of various countries. Due to their operational characteristics, they can generate time-varying harmonic and inter-harmonic distortions. For this reason, the application of methods of measurement based on traditional Fourier analysis, as proposed by IEC 61000-4-7, can provide inaccurate results. Considering the aspects mentioned herein, came the idea of the development of this work which aims to present the results of a comparative evaluation between a methodology arising from the combination of the Prony method with the Kalman filter and another method based on the IEC 61000-4-30 and IEC 61000-4-7 standards. Employed in this study were synthetic signals and data acquired through measurements in a 50kWp photovoltaic installation.

Keywords: harmonics, inter-harmonics, iec61000-4-7, parametric estimators, photovoltaic generation

Procedia PDF Downloads 474
19794 PathoPy2.0: Application of Fractal Geometry for Early Detection and Histopathological Analysis of Lung Cancer

Authors: Rhea Kapoor

Abstract:

Fractal dimension provides a way to characterize non-geometric shapes like those found in nature. The purpose of this research is to estimate Minkowski fractal dimension of human lung images for early detection of lung cancer. Lung cancer is the leading cause of death among all types of cancer and an early histopathological analysis will help reduce deaths primarily due to late diagnosis. A Python application program, PathoPy2.0, was developed for analyzing medical images in pixelated format and estimating Minkowski fractal dimension using a new box-counting algorithm that allows windowing of images for more accurate calculation in the suspected areas of cancerous growth. Benchmark geometric fractals were used to validate the accuracy of the program and changes in fractal dimension of lung images to indicate the presence of issues in the lung. The accuracy of the program for the benchmark examples was between 93-99% of known values of the fractal dimensions. Fractal dimension values were then calculated for lung images, from National Cancer Institute, taken over time to correctly detect the presence of cancerous growth. For example, as the fractal dimension for a given lung increased from 1.19 to 1.27 due to cancerous growth, it represents a significant change in fractal dimension which lies between 1 and 2 for 2-D images. Based on the results obtained on many lung test cases, it was concluded that fractal dimension of human lungs can be used to diagnose lung cancer early. The ideas behind PathoPy2.0 can also be applied to study patterns in the electrical activity of the human brain and DNA matching.

Keywords: fractals, histopathological analysis, image processing, lung cancer, Minkowski dimension

Procedia PDF Downloads 162
19793 Approximation of Periodic Functions Belonging to Lipschitz Classes by Product Matrix Means of Fourier Series

Authors: Smita Sonker, Uaday Singh

Abstract:

Various investigators have determined the degree of approximation of functions belonging to the classes W(L r , ξ(t)), Lip(ξ(t), r), Lip(α, r), and Lipα using different summability methods with monotonocity conditions. Recently, Lal has determined the degree of approximation of the functions belonging to Lipα and W(L r , ξ(t)) classes by using Ces`aro-N¨orlund (C 1 .Np)- summability with non-increasing weights {pn}. In this paper, we shall determine the degree of approximation of 2π - periodic functions f belonging to the function classes Lipα and W(L r , ξ(t)) by C 1 .T - means of Fourier series of f. Our theorems generalize the results of Lal and we also improve these results in the light off. From our results, we also derive some corollaries.

Keywords: Lipschitz classes, product matrix operator, signals, trigonometric Fourier approximation

Procedia PDF Downloads 468
19792 Graph Similarity: Algebraic Model and Its Application to Nonuniform Signal Processing

Authors: Nileshkumar Vishnav, Aditya Tatu

Abstract:

A recent approach of representing graph signals and graph filters as polynomials is useful for graph signal processing. In this approach, the adjacency matrix plays pivotal role; instead of the more common approach involving graph-Laplacian. In this work, we follow the adjacency matrix based approach and corresponding algebraic signal model. We further expand the theory and introduce the concept of similarity of two graphs. The similarity of graphs is useful in that key properties (such as filter-response, algebra related to graph) get transferred from one graph to another. We demonstrate potential applications of the relation between two similar graphs, such as nonuniform filter design, DTMF detection and signal reconstruction.

Keywords: graph signal processing, algebraic signal processing, graph similarity, isospectral graphs, nonuniform signal processing

Procedia PDF Downloads 336
19791 Lightweight Sheet Molding Compound Composites by Coating Glass Fiber with Cellulose Nanocrystals

Authors: Amir Asadi, Karim Habib, Robert J. Moon, Kyriaki Kalaitzidou

Abstract:

There has been considerable interest in cellulose nanomaterials (CN) as polymer and polymer composites reinforcement due to their high specific modulus and strength, low density and toxicity, and accessible hydroxyl side groups that can be readily chemically modified. The focus of this study is making lightweight composites for better fuel efficiency and lower CO2 emission in auto industries with no compromise on mechanical performance using a scalable technique that can be easily integrated in sheet molding compound (SMC) manufacturing lines. Light weighting will be achieved by replacing part of the heavier components, i.e. glass fibers (GF), with a small amount of cellulose nanocrystals (CNC) in short GF/epoxy composites made using SMC. CNC will be introduced as coating of the GF rovings prior to their use in the SMC line. The employed coating method is similar to the fiber sizing technique commonly used and thus it can be easily scaled and integrated to industrial SMC lines. This will be an alternative route to the most techniques that involve dispersing CN in polymer matrix, in which the nanomaterials agglomeration limits the capability for scaling up in an industrial production. We have demonstrated that incorporating CNC as a coating on GF surface by immersing the GF in CNC aqueous suspensions, a simple and scalable technique, increases the interfacial shear strength (IFSS) by ~69% compared to the composites produced by uncoated GF, suggesting an enhancement of stress transfer across the GF/matrix interface. As a result of IFSS enhancement, incorporation of 0.17 wt% CNC in the composite results in increases of ~10% in both elastic modulus and tensile strength, and 40 % and 43 % in flexural modulus and strength respectively. We have also determined that dispersing 1.4 and 2 wt% CNC in the epoxy matrix of short GF/epoxy SMC composites by sonication allows removing 10 wt% GF with no penalty on tensile and flexural properties leading to 7.5% lighter composites. Although sonication is a scalable technique, it is not quite as simple and inexpensive as coating the GF by passing through an aqueous suspension of CNC. In this study, the above findings are integrated to 1) investigate the effect of CNC content on mechanical properties by passing the GF rovings through CNC aqueous suspension with various concentrations (0-5%) and 2) determine the optimum ratio of the added CNC to the removed GF to achieve the maximum possible weight reduction with no cost on mechanical performance of the SMC composites. The results of this study are of industrial relevance, providing a path toward producing high volume lightweight and mechanically enhanced SMC composites using cellulose nanomaterials.

Keywords: cellulose nanocrystals, light weight polymer-matrix composites, mechanical properties, sheet molding compound (SMC)

Procedia PDF Downloads 213
19790 Maker-Based Learning in Secondary Mathematics: Investigating Students’ Proportional Reasoning Understanding through Digital Making

Authors: Juan Torralba

Abstract:

Student digital artifacts were investigated, utilizing a qualitative exploratory research design to understand the ways in which students represented their knowledge of seventh-grade proportionality concepts as they participated in maker-based activities that culminated in the creation of digital 3-dimensional models of their dream homes. Representations of the geometric and numeric dimensions of proportionality were analyzed in the written, verbal, and visual data collected from the students. A directed content analysis approach was utilized in the data analysis, as this work aimed to build upon existing research in the field of maker-based STEAM Education. The results from this work show that students can represent their understanding of proportional reasoning through open-ended written responses more accurately than through verbal descriptions or digital artifacts. The geometric and numeric dimensions of proportionality and their respective components of attributes of similarity representation and percents, rates, and ratios representations were the most represented by the students than any other across the data, suggesting a maker-based instructional approach to teaching proportionality in the middle grades may be promising in helping students gain a solid foundation in those components. Recommendations for practice and research are discussed.

Keywords: learning through making, maker-based education, maker education in the middle grades, making in mathematics, the maker movement

Procedia PDF Downloads 60
19789 A Monolithic Arbitrary Lagrangian-Eulerian Finite Element Strategy for Partly Submerged Solid in Incompressible Fluid with Mortar Method for Modeling the Contact Surface

Authors: Suman Dutta, Manish Agrawal, C. S. Jog

Abstract:

Accurate computation of hydrodynamic forces on floating structures and their deformation finds application in the ocean and naval engineering and wave energy harvesting. This manuscript presents a monolithic, finite element strategy for fluid-structure interaction involving hyper-elastic solids partly submerged in an incompressible fluid. A velocity-based Arbitrary Lagrangian-Eulerian (ALE) formulation has been used for the fluid and a displacement-based Lagrangian approach has been used for the solid. The flexibility of the ALE technique permits us to treat the free surface of the fluid as a Lagrangian entity. At the interface, the continuity of displacement, velocity and traction are enforced using the mortar method. In the mortar method, the constraints are enforced in a weak sense using the Lagrange multiplier method. In the literature, the mortar method has been shown to be robust in solving various contact mechanics problems. The time-stepping strategy used in this work reduces to the generalized trapezoidal rule in the Eulerian setting. In the Lagrangian limit, in the absence of external load, the algorithm conserves the linear and angular momentum and the total energy of the system. The use of monolithic coupling with an energy-conserving time-stepping strategy gives an unconditionally stable algorithm and allows the user to take large time steps. All the governing equations and boundary conditions have been mapped to the reference configuration. The use of the exact tangent stiffness matrix ensures that the algorithm converges quadratically within each time step. The robustness and good performance of the proposed method are demonstrated by solving benchmark problems from the literature.

Keywords: ALE, floating body, fluid-structure interaction, monolithic, mortar method

Procedia PDF Downloads 268
19788 Contextual SenSe Model: Word Sense Disambiguation using Sense and Sense Value of Context Surrounding the Target

Authors: Vishal Raj, Noorhan Abbas

Abstract:

Ambiguity in NLP (Natural language processing) refers to the ability of a word, phrase, sentence, or text to have multiple meanings. This results in various kinds of ambiguities such as lexical, syntactic, semantic, anaphoric and referential am-biguities. This study is focused mainly on solving the issue of Lexical ambiguity. Word Sense Disambiguation (WSD) is an NLP technique that aims to resolve lexical ambiguity by determining the correct meaning of a word within a given context. Most WSD solutions rely on words for training and testing, but we have used lemma and Part of Speech (POS) tokens of words for training and testing. Lemma adds generality and POS adds properties of word into token. We have designed a novel method to create an affinity matrix to calculate the affinity be-tween any pair of lemma_POS (a token where lemma and POS of word are joined by underscore) of given training set. Additionally, we have devised an al-gorithm to create the sense clusters of tokens using affinity matrix under hierar-chy of POS of lemma. Furthermore, three different mechanisms to predict the sense of target word using the affinity/similarity value are devised. Each contex-tual token contributes to the sense of target word with some value and whichever sense gets higher value becomes the sense of target word. So, contextual tokens play a key role in creating sense clusters and predicting the sense of target word, hence, the model is named Contextual SenSe Model (CSM). CSM exhibits a noteworthy simplicity and explication lucidity in contrast to contemporary deep learning models characterized by intricacy, time-intensive processes, and chal-lenging explication. CSM is trained on SemCor training data and evaluated on SemEval test dataset. The results indicate that despite the naivety of the method, it achieves promising results when compared to the Most Frequent Sense (MFS) model.

Keywords: word sense disambiguation (wsd), contextual sense model (csm), most frequent sense (mfs), part of speech (pos), natural language processing (nlp), oov (out of vocabulary), lemma_pos (a token where lemma and pos of word are joined by underscore), information retrieval (ir), machine translation (mt)

Procedia PDF Downloads 90
19787 Effects of Oxidized LDL in M2 Macrophages: Implications in Atherosclerosis

Authors: Fernanda Gonçalves, Karla Alcântara, Vanessa Moura, Patrícia Nolasco, Jorge Kalil, Maristela Hernandez

Abstract:

Introduction: Atherosclerosis is a chronic disease where two striking features are observed: retention of lipids and inflammation. Understanding the interaction between immune cells and lipoproteins involved in atherogenesis are urgent challenges, since cardiovascular diseases are the leading cause of death worldwide. Macrophages are critical to the development of atherosclerotic plaques and in the perpetuation of inflammation in these lesions. These cells are also directly involved in unstable plaque rupture. Recently different populations of macrophages are being identified in atherosclerotic lesions. Although the presence of M2 macrophages (macrophages activated by the alternative pathway, eg. The IL-4) has been identified, the function of these cells in atherosclerosis is not yet defined. M2 macrophages have a high endocytic capacity, they promote remodeling of tissues and to have anti-inflammatory activity. However, in atherosclerosis, especially unstable plaques, severe inflammatory reaction, accumulation of cellular debris and intense degradation of the tissue is observed. Thus, it is possible that the M2 macrophages have altered function (phenotype) in atherosclerosis. Objective: Our aim is to evaluate if the presence of oxidized LDL alters the phenotype and function of M2 macrophages in vitro. Methods: For this, we will evaluate whether the addition of lipoprotein in M2 macrophages differentiated in vitro with IL -4 induces 1) a reduction in the secretion of anti-inflammatory cytokines (CBA and ELISA), 2) secretion of inflammatory cytokines (CBA and ELISA), 3) expression of cell activation markers (Flow cytometry), 4) alteration in gene expression of molecules adhesion and extracellular matrix (Real-Time PCR) and 5) Matrix degradation (confocal microscopy). Results: In oxLDL stimulated M2 macrophages cultures we did not find any differences in the expression of the cell surface markers tested, including: HLA-DR, CD80, CD86, CD206, CD163 and CD36. Also, cultures stimulated with oxLDL had similar phagocytic capacity when compared to unstimulated cells. However, in the supernatant of these cultures an increase in the secretion of the pro-inflammatory cytokine IL-8 was detected. No significant changes where observed in IL-6, IL-10, IL-12 and IL-1b levels. The culture supernatant also induced massive extracellular matrix (produced by mouse embryo fibroblast) filaments degradation. When evaluating the expression of 84 extracellular matrix and adhesion molecules genes, we observed that the stimulation of oxLDL in M2 macrophages decreased 47% of the genes and increased the expression of only 3% of the genes. In particular we noted that oxLDL inhibit the expression of 60% of the genes constituents of extracellular matrix and collagen expressed by these cells, including fibronectin1 and collagen VI. We also observed a decrease in the expression of matrix protease inhibitors, such as TIMP 2. On the opposite, the matricellular protein thrombospondin had a 12 fold increase in gene expression. In the presence of native LDL 90% of the genes had no altered expression. Conclusion: M2 macrophages stimulated with oxLDL secrete the pro-inflammatory cytokine IL-8, have an altered extracellular matrix constituents gene expression, and promote the degradation of extracellular matrix. M2 macrophages may contribute to the perpetuation of inflammation in atherosclerosis and to plaque rupture.

Keywords: atherosclerosis, LDL, macrophages, m2

Procedia PDF Downloads 325
19786 Effect of the Applied Bias on Mini-Band Structures in Dimer Fibonacci InAs/Ga1-XInXAs Superlattices

Authors: Z. Aziz, S. Terkhi, Y. Sefir, R. Djelti, S. Bentata

Abstract:

The effect of a uniform electric field across multi-barrier systems (InAs/InxGa1-xAs) is exhaustively explored by a computational model using exact Airy function formalism and the transfer-matrix technique. In the case of biased DFHBSL structure a strong reduction in transmission properties was observed and the width of the mini-band structure linearly decreases with the increase of the applied bias. This is due to the confinement of the states in the mini-band structure, which becomes increasingly important (Wannier-Stark Effect).

Keywords: dimer fibonacci height barrier superlattices, singular extended state, exact Airy function and transfer matrix formalism, bioinformatics

Procedia PDF Downloads 275
19785 Formation of Miniband Structure in Dimer Fibonacci GaAs/Ga1-XAlXAs Superlattices

Authors: Aziz Zoubir, Sefir Yamina, Djelti Redouan, Bentata Samir

Abstract:

The effect of a uniform electric field across multibarrier systems (GaAs/AlxGa1-xAs) is exhaustively explored by a computational model using exact Airy function formalism and the transfer-matrix technique. In the case of biased Dimer Fibonacci Height Barrier superlattices (DFHBSL) structure a strong reduction in transmission properties was observed and the width of the miniband structure linearly decreases with the increase of the applied bias. This is due to the confinement of the states in the miniband structure, which becomes increasingly important (Wannier-Stark effect).

Keywords: Dimer Fibonacci Height Barrier superlattices, singular extended states, exact Airy function, transfer matrix formalism

Procedia PDF Downloads 497
19784 The Study of Aluminum Effects Layer Austenite Twins Adjacent to K-Carbide Plates in the Cellular Structure of a Mn-Al Alloy Steel

Authors: Wu Wei-Ting, Liu Po-Yen, Chang Chin-Tzu, Cheng Wei-Chun

Abstract:

Three types of low-temperature phase transformations in an Fe-12.5 Mn-6.53 Al-1.28 C (wt %) alloy have been studied. The steel underwent solution heat treatment at 1100℃ and isothermal holding at low temperatures. γ’ phase appears in the austenite matrix in the air-cooled steel. Coherent ultra-fine particles of γ’ phase precipitated uniformly in the austenite matrix after the air-cooling process. These ultra-fine particles were very small and only could be detected by TEM through dark-field images. After short periods of isothermal holding at low temperatures these particles of γ’ phase grew and could be easily detected by TEM. A pro-eutectoid reaction happened after isothermal holding at temperatures below 875 ℃. Proeutectoid κ-carbide and ferrite appear in the austenite matrix as grain boundary precipitates and cellular precipitates. The cellular precipitates are composed of lamellar κ-carbide and austenite. The lamellar κ-carbide grains are always accompanied by layers of austenite twins. The presence of twin layers adhering to the κ-carbide plates might be attributed to the lower activation energy for the precipitation of κ-carbide plates in the austenite. The final form of phase transformation is the eutectoid reaction for the decomposition of supersaturated austenite into stable κ-carbide and ferrite phases at temperatures below 700℃. The ferrite and κ-carbide are in the form of pearlite lamellae.

Keywords: austenite, austenite twin layers, κ-carbide, twins

Procedia PDF Downloads 216
19783 The Study of Intangible Assets at Various Firm States

Authors: Gulnara Galeeva, Yulia Kasperskaya

Abstract:

The study deals with the relevant problem related to the formation of the efficient investment portfolio of an enterprise. The structure of the investment portfolio is connected to the degree of influence of intangible assets on the enterprise’s income. This determines the importance of research on the content of intangible assets. However, intangible assets studies do not take into consideration how the enterprise state can affect the content and the importance of intangible assets for the enterprise`s income. This affects accurateness of the calculations. In order to study this problem, the research was divided into several stages. In the first stage, intangible assets were classified based on their synergies as the underlying intangibles and the additional intangibles. In the second stage, this classification was applied. It showed that the lifecycle model and the theory of abrupt development of the enterprise, that are taken into account while designing investment projects, constitute limit cases of a more general theory of bifurcations. The research identified that the qualitative content of intangible assets significant depends on how close the enterprise is to being in crisis. In the third stage, the author developed and applied the Wide Pairwise Comparison Matrix method. This allowed to establish that using the ratio of the standard deviation to the mean value of the elements of the vector of priority of intangible assets makes it possible to estimate the probability of a full-blown crisis of the enterprise. The author has identified a criterion, which allows making fundamental decisions on investment feasibility. The study also developed an additional rapid method of assessing the enterprise overall status based on using the questionnaire survey with its Director. The questionnaire consists only of two questions. The research specifically focused on the fundamental role of stochastic resonance in the emergence of bifurcation (crisis) in the economic development of the enterprise. The synergetic approach made it possible to describe the mechanism of the crisis start in details and also to identify a range of universal ways of overcoming the crisis. It was outlined that the structure of intangible assets transforms into a more organized state with the strengthened synchronization of all processes as a result of the impact of the sporadic (white) noise. Obtained results offer managers and business owners a simple and an affordable method of investment portfolio optimization, which takes into account how close the enterprise is to a state of a full-blown crisis.

Keywords: analytic hierarchy process, bifurcation, investment portfolio, intangible assets, wide matrix

Procedia PDF Downloads 197
19782 Analysis of Two-Phase Flow Instabilities in Conventional Channel of Nuclear Power Reactor

Authors: M. Abdur Rashid Sarkar, Riffat Mahmud

Abstract:

Boiling heat transfer plays a crucial role in cooling nuclear reactor for safe electricity generation. A two phase flow is susceptible to thermal-hydrodynamic instabilities, which may cause flow oscillations of constant amplitude or diverging amplitude. These oscillations may induce boiling crisis, disturb control systems, or cause mechanical damage. Based on their mechanisms, various types of instabilities can be classified for a nuclear reactor. From a practical engineering point of view one of the major design difficulties in dealing with multiphase flow is that the mass, momentum, and energy transfer rates and processes may be quite sensitive to the geometric configuration of the heat transfer surface. Moreover, the flow within each phase or component will clearly depend on that geometric configuration. The complexity of this two-way coupling presents a major challenge in the study of multiphase flows and there is much that remains to be done. Yet, the parametric effects on flow instability such as the effect of aspect ratio, pressure drop, channel length, its orientation inlet subcooling and surface roughness etc. have been analyzed. Another frequently occurring instability, known as the Kelvin–Helmholtz instability has been briefly reviewed. Various analytical techniques for predicting parametric effect on the instability are analyzed in terms of their applicability and accuracy.

Keywords: two phase flows, boiling crisis, thermal-hydrodynamic instabilities, water cooled nuclear reactors, kelvin–helmholtz instability

Procedia PDF Downloads 388
19781 Augmented Reality: New Relations with the Architectural Heritage Education

Authors: Carla Maria Furuno Rimkus

Abstract:

The technologies related to virtual reality and augmented reality in combination with mobile technologies, are being more consolidated and used each day. The increasing technological availability along with the decrease of their acquisition and maintenance costs, have favored the expansion of its use in the field of historic heritage. In this context it is focused, in this article, on the potential of mobile applications in the dissemination of the architectural heritage, using the technology of Augmented Reality. From this perspective approach, it is discussed about the process of producing an application for mobile devices on the Android platform, which combines the technologies of geometric modeling with augmented reality (AR) and access to interactive multimedia contents with cultural, social and historic information of the historic building that we take as the object of study: a block with a set of buildings built in the XVIII century, known as "Quarteirão dos Trapiches", which was modeled in 3D, coated with the original texture of its facades and displayed on AR. From this perspective approach, this paper discusses about methodological aspects of the development of this application regarding to the process and the project development tools, and presents our considerations on methodological aspects of developing an application for the Android system, focused on the dissemination of the architectural heritage, in order to encourage the tourist potential of the city in a sustainable way and to contribute to develop the digital documentation of the heritage of the city, meeting a demand of tourists visiting the city and the professionals who work in the preservation and restoration of it, consisting of architects, historians, archaeologists, museum specialists, among others.

Keywords: augmented reality, architectural heritage, geometric modeling, mobile applications

Procedia PDF Downloads 469
19780 Dynamic Response around Inclusions in Infinitely Inhomogeneous Media

Authors: Jinlai Bian, Zailin Yang, Guanxixi Jiang, Xinzhu Li

Abstract:

The problem of elastic wave propagation in inhomogeneous medium has always been a classic problem. Due to the frequent occurrence of earthquakes, many economic losses and casualties have been caused, therefore, to prevent earthquake damage to people and reduce damage, this paper studies the dynamic response around the circular inclusion in the whole space with inhomogeneous modulus, the inhomogeneity of the medium is reflected in the shear modulus of the medium with the spatial position, and the density is constant, this method can be used to solve the problem of the underground buried pipeline. Stress concentration phenomena are common in aerospace and earthquake engineering, and the dynamic stress concentration factor (DSCF) is one of the main factors leading to material damage, one of the important applications of the theory of elastic dynamics is to determine the stress concentration in the body with discontinuities such as cracks, holes, and inclusions. At present, the methods include wave function expansion method, integral transformation method, integral equation method and so on. Based on the complex function method, the Helmholtz equation with variable coefficients is standardized by using conformal transformation method and wave function expansion method, the displacement and stress fields in the whole space with circular inclusions are solved in the complex coordinate system, the unknown coefficients are solved by using boundary conditions, by comparing with the existing results, the correctness of this method is verified, based on the superiority of the complex variable function theory to the conformal transformation, this method can be extended to study the inclusion problem of arbitrary shapes. By solving the dynamic stress concentration factor around the inclusions, the influence of the inhomogeneous parameters of the medium and the wavenumber ratio of the inclusions to the matrix on the dynamic stress concentration factor is analyzed. The research results can provide some reference value for the evaluation of nondestructive testing (NDT), oil exploration, seismic monitoring, and soil-structure interaction.

Keywords: circular inclusions, complex variable function, dynamic stress concentration factor (DSCF), inhomogeneous medium

Procedia PDF Downloads 127
19779 A Framework for Designing Complex Product-Service Systems with a Multi-Domain Matrix

Authors: Yoonjung An, Yongtae Park

Abstract:

Offering a Product-Service System (PSS) is a well-accepted strategy that companies may adopt to provide a set of systemic solutions to customers. PSSs were initially provided in a simple form but now take diversified and complex forms involving multiple services, products and technologies. With the growing interest in the PSS, frameworks for the PSS development have been introduced by many researchers. However, most of the existing frameworks fail to examine various relations existing in a complex PSS. Since designing a complex PSS involves full integration of multiple products and services, it is essential to identify not only product-service relations but also product-product/ service-service relations. It is also equally important to specify how they are related for better understanding of the system. Moreover, as customers tend to view their purchase from a more holistic perspective, a PSS should be developed based on the whole system’s requirements, rather than focusing only on the product requirements or service requirements. Thus, we propose a framework to develop a complex PSS that is coordinated fully with the requirements of both worlds. Specifically, our approach adopts a multi-domain matrix (MDM). A MDM identifies not only inter-domain relations but also intra-domain relations so that it helps to design a PSS that includes highly desired and closely related core functions/ features. Also, various dependency types and rating schemes proposed in our approach would help the integration process.

Keywords: inter-domain relations, intra-domain relations, multi-domain matrix, product-service system design

Procedia PDF Downloads 633
19778 Carbon Coated Silicon Nanoparticles Embedded MWCNT/Graphene Matrix Anode Material for Li-Ion Batteries

Authors: Ubeyd Toçoğlu, Miraç Alaf, Hatem Akbulut

Abstract:

We present a work which was conducted in order to improve the cycle life of silicon based lithium ion battery anodes by utilizing novel composite structure. In this study, carbon coated nano sized (50-100 nm) silicon particles were embedded into Graphene/MWCNT silicon matrix to produce free standing silicon based electrodes. Also, conventional Si powder anodes were produced from Si powder slurry on copper current collectors in order to make comparison of composite and conventional anode structures. Free –standing composite anodes (binder-free) were produced via vacuum filtration from a well dispersion of Graphene, MWCNT and carbon coated silicon powders. Carbon coating process of silicon powders was carried out via microwave reaction system. The certain amount of silicon powder and glucose was mixed under ultrasonication and then coating was conducted at 200 °C for two hours in Teflon lined autoclave reaction chamber. Graphene which was used in this study was synthesized from well-known Hummers method and hydrazine reduction of graphene oxide. X-Ray diffraction analysis and RAMAN spectroscopy techniques were used for phase characterization of anodes. Scanning electron microscopy analyses were conducted for morphological characterization. The electrochemical performance tests were carried out by means of galvanostatic charge/discharge, cyclic voltammetry and electrochemical impedance spectroscopy.

Keywords: graphene, Li-Ion, MWCNT, silicon

Procedia PDF Downloads 241
19777 Space Time Adaptive Algorithm in Bi-Static Passive Radar Systems for Clutter Mitigation

Authors: D. Venu, N. V. Koteswara Rao

Abstract:

Space – time adaptive processing (STAP) is an effective tool for detecting a moving target in spaceborne or airborne radar systems. Since airborne passive radar systems utilize broadcast, navigation and excellent communication signals to perform various surveillance tasks and also has attracted significant interest from the distinct past, therefore the need of the hour is to have cost effective systems as compared to conventional active radar systems. Moreover, requirements of small number of secondary samples for effective clutter suppression in bi-static passive radar offer abundant illuminator resources for passive surveillance radar systems. This paper presents a framework for incorporating knowledge sources directly in the space-time beam former of airborne adaptive radars. STAP algorithm for clutter mitigation for passive bi-static radar has better quantitation of the reduction in sample size thereby amalgamating the earlier data bank with existing radar data sets. Also, we proposed a novel method to estimate the clutter matrix and perform STAP for efficient clutter suppression based on small sample size. Furthermore, the effectiveness of the proposed algorithm is verified using MATLAB simulations in order to validate STAP algorithm for passive bi-static radar. In conclusion, this study highlights the importance for various applications which augments traditional active radars using cost-effective measures.

Keywords: bistatic radar, clutter, covariance matrix passive radar, STAP

Procedia PDF Downloads 286
19776 Increase of Energy Efficiency by Means of Application of Active Bearings

Authors: Alexander Babin, Leonid Savin

Abstract:

In the present paper, increasing of energy efficiency of a thrust hybrid bearing with a central feeding chamber is considered. The mathematical model was developed to determine the pressure distribution and the reaction forces, based on the Reynolds equation and static characteristics’ equations. The boundary problem of pressure distribution calculation was solved using the method of finite differences. For various types of lubricants, geometry and operational characteristics, axial gaps can be determined, where the minimal friction coefficient is provided. The next part of the study considers the application of servovalves in order to maintain the desired position of the rotor. The report features the calculation results and the analysis of the influence of the operational and geometric parameters on the energy efficiency of mechatronic fluid-film bearings.

Keywords: active bearings, energy efficiency, mathematical model, mechatronics, thrust multipad bearing

Procedia PDF Downloads 275
19775 Surface Flattening Assisted with 3D Mannequin Based on Minimum Energy

Authors: Shih-Wen Hsiao, Rong-Qi Chen, Chien-Yu Lin

Abstract:

The topic of surface flattening plays a vital role in the field of computer aided design and manufacture. Surface flattening enables the production of 2D patterns and it can be used in design and manufacturing for developing a 3D surface to a 2D platform, especially in fashion design. This study describes surface flattening based on minimum energy methods according to the property of different fabrics. Firstly, through the geometric feature of a 3D surface, the less transformed area can be flattened on a 2D platform by geodesic. Then, strain energy that has accumulated in mesh can be stably released by an approximate implicit method and revised error function. In some cases, cutting mesh to further release the energy is a common way to fix the situation and enhance the accuracy of the surface flattening, and this makes the obtained 2D pattern naturally generate significant cracks. When this methodology is applied to a 3D mannequin constructed with feature lines, it enhances the level of computer-aided fashion design. Besides, when different fabrics are applied to fashion design, it is necessary to revise the shape of a 2D pattern according to the properties of the fabric. With this model, the outline of 2D patterns can be revised by distributing the strain energy with different results according to different fabric properties. Finally, this research uses some common design cases to illustrate and verify the feasibility of this methodology.

Keywords: surface flattening, strain energy, minimum energy, approximate implicit method, fashion design

Procedia PDF Downloads 324
19774 Using Cyclic Structure to Improve Inference on Network Community Structure

Authors: Behnaz Moradijamei, Michael Higgins

Abstract:

Identifying community structure is a critical task in analyzing social media data sets often modeled by networks. Statistical models such as the stochastic block model have proven to explain the structure of communities in real-world network data. In this work, we develop a goodness-of-fit test to examine community structure's existence by using a distinguishing property in networks: cyclic structures are more prevalent within communities than across them. To better understand how communities are shaped by the cyclic structure of the network rather than just the number of edges, we introduce a novel method for deciding on the existence of communities. We utilize these structures by using renewal non-backtracking random walk (RNBRW) to the existing goodness-of-fit test. RNBRW is an important variant of random walk in which the walk is prohibited from returning back to a node in exactly two steps and terminates and restarts once it completes a cycle. We investigate the use of RNBRW to improve the performance of existing goodness-of-fit tests for community detection algorithms based on the spectral properties of the adjacency matrix. Our proposed test on community structure is based on the probability distribution of eigenvalues of the normalized retracing probability matrix derived by RNBRW. We attempt to make the best use of asymptotic results on such a distribution when there is no community structure, i.e., asymptotic distribution under the null hypothesis. Moreover, we provide a theoretical foundation for our statistic by obtaining the true mean and a tight lower bound for RNBRW edge weights variance.

Keywords: hypothesis testing, RNBRW, network inference, community structure

Procedia PDF Downloads 139
19773 Development of a Psychometric Testing Instrument Using Algorithms and Combinatorics to Yield Coupled Parameters and Multiple Geometric Arrays in Large Information Grids

Authors: Laith F. Gulli, Nicole M. Mallory

Abstract:

The undertaking to develop a psychometric instrument is monumental. Understanding the relationship between variables and events is important in structural and exploratory design of psychometric instruments. Considering this, we describe a method used to group, pair and combine multiple Philosophical Assumption statements that assisted in development of a 13 item psychometric screening instrument. We abbreviated our Philosophical Assumptions (PA)s and added parameters, which were then condensed and mathematically modeled in a specific process. This model produced clusters of combinatorics which was utilized in design and development for 1) information retrieval and categorization 2) item development and 3) estimation of interactions among variables and likelihood of events. The psychometric screening instrument measured Knowledge, Assessment (education) and Beliefs (KAB) of New Addictions Research (NAR), which we called KABNAR. We obtained an overall internal consistency for the seven Likert belief items as measured by Cronbach’s α of .81 in the final study of 40 Clinicians, calculated by SPSS 14.0.1 for Windows. We constructed the instrument to begin with demographic items (degree/addictions certifications) for identification of target populations that practiced within Outpatient Substance Abuse Counseling (OSAC) settings. We then devised education items, beliefs items (seven items) and a modifiable “barrier from learning” item that consisted of six “choose any” choices. We also conceptualized a close relationship between identifying various degrees and certifications held by Outpatient Substance Abuse Therapists (OSAT) (the demographics domain) and all aspects of their education related to EB-NAR (past and present education and desired future training). We placed a descriptive (PA)1tx in both demographic and education domains to trace relationships of therapist education within these two domains. The two perceptions domains B1/b1 and B2/b2 represented different but interrelated perceptions from the therapist perspective. The belief items measured therapist perceptions concerning EB-NAR and therapist perceptions using EB-NAR during the beginning of outpatient addictions counseling. The (PA)s were written in simple words and descriptively accurate and concise. We then devised a list of parameters and appropriately matched them to each PA and devised descriptive parametric (PA)s in a domain categorized information grid. Descriptive parametric (PA)s were reduced to simple mathematical symbols. This made it easy to utilize parametric (PA)s into algorithms, combinatorics and clusters to develop larger information grids. By using matching combinatorics we took paired demographic and education domains with a subscript of 1 and matched them to the column with each B domain with subscript 1. Our algorithmic matching formed larger information grids with organized clusters in columns and rows. We repeated the process using different demographic, education and belief domains and devised multiple information grids with different parametric clusters and geometric arrays. We found benefit combining clusters by different geometric arrays, which enabled us to trace parametric variables and concepts. We were able to understand potential differences between dependent and independent variables and trace relationships of maximum likelihoods.

Keywords: psychometric, parametric, domains, grids, therapists

Procedia PDF Downloads 268