Search results for: LMTD correction factor
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1731

Search results for: LMTD correction factor

1671 Temperature-Dependent Barrier Characteristics of Inhomogeneous Pd/n-GaN Schottky Barrier Diodes Surface

Authors: K. Al-Heuseen, M. R. Hashim

Abstract:

The current-voltage (I-V) characteristics of Pd/n-GaN Schottky barrier were studied at temperatures over room temperature (300-470K). The values of ideality factor (n), zero-bias barrier height (φB0), flat barrier height (φBF) and series resistance (Rs) obtained from I-V-T measurements were found to be strongly temperature dependent while (φBo) increase, (n), (φBF) and (Rs) decrease with increasing temperature. The apparent Richardson constant was found to be 2.1x10-9 Acm-2K-2 and mean barrier height of 0.19 eV. After barrier height inhomogeneities correction, by assuming a Gaussian distribution (GD) of the barrier heights, the Richardson constant and the mean barrier height were obtained as 23 Acm-2K-2 and 1.78eV, respectively. The corrected Richardson constant was very closer to theoretical value of 26 Acm-2K-2. 

Keywords: Electrical properties, Gaussian distribution, Pd-GaN Schottky diodes, thermionic emission.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2134
1670 Cognitive Weighted Polymorphism Factor: A Comprehension Augmented Complexity Metric

Authors: T. Francis Thamburaj, A. Aloysius

Abstract:

Polymorphism is one of the main pillars of objectoriented paradigm. It induces hidden forms of class dependencies which may impact software quality, resulting in higher cost factor for comprehending, debugging, testing, and maintaining the software. In this paper, a new cognitive complexity metric called Cognitive Weighted Polymorphism Factor (CWPF) is proposed. Apart from the software structural complexity, it includes the cognitive complexity on the basis of type. The cognitive weights are calibrated based on 27 empirical studies with 120 persons. A case study and experimentation of the new software metric shows positive results. Further, a comparative study is made and the correlation test has proved that CWPF complexity metric is a better, more comprehensive, and more realistic indicator of the software complexity than Abreu’s Polymorphism Factor (PF) complexity metric.

Keywords: Cognitive complexity metric, cognitive weighted polymorphism factor, object-oriented metrics, polymorphism factor, software metrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2177
1669 Computational Evaluation of a C-A Heat Pump

Authors: Young-Jin Baik, Minsung Kim, Young-Soo Lee, Ki-Chang Chang, Seong-Ryong Park

Abstract:

The compression-absorption heat pump (C-A HP), one of the promising heat recovery equipments that make process hot water using low temperature heat of wastewater, was evaluated by computer simulation. A simulation program was developed based on the continuity and the first and second laws of thermodynamics. Both the absorber and desorber were modeled using UA-LMTD method. In order to prevent an unfeasible temperature profile and to reduce calculation errors from the curved temperature profile of a mixture, heat loads were divided into lots of segments. A single-stage compressor was considered. A compressor cooling load was also taken into account. An isentropic efficiency was computed from the map data. Simulation conditions were given based on the system consisting of ordinarily designed components. The simulation results show that most of the total entropy generation occurs during the compression and cooling process, thus suggesting the possibility that system performance can be enhanced if a rectifier is introduced.

Keywords: Waste heat recovery, Heat Pump.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1673
1668 A High Quality Factor Filter Based on Quasi-Periodic Photonic Structure

Authors: Hamed Alipour-Banaei, Farhad Mehdizadeh

Abstract:

We report the design and characterization of ultra high quality factor filter based on one-dimensional photonic-crystal Thue- Morse sequence structure. The behavior of aperiodic array of photonic crystal structure is numerically investigated and we show that by changing the angle of incident wave, desired wavelengths could be tuned and a tunable filter is realized. Also it is shown that high quality factor filter be achieved in the telecommunication window around 1550 nm, with a device based on Thue-Morse structure. Simulation results show that the proposed structure has a quality factor more than 100000 and it is suitable for DWDM communication applications.

Keywords: Thue-Morse, filter, quality factor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1978
1667 The Long Run Relationship between Exports and Imports in South Africa: Evidence from Cointegration Analysis

Authors: Sagaren Pillay

Abstract:

This study empirically examines the long run equilibrium relationship between South Africa’s exports and imports using quarterly data from 1985 to 2012. The theoretical framework used for the study is based on Johansen’s Maximum Likelihood cointegration technique which tests for both the existence and number of cointegration vectors that exists. The study finds that both the series are integrated of order one and are cointegrated. A statistically significant cointegrating relationship is found to exist between exports and imports. The study models this unique linear and lagged relationship using a Vector Error Correction Model (VECM). The findings of the study confirm the existence of a long run equilibrium relationship between exports and imports.

Keywords: Cointegration lagged, linear, maximum likelihood, vector error correction model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2730
1666 Affine Projection Algorithm with Variable Data-Reuse Factor

Authors: ChangWoo Lee, Young Kow Lee, Sung Jun Ban, SungHoo Choi, Sang Woo Kim

Abstract:

This paper suggests a new Affine Projection (AP) algorithm with variable data-reuse factor using the condition number as a decision factor. To reduce computational burden, we adopt a recently reported technique which estimates the condition number of an input data matrix. Several simulations show that the new algorithm has better performance than that of the conventional AP algorithm.

Keywords: Affine projection algorithm, variable data-reuse factor, condition number, convergence rate, misalignment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1506
1665 Evidence of the Long-run Equilibrium between Money Demand Determinants in Croatia

Authors: B. Skrabic, N. Tomic-Plazibat

Abstract:

In this paper real money demand function is analyzed within multivariate time-series framework. Cointegration approach is used (Johansen procedure) assuming interdependence between money demand determinants, which are nonstationary variables. This will help us to understand the behavior of money demand in Croatia, revealing the significant influence between endogenous variables in vector autoregrression system (VAR), i.e. vector error correction model (VECM). Exogeneity of the explanatory variables is tested. Long-run money demand function is estimated indicating slow speed of adjustment of removing the disequilibrium. Empirical results provide the evidence that real industrial production and exchange rate explains the most variations of money demand in the long-run, while interest rate is significant only in short-run.

Keywords: Cointegration, Long-run equilibrium, Money demand function, Vector error correction model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2108
1664 Establishing a New Simple Formula for Buckling Length Factor (K) of Rigid Frames Columns

Authors: Ehab Hasan Ahmed Hasan Ali

Abstract:

The calculation of buckling length factor (K) for steel frames columns is a major and governing processes to determine the dimensions steel frame columns cross sections during design. The buckling length of steel frames columns has a direct effect on the cost (weight) of using cross section. A new formula is required to determine buckling length factor (K) by simplified way. In this research a new formula for buckling length factor (K) was established to determine by accurate method for a limited interval of columns ends rigidity (GA, GB). The new formula can be used ease to evaluate the buckling length factor without needing to complicated equations or difficult charts.

Keywords: Buckling length, New formula, Curve fitting, Simplification, Steel column design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2193
1663 Electron Filling Factor and Sunlight Concentration Effects on the Efficiency of Intermediate Band Solar Cell

Authors: Nima Es'haghi Gorji, Hossein Movla, Foozieh Sohrabi, Alireza Mottaghizadeh, Mohammad Houshmand, Hassan Babaei, Arash Nikniazi

Abstract:

For a determined intermediate band position, the effects of electron filling factor and sunlight concentration on the active region thickness and efficiency of the quantum-dot intermediate band solar cell are calculated. For each value of electron filling factor, the maximum point of efficiency obtained and resulted in the optimum thickness of the cell under three different sunlight concentrations. We show the importance of filling factor as a parameter to be more considered. The photon recycling effect eliminated in all calculations.

Keywords: Intermediate band, Sunlight concentration, Efficiency limits, Electron filling factor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1495
1662 Decoder Design for a New Single Error Correcting/Double Error Detecting Code

Authors: M. T. Anwar, P. K. Lala, P. Thenappan

Abstract:

This paper presents the decoder design for the single error correcting and double error detecting code proposed by the authors in an earlier paper. The speed of error detection and correction of a code is largely dependent upon the associated encoder and decoder circuits. The complexity and the speed of such circuits are determined by the number of 1?s in the parity check matrix (PCM). The number of 1?s in the parity check matrix for the code proposed by the authors are fewer than in any currently known single error correcting/double error detecting code. This results in simplified encoding and decoding circuitry for error detection and correction.

Keywords: Decoder, Hsiao code, Parity Check Matrix, Syndrome Pattern.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2030
1661 Forecasting Electricity Spot Price with Generalized Long Memory Modeling: Wavelet and Neural Network

Authors: Souhir Ben Amor, Heni Boubaker, Lotfi Belkacem

Abstract:

This aims of this paper is to forecast the electricity spot prices. First, we focus on modeling the conditional mean of the series so we adopt a generalized fractional -factor Gegenbauer process (k-factor GARMA). Secondly, the residual from the -factor GARMA model has used as a proxy for the conditional variance; these residuals were predicted using two different approaches. In the first approach, a local linear wavelet neural network model (LLWNN) has developed to predict the conditional variance using the Back Propagation learning algorithms. In the second approach, the Gegenbauer generalized autoregressive conditional heteroscedasticity process (G-GARCH) has adopted, and the parameters of the k-factor GARMA-G-GARCH model has estimated using the wavelet methodology based on the discrete wavelet packet transform (DWPT) approach. The empirical results have shown that the k-factor GARMA-G-GARCH model outperform the hybrid k-factor GARMA-LLWNN model, and find it is more appropriate for forecasts.

Keywords: k-factor, GARMA, LLWNN, G-GARCH, electricity price, forecasting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 939
1660 Mind Your Product-Market Strategy on Selecting Marketing Inputs: An Uncertainty Approach in Indian Context

Authors: Susmita Ghosh, Bhaskar Bhowmick

Abstract:

Market is an important factor for start-ups to look into during decision-making in product development and related areas. Emerging country markets are more uncertain in terms of information availability and institutional supports. The literature review of market uncertainty reveals the need for identifying factors representing the market uncertainty. This paper identifies factors for market uncertainty using Exploratory Factor Analysis (EFA) and confirmed the number of factor retention using an alternative factor retention criterion ‘Parallel Analysis’. 500 entrepreneurs, engaged in start-ups from all over India participated in the study. This paper concludes with the factor structure of ‘market uncertainty’ having dimensions of uncertainty in industry orientation, uncertainty in customer orientation and uncertainty in marketing orientation.

Keywords: Uncertainty, market, orientation, competitor, demand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1582
1659 Face Recognition with Image Rotation Detection, Correction and Reinforced Decision using ANN

Authors: Hemashree Bordoloi, Kandarpa Kumar Sarma

Abstract:

Rotation or tilt present in an image capture by digital means can be detected and corrected using Artificial Neural Network (ANN) for application with a Face Recognition System (FRS). Principal Component Analysis (PCA) features of faces at different angles are used to train an ANN which detects the rotation for an input image and corrected using a set of operations implemented using another system based on ANN. The work also deals with the recognition of human faces with features from the foreheads, eyes, nose and mouths as decision support entities of the system configured using a Generalized Feed Forward Artificial Neural Network (GFFANN). These features are combined to provide a reinforced decision for verification of a person-s identity despite illumination variations. The complete system performing facial image rotation detection, correction and recognition using re-enforced decision support provides a success rate in the higher 90s.

Keywords: Rotation, Face, Recognition, ANN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2022
1658 Physical Properties of Uranium Dinitride UN2 by Using Density Functional Theory (DFT and DFT+U)

Authors: T. Zergoug, S.H. Abaidia, A. Nedjar, M. Y. Mokeddem

Abstract:

Physical properties of uranium dinitride (UN2) were investigated in detail using first principle calculations based on density functional theory (DFT). To study the strong correlation effects due to 5f uranium valence electrons, the on-site coulomb interaction correction U via the Hubbard-like term (DFT+U) was employed. The UN2 structural, mechanical and thermodynamic properties were calculated within DFT and Various U of DFT+U approach. The Perdew–Burke–Ernzerhof (PBE.5.2) version of the generalized gradient approximation (GGA) is used to describe the exchange-correlation with the projector-augmented wave (PAW) pseudo potentials. A comparative study shows that results are improved by using the Hubbard formalism for a certain U value correction like the structural parameter. For some physical properties the variation versus Hubbard-U is strong like Young modulus but for others it is weakly noticeable such as bulk modulus. We noticed also that from U=7.5 eV, elastic results don’t agree with the cubic cell because of the C44 values which turn out to be negative.

Keywords: Ab initio, bulk modulus, DFT, DFT + U.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2488
1657 Enhanced Approaches to Rectify the Noise, Illumination and Shadow Artifacts

Authors: M. Sankari, C. Meena

Abstract:

Enhancing the quality of two dimensional signals is one of the most important factors in the fields of video surveillance and computer vision. Usually in real-life video surveillance, false detection occurs due to the presence of random noise, illumination and shadow artifacts. The detection methods based on background subtraction faces several problems in accurately detecting objects in realistic environments: In this paper, we propose a noise removal algorithm using neighborhood comparison method with thresholding. The illumination variations correction is done in the detected foreground objects by using an amalgamation of techniques like homomorphic decomposition, curvelet transformation and gamma adjustment operator. Shadow is removed using chromaticity estimator with local relation estimator. Results are compared with the existing methods and prove as high robustness in the video surveillance.

Keywords: Chromaticity Estimator, Curvelet Transformation, Denoising, Gamma correction, Homomorphic, Neighborhood Assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1916
1656 [a, b]-Factors Excluding Some Specified Edges In Graphs

Authors: Sizhong Zhou, Bingyuan Pu

Abstract:

Let G be a graph of order n, and let a, b and m be positive integers with 1 ≤ a<b. An [a, b]-factor of G is defined as a spanning subgraph F of G such that a ≤ dF (x) ≤ b for each x ∈ V (G). In this paper, it is proved that if n ≥ (a+b−1+√(a+b+1)m−2)2−1 b and δ(G) > n + a + b − 2 √bn+ 1, then for any subgraph H of G with m edges, G has an [a, b]-factor F such that E(H)∩ E(F) = ∅. This result is an extension of thatof Egawa [2].

Keywords: graph, minimum degree, [a, b]-factor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1049
1655 A Novel Forgetting Factor Recursive Least Square Algorithm Applied to the Human Motion Analysis

Authors: Hadi Sadoghi Yazdi, Mehri Sadoghi Yazdi, Mohammad Reza Mohammadi

Abstract:

This paper is concerned with studying the forgetting factor of the recursive least square (RLS). A new dynamic forgetting factor (DFF) for RLS algorithm is presented. The proposed DFF-RLS is compared to other methods. Better performance at convergence and tracking of noisy chirp sinusoid is achieved. The control of the forgetting factor at DFF-RLS is based on the gradient of inverse correlation matrix. Compared with the gradient of mean square error algorithm, the proposed approach provides faster tracking and smaller mean square error. In low signal-to-noise ratios, the performance of the proposed method is superior to other approaches.

Keywords: Forgetting factor, RLS, Inverse correlation matrix, human motion analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2202
1654 An Investigation on Overstrength Factor (Ω) of Reinforced Concrete Buildings in Turkish Earthquake Draft Code (TEC-2016)

Authors: M. Hakan Arslan, I. Hakkı Erkan

Abstract:

Overstrength factor is an important parameter of load reduction factor. In this research, the overstrength factor (Ω) of reinforced concrete (RC) buildings and the parameters of Ω in TEC-2016 draft version have been explored. For this aim, 48 RC buildings have been modeled according to the current seismic code TEC-2007 and Turkish Building Code-500-2000 criteria. After modelling step, nonlinear static pushover analyses have been applied to these buildings by using TEC-2007 Section 7. After the nonlinear pushover analyses, capacity curves (lateral load-lateral top displacement curves) have been plotted for 48 RC buildings. Using capacity curves, overstrength factors (Ω) have been derived for each building. The obtained overstrength factor (Ω) values have been compared with TEC-2016 values for related building types, and the results have been interpreted. According to the obtained values from the study, overstrength factor (Ω) given in TEC-2016 draft code is found quite suitable.

Keywords: Reinforced concrete buildings, overstrength factor, earthquake, static pushover analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1977
1653 GPU Based High Speed Error Protection for Watermarked Medical Image Transmission

Authors: Md Shohidul Islam, Jongmyon Kim, Ui-pil Chong

Abstract:

Medical image is an integral part of e-health care and e-diagnosis system. Medical image watermarking is widely used to protect patients’ information from malicious alteration and manipulation. The watermarked medical images are transmitted over the internet among patients, primary and referred physicians. The images are highly prone to corruption in the wireless transmission medium due to various noises, deflection, and refractions. Distortion in the received images leads to faulty watermark detection and inappropriate disease diagnosis. To address the issue, this paper utilizes error correction code (ECC) with (8, 4) Hamming code in an existing watermarking system. In addition, we implement the high complex ECC on a graphics processing units (GPU) to accelerate and support real-time requirement. Experimental results show that GPU achieves considerable speedup over the sequential CPU implementation, while maintaining 100% ECC efficiency.

Keywords: Medical Image Watermarking (MIW), e-health system, error correction, Hamming code, GPU.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703
1652 Fracture Toughness Characterization of Carbon-Epoxy Composite using Arcan Specimen

Authors: M. Nikbakht, N. Choupani

Abstract:

In this study the behavior of interlaminar fracture of carbon-epoxy thermoplastic laminated composite is investigated numerically and experimentally. Tests are performed with Arcan specimens. Testing with Arcan specimen gives the opportunity of utilizing just one kind of specimen for extracting fracture properties for mode I, mode II and different mixed mode ratios of materials with exerting load via different loading angles. Variation of loading angles in range of 0-90° made possible to achieve different mixed mode ratios. Correction factors for various conditions are obtained from ABAQUS 2D finite element models which demonstrate the finite shape of Arcan specimens used in this study. Finally, applying the correction factors to critical loads obtained experimentally, critical interlaminar fracture toughness of this type of carbon- epoxy composite has been attained.

Keywords: Fracture Mechanics, Mixed Mode, Arcan Specimen, Finite Element.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2865
1651 Determining the Maximum Lateral Displacement Due to Sever Earthquakes without Using Nonlinear Analysis

Authors: Mussa Mahmoudi

Abstract:

For Seismic design, it is important to estimate, maximum lateral displacement (inelastic displacement) of the structures due to sever earthquakes for several reasons. Seismic design provisions estimate the maximum roof and storey drifts occurring in major earthquakes by amplifying the drifts of the structures obtained by elastic analysis subjected to seismic design load, with a coefficient named “displacement amplification factor" which is greater than one. Here, this coefficient depends on various parameters, such as ductility and overstrength factors. The present research aims to evaluate the value of the displacement amplification factor in seismic design codes and then tries to propose a value to estimate the maximum lateral structural displacement from sever earthquakes, without using non-linear analysis. In seismic codes, since the displacement amplification is related to “force reduction factor" hence; this aspect has been accepted in the current study. Meanwhile, two methodologies are applied to evaluate the value of displacement amplification factor and its relation with the force reduction factor. In the first methodology, which is applied for all structures, the ratio of displacement amplification and force reduction factors is determined directly. Whereas, in the second methodology that is applicable just for R/C moment resisting frame, the ratio is obtained by calculating both factors, separately. The acquired results of these methodologies are alike and estimate the ratio of two factors from 1 to 1.2. The results indicate that the ratio of the displacement amplification factor and the force reduction factor differs to those proposed by seismic provisions such as NEHRP, IBC and Iranian seismic code (standard no. 2800).

Keywords: Displacement amplification factor, Ductility factor, Force reduction factor, Maximum lateral displacement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2844
1650 The Link between Money Market and Economic Growth in Nigeria: Vector Error Correction Model Approach

Authors: Ehigiamusoe, Uyi Kizito

Abstract:

The paper examines the impact of money market on economic growth in Nigeria using data for the period 1980-2012. Econometrics techniques such as Ordinary Least Squares Method, Johanson’s Co-integration Test and Vector Error Correction Model were used to examine both the long-run and short-run relationship. Evidence from the study suggest that though a long-run relationship exists between money market and economic growth, but the present state of the Nigerian money market is significantly and negatively related to economic growth. The link between the money market and the real sector of the economy remains very weak. This implies that the market is not yet developed enough to produce the needed growth that will propel the Nigerian economy because of several challenges. It was therefore recommended that government should create the appropriate macroeconomic policies, legal framework and sustain the present reforms with a view to developing the market so as to promote productive activities, investments, and ultimately economic growth.

Keywords: Economic Growth, Investments, Money Market, Money Market Challenges, Money Market Instruments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8442
1649 Data Gathering and Analysis for Arabic Historical Documents

Authors: Ali Dulla

Abstract:

This paper introduces a new dataset (and the methodology used to generate it) based on a wide range of historical Arabic documents containing clean data simple and homogeneous-page layouts. The experiments are implemented on printed and handwritten documents obtained respectively from some important libraries such as Qatar Digital Library, the British Library and the Library of Congress. We have gathered and commented on 150 archival document images from different locations and time periods. It is based on different documents from the 17th-19th century. The dataset comprises differing page layouts and degradations that challenge text line segmentation methods. Ground truth is produced using the Aletheia tool by PRImA and stored in an XML representation, in the PAGE (Page Analysis and Ground truth Elements) format. The dataset presented will be easily available to researchers world-wide for research into the obstacles facing various historical Arabic documents such as geometric correction of historical Arabic documents.

Keywords: Dataset production, ground truth production, historical documents, arbitrary warping, geometric correction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 820
1648 Loading Factor Performance of a Centrifugal Compressor Impeller: Specific Features and Way of Modeling

Authors: K. Soldatova, Y. Galerkin

Abstract:

A loading factor performance is necessary for the modeling of centrifugal compressor gas dynamic performance curve. Measured loading factors are linear function of a flow coefficient at an impeller exit. The performance does not depend on the compressibility criterion. To simulate loading factor performances, the authors present two parameters: a loading factor at zero flow rate and an angle between an ordinate and performance line. The calculated loading factor performances of non-viscous are linear too and close to experimental performances. Loading factor performances of several dozens of impellers with different blade exit angles, blade thickness and number, ratio of blade exit/inlet height, and two different type of blade mean line configuration. There are some trends of influence, which are evident – comparatively small blade thickness influence, and influence of geometry parameters is more for impellers with bigger blade exit angles, etc. Approximating equations for both parameters are suggested. The next phase of work will be simulating of experimental performances with the suggested approximation equations as a base.

Keywords: Centrifugal compressor stage, centrifugal compressor, loading factor, gas dynamic performance curve.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2052
1647 Degradation of Irradiated UO2 Fuel Thermal Conductivity Calculated by FRAPCON Model Due to Porosity Evolution at High Burn-Up

Authors: B. Roostaii, H. Kazeminejad, S. Khakshournia

Abstract:

The evolution of volume porosity previously obtained by using the existing low temperature high burn-up gaseous swelling model with progressive recrystallization for UO2 fuel is utilized to study the degradation of irradiated UO2 thermal conductivity calculated by the FRAPCON model of thermal conductivity. A porosity correction factor is developed based on the assumption that the fuel morphology is a three-phase type, consisting of the as-fabricated pores and pores due to intergranular bubbles whitin UO2 matrix and solid fission products. The predicted thermal conductivity demonstrates an additional degradation of 27% due to porosity formation at burn-up levels around 120 MWd/kgU which would cause an increase in the fuel temperature accordingly. Results of the calculations are compared with available data.

Keywords: Irradiation-induced recrystallization, matrix swelling, porosity evolution, UO2 thermal conductivity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1198
1646 Optimization of the Characteristic Straight Line Method by a “Best Estimate“ of Observed, Normal Orthometric Elevation Differences

Authors: Mahmoud M. S. Albattah

Abstract:

In this paper, to optimize the “Characteristic Straight Line Method" which is used in the soil displacement analysis, a “best estimate" of the geodetic leveling observations has been achieved by taking in account the concept of 'Height systems'. This concept has been discussed in detail and consequently the concept of “height". In landslides dynamic analysis, the soil is considered as a mosaic of rigid blocks. The soil displacement has been monitored and analyzed by using the “Characteristic Straight Line Method". Its characteristic components have been defined constructed from a “best estimate" of the topometric observations. In the measurement of elevation differences, we have used the most modern leveling equipment available. Observational procedures have also been designed to provide the most effective method to acquire data. In addition systematic errors which cannot be sufficiently controlled by instrumentation or observational techniques are minimized by applying appropriate corrections to the observed data: the level collimation correction minimizes the error caused by nonhorizontality of the leveling instrument's line of sight for unequal sight lengths, the refraction correction is modeled to minimize the refraction error caused by temperature (density) variation of air strata, the rod temperature correction accounts for variation in the length of the leveling rod' s Invar/LO-VAR® strip which results from temperature changes, the rod scale correction ensures a uniform scale which conforms to the international length standard and the introduction of the concept of the 'Height systems' where all types of height (orthometric, dynamic, normal, gravity correction, and equipotential surface) have been investigated. The “Characteristic Straight Line Method" is slightly more convenient than the “Characteristic Circle Method". It permits to evaluate a displacement of very small magnitude even when the displacement is of an infinitesimal quantity. The inclination of the landslide is given by the inverse of the distance reference point O to the “Characteristic Straight Line". Its direction is given by the bearing of the normal directed from point O to the Characteristic Straight Line (Fig..6). A “best estimate" of the topometric observations was used to measure the elevation of points carefully selected, before and after the deformation. Gross errors have been eliminated by statistical analyses and by comparing the heights within local neighborhoods. The results of a test using an area where very interesting land surface deformation occurs are reported. Monitoring with different options and qualitative comparison of results based on a sufficient number of check points are presented.

Keywords: Characteristic straight line method, dynamic height, landslides, orthometric height, systematic errors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1526
1645 DIAL Measurements of Vertical Distribution of Ozone at the Siberian Lidar Station in Tomsk

Authors: Oleg A. Romanovskii, Vladimir D. Burlakov, Sergey I. Dolgii, Olga V. Kharchenko, Alexey A. Nevzorov, Alexey V. Nevzorov

Abstract:

The paper presents the results of DIAL measurements of the vertical ozone distribution. The ozone lidar operate as part of the measurement complex at Siberian Lidar Station (SLS) of V.E. Zuev Institute of Atmospheric Optics SB RAS, Tomsk (56.5ºN; 85.0ºE) and designed for study of the vertical ozone distribution in the upper troposphere–lower stratosphere. Most suitable wavelengths for measurements of ozone profiles are selected. We present an algorithm for retrieval of vertical distribution of ozone with temperature and aerosol correction during DIAL lidar sounding of the atmosphere. The temperature correction of ozone absorption coefficients is introduced in the software to reduce the retrieval errors. Results of lidar measurement at wavelengths of 299 and 341 nm agree with model estimates, which point to acceptable accuracy of ozone sounding in the 6–18 km altitude range.

Keywords: Lidar, ozone distribution, atmosphere, DIAL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1241
1644 Connectivity Characteristic of Transcription Factor

Authors: T. Mahalakshmi, Aswathi B. L., Achuthsankar S. Nair

Abstract:

Transcription factors are a group of proteins that helps for interpreting the genetic information in DNA. Protein-protein interactions play a major role in the execution of key biological functions of a cell. These interactions are represented in the form of a graph with nodes and edges. Studies have showed that some nodes have high degree of connectivity and such nodes, known as hub nodes, are the inevitable parts of the network. In the present paper a method is proposed to identify hub transcription factor proteins using sequence information. On a complete data set of transcription factor proteins available from the APID database, the proposed method showed an accuracy of 77%, sensitivity of 79% and specificity of 76%.

Keywords: Transcription Factor Proteins, Hub Proteins, Shannon Index, Transfer Free Energy to Surface (TFES).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1344
1643 Robust Parameter and Scale Factor Estimation in Nonstationary and Impulsive Noise Environment

Authors: Zoran D. Banjac, Branko D. Kovacevic

Abstract:

The problem of FIR system parameter estimation has been considered in the paper. A new robust recursive algorithm for simultaneously estimation of parameters and scale factor of prediction residuals in non-stationary environment corrupted by impulsive noise has been proposed. The performance of derived algorithm has been tested by simulations.

Keywords: Adaptive filtering, Non-Gaussian filtering, Robustestimation, Scale factor estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1659
1642 Design and Implementation of Reed Solomon Encoder on FPGA

Authors: Amandeep Singh, Mandeep Kaur

Abstract:

Error correcting codes are used for detection and correction of errors in digital communication system. Error correcting coding is based on appending of redundancy to the information message according to a prescribed algorithm. Reed Solomon codes are part of channel coding and withstand the effect of noise, interference and fading. Galois field arithmetic is used for encoding and decoding reed Solomon codes. Galois field multipliers and linear feedback shift registers are used for encoding the information data block. The design of Reed Solomon encoder is complex because of use of LFSR and Galois field arithmetic. The purpose of this paper is to design and implement Reed Solomon (255, 239) encoder with optimized and lesser number of Galois Field multipliers. Symmetric generator polynomial is used to reduce the number of GF multipliers. To increase the capability toward error correction, convolution interleaving will be used with RS encoder. The Design will be implemented on Xilinx FPGA Spartan II.

Keywords: Galois Field, Generator polynomial, LFSR, Reed Solomon.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4799