Search results for: Knowledge based system
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16911

Search results for: Knowledge based system

3591 XPM Response of Multiple Quantum Well chirped DFB-SOA All Optical Flip-Flop Switching

Authors: Masoud Jabbari, Mohammad Kazem Moravvej-Farshi, Rahim Ghayour, Abbas Zarifkar

Abstract:

In this paper, based on the coupled-mode and carrier rate equations, derivation of a dynamic model and numerically analysis of a MQW chirped DFB-SOA all-optical flip-flop is done precisely. We have analyzed the effects of strains of QW and MQW and cross phase modulation (XPM) on the dynamic response, and rise and fall times of the DFB-SOA all optical flip flop. We have shown that strained MQW active region in under an optimized condition into a DFB-SOA with chirped grating can improve the switching ON speed limitation in such a of the device, significantly while the fall time is increased. The values of the rise times for such an all optical flip-flop, are obtained in an optimized condition, areas tr=255ps.

Keywords: All-Optical Flip-Flop (AO-FF), Distributed feedback semiconductor optical amplifier (DFB-SOA), Optical Bistability, Multi quantum well (MQW)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1543
3590 Normalizing Logarithms of Realized Volatility in an ARFIMA Model

Authors: G. L. C. Yap

Abstract:

Modelling realized volatility with high-frequency returns is popular as it is an unbiased and efficient estimator of return volatility. A computationally simple model is fitting the logarithms of the realized volatilities with a fractionally integrated long-memory Gaussian process. The Gaussianity assumption simplifies the parameter estimation using the Whittle approximation. Nonetheless, this assumption may not be met in the finite samples and there may be a need to normalize the financial series. Based on the empirical indices S&P500 and DAX, this paper examines the performance of the linear volatility model pre-treated with normalization compared to its existing counterpart. The empirical results show that by including normalization as a pre-treatment procedure, the forecast performance outperforms the existing model in terms of statistical and economic evaluations.

Keywords: Long-memory, Gaussian process, Whittle estimator, normalization, volatility, value-at-risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1660
3589 Micropower Fuzzy Linguistic-Hedges Circuit in Current-Mode Approach

Authors: E. Farshidi

Abstract:

In this paper, based on a novel synthesis, a set of new simplified circuit design to implement the linguistic-hedge operations for adjusting the fuzzy membership function set is presented. The circuits work in current-mode and employ floating-gate MOS (FGMOS) transistors that operate in weak inversion region. Compared to the other proposed circuits, these circuits feature severe reduction of the elements number, low supply voltage (0.7V), low power consumption (<200nW), immunity from body effect and wide input dynamic range (>60dB). In this paper, a set of fuzzy linguistic hedge circuits, including absolutely, very, much more, more, plus minus, more or less and slightly, has been implemented in 0.18 mm CMOS process. Simulation results by Hspice confirm the validity of the proposed design technique and show high performance of the circuits.

Keywords: Current-mode, Linguistic-Hedge, Fuzzy Logic, lowpower

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1742
3588 Low Temperature Ethanol Gas Sensor based on SnO2/MWNTs Nanocomposite

Authors: O. Alizadeh Sahraei, A. Khodadadi, Y. Mortazavi, M. Vesali Naseh, S. Mosadegh

Abstract:

A composite made of plasma functionalized multiwall carbon nanotubes (MWNTs) coated with SnO2 was synthesized by sonochemical precipitation method. Thick layer of this nanocomposite material was used as ethanol sensor at low temperatures. The composite sensitivity for ethanol has increased by a factor of 2 at room temperature and by a factor of 13 at 250°C in comparison to that of pure SnO2. SEM image of nanocomposite material showed MWNTs were embedded in SnO2 matrix and also a higher surface area was observed in the presence of functionalized MWNTs. Greatly improved sensitivity of the composite material to ethanol can be attributed to new gas accessing passes through MWNTs and higher specific surface area.

Keywords: Carbon nanotube, Functionalized, Gas sensor, Low temperature, Nanocomposite, Tin oxide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2316
3587 Secondary Organic Contribution to Particles Formed on the Ice Melted Arctic Ocean

Authors: Petri Vaattovaara, Zoran D. Ristovski, Martin Graus, Marcus Müller, EijaAsmi, Luca Di Liberto, StaffanSjögren, Douglas Orsini, Caroline Leck, Ari Laaksonen

Abstract:

Due to climate warming and consequently due to ice and snow melting of the Arctic Ocean, the highly biologically active ocean surface area has been expanding quickly making possible longer marine biota growth seasons during polar summers. That increase the probability of the remote marine environment secondary contribution, especially secondary organic contribution, to the particle production and particle growth events and particle properties, consequently effecting on the open ocean, pack ice and ground based regions radiation budget and thus on the feedbacks between arctic biota, particles, clouds, and climate.

Keywords: Arctic Ocean, ice melting, nucleation, secondary organics, clouds, climate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1460
3586 Comparison of Imputation Techniques for Efficient Prediction of Software Fault Proneness in Classes

Authors: Geeta Sikka, Arvinder Kaur Takkar, Moin Uddin

Abstract:

Missing data is a persistent problem in almost all areas of empirical research. The missing data must be treated very carefully, as data plays a fundamental role in every analysis. Improper treatment can distort the analysis or generate biased results. In this paper, we compare and contrast various imputation techniques on missing data sets and make an empirical evaluation of these methods so as to construct quality software models. Our empirical study is based on NASA-s two public dataset. KC4 and KC1. The actual data sets of 125 cases and 2107 cases respectively, without any missing values were considered. The data set is used to create Missing at Random (MAR) data Listwise Deletion(LD), Mean Substitution(MS), Interpolation, Regression with an error term and Expectation-Maximization (EM) approaches were used to compare the effects of the various techniques.

Keywords: Missing data, Imputation, Missing Data Techniques.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1646
3585 Predicting Groundwater Areas Using Data Mining Techniques: Groundwater in Jordan as Case Study

Authors: Faisal Aburub, Wael Hadi

Abstract:

Data mining is the process of extracting useful or hidden information from a large database. Extracted information can be used to discover relationships among features, where data objects are grouped according to logical relationships; or to predict unseen objects to one of the predefined groups. In this paper, we aim to investigate four well-known data mining algorithms in order to predict groundwater areas in Jordan. These algorithms are Support Vector Machines (SVMs), Naïve Bayes (NB), K-Nearest Neighbor (kNN) and Classification Based on Association Rule (CBA). The experimental results indicate that the SVMs algorithm outperformed other algorithms in terms of classification accuracy, precision and F1 evaluation measures using the datasets of groundwater areas that were collected from Jordanian Ministry of Water and Irrigation.

Keywords: Classification, data mining, evaluation measures, groundwater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2570
3584 Skin Detection using Histogram depend on the Mean Shift Algorithm

Authors: Soo- Young Ye, Ki-Gon Nam, Ki-Won Byun

Abstract:

In this paper, we were introduces a skin detection method using a histogram approximation based on the mean shift algorithm. The proposed method applies the mean shift procedure to a histogram of a skin map of the input image, generated by comparison with standard skin colors in the CbCr color space, and divides the background from the skin region by selecting the maximum value according to brightness level. The proposed method detects the skin region using the mean shift procedure to determine a maximum value that becomes the dividing point, rather than using a manually selected threshold value, as in existing techniques. Even when skin color is contaminated by illumination, the procedure can accurately segment the skin region and the background region. The proposed method may be useful in detecting facial regions as a pretreatment for face recognition in various types of illumination.

Keywords: Skin region detection, mean shift, histogram approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2238
3583 Modeling and Simulation for 3D Eddy Current Testing in Conducting Materials

Authors: S. Bennoud, M. Zergoug

Abstract:

The numerical simulation of electromagnetic interactions is still a challenging problem, especially in problems that result in fully three dimensional mathematical models.

The goal of this work is to use mathematical modeling to characterize the reliability and capacity of eddy current technique to detect and characterize defects embedded in aeronautical in-service pieces.

The finite element method is used for describing the eddy current technique in a mathematical model by the prediction of the eddy current interaction with defects. However, this model is an approximation of the full Maxwell equations.

In this study, the analysis of the problem is based on a three dimensional finite element model that computes directly the electromagnetic field distortions due to defects.

Keywords: Eddy current, Finite element method, Non destructive testing, Numerical simulations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3100
3582 Slice Bispectrogram Analysis-Based Classification of Environmental Sounds Using Convolutional Neural Network

Authors: Katsumi Hirata

Abstract:

Certain systems can function well only if they recognize the sound environment as humans do. In this research, we focus on sound classification by adopting a convolutional neural network and aim to develop a method that automatically classifies various environmental sounds. Although the neural network is a powerful technique, the performance depends on the type of input data. Therefore, we propose an approach via a slice bispectrogram, which is a third-order spectrogram and is a slice version of the amplitude for the short-time bispectrum. This paper explains the slice bispectrogram and discusses the effectiveness of the derived method by evaluating the experimental results using the ESC‑50 sound dataset. As a result, the proposed scheme gives high accuracy and stability. Furthermore, some relationship between the accuracy and non-Gaussianity of sound signals was confirmed.

Keywords: Bispectrum, convolutional neural network, environmental sound, slice bispectrogram, spectrogram.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 589
3581 A CUSUM Control Chart to Monitor Wafer Quality

Authors: Sheng-Shu Cheng, Fong-Jung Yu

Abstract:

C-control chart assumes that process nonconformities follow a Poisson distribution. In actuality, however, this Poisson distribution does not always occur. A process control for semiconductor based on a Poisson distribution always underestimates the true average amount of nonconformities and the process variance. Quality is described more accurately if a compound Poisson process is used for process control at this time. A cumulative sum (CUSUM) control chart is much better than a C control chart when a small shift will be detected. This study calculates one-sided CUSUM ARLs using a Markov chain approach to construct a CUSUM control chart with an underlying Poisson-Gamma compound distribution for the failure mechanism. Moreover, an actual data set from a wafer plant is used to demonstrate the operation of the proposed model. The results show that a CUSUM control chart realizes significantly better performance than EWMA.

Keywords: Nonconformities, Compound Poisson distribution, CUSUM control chart.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2700
3580 Review and Experiments on SDMSCue

Authors: Ashraf Anwar

Abstract:

In this work, I present a review on Sparse Distributed Memory for Small Cues (SDMSCue), a variant of Sparse Distributed Memory (SDM) that is capable of handling small cues. I then conduct and show some cognitive experiments on SDMSCue to test its cognitive soundness compared to SDM. Small cues refer to input cues that are presented to memory for reading associations; but have many missing parts or fields from them. The original SDM failed to handle such a problem. SDMSCue handles and overcomes this pitfall. The main idea in SDMSCue; is the repeated projection of the semantic space on smaller subspaces; that are selected based on the input cue length and pattern. This process allows for Read/Write operations using an input cue that is missing a large portion. SDMSCue is augmented with the use of genetic algorithms for memory allocation and initialization. I claim that SDM functionality is a subset of SDMSCue functionality.

Keywords: Artificial intelligence, recall, recognition, SDM, SDMSCue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1354
3579 Comparative Performance Analysis of Fiber Delay Line Based Buffer Architectures for Contention Resolution in Optical WDM Networks

Authors: Manoj Kumar Dutta

Abstract:

Wavelength Division Multiplexing (WDM) technology is the most promising technology for the proper utilization of huge raw bandwidth provided by an optical fiber. One of the key problems in implementing the all-optical WDM network is the packet contention. This problem can be solved by several different techniques. In time domain approach the packet contention can be reduced by incorporating Fiber Delay Lines (FDLs) as optical buffer in the switch architecture. Different types of buffering architectures are reported in literatures. In the present paper a comparative performance analysis of three most popular FDL architectures are presented in order to obtain the best contention resolution performance. The analysis is further extended to consider the effect of different fiber non-linearities on the network performance.

Keywords: WDM network, contention resolution, optical buffering, non-linearity, throughput.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1769
3578 Development of Risk Assessment and Occupational Safety Management Model for Building Construction Projects

Authors: Preeda Sansakorn, Min An

Abstract:

In order to be capable of dealing with uncertainties, subjectivities, including vagueness arising in building construction projects, the application of fuzzy reasoning technique based on fuzzy set theory is proposed. This study contributes significantly to the development of a fuzzy reasoning safety risk assessment model for building construction projects that could be employed to assess the risk magnitude of each hazardous event identified during construction, and a third parameter of probability of consequence is incorporated in the model. By using the proposed safety risk analysis methodology, more reliable and less ambiguities, which provide the safety risk management project team for decision-making purposes.

Keywords: Safety risks assessment, building construction safety, fuzzy reasoning, construction risk assessment model, building construction projects.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2314
3577 Linguistic Competence Analysis and the Development of Speaking Instructional Material

Authors: Felipa M. Rico

Abstract:

Linguistic oral competence plays a vital role in attaining effective communication. Since the English language is considered as universally used language and has a high demand skill needed in the work-place, mastery is the expected output from learners. To achieve this, learners should be given integrated differentiated tasks which help them develop and strengthen the expected skills. This study aimed to develop speaking instructional supplementary material to enhance the English linguistic competence of Grade 9 students in areas of pronunciation, intonation and stress, voice projection, diction and fluency. A descriptive analysis was utilized to analyze the speaking level of performance of the students in order to employ appropriate strategies. There were two sets of respondents: 178 Grade 9 students selected through a stratified sampling and chosen at random. The other set comprised English teachers who evaluated the usefulness of the devised teaching materials. A teacher conducted a speaking test and activities were employed to analyze the speaking needs of students. Observation and recordings were also used to evaluate the students’ performance. The findings revealed that the English pronunciation of the students was slightly unclear at times, but generally fair. There were lapses but generally they rated moderate in intonation and stress, because of other language interference. In terms of voice projection, students have erratic high volume pitch. For diction, the students’ ability to produce comprehensible language is limited, and as to fluency, the choice of vocabulary and use of structure were severely limited. Based on the students’ speaking needs analyses, the supplementary material devised was based on Nunan’s IM model, incorporating context of daily life and global work settings, considering the principle that language is best learned in the actual meaningful situation. To widen the mastery of skill, a rich learning environment, filled with a variety instructional material tends to foster faster acquisition of the requisite skills for sustained learning and development. The role of IM is to encourage information to stick in the learners’ mind, as what is seen is understood more than what is heard. Teachers say they found the IM “very useful.” This implied that English teachers could adopt the materials to improve the speaking skills of students. Further, teachers should provide varied opportunities for students to get involved in real life situations where they could take turns in asking and answering questions and share information related to the activities. This would minimize anxiety among students in the use of the English language.

Keywords: Fluency, intonation, instructional materials, linguistic competence, pronunciation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1624
3576 ICT for Smart Appliances: Current Technology and Identification of Future ICT Trend

Authors: Abubakar Uba Ibrahim, Ibrahim Haruna Shanono

Abstract:

Smart metering and demand response are gaining ground in industrial and residential applications. Smart Appliances have been given concern towards achieving Smart home. The success of Smart grid development relies on the successful implementation of Information and Communication Technology (ICT) in power sector. Smart Appliances have been the technology under development and many new contributions to its realization have been reported in the last few years. The role of ICT here is to capture data in real time, thereby allowing bi-directional flow of information/data between producing and utilization point; that lead a way for the attainment of Smart appliances where home appliances can communicate between themselves and provide a self-control (switch on and off) using the signal (information) obtained from the grid. This paper depicts the background on ICT for smart appliances paying a particular attention to the current technology and identifying the future ICT trends for load monitoring through which smart appliances can be achieved to facilitate an efficient smart home system which promote demand response program. This paper grouped and reviewed the recent contributions, in order to establish the current state of the art and trends of the technology, so that the reader can be provided with a comprehensive and insightful review of where ICT for smart appliances stands and is heading to. The paper also presents a brief overview of communication types, and then narrowed the discussion to the load monitoring (Non-intrusive Appliances Load Monitoring ‘NALM’). Finally, some future trends and challenges in the further development of the ICT framework are discussed to motivate future contributions that address open problems and explore new possibilities.

Keywords: Communication technology between appliances, demand response, load monitoring, smart appliances and smart grid.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2527
3575 A Novel Forgetting Factor Recursive Least Square Algorithm Applied to the Human Motion Analysis

Authors: Hadi Sadoghi Yazdi, Mehri Sadoghi Yazdi, Mohammad Reza Mohammadi

Abstract:

This paper is concerned with studying the forgetting factor of the recursive least square (RLS). A new dynamic forgetting factor (DFF) for RLS algorithm is presented. The proposed DFF-RLS is compared to other methods. Better performance at convergence and tracking of noisy chirp sinusoid is achieved. The control of the forgetting factor at DFF-RLS is based on the gradient of inverse correlation matrix. Compared with the gradient of mean square error algorithm, the proposed approach provides faster tracking and smaller mean square error. In low signal-to-noise ratios, the performance of the proposed method is superior to other approaches.

Keywords: Forgetting factor, RLS, Inverse correlation matrix, human motion analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2225
3574 Categorical Data Modeling: Logistic Regression Software

Authors: Abdellatif Tchantchane

Abstract:

A Matlab based software for logistic regression is developed to enhance the process of teaching quantitative topics and assist researchers with analyzing wide area of applications where categorical data is involved. The software offers an option of performing stepwise logistic regression to select the most significant predictors. The software includes a feature to detect influential observations in data, and investigates the effect of dropping or misclassifying an observation on a predictor variable. The input data may consist either as a set of individual responses (yes/no) with the predictor variables or as grouped records summarizing various categories for each unique set of predictor variables' values. Graphical displays are used to output various statistical results and to assess the goodness of fit of the logistic regression model. The software recognizes possible convergence constraints when present in data, and the user is notified accordingly.

Keywords: Logistic regression, Matlab, Categorical data, Influential observation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1865
3573 A Decision Support Model for Bank Branch Location Selection

Authors: Nihan Cinar

Abstract:

Location selection is one of the most important decision making process which requires to consider several criteria based on the mission and the strategy. This study-s object is to provide a decision support model in order to help the bank selecting the most appropriate location for a bank-s branch considering a case study in Turkey. The object of the bank is to select the most appropriate city for opening a branch among six alternatives in the South-Eastern of Turkey. The model in this study was consisted of five main criteria which are Demographic, Socio-Economic, Sectoral Employment, Banking and Trade Potential and twenty one subcriteria which represent the bank-s mission and strategy. Because of the multi-criteria structure of the problem and the fuzziness in the comparisons of the criteria, fuzzy AHP is used and for the ranking of the alternatives, TOPSIS method is used.

Keywords: MCDM, bank branch location, fuzzy AHP, TOPSIS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4939
3572 Study of Structural and Electronic Properties of Ternary PdMnGe Half-Heusler Alloy

Authors: F. Bendahma, M. Mana, B. Bestani, S. Bentata

Abstract:

This study deals with the structural and electronic properties of ternary PdMnGe Half-Heusler alloy using the full potential linearized augmented plane wave (FP-LAPW) method based on the density functional theory (DFT) as implemented in the WIEN2k package, within the framework of generalized gradient approximation (GGA). Structural parameters, total and partial densities of states were also analyzed. The obtained result shows that the studied material is metallic in GGA treatment. The elastic constants (Cij) show that our compound is ductile, stiff and anisotropic.

Keywords: Full potential linearized augmented plane wave, generalized gradient approximation treatment, Half-Heusler, structural and electronic properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 618
3571 Hidden State Probabilistic Modeling for Complex Wavelet Based Image Registration

Authors: F. C. Calnegru

Abstract:

This article presents a computationally tractable probabilistic model for the relation between the complex wavelet coefficients of two images of the same scene. The two images are acquisitioned at distinct moments of times, or from distinct viewpoints, or by distinct sensors. By means of the introduced probabilistic model, we argue that the similarity between the two images is controlled not by the values of the wavelet coefficients, which can be altered by many factors, but by the nature of the wavelet coefficients, that we model with the help of hidden state variables. We integrate this probabilistic framework in the construction of a new image registration algorithm. This algorithm has sub-pixel accuracy and is robust to noise and to other variations like local illumination changes. We present the performance of our algorithm on various image types.

Keywords: Complex wavelet transform, image registration, modeling using hidden state variables, probabilistic similaritymeasure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1361
3570 Application Reliability Method for Concrete Dams

Authors: Mustapha Kamel Mihoubi, Mohamed Essadik Kerkar

Abstract:

Probabilistic risk analysis models are used to provide a better understanding of the reliability and structural failure of works, including when calculating the stability of large structures to a major risk in the event of an accident or breakdown. This work is interested in the study of the probability of failure of concrete dams through the application of reliability analysis methods including the methods used in engineering. It is in our case, the use of level 2 methods via the study limit state. Hence, the probability of product failures is estimated by analytical methods of the type first order risk method (FORM) and the second order risk method (SORM). By way of comparison, a level three method was used which generates a full analysis of the problem and involves an integration of the probability density function of random variables extended to the field of security using the Monte Carlo simulation method. Taking into account the change in stress following load combinations: normal, exceptional and extreme acting on the dam, calculation of the results obtained have provided acceptable failure probability values which largely corroborate the theory, in fact, the probability of failure tends to increase with increasing load intensities, thus causing a significant decrease in strength, shear forces then induce a shift that threatens the reliability of the structure by intolerable values of the probability of product failures. Especially, in case the increase of uplift in a hypothetical default of the drainage system.

Keywords: Dam, failure, limit-state, Monte Carlo simulation, reliability, probability, simulation, sliding, Taylor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1204
3569 An Experimental Method for Measuring Clamping Force in Bolted Connections and Effect of Bolt Threads Lubrication on Its Value

Authors: E. Hemmati Vand, R. H. Oskouei, T. N. Chakherlou

Abstract:

In this paper, the details of an experimental method to measure the clamping force value at bolted connections due to application of wrenching torque to tighten the nut have been presented. A simplified bolted joint including a holed plate with a single bolt was considered to carry out the experiments. This method was designed based on Hooke-s law by measuring compressive axial strain of a steel bush placed between the nut and the plate. In the experimental procedure, the values of clamping force were calculated for seven different levels of applied torque, and this process was repeated three times for each level of the torque. Moreover, the effect of lubrication of threads on the clamping value was studied using the same method. In both conditions (dry and lubricated threads), relation between the torque and the clamping force have been displayed in graphs.

Keywords: Clamping force, Bolted joints, Experimental method, Lubrication.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7638
3568 Calibration of the Radical Installation Limit Error of the Accelerometer in the Gravity Gradient Instrument

Authors: Danni Cong, Meiping Wu, Xiaofeng He, Junxiang Lian, Juliang Cao, Shaokuncai, Hao Qin

Abstract:

Gravity gradient instrument (GGI) is the core of the gravity gradiometer, so the structural error of the sensor has a great impact on the measurement results. In order not to affect the aimed measurement accuracy, limit error is required in the installation of the accelerometer. In this paper, based on the established measuring principle model, the radial installation limit error is calibrated, which is taken as an example to provide a method to calculate the other limit error of the installation under the premise of ensuring the accuracy of the measurement result. This method provides the idea for deriving the limit error of the geometry structure of the sensor, laying the foundation for the mechanical precision design and physical design.

Keywords: Gravity gradient sensor, radial installation limit error, accelerometer, uniaxial rotational modulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 907
3567 Appraisal of Methods for Identifying, Mapping, and Modelling of Fluvial Erosion in a Mining Environment

Authors: F. F. Howard, I. Yakubu, C. B. Boye, J. S. Y. Kuma

Abstract:

Natural and human activities, such as mining operations, expose the natural soil to adverse environmental conditions, leading to contamination of soil, groundwater, and surface water, which has negative effects on humans, flora, and fauna. Bare or partly exposed soil is most liable to fluvial erosion. This paper enumerates various methods used to identify, map, and model fluvial erosion in a mining environment. Classical, Artificial Intelligence (AI), and GIS methods have been reviewed. One of the many classical methods used to estimate river erosion is the Revised Universal Soil Loss Equation (RUSLE) model. The RUSLE model is easy to use. Its reliance on empirical relationships that may not always be applicable to specific circumstances or locations is a flaw. Other classical models for estimating fluvial erosion are the Soil and Water Assessment Tool (SWAT) and the Universal Soil Loss Equation (USLE). These models offer a more complete understanding of the underlying physical processes and encompass a wider range of situations. Although more difficult to utilise, they depend on the availability and dependability of input data for correctness. AI can help deal with multivariate and complex difficulties and predict soil loss with higher accuracy than traditional methods, and also be used to build unique models for identifying degraded areas. AI techniques have become popular as an alternative predictor for degraded environments. However, this research proposed a hybrid of classical, AI, and GIS methods for efficient and effective modelling of fluvial erosion.

Keywords: Fluvial erosion, classical methods, Artificial Intelligence, Geographic Information System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 140
3566 Determination of Non Uniform Sinusoidal Microstrip Leaky-Wave Antenna Radiating Performances in Millimeter Band

Authors: Zahéra Mekkioui

Abstract:

Here we have considered non uniform microstrip leaky-wave antenna implemented on a dielectric waveguide by a sinusoidal profile of periodic metallic grating. The non distribution of the attenuation constant α along propagation axis, optimize the radiating characteristics and performances of such antennas. The method developped here is based on an integral method where the formalism of the admittance operator is combined to a BKW approximation. First, the effect of the modeling in the modal analysis of complex waves is studied in detail. Then, the BKW model is used for the dispersion analysis of the antenna of interest. According to antenna theory, a forced continuity of the leaky-wave magnitude at discontinuities of the non uniform structure is established. To test the validity of our dispersion analysis, computed radiation patterns are presented and compared in the millimeter band.

Keywords: antenna, leaky-wave, performances, sinusoidal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1747
3565 Optical Characterization of a Microwave Plasma Torch for Hydrogen Production

Authors: Babajide O. Ogungbesan, Rajneesh Kumar, Mohamed Sassi

Abstract:

Hydrogen sulfide (H2S) is a very toxic gas that is produced in very large quantities in the oil and gas industry. It cannot be flared to the atmosphere and Claus process based gas plants are used to recover the sulfur and convert the hydrogen to water. In this paper, we present optical characterization of an atmospheric pressure microwave plasma torch for H2S dissociation into hydrogen and sulfur. The torch is operated at 2.45 GHz with power up to 2 kW. Three different gases can simultaneously be injected in the plasma torch. Visual imaging and optical emission spectroscopy are used to characterize the plasma for varying gas flow rates and microwave power. The plasma length, emission spectra and temperature are presented. The obtained experimental results validate our earlier published simulation results of plasma torch.

Keywords: Atmospheric pressure microwave plasma, gas dissociation, optical emission spectroscopy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3253
3564 Relation between Roots and Tangent Lines of Function in Fractional Dimensions: A Method for Optimization Problems

Authors: Ali Dorostkar

Abstract:

In this paper, a basic schematic of fractional dimensional optimization problem is presented. As will be shown, a method is performed based on a relation between roots and tangent lines of function in fractional dimensions for an arbitrary initial point. It is shown that for each polynomial function with order N at least N tangent lines must be existed in fractional dimensions of 0 < α < N+1 which pass exactly through the all roots of the proposed function. Geometrical analysis of tangent lines in fractional dimensions is also presented to clarify more intuitively the proposed method. Results show that with an appropriate selection of fractional dimensions, we can directly find the roots. Method is presented for giving a different direction of optimization problems by the use of fractional dimensions.

Keywords: Tangent line, fractional dimension, root, optimization problem.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 539
3563 Microcrystalline Cellulose (MCC) From Oil Palm Empty Fruit Bunch (EFB) Fiber via Simultaneous Ultrasonic and Alkali Treatment

Authors: Ridzuan Ramli, Norhafzan Junadi, Mohammad D.H. Beg, Rosli M. Yunus

Abstract:

In this study, microcrystalline cellulose (MCC) was extracted from oil palm empty fruit bunch (EFB) cellulose which was earlier isolated from oil palm EFB fibre. In order to isolate the cellulose, the chlorination method was carried out. Then, the MCC was prepared by simultaneous ultrasonic and alkali treatment from the isolated α-cellulose. Based on mass balance calculation, the yields for MCC obtained from EFB was 44%. For fiber characterization, it is observed that the chemical composition of the hemicellulose and lignin for all samples decreased while composition for cellulose increased. The structural property of the MCC was studied by X-ray diffraction (XRD) method and the result shows that the MCC produced is a cellulose-I polymorph, with 73% crystallinity.

Keywords: Oil palm empty fruit bunch, microcrystalline cellulose, ultrasonic, alkali treatment, X-ray diffraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3937
3562 A New Edit Distance Method for Finding Similarity in Dna Sequence

Authors: Patsaraporn Somboonsak, Mud-Armeen Munlin

Abstract:

The P-Bigram method is a string comparison methods base on an internal two characters-based similarity measure. The edit distance between two strings is the minimal number of elementary editing operations required to transform one string into the other. The elementary editing operations include deletion, insertion, substitution two characters. In this paper, we address the P-Bigram method to sole the similarity problem in DNA sequence. This method provided an efficient algorithm that locates all minimum operation in a string. We have been implemented algorithm and found that our program calculated that smaller distance than one string. We develop PBigram edit distance and show that edit distance or the similarity and implementation using dynamic programming. The performance of the proposed approach is evaluated using number edit and percentage similarity measures.

Keywords: Edit distance, String Matching, String Similarity

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3291