Search results for: Fuzzy Logic Speed Controller
130 Impact of Regulation on Trading in Financial Derivatives in Europe
Authors: H. Florianová, J. Nešleha
Abstract:
Financial derivatives are considered to be risky investment instruments which could possibly bring another financial crisis. As prevention, European Union and its member states have released new legal acts adjusting this area of law in recent years. There have been several cases in history of capital markets worldwide where it was shown that legislature may affect behavior of subjects on capital markets. In our paper we analyze main events on selected European stock exchanges in order to apply them on three chosen markets - Czech capital market represented by Prague Stock Exchange, German capital market represented by Deutsche Börse and Polish capital market represented by Warsaw Stock Exchange. We follow time series of development of the sum of listed derivatives on these three stock exchanges in order to evaluate popularity of those exchanges. Afterwards we compare newly listed derivatives in relation to the speed of development of these exchanges. We also make a comparison between trends in derivatives and shares development. We explain how a legal regulation may affect situation on capital markets. If the regulation is too strict, potential investors or traders are not willing to undertake it and move to other markets. On the other hand, if the regulation is too vague, trading scandals occur and the market is not reliable from the prospect of potential investors or issuers. We see that making the regulation stricter usually discourages subjects to stay on the market immediately although making the regulation vaguer to interest more subjects is usually much slower process.
Keywords: Capital markets, financial derivatives, investors' behavior, regulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 921129 Dynamic Stability Assessment of Different Wheel Sized Bicycles Based on Current Frame Design Practice with ISO Requirement for Bicycle Safety
Authors: Milan Paudel, Fook Fah Yap, Anil K. Bastola
Abstract:
The difficulties in riding small wheel bicycles and their lesser stability have been perceived for a long time. Although small wheel bicycles are designed using the similar approach and guidelines that have worked well for big wheel bicycles, the performance of the big wheelers and the smaller wheelers are markedly different. Since both the big wheelers and small wheelers have same fundamental geometry, most blame the small wheel for this discrepancy in the performance. This paper reviews existing guidelines for bicycle design, especially the front steering geometry for the bicycle, and provides a systematic and quantitative analysis of different wheel sized bicycles. A validated mathematical model has been used as a tool to assess the dynamic performance of the bicycles in term of their self-stability. The results obtained were found to corroborate the subjective perception of cyclists for small wheel bicycles. The current approach for small wheel bicycle design requires higher speed to be self-stable. However, it was found that increasing the headtube angle and selecting a proper trail could improve the dynamic performance of small wheel bicycles. A range of parameters for front steering geometry has been identified for small wheel bicycles that have comparable stability as big wheel bicycles. Interestingly, most of the identified geometries are found to be beyond the ISO recommended range and seem to counter the current approach of small wheel bicycle design. Therefore, it was successfully shown that the guidelines for big wheelers do not translate directly to small wheelers, but careful selection of the front geometry could make small wheel bicycles as stable as big wheel bicycles.
Keywords: Big wheel bicycle, design approach, ISO requirements, small wheel bicycle, stability and performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 851128 Microwave Assisted Solvent-Free Catalytic Transesterification of Glycerol to Glycerol Carbonate
Authors: Wai Keng Teng, Gek Cheng Ngoh, Rozita Yusoff, Mohamed Kheireddine Aroua, Joe Shen Heng
Abstract:
As a by-product of the biodiesel industries, glycerol has been vastly generated which surpasses the market demand. It is imperative to develop an efficient glycerol valorization processes in minimizing the net energy requirement and intensifying the biodiesel production. In this study, base-catalyzed transesterification of glycerol with dimethyl carbonate using microwave irradiation as heating method to produce glycerol carbonate was conducted by varying grades of glycerol, i.e. 70%, 86% and 99% purity, that is obtained from biodiesel plant. Metal oxide catalysts were used with varying operating parameters including reaction time, DMC/glycerol molar ratio, catalyst weight %, temperature and stirring speed. From the study on the effect of different operating parameters it was found that the type of catalyst used has the most significant effect on the transesterification reaction. Amidst the metal oxide catalysts examined, CaO gave the best performance. This study indicates the feasibility of producing glycerol carbonate using different grade of glycerol in both conventional thermal activation and microwave irradiation with CaO as catalyst. Microwave assisted transesterification (MAT) of glycerol into glycerol carbonate has demonstrated itself as an energy efficient route by achieving 94.2% yield of GC at 65°C, 5 minutes reaction time, 1 wt% CaO and DMC/glycerol molar ratio of 2. The advantages of MAT transesterification route has made the direct utilization of bioglycerol from biodiesel production without the need of purification. This has marked a more economical and less-energy intensive glycerol carbonate synthesis route.Keywords: Biodiesel, glycerol, glycerol carbonate, microwave irradiation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2903127 Effect of Peak-to-Average Power Ratio Reduction on the Multicarrier Communication System Performance Parameters
Authors: Sanjay Singh, M Sathish Kumar, H. S Mruthyunjaya
Abstract:
Multicarrier transmission system such as Orthogonal Frequency Division Multiplexing (OFDM) is a promising technique for high bit rate transmission in wireless communication system. OFDM is a spectrally efficient modulation technique that can achieve high speed data transmission over multipath fading channels without the need for powerful equalization techniques. However the price paid for this high spectral efficiency and less intensive equalization is low power efficiency. OFDM signals are very sensitive to nonlinear effects due to the high Peak-to-Average Power Ratio (PAPR), which leads to the power inefficiency in the RF section of the transmitter. This paper investigates the effect of PAPR reduction on the performance parameter of multicarrier communication system. Performance parameters considered are power consumption of Power Amplifier (PA) and Digital-to-Analog Converter (DAC), power amplifier efficiency, SNR of DAC and BER performance of the system. From our analysis it is found that irrespective of PAPR reduction technique being employed, the power consumption of PA and DAC reduces and power amplifier efficiency increases due to reduction in PAPR. Moreover, it has been shown that for a given BER performance the requirement of Input-Backoff (IBO) reduces with reduction in PAPR.Keywords: BER, Crest Factor (CF), Digital-to-Analog Converter(DAC), Input-Backoff (IBO), Orthogonal Frequency Division Multiplexing(OFDM), Peak-to-Average Power Ratio (PAPR), PowerAmplifier efficiency, SNR
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3288126 Study of Temperature Changes in Fars Province
Authors: A. Gandomkar, R. Dehghani
Abstract:
Climate change is a phenomenon has been based on the available evidence from a very long time ago and now its existence is very probable. The speed and nature of climate parameters changes at the middle of twentieth century has been different and its quickness more than the before and its trend changed to some extent comparing to the past. Climate change issue now regarded as not only one of the most common scientific topic but also a social political one, is not a new issue. Climate change is a complicated atmospheric oceanic phenomenon on a global scale and long-term. Precipitation pattern change, fast decrease of snowcovered resources and its rapid melting, increased evaporation, the occurrence of destroying floods, water shortage crisis, severe reduction at the rate of harvesting agricultural products and, so on are all the significant of climate change. To cope with this phenomenon, its consequences and events in which public instruction is the most important but it may be climate that no significant cant and effective action has been done so far. The present article is included a part of one surrey about climate change in Fars. The study area having annually mean temperature 14 and precipitation 320 mm .23 stations inside the basin with a common 37 year statistical period have been applied to the meteorology data (1974-2010). Man-kendal and change factor methods are two statistical methods, applying them, the trend of changes and the annual mean average temperature and the annual minimum mean temperature were studied by using them. Based on time series for each parameter, the annual mean average temperature and the mean of annual maximum temperature have a rising trend so that this trend is clearer to the mean of annual maximum temperature.Keywords: Climate change, Coefficient Variation, Fars province, Man-Kendal method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1916125 A Comparative Study of Rigid and Modified Simplex Methods for Optimal Parameter Settings of ACO for Noisy Non-Linear Surfaces
Authors: Seksan Chunothaisawat, Pongchanun Luangpaiboon
Abstract:
There are two common types of operational research techniques, optimisation and metaheuristic methods. The latter may be defined as a sequential process that intelligently performs the exploration and exploitation adopted by natural intelligence and strong inspiration to form several iterative searches. An aim is to effectively determine near optimal solutions in a solution space. In this work, a type of metaheuristics called Ant Colonies Optimisation, ACO, inspired by a foraging behaviour of ants was adapted to find optimal solutions of eight non-linear continuous mathematical models. Under a consideration of a solution space in a specified region on each model, sub-solutions may contain global or multiple local optimum. Moreover, the algorithm has several common parameters; number of ants, moves, and iterations, which act as the algorithm-s driver. A series of computational experiments for initialising parameters were conducted through methods of Rigid Simplex, RS, and Modified Simplex, MSM. Experimental results were analysed in terms of the best so far solutions, mean and standard deviation. Finally, they stated a recommendation of proper level settings of ACO parameters for all eight functions. These parameter settings can be applied as a guideline for future uses of ACO. This is to promote an ease of use of ACO in real industrial processes. It was found that the results obtained from MSM were pretty similar to those gained from RS. However, if these results with noise standard deviations of 1 and 3 are compared, MSM will reach optimal solutions more efficiently than RS, in terms of speed of convergence.
Keywords: Ant colony optimisation, metaheuristics, modified simplex, non-linear, rigid simplex.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1625124 A Vehicular Visual Tracking System Incorporating Global Positioning System
Authors: Hsien-Chou Liao, Yu-Shiang Wang
Abstract:
Surveillance system is widely used in the traffic monitoring. The deployment of cameras is moving toward a ubiquitous camera (UbiCam) environment. In our previous study, a novel service, called GPS-VT, was firstly proposed by incorporating global positioning system (GPS) and visual tracking techniques for the UbiCam environment. The first prototype is called GODTA (GPS-based Moving Object Detection and Tracking Approach). For a moving person carried GPS-enabled mobile device, he can be tracking when he enters the field-of-view (FOV) of a camera according to his real-time GPS coordinate. In this paper, GPS-VT service is applied to the tracking of vehicles. The moving speed of a vehicle is much faster than a person. It means that the time passing through the FOV is much shorter than that of a person. Besides, the update interval of GPS coordinate is once per second, it is asynchronous with the frame rate of the real-time image. The above asynchronous is worsen by the network transmission delay. These factors are the main challenging to fulfill GPS-VT service on a vehicle.In order to overcome the influence of the above factors, a back-propagation neural network (BPNN) is used to predict the possible lane before the vehicle enters the FOV of a camera. Then, a template matching technique is used for the visual tracking of a target vehicle. The experimental result shows that the target vehicle can be located and tracking successfully. The success location rate of the implemented prototype is higher than that of the previous GODTA.Keywords: visual surveillance, visual tracking, globalpositioning system, intelligent transportation system
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1918123 Near Field Focusing Behaviour of Airborne Ultrasonic Phased Arrays Influenced by Airflows
Authors: D. Sun, T. F. Lu, A. Zander, M. Trinkle
Abstract:
This paper investigates the potential use of airborne ultrasonic phased arrays for imaging in outdoor environments as a means of overcoming the limitations experienced by kinect sensors, which may fail to work in the outdoor environments due to the oversaturation of the infrared photo diodes. Ultrasonic phased arrays have been well studied for static media, yet there appears to be no comparable examination in the literature of the impact of a flowing medium on the focusing behaviour of near field focused ultrasonic arrays. This paper presents a method for predicting the sound pressure fields produced by a single ultrasound element or an ultrasonic phased array influenced by airflows. The approach can be used to determine the actual focal point location of an array exposed in a known flow field. From the presented simulation results based upon this model, it can be concluded that uniform flows in the direction orthogonal to the acoustic propagation have a noticeable influence on the sound pressure field, which is reflected in the twisting of the steering angle of the array. Uniform flows in the same direction as the acoustic propagation have negligible influence on the array. For an array impacted by a turbulent flow, determining the location of the focused sound field becomes difficult due to the irregularity and continuously changing direction and the speed of the turbulent flow. In some circumstances, ultrasonic phased arrays impacted by turbulent flows may not be capable of producing a focused sound field.
Keywords: Airborne, airflow, focused sound field, ultrasonic phased array.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1626122 Vehicle Gearbox Fault Diagnosis Based On Cepstrum Analysis
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
Research on damage of gears and gear pairs using vibration signals remains very attractive, because vibration signals from a gear pair are complex in nature and not easy to interpret. Predicting gear pair defects by analyzing changes in vibration signal of gears pairs in operation is a very reliable method. Therefore, a suitable vibration signal processing technique is necessary to extract defect information generally obscured by the noise from dynamic factors of other gear pairs.This article presents the value of cepstrum analysis in vehicle gearbox fault diagnosis. Cepstrum represents the overall power content of a whole family of harmonics and sidebands when more than one family of sidebands is present at the same time. The concept for the measurement and analysis involved in using the technique are briefly outlined. Cepstrum analysis is used for detection of an artificial pitting defect in a vehicle gearbox loaded with different speeds and torques. The test stand is equipped with three dynamometers; the input dynamometer serves asthe internal combustion engine, the output dynamometers introduce the load on the flanges of the output joint shafts. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. Also, a method for fault diagnosis of gear faults is presented based on order Cepstrum. The procedure is illustrated with the experimental vibration data of the vehicle gearbox. The results show the effectiveness of Cepstrum analysis in detection and diagnosis of the gear condition.
Keywords: Cepstrum analysis, fault diagnosis, gearbox.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3311121 Effect of the Machine Frame Structures on the Frequency Responses of Spindle Tool
Authors: Yuan L. Lai, Yong R. Chen, Jui P. Hung, Tzuo L. Luo, Hsi H. Hsiao
Abstract:
Chatter vibration has been a troublesome problem for a machine tool toward the high precision and high speed machining. Essentially, the machining performance is determined by the dynamic characteristics of the machine tool structure and dynamics of cutting process. Therefore the dynamic vibration behavior of spindle tool system greatly determines the performance of machine tool. The purpose of this study is to investigate the influences of the machine frame structure on the dynamic frequency of spindle tool unit through finite element modeling approach. To this end, a realistic finite element model of the vertical milling system was created by incorporated the spindle-bearing model into the spindle head stock of the machine frame. Using this model, the dynamic characteristics of the milling machines with different structural designs of spindle head stock and identical spindle tool unit were demonstrated. The results of the finite element modeling reveal that the spindle tool unit behaves more compliant when the excited frequency approaches the natural mode of the spindle tool; while the spindle tool show a higher dynamic stiffness at lower frequency that may be initiated by the structural mode of milling head. Under this condition, it is concluded that the structural configuration of spindle head stock associated with the vertical column of milling machine plays an important role in determining the machining dynamics of the spindle unit.Keywords: Machine tools, Compliance, Frequency response function, Machine frame structure, Spindle unit
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2819120 Non-Burn Treatment of Health Care Risk Waste
Authors: Jefrey Pilusa, Tumisang Seodigeng
Abstract:
This research discusses a South African case study for the potential of utilizing refuse-derived fuel (RDF) obtained from non-burn treatment of health care risk waste (HCRW) as potential feedstock for green energy production. This specific waste stream can be destroyed via non-burn treatment technology involving high-speed mechanical shredding followed by steam or chemical injection to disinfect the final product. The RDF obtained from this process is characterised by a low moisture, low ash, and high calorific value which means it can be potentially used as high-value solid fuel. Due to the raw feed of this RDF being classified as hazardous, the final RDF has been reported to be non-infectious and can blend with other combustible wastes such as rubber and plastic for waste to energy applications. This study evaluated non-burn treatment technology as a possible solution for on-site destruction of HCRW in South African private and public health care centres. Waste generation quantities were estimated based on the number of registered patient beds, theoretical bed occupancy. Time and motion study was conducted to evaluate the logistics viability of on-site treatment. Non-burn treatment technology for HCRW is a promising option for South Africa, and successful implementation of this method depends upon the initial capital investment, operational cost and environmental permitting of such technology; there are other influencing factors such as the size of the waste stream, product off-take price as well as product demand.
Keywords: Autoclave, disposal, fuel, incineration, medical waste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1165119 Modal Analysis of Machine Tool Column Using Finite Element Method
Authors: Migbar Assefa
Abstract:
The performance of a machine tool is eventually assessed by its ability to produce a component of the required geometry in minimum time and at small operating cost. It is customary to base the structural design of any machine tool primarily upon the requirements of static rigidity and minimum natural frequency of vibration. The operating properties of machines like cutting speed, feed and depth of cut as well as the size of the work piece also have to be kept in mind by a machine tool structural designer. This paper presents a novel approach to the design of machine tool column for static and dynamic rigidity requirement. Model evaluation is done effectively through use of General Finite Element Analysis software ANSYS. Studies on machine tool column are used to illustrate finite element based concept evaluation technique. This paper also presents results obtained from the computations of thin walled box type columns that are subjected to torsional and bending loads in case of static analysis and also results from modal analysis. The columns analyzed are square and rectangle based tapered open column, column with cover plate, horizontal partitions and with apertures. For the analysis purpose a total of 70 columns were analyzed for bending, torsional and modal analysis. In this study it is observed that the orientation and aspect ratio of apertures have no significant effect on the static and dynamic rigidity of the machine tool structure.
Keywords: Finite Element Modeling, Modal Analysis, Machine tool structure, Static Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5039118 A Novel Hopfield Neural Network for Perfect Calculation of Magnetic Resonance Spectroscopy
Authors: Hazem M. El-Bakry
Abstract:
In this paper, an automatic determination algorithm for nuclear magnetic resonance (NMR) spectra of the metabolites in the living body by magnetic resonance spectroscopy (MRS) without human intervention or complicated calculations is presented. In such method, the problem of NMR spectrum determination is transformed into the determination of the parameters of a mathematical model of the NMR signal. To calculate these parameters efficiently, a new model called modified Hopfield neural network is designed. The main achievement of this paper over the work in literature [30] is that the speed of the modified Hopfield neural network is accelerated. This is done by applying cross correlation in the frequency domain between the input values and the input weights. The modified Hopfield neural network can accomplish complex dignals perfectly with out any additinal computation steps. This is a valuable advantage as NMR signals are complex-valued. In addition, a technique called “modified sequential extension of section (MSES)" that takes into account the damping rate of the NMR signal is developed to be faster than that presented in [30]. Simulation results show that the calculation precision of the spectrum improves when MSES is used along with the neural network. Furthermore, MSES is found to reduce the local minimum problem in Hopfield neural networks. Moreover, the performance of the proposed method is evaluated and there is no effect on the performance of calculations when using the modified Hopfield neural networks.
Keywords: Hopfield Neural Networks, Cross Correlation, Nuclear Magnetic Resonance, Magnetic Resonance Spectroscopy, Fast Fourier Transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1846117 Cascaded ANN for Evaluation of Frequency and Air-gap Voltage of Self-Excited Induction Generator
Authors: Raja Singh Khela, R. K. Bansal, K. S. Sandhu, A. K. Goel
Abstract:
Self-Excited Induction Generator (SEIG) builds up voltage while it enters in its magnetic saturation region. Due to non-linear magnetic characteristics, the performance analysis of SEIG involves cumbersome mathematical computations. The dependence of air-gap voltage on saturated magnetizing reactance can only be established at rated frequency by conducting a laboratory test commonly known as synchronous run test. But, there is no laboratory method to determine saturated magnetizing reactance and air-gap voltage of SEIG at varying speed, terminal capacitance and other loading conditions. For overall analysis of SEIG, prior information of magnetizing reactance, generated frequency and air-gap voltage is essentially required. Thus, analytical methods are the only alternative to determine these variables. Non-existence of direct mathematical relationship of these variables for different terminal conditions has forced the researchers to evolve new computational techniques. Artificial Neural Networks (ANNs) are very useful for solution of such complex problems, as they do not require any a priori information about the system. In this paper, an attempt is made to use cascaded neural networks to first determine the generated frequency and magnetizing reactance with varying terminal conditions and then air-gap voltage of SEIG. The results obtained from the ANN model are used to evaluate the overall performance of SEIG and are found to be in good agreement with experimental results. Hence, it is concluded that analysis of SEIG can be carried out effectively using ANNs.Keywords: Self-Excited Induction Generator, Artificial NeuralNetworks, Exciting Capacitance and Saturated magnetizingreactance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1693116 The Effect of Motor Learning Based Computer-Assisted Practice for Children with Handwriting Deficit – Comparing with the Effect of Traditional Sensorimotor Approach
Authors: Shao-Hsia Chang, Nan-Ying Yu
Abstract:
The objective of this study was to test how advanced digital technology enables a more effective training on the handwriting of children with handwriting deficit. This study implemented the graphomotor apparatuses to a computer-assisted instruction system. In a randomized controlled trial, the experiments for verifying the intervention effect were conducted. Forty two children with handwriting deficit were assigned to computer-assisted instruction, sensorimotor training or control (no intervention) group. Handwriting performance was measured using the Elementary reading/writing test and computerized handwriting evaluation before and after 6 weeks of intervention. Analysis of variance of change scores were conducted to show whether statistically significant difference across the three groups. Significant difference was found among three groups. Computer group shows significant difference from the other two groups. Significance was denoted in near-point, far-point copy, dictation test, and writing from phonetic symbols. Writing speed and mean stroke velocity in near-, far-point and short paragraph copy were found significantly difference among three groups. Computer group shows significant improvement from the other groups. For clinicians and school teachers, the results of this study provide a motor control based insight for the improvement of handwriting difficulties.
Keywords: Dysgraphia, computerized handwriting evaluation, sensorimotor program, computer assisted program.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2090115 Probabilistic Crash Prediction and Prevention of Vehicle Crash
Authors: Lavanya Annadi, Fahimeh Jafari
Abstract:
Transportation brings immense benefits to society, but it also has its costs. Costs include the cost of infrastructure, personnel, and equipment, but also the loss of life and property in traffic accidents on the road, delays in travel due to traffic congestion, and various indirect costs in terms of air transport. This research aims to predict the probabilistic crash prediction of vehicles using Machine Learning due to natural and structural reasons by excluding spontaneous reasons, like overspeeding, etc., in the United States. These factors range from meteorological elements such as weather conditions, precipitation, visibility, wind speed, wind direction, temperature, pressure, and humidity, to human-made structures, like road structure components such as Bumps, Roundabouts, No Exit, Turning Loops, Give Away, etc. The probabilities are categorized into ten distinct classes. All the predictions are based on multiclass classification techniques, which are supervised learning. This study considers all crashes in all states collected by the US government. The probability of the crash was determined by employing Multinomial Expected Value, and a classification label was assigned accordingly. We applied three classification models, including multiclass Logistic Regression, Random Forest and XGBoost. The numerical results show that XGBoost achieved a 75.2% accuracy rate which indicates the part that is being played by natural and structural reasons for the crash. The paper has provided in-depth insights through exploratory data analysis.
Keywords: Road safety, crash prediction, exploratory analysis, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 88114 Fast Factored DCT-LMS Speech Enhancement for Performance Enhancement of Digital Hearing Aid
Authors: Sunitha. S.L., V. Udayashankara
Abstract:
Background noise is particularly damaging to speech intelligibility for people with hearing loss especially for sensorineural loss patients. Several investigations on speech intelligibility have demonstrated sensorineural loss patients need 5-15 dB higher SNR than the normal hearing subjects. This paper describes Discrete Cosine Transform Power Normalized Least Mean Square algorithm to improve the SNR and to reduce the convergence rate of the LMS for Sensory neural loss patients. Since it requires only real arithmetic, it establishes the faster convergence rate as compare to time domain LMS and also this transformation improves the eigenvalue distribution of the input autocorrelation matrix of the LMS filter. The DCT has good ortho-normal, separable, and energy compaction property. Although the DCT does not separate frequencies, it is a powerful signal decorrelator. It is a real valued function and thus can be effectively used in real-time operation. The advantages of DCT-LMS as compared to standard LMS algorithm are shown via SNR and eigenvalue ratio computations. . Exploiting the symmetry of the basis functions, the DCT transform matrix [AN] can be factored into a series of ±1 butterflies and rotation angles. This factorization results in one of the fastest DCT implementation. There are different ways to obtain factorizations. This work uses the fast factored DCT algorithm developed by Chen and company. The computer simulations results show superior convergence characteristics of the proposed algorithm by improving the SNR at least 10 dB for input SNR less than and equal to 0 dB, faster convergence speed and better time and frequency characteristics.Keywords: Hearing Impairment, DCT Adaptive filter, Sensorineural loss patients, Convergence rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2171113 Probability-Based Damage Detection of Structures Using Model Updating with Enhanced Ideal Gas Molecular Movement Algorithm
Authors: M. R. Ghasemi, R. Ghiasi, H. Varaee
Abstract:
Model updating method has received increasing attention in damage detection structures based on measured modal parameters. Therefore, a probability-based damage detection (PBDD) procedure based on a model updating procedure is presented in this paper, in which a one-stage model-based damage identification technique based on the dynamic features of a structure is investigated. The presented framework uses a finite element updating method with a Monte Carlo simulation that considers the uncertainty caused by measurement noise. Enhanced ideal gas molecular movement (EIGMM) is used as the main algorithm for model updating. Ideal gas molecular movement (IGMM) is a multiagent algorithm based on the ideal gas molecular movement. Ideal gas molecules disperse rapidly in different directions and cover all the space inside. This is embedded in the high speed of molecules, collisions between them and with the surrounding barriers. In IGMM algorithm to accomplish the optimal solutions, the initial population of gas molecules is randomly generated and the governing equations related to the velocity of gas molecules and collisions between those are utilized. In this paper, an enhanced version of IGMM, which removes unchanged variables after specified iterations, is developed. The proposed method is implemented on two numerical examples in the field of structural damage detection. The results show that the proposed method can perform well and competitive in PBDD of structures.Keywords: Enhanced ideal gas molecular movement, ideal gas molecular movement, model updating method, probability-based damage detection, uncertainty quantification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1077112 An Evaluation Method for Two-Dimensional Position Errors and Assembly Errors of a Rotational Table on a 4 Axis Machine Tool
Authors: Jooho Hwang, Chang-Kyu Song, Chun-Hong Park
Abstract:
This paper describes a method to measure and compensate a 4 axes ultra-precision machine tool that generates micro patterns on the large surfaces. The grooving machine is usually used for making a micro mold for many electrical parts such as a light guide plate for LCD and fuel cells. The ultra precision machine tool has three linear axes and one rotational table. Shaping is usually used to generate micro patterns. In the case of 50 μm pitch and 25 μm height pyramid pattern machining with a 90° wedge angle bite, one of linear axis is used for long stroke motion for high cutting speed and other linear axis are used for feeding. The triangular patterns can be generated with many times of long stroke of one axis. Then 90° rotation of work piece is needed to make pyramid patterns with superposition of machined two triangular patterns. To make a two dimensional positioning error, straightness of two axes in out of plane, squareness between the each axis are important. Positioning errors, straightness and squarness were measured by laser interferometer system. Those were compensated and confirmed by ISO230-6. One of difficult problem to measure the error motions is squareness or parallelism of axis between the rotational table and linear axis. It was investigated by simultaneous moving of rotary table and XY axes. This compensation method is introduced in this paper.Keywords: Ultra-precision machine tool, muti-axis errors, squraness, positioning errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1582111 Endeavor in Management Process by Executive Dashboards: The Case of the Financial Directorship in Brazilian Navy
Authors: R. S. Quintal, J. L. Tesch Santos, M. D. Davis, E. C. de Santana, M. de F. Bandeira dos Santos
Abstract:
The objective is to identify the contributions from the introduction of the computerized system deal within the Accounting Department of Brazilian Navy Financial Directorship and its possible effects on the budgetary and financial harvest of Brazilian Navy. The relevance lies in the fact that the management process is responsible for the continuous improvement of organizational performance through higher levels of quality in their activities. Improvements in organizational processes have direct effects on crops cost, quality, reliability, flexibility and speed. The method of study of this research is the case study. The choice of case study attended, among other demands, a need for greater flexibility to study processes related to a computerized system. The sources of evidence were used literature, documentary and direct observation. Direct observation was made by monitoring the implementation of the computerized system in the Division of Management Analysis. The main findings of the study point to the fact that the computerized system may contribute significantly to the standardization of information. There was improvement of internal processes in the division of management analysis, made possible the consolidation of a standard management and performance analysis that contribute to global homogeneity in the treatment of information essential to the process of decision making. This study has limitations related to the fact the search result be subject exclusively to the case studied, and it is impossible to generalize to other organs of government.
Keywords: Process Management, Management Control, Business Intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1986110 Modeling the Hybrid Battery/Super-Storage System for a Solar Standalone Microgrid
Authors: Astiaj Khoramshahi, Hossein Ahmadi Danesh Ashtiani, Ahmad Khoshgard, Hamidreza Damghani, Leila Damghani
Abstract:
Solar energy systems using various storages are required to be evaluated based on energy requirements and applications. Also, modeling and analysis of storage systems are necessary to increase the effectiveness of combinations of these systems. In this paper, analysis based on the MATLAB software has been analyzed to evaluate the response of the hybrid energy system considering various technologies of renewable energy and energy storage. In the present study, three different simulation scenarios are presented. Simulation output results using software for the first scenario show that the battery is effective in smoothing the overall power demand to the consumer studied during a day, but temporary loads on the grid with high frequencies, effectively cannot be canceled due to the limited response speed of battery control. Simulation outputs for the second scenario using the energy storage system show that sudden changes in demand power are paved by super saving. The majority of these sudden changes in power demand are caused by sewing consumers and receiving variable solar power (due to clouds passing through the solar array). Simulation outputs for the third scenario show the effects of the hybrid system for the same consumer and the output of the solar array, leading to the smallest amount of power demand fed into the grid, as well as demand at peak times. According to the "battery only" scenario, the displacement technique of the peak load has been significantly reduced.
Keywords: Storage system, super storage, standalone, microgrid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 336109 A Study on Algorithm Fusion for Recognition and Tracking of Moving Robot
Authors: Jungho Choi, Youngwan Cho
Abstract:
This paper presents an algorithm for the recognition and tracking of moving objects, 1/10 scale model car is used to verify performance of the algorithm. Presented algorithm for the recognition and tracking of moving objects in the paper is as follows. SURF algorithm is merged with Lucas-Kanade algorithm. SURF algorithm has strong performance on contrast, size, rotation changes and it recognizes objects but it is slow due to many computational complexities. Processing speed of Lucas-Kanade algorithm is fast but the recognition of objects is impossible. Its optical flow compares the previous and current frames so that can track the movement of a pixel. The fusion algorithm is created in order to solve problems which occurred using the Kalman Filter to estimate the position and the accumulated error compensation algorithm was implemented. Kalman filter is used to create presented algorithm to complement problems that is occurred when fusion two algorithms. Kalman filter is used to estimate next location, compensate for the accumulated error. The resolution of the camera (Vision Sensor) is fixed to be 640x480. To verify the performance of the fusion algorithm, test is compared to SURF algorithm under three situations, driving straight, curve, and recognizing cars behind the obstacles. Situation similar to the actual is possible using a model vehicle. Proposed fusion algorithm showed superior performance and accuracy than the existing object recognition and tracking algorithms. We will improve the performance of the algorithm, so that you can experiment with the images of the actual road environment.Keywords: SURF, Optical Flow Lucas-Kanade, Kalman Filter, object recognition, object tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2294108 Multi Response Optimization in Drilling Al6063/SiC/15% Metal Matrix Composite
Authors: Hari Singh, Abhishek Kamboj, Sudhir Kumar
Abstract:
This investigation proposes a grey-based Taguchi method to solve the multi-response problems. The grey-based Taguchi method is based on the Taguchi’s design of experimental method, and adopts grey relational analysis (GRA) to transfer multi-response problems into single-response problems. In this investigation, an attempt has been made to optimize the drilling process parameters considering weighted output response characteristics using grey relational analysis. The output response characteristics considered are surface roughness, burr height and hole diameter error under the experimental conditions of cutting speed, feed rate, step angle, and cutting environment. The drilling experiments were conducted using L27 orthogonal array. A combination of orthogonal array, design of experiments and grey relational analysis was used to ascertain best possible drilling process parameters that give minimum surface roughness, burr height and hole diameter error. The results reveal that combination of Taguchi design of experiment and grey relational analysis improves surface quality of drilled hole.
Keywords: Metal matrix composite, Drilling, Optimization, step drill, Surface roughness, burr height, hole diameter error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3255107 The Design of a Vehicle Traffic Flow Prediction Model for a Gauteng Freeway Based on an Ensemble of Multi-Layer Perceptron
Authors: Tebogo Emma Makaba, Barnabas Ndlovu Gatsheni
Abstract:
The cities of Johannesburg and Pretoria both located in the Gauteng province are separated by a distance of 58 km. The traffic queues on the Ben Schoeman freeway which connects these two cities can stretch for almost 1.5 km. Vehicle traffic congestion impacts negatively on the business and the commuter’s quality of life. The goal of this paper is to identify variables that influence the flow of traffic and to design a vehicle traffic prediction model, which will predict the traffic flow pattern in advance. The model will unable motorist to be able to make appropriate travel decisions ahead of time. The data used was collected by Mikro’s Traffic Monitoring (MTM). Multi-Layer perceptron (MLP) was used individually to construct the model and the MLP was also combined with Bagging ensemble method to training the data. The cross—validation method was used for evaluating the models. The results obtained from the techniques were compared using predictive and prediction costs. The cost was computed using combination of the loss matrix and the confusion matrix. The predicted models designed shows that the status of the traffic flow on the freeway can be predicted using the following parameters travel time, average speed, traffic volume and day of month. The implications of this work is that commuters will be able to spend less time travelling on the route and spend time with their families. The logistics industry will save more than twice what they are currently spending.Keywords: Bagging ensemble methods, confusion matrix, multi-layer perceptron, vehicle traffic flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1777106 Perforation Analysis of the Aluminum Alloy Sheets Subjected to High Rate of Loading and Heated Using Thermal Chamber: Experimental and Numerical Approach
Authors: A. Bendarma, T. Jankowiak, A. Rusinek, T. Lodygowski, M. Klósak, S. Bouslikhane
Abstract:
The analysis of the mechanical characteristics and dynamic behavior of aluminum alloy sheet due to perforation tests based on the experimental tests coupled with the numerical simulation is presented. The impact problems (penetration and perforation) of the metallic plates have been of interest for a long time. Experimental, analytical as well as numerical studies have been carried out to analyze in details the perforation process. Based on these approaches, the ballistic properties of the material have been studied. The initial and residual velocities laser sensor is used during experiments to obtain the ballistic curve and the ballistic limit. The energy balance is also reported together with the energy absorbed by the aluminum including the ballistic curve and ballistic limit. The high speed camera helps to estimate the failure time and to calculate the impact force. A wide range of initial impact velocities from 40 up to 180 m/s has been covered during the tests. The mass of the conical nose shaped projectile is 28 g, its diameter is 12 mm, and the thickness of the aluminum sheet is equal to 1.0 mm. The ABAQUS/Explicit finite element code has been used to simulate the perforation processes. The comparison of the ballistic curve was obtained numerically and was verified experimentally, and the failure patterns are presented using the optimal mesh densities which provide the stability of the results. A good agreement of the numerical and experimental results is observed.Keywords: Aluminum alloy, ballistic behavior, failure criterion, numerical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 934105 A Hybrid Ontology Based Approach for Ranking Documents
Authors: Sarah Motiee, Azadeh Nematzadeh, Mehrnoush Shamsfard
Abstract:
Increasing growth of information volume in the internet causes an increasing need to develop new (semi)automatic methods for retrieval of documents and ranking them according to their relevance to the user query. In this paper, after a brief review on ranking models, a new ontology based approach for ranking HTML documents is proposed and evaluated in various circumstances. Our approach is a combination of conceptual, statistical and linguistic methods. This combination reserves the precision of ranking without loosing the speed. Our approach exploits natural language processing techniques to extract phrases from documents and the query and doing stemming on words. Then an ontology based conceptual method will be used to annotate documents and expand the query. To expand a query the spread activation algorithm is improved so that the expansion can be done flexible and in various aspects. The annotated documents and the expanded query will be processed to compute the relevance degree exploiting statistical methods. The outstanding features of our approach are (1) combining conceptual, statistical and linguistic features of documents, (2) expanding the query with its related concepts before comparing to documents, (3) extracting and using both words and phrases to compute relevance degree, (4) improving the spread activation algorithm to do the expansion based on weighted combination of different conceptual relationships and (5) allowing variable document vector dimensions. A ranking system called ORank is developed to implement and test the proposed model. The test results will be included at the end of the paper.Keywords: Document ranking, Ontology, Spread activation algorithm, Annotation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1632104 ORank: An Ontology Based System for Ranking Documents
Authors: Mehrnoush Shamsfard, Azadeh Nematzadeh, Sarah Motiee
Abstract:
Increasing growth of information volume in the internet causes an increasing need to develop new (semi)automatic methods for retrieval of documents and ranking them according to their relevance to the user query. In this paper, after a brief review on ranking models, a new ontology based approach for ranking HTML documents is proposed and evaluated in various circumstances. Our approach is a combination of conceptual, statistical and linguistic methods. This combination reserves the precision of ranking without loosing the speed. Our approach exploits natural language processing techniques for extracting phrases and stemming words. Then an ontology based conceptual method will be used to annotate documents and expand the query. To expand a query the spread activation algorithm is improved so that the expansion can be done in various aspects. The annotated documents and the expanded query will be processed to compute the relevance degree exploiting statistical methods. The outstanding features of our approach are (1) combining conceptual, statistical and linguistic features of documents, (2) expanding the query with its related concepts before comparing to documents, (3) extracting and using both words and phrases to compute relevance degree, (4) improving the spread activation algorithm to do the expansion based on weighted combination of different conceptual relationships and (5) allowing variable document vector dimensions. A ranking system called ORank is developed to implement and test the proposed model. The test results will be included at the end of the paper.Keywords: Document ranking, Ontology, Spread activation algorithm, Annotation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1890103 Solid State Drive End to End Reliability Prediction, Characterization and Control
Authors: Mohd Azman Abdul Latif, Erwan Basiron
Abstract:
A flaw or drift from expected operational performance in one component (NAND, PMIC, controller, DRAM, etc.) may affect the reliability of the entire Solid State Drive (SSD) system. Therefore, it is important to ensure the required quality of each individual component through qualification testing specified using standards or user requirements. Qualification testing is time-consuming and comes at a substantial cost for product manufacturers. A highly technical team, from all the eminent stakeholders is embarking on reliability prediction from beginning of new product development, identify critical to reliability parameters, perform full-blown characterization to embed margin into product reliability and establish control to ensure the product reliability is sustainable in the mass production. The paper will discuss a comprehensive development framework, comprehending SSD end to end from design to assembly, in-line inspection, in-line testing and will be able to predict and to validate the product reliability at the early stage of new product development. During the design stage, the SSD will go through intense reliability margin investigation with focus on assembly process attributes, process equipment control, in-process metrology and also comprehending forward looking product roadmap. Once these pillars are completed, the next step is to perform process characterization and build up reliability prediction modeling. Next, for the design validation process, the reliability prediction specifically solder joint simulator will be established. The SSD will be stratified into Non-Operating and Operating tests with focus on solder joint reliability and connectivity/component latent failures by prevention through design intervention and containment through Temperature Cycle Test (TCT). Some of the SSDs will be subjected to the physical solder joint analysis called Dye and Pry (DP) and Cross Section analysis. The result will be feedbacked to the simulation team for any corrective actions required to further improve the design. Once the SSD is validated and is proven working, it will be subjected to implementation of the monitor phase whereby Design for Assembly (DFA) rules will be updated. At this stage, the design change, process and equipment parameters are in control. Predictable product reliability at early product development will enable on-time sample qualification delivery to customer and will optimize product development validation, effective development resource and will avoid forced late investment to bandage the end-of-life product failures. Understanding the critical to reliability parameters earlier will allow focus on increasing the product margin that will increase customer confidence to product reliability.
Keywords: e2e reliability prediction, SSD, TCT, Solder Joint Reliability, NUDD, connectivity issues, qualifications, characterization and control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 402102 Entropy Generation and Heat Transfer of Cu–Water Nanofluid Mixed Convection in a Cavity
Authors: Mliki Bouchmel, Belgacem Nabil, Abbassi Mohamed Ammar, Geudri Kamel, Omri Ahmed
Abstract:
In this numerical work, mixed convection and entropy generation of Cu–water nanofluid in a lid-driven square cavity have been investigated numerically using the Lattice Boltzmann Method. Horizontal walls of the cavity are adiabatic and vertical walls have constant temperature but different values. The top wall has been considered as moving from left to right at a constant speed, U0. The effects of different parameters such as nanoparticle volume concentration (0–0.05), Rayleigh number (104–106) and Reynolds numbers (1, 10 and 100) on the entropy generation, flow and temperature fields are studied. The results have shown that addition of nanoparticles to the base fluid affects the entropy generation, flow pattern and thermal behavior especially at higher Rayleigh and low Reynolds numbers. For pure fluid as well as nanofluid, the increase of Reynolds number increases the average Nusselt number and the total entropy generation, linearly. The maximum entropy generation occurs in nanofluid at low Rayleigh number and at high Reynolds number. The minimum entropy generation occurs in pure fluid at low Rayleigh and Reynolds numbers. Also at higher Reynolds number, the effect of Cu nanoparticles on enhancement of heat transfer was decreased because the effect of lid-driven cavity was increased. The present results are validated by favorable comparisons with previously published results. The results of the problem are presented in graphical and tabular forms and discussed.Keywords: Entropy generation, mixed convection, nanofluid, lattice Boltzmann method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1951101 Nature of Cities: Ontological Dimension of the Urban
Authors: Ana Cristina García-Luna Romero
Abstract:
This document seeks to reflect on the urban project from its conceptual identity root. In the first instance, a proposal is made on how the city project is sustained from the conceptual root, from the logos: it opens a way to assimilate the imagination; what we imagine becomes a reality. In this way, firstly, the need to use language as a vehicle for transmitting the stories that sustain us as humanity can be deemed as an important social factor that enables us to social behavior. Secondly, the need to attend to the written language as a mechanism of power, as a means to consolidate a dominant ideology or a political position, is raised; as it served to carry out the modernization project, it is therefore addressed differences between the real and the literate city. Thus, the consolidated urban-architectural project is based on logos, the project, and planning. Considering the importance of materiality and its relation to subjective well-being contextualized from a socio-urban approach, we question ourselves into how we can look at something that is doubtful. From a philosophy perspective, the truth is considered to be nothing more than a matter of correspondence between the observer and the observed. To understand beyond the relative of the gaze, it is necessary to expose different perspectives since it depends on the understanding of what is observed and how it is critically analyzed. Therefore, the analysis of materiality, as a political field, takes a proposal based on this research in the principles in transgenesis: principle of communication, representativeness, security, health, malleability, availability of potentiality or development, conservation, sustainability, economy, harmony, stability, accessibility, justice, legibility, significance, consistency, joint responsibility, connectivity, beauty, among others. The (urban) human being acts because he wants to live in a certain way: in a community, in a fair way, with opportunity for development, with the possibility of managing the environment according to their needs, etc. In order to comply with this principle, it is necessary to design strategies from the principles in transgenesis, which must be named, defined, understood, and socialized by the urban being, the companies, and from themselves. In this way, the technical status of the city in the neoliberal present determines extraordinary conditions for reflecting on an almost emergency scenario created by the impact of cities that, far from being limited to resilient proposals, must aim at the reflection of the urban process that the present social model has generated. Therefore, can we rethink the paradigm of the perception of life quality in the current neoliberal model in the production of the character of public space related to the practices of being urban. What we are trying to do within this document is to build a framework to study under what logic the practices of the social system that make sense of the public space are developed, what the implications of the phenomena of the inscription of action and materialization (and its results over political action between the social and the technical system) are and finally, how we can improve the quality of life of individuals from the urban space.
Keywords: Cities, nature, society, urban quality of life.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 564