Search results for: average error.
2037 Performance Analysis of MIMO-OFDM Using Convolution Codes with QAM Modulation
Authors: I Gede Puja Astawa, Yoedy Moegiharto, Ahmad Zainudin, Imam Dui Agus Salim, Nur Annisa Anggraeni
Abstract:
Performance of Orthogonal Frequency Division Multiplexing (OFDM) system can be improved by adding channel coding (error correction code) to detect and correct errors that occur during data transmission. One can use the convolution code. This paper present performance of OFDM using Space Time Block Codes (STBC) diversity technique use QAM modulation with code rate ½. The evaluation is done by analyzing the value of Bit Error Rate (BER) vs. Energy per Bit to Noise Power Spectral Density Ratio (Eb/No). This scheme is conducted 256 subcarrier transmits Rayleigh multipath channel in OFDM system. To achieve a BER of 10-3 is required 10dB SNR in SISO-OFDM scheme. For 2x2 MIMO-OFDM scheme requires 10 dB to achieve a BER of 10-3. For 4x4 MIMO-OFDM scheme requires 5 dB while adding convolution in a 4x4 MIMO-OFDM can improve performance up to 0 dB to achieve the same BER. This proves the existence of saving power by 3 dB of 4x4 MIMO-OFDM system without coding, power saving 7dB of 2x2 MIMO-OFDM and significant power savings from SISO-OFDM system
Keywords: Convolution code, OFDM, MIMO, QAM, BER.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33872036 HaskellFL: A Tool for Detecting Logical Errors in Haskell
Authors: Vanessa Vasconcelos, Mariza A. S. Bigonha
Abstract:
Understanding and using the functional paradigm is a challenge for many programmers. Looking for logical errors in code may take a lot of a developer’s time when a program grows in size. In order to facilitate both processes, this paper presents HaskellFL, a tool that uses fault localization techniques to locate a logical error in Haskell code. The Haskell subset used in this work is sufficiently expressive for those studying Functional Programming to get immediate help debugging their code and to answer questions about key concepts associated with the functional paradigm. HaskellFL was tested against Functional Programming assignments submitted by students enrolled at the Functional Programming class at the Federal University of Minas Gerais and against exercises from the Exercism Haskell track that are publicly available in GitHub. This work also evaluated the effectiveness of two fault localization techniques, Tarantula and Ochiai, in the Haskell context. Furthermore, the EXAM score was chosen to evaluate the tool’s effectiveness, and results showed that HaskellFL reduced the effort needed to locate an error for all tested scenarios. The results also showed that the Ochiai method was more effective than Tarantula.
Keywords: Debug, fault localization, functional programming, Haskell.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7312035 Appraisal of Relativistic Effects on GNSS Receiver Positioning
Authors: I. Yakubu, Y. Y. Ziggah, E. A. Gyamera
Abstract:
The Global Navigation Satellite System (GNSS) started with the launch of the United State Department of Defense Global Positioning System (GPS). GNSS systems has grown over the years to include: GLONASS (Russia); Galileo (European Union); BeiDou (China). Any GNSS architecture consists of three major segments: Space, Control and User Segments. Errors such as; multipath, ionospheric and tropospheric effects, satellite clocks, receiver noise and orbit errors (relativity effect) have significant effects on GNSS positioning. To obtain centimeter level accuracy, the impacts of the relative motion of the satellites and earth need to be taken into account. This paper discusses the relevance of the theory of relativity as a source of error for GNSS receivers for position fix based on available relevant literature. Review of relevant literature reveals that due to relativity; Time dilation, Gravitational frequency shift and Sagnac effect cause significant influence on the use of GNSS receivers for positioning by an error range of ± 2.5 m based on pseudo-range computation.
Keywords: GNSS, relativistic effects, pseudo-range, accuracy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3882034 Performance Evaluation of a Minimum Mean Square Error-Based Physical Sidelink Share Channel Receiver under Fading Channel
Authors: Yang Fu, Jaime Rodrigo Navarro, Jose F. Monserrat, Faiza Bouchmal, Oscar Carrasco Quilis
Abstract:
Cellular Vehicle to Everything (C-V2X) is considered a promising solution for future autonomous driving. From Release 16 to Release 17, the Third Generation Partnership Project (3GPP) has introduced the definitions and services for 5G New Radio (NR) V2X. Since establishing a simulator for C-V2X communications is an essential preliminary step to achieve reliable and stable communication links, this paper proposes a complete framework of a link-level simulator based on the 3GPP specifications for the Physical Sidelink Share Channel (PSSCH) of the 5G NR Physical Layer (PHY). In this framework, several algorithms in the receiver part, i.e., sliding window in channel estimation and Minimum Mean Square Error (MMSE)-based equalization, are developed. Finally, the performance of the developed PSSCH receiver is validated through extensive simulations under different assumptions.
Keywords: Yang Fu, Jaime Rodrigo Navarro, Jose F. Monserrat, Faiza Bouchmal, Oscar Carrasco Quilis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1502033 Comparison of Different k-NN Models for Speed Prediction in an Urban Traffic Network
Authors: Seyoung Kim, Jeongmin Kim, Kwang Ryel Ryu
Abstract:
A database that records average traffic speeds measured at five-minute intervals for all the links in the traffic network of a metropolitan city. While learning from this data the models that can predict future traffic speed would be beneficial for the applications such as the car navigation system, building predictive models for every link becomes a nontrivial job if the number of links in a given network is huge. An advantage of adopting k-nearest neighbor (k-NN) as predictive models is that it does not require any explicit model building. Instead, k-NN takes a long time to make a prediction because it needs to search for the k-nearest neighbors in the database at prediction time. In this paper, we investigate how much we can speed up k-NN in making traffic speed predictions by reducing the amount of data to be searched for without a significant sacrifice of prediction accuracy. The rationale behind this is that we had a better look at only the recent data because the traffic patterns not only repeat daily or weekly but also change over time. In our experiments, we build several different k-NN models employing different sets of features which are the current and past traffic speeds of the target link and the neighbor links in its up/down-stream. The performances of these models are compared by measuring the average prediction accuracy and the average time taken to make a prediction using various amounts of data.Keywords: Big data, k-NN, machine learning, traffic speed prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13762032 Investigating the Process Kinetics and Nitrogen Gas Production in Anammox Hybrid Reactor with Special Emphasis on the Role of Filter Media
Authors: Swati Tomar, Sunil Kumar Gupta
Abstract:
Anammox is a novel and promising technology that has changed the traditional concept of biological nitrogen removal. The process facilitates direct oxidation of ammonical nitrogen under anaerobic conditions with nitrite as an electron acceptor without addition of external carbon sources. The present study investigated the feasibility of Anammox Hybrid Reactor (AHR) combining the dual advantages of suspended and attached growth media for biodegradation of ammonical nitrogen in wastewater. Experimental unit consisted of 4 nos. of 5L capacity AHR inoculated with mixed seed culture containing anoxic and activated sludge (1:1). The process was established by feeding the reactors with synthetic wastewater containing NH4-H and NO2-N in the ratio 1:1 at HRT (hydraulic retention time) of 1 day. The reactors were gradually acclimated to higher ammonium concentration till it attained pseudo steady state removal at a total nitrogen concentration of 1200 mg/l. During this period, the performance of the AHR was monitored at twelve different HRTs varying from 0.25-3.0 d with increasing NLR from 0.4 to 4.8 kg N/m3d. AHR demonstrated significantly higher nitrogen removal (95.1%) at optimal HRT of 1 day. Filter media in AHR contributed an additional 27.2% ammonium removal in addition to 72% reduction in the sludge washout rate. This may be attributed to the functional mechanism of filter media which acts as a mechanical sieve and reduces the sludge washout rate many folds. This enhances the biomass retention capacity of the reactor by 25%, which is the key parameter for successful operation of high rate bioreactors. The effluent nitrate concentration, which is one of the bottlenecks of anammox process was also minimised significantly (42.3-52.3 mg/L). Process kinetics was evaluated using first order and Grau-second order models. The first-order substrate removal rate constant was found as 13.0 d-1. Model validation revealed that Grau second order model was more precise and predicted effluent nitrogen concentration with least error (1.84±10%). A new mathematical model based on mass balance was developed to predict N2 gas in AHR. The mass balance model derived from total nitrogen dictated significantly higher correlation (R2=0.986) and predicted N2 gas with least error of precision (0.12±8.49%). SEM study of biomass indicated the presence of heterogeneous population of cocci and rod shaped bacteria of average diameter varying from 1.2-1.5 mm. Owing to enhanced NRE coupled with meagre production of effluent nitrate and its ability to retain high biomass, AHR proved to be the most competitive reactor configuration for dealing with nitrogen laden wastewater.
Keywords: Anammox, filter media, kinetics, nitrogen removal.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25512031 Mathematical Analysis of Stock Prices Prediction in a Financial Market Using Geometric Brownian Motion Model
Authors: Edikan E. Akpanibah, Ogunmodimu Dupe Catherine
Abstract:
The relevance of geometric Brownian motion (GBM) in modelling the behaviour of stock market prices (SMP) cannot be over emphasized taking into consideration the volatility of the SMP. Consequently, there is need to investigate how GBM models are being estimated and used in financial market to predict SMP. To achieve this, the GBM estimation and its application to the SMP of some selected companies are studied. The normal and log-normal distributions were used to determine the expected value, variance and co-variance. Furthermore, the GBM model was used to predict the SMP of some selected companies over a period of time and the mean absolute percentage error (MAPE) were calculated and used to determine the accuracy of the GBM model in predicting the SMP of the four companies under consideration. It was observed that for all the four companies, their MAPE values were within the region of acceptance. Also, the MAPE values of our data were compared to an existing literature to test the accuracy of our prediction with respect to time of investment. Finally, some numerical simulations of the graphs of the SMP, expectations and variance of the four companies over a period of time were presented using MATLAB programming software.
Keywords: Stock Market, Geometric Brownian Motion, normal and log-normal distribution, mean absolute percentage error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2682030 An Investigation into Ozone Concentration at Urban and Rural Monitoring Stations in Malaysia
Authors: Negar Banan, Mohd Talib Latif
Abstract:
This study investigated the relationship between urban and rural ozone concentrations and quantified the extent to which ambient rural conditions and the concentrations of other pollutants can be used to predict urban ozone concentrations. The study describes the variations of ozone in weekday and weekends as well as the daily maximum recorded at selected monitoring stations. The results showed that Putrajaya station had the highest concentrations of O3 on weekend due the titration of NO during the weekday. Additionally, Jerantut had the lowest average concentration with a reading value high on Wednesdays. The comparisons of average and maximum concentrations of ozone for the three stations showed that the strongest significant correlation is recorded in Jerantut station with the value R2= 0.769. Ozone concentrations originating from a neighbouring urban site form a better predictor to the urban ozone concentrations than widespread rural ozone at some levels of temporal averaging. It is found that in urban and rural of Malaysian peninsular, the concentration of ozone depends on the concentration of NOx and seasonal meteorological factors. The HYSPLIT Model (the northeast monsoon) showed that the wind direction can also influence the concentration of ozone in the atmosphere in the studied areas.Keywords: Ozone, Hysplit model, Weekend effect, Daily Average and Daily maximum, Malaysia
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21332029 On Hyperbolic Gompertz Growth Model
Authors: Angela Unna Chukwu, Samuel Oluwafemi Oyamakin
Abstract:
We proposed a Hyperbolic Gompertz Growth Model (HGGM), which was developed by introducing a shape parameter (allometric). This was achieved by convoluting hyperbolic sine function on the intrinsic rate of growth in the classical gompertz growth equation. The resulting integral solution obtained deterministically was reprogrammed into a statistical model and used in modeling the height and diameter of Pines (Pinus caribaea). Its ability in model prediction was compared with the classical gompertz growth model, an approach which mimicked the natural variability of height/diameter increment with respect to age and therefore provides a more realistic height/diameter predictions using goodness of fit tests and model selection criteria. The Kolmogorov Smirnov test and Shapiro-Wilk test was also used to test the compliance of the error term to normality assumptions while the independence of the error term was confirmed using the runs test. The mean function of top height/Dbh over age using the two models under study predicted closely the observed values of top height/Dbh in the hyperbolic gompertz growth models better than the source model (classical gompertz growth model) while the results of R2, Adj. R2, MSE and AIC confirmed the predictive power of the Hyperbolic Gompertz growth models over its source model.Keywords: Height, Dbh, forest, Pinus caribaea, hyperbolic, gompertz.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27072028 The Relationship between Representational Conflicts, Generalization, and Encoding Requirements in an Instance Memory Network
Authors: Mathew Wakefield, Matthew Mitchell, Lisa Wise, Christopher McCarthy
Abstract:
This paper aims to provide an interpretation of artificial neural networks (ANNs) and explore some of its implications. The interpretation views ANNs as a memory which encodes instances of experience. An experiment explores the behavior of encoding and retrieval of instances from memory. A localised representation ANN is created that allows control over encoding and retrieved memory sample size and is experimented with using the MNIST digits dataset. The relationship between input familiarity, conflict within retrieved samples, and error rates is described and demonstrated to be an effective driver for memory encoding. Results indicate that selective encoding and retrieval samples that allow detection of memory conflicts produce optimal performance, and that error rates are normally distributed with input familiarity and conflict. By using input familiarity and sample consistency to guide memory encoding, the number of encoding trials on the dataset were reduced to 18.33% of the training data while maintaining good recognition performance on the test data.
Keywords: Artificial Neural Networks, ANNs, representation, memory, conflict monitoring, confidence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5072027 Model Reduction of Linear Systems by Conventional and Evolutionary Techniques
Authors: S. Panda, S. K. Tomar, R. Prasad, C. Ardil
Abstract:
Reduction of Single Input Single Output (SISO) continuous systems into Reduced Order Model (ROM), using a conventional and an evolutionary technique is presented in this paper. In the conventional technique, the mixed advantages of Mihailov stability criterion and continued fraction expansions (CFE) technique is employed where the reduced denominator polynomial is derived using Mihailov stability criterion and the numerator is obtained by matching the quotients of the Cauer second form of Continued fraction expansions. In the evolutionary technique method Particle Swarm Optimization (PSO) is employed to reduce the higher order model. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example.
Keywords: Reduced Order Modeling, Stability, Continued Fraction Expansions, Mihailov Stability Criterion, Particle Swarm Optimization, Integral Squared Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19272026 Reducing Unplanned Extubation in Psychiatric LTC
Authors: Jih-Rue Pan, Feng-Chuan Pan
Abstract:
Today-s healthcare industries had become more patient-centric than profession-centric, from which the issues of quality of healthcare and the patient safety are the major concerns in the modern healthcare facilities. An unplanned extubation (UE) may be detrimental to the patient-s life, and thus is one of the major indexes of patient safety and healthcare quality. A high UE rate not only defeated the healthcare quality as well as the patient safety policy but also the nurses- morality, and job satisfaction. The UE problem in a psychiatric hospital is unique and may be a tough challenge for the healthcare professionals for the patients were mostly lacking communication capabilities. We reported with this essay a particular project that was organized to reduce the UE rate from the current 2.3% to a lower and satisfactory level in the long-term care units of a psychiatric hospital. The project was conducted between March 1st, 2011 and August 31st, 2011. Based on the error information gathered from varied units of the hospital, the team analyzed the root causes with possible solutions proposed to the meetings. Four solutions were then concluded with consensus and launched to the units in question. The UE rate was now reduced to a level of 0.17%. Experience from this project, the procedure and the tools adopted would be good reference to other hospitals.Keywords: Unplanned extubation, patient safety, error information
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18392025 An Improved Performance of the SRM Drives Using Z-Source Inverter with the Simplified Fuzzy Logic Rule Base
Authors: M. Hari Prabhu
Abstract:
This paper is based on the performance of the Switched Reluctance Motor (SRM) drives using Z-Source Inverter with the simplified rule base of Fuzzy Logic Controller (FLC) with the output scaling factor (SF) self-tuning mechanism are proposed. The aim of this paper is to simplify the program complexity of the controller by reducing the number of fuzzy sets of the membership functions (MFs) without losing the system performance and stability via the adjustable controller gain. ZSI exhibits both voltage-buck and voltage-boost capability. It reduces line harmonics, improves reliability, and extends output voltage range. The output SF of the controller can be tuned continuously by a gain updating factor, whose value is derived from fuzzy logic, with the plant error and error change ratio as input variables. Then the results, carried out on a four-phase 6/8 pole SRM based on the dSPACEDS1104 platform, to show the feasibility and effectiveness of the devised methods and also performance of the proposed controllers will be compared with conventional counterpart.
Keywords: Fuzzy logic controller, scaling factor (SF), switched reluctance motor (SRM), variable-speed drives.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24282024 Adaptive Pulse Coupled Neural Network Parameters for Image Segmentation
Authors: Thejaswi H. Raya, Vineetha Bettaiah, Heggere S. Ranganath
Abstract:
For over a decade, the Pulse Coupled Neural Network (PCNN) based algorithms have been successfully used in image interpretation applications including image segmentation. There are several versions of the PCNN based image segmentation methods, and the segmentation accuracy of all of them is very sensitive to the values of the network parameters. Most methods treat PCNN parameters like linking coefficient and primary firing threshold as global parameters, and determine them by trial-and-error. The automatic determination of appropriate values for linking coefficient, and primary firing threshold is a challenging problem and deserves further research. This paper presents a method for obtaining global as well as local values for the linking coefficient and the primary firing threshold for neurons directly from the image statistics. Extensive simulation results show that the proposed approach achieves excellent segmentation accuracy comparable to the best accuracy obtainable by trial-and-error for a variety of images.Keywords: Automatic Selection of PCNN Parameters, Image Segmentation, Neural Networks, Pulse Coupled Neural Network
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22882023 Geometric Simplification Method of Building Energy Model Based on Building Performance Simulation
Authors: Yan Lyu, Yiqun Pan, Zhizhong Huang
Abstract:
In the design stage of a new building, the energy model of this building is often required for the analysis of the performance on energy efficiency. In practice, a certain degree of geometric simplification should be done in the establishment of building energy models, since the detailed geometric features of a real building are hard to be described perfectly in most energy simulation engine, such as ESP-r, eQuest or EnergyPlus. Actually, the detailed description is not necessary when the result with extremely high accuracy is not demanded. Therefore, this paper analyzed the relationship between the error of the simulation result from building energy models and the geometric simplification of the models. Finally, the following two parameters are selected as the indices to characterize the geometric feature of in building energy simulation: the southward projected area and total side surface area of the building. Based on the parameterization method, the simplification from an arbitrary column building to a typical shape (a cuboid) building can be made for energy modeling. The result in this study indicates that no more than 7% prediction error of annual cooling/heating load will be caused by the geometric simplification for those buildings with the ratio of southward projection length to total perimeter of the bottom of 0.25~0.35, which means this method is applicable for building performance simulation.
Keywords: building energy model, simulation, geometric simplification, design, regression
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6242022 Women's Employment Issues in Georgia and Solutions Based on European Experience
Authors: N. Damenia, E. Kharaishvili, N. Sagareishvili, M. Saghareishvili
Abstract:
Women's Employment is one of the most important issues in the global economy. The article discusses the stated topic in Georgia, through historical content, Soviet experience, and modern perspectives. The paper discusses segmentation insa terms of employment and related problems. Based on statistical analysis, women's unemployment rate and its factors are analyzed. The level of employment of women in Transcaucasia (Georgia, Armenia, and Azerbaijan) is discussed and is compared with Baltic countries (Lithuania, Latvia, and Estonia). The study analyzes women’s level of development, according to the average age of marriage and migration level. The focus is on Georgia's Association Agreement with the EU in 2014, which includes economic, social, trade and political issues. One part of it is gender equality at workplaces. According to the research, the average monthly remuneration of women managers in the financial and insurance sector equaled to 1044.6 Georgian Lari, while in overall business sector average monthly remuneration equaled to 961.1 GEL. Average salaries are increasing; however, the employment rate remains problematic. For example, in 2017, 74.6% of men and 50.8% of women were employed from a total workforce. It is also interesting that the proportion of men and women at managerial positions is 29% (women) to 71% (men). Based on the results, the main recommendation for government and civil society is to consider women as a part of the country’s economic development. In this aspect, the experience of developed countries should be considered. It is important to create additional jobs in urban or rural areas and help migrant women return and use their working resources properly.
Keywords: Employment of women, segregation in terms of employment, women's employment level in Transcaucasia, migration level.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7252021 General Regression Neural Network and Back Propagation Neural Network Modeling for Predicting Radial Overcut in EDM: A Comparative Study
Authors: Raja Das, M. K. Pradhan
Abstract:
This paper presents a comparative study between two neural network models namely General Regression Neural Network (GRNN) and Back Propagation Neural Network (BPNN) are used to estimate radial overcut produced during Electrical Discharge Machining (EDM). Four input parameters have been employed: discharge current (Ip), pulse on time (Ton), Duty fraction (Tau) and discharge voltage (V). Recently, artificial intelligence techniques, as it is emerged as an effective tool that could be used to replace time consuming procedures in various scientific or engineering applications, explicitly in prediction and estimation of the complex and nonlinear process. The both networks are trained, and the prediction results are tested with the unseen validation set of the experiment and analysed. It is found that the performance of both the networks are found to be in good agreement with average percentage error less than 11% and the correlation coefficient obtained for the validation data set for GRNN and BPNN is more than 91%. However, it is much faster to train GRNN network than a BPNN and GRNN is often more accurate than BPNN. GRNN requires more memory space to store the model, GRNN features fast learning that does not require an iterative procedure, and highly parallel structure. GRNN networks are slower than multilayer perceptron networks at classifying new cases.
Keywords: Electrical-discharge machining, General Regression Neural Network, Back-propagation Neural Network, Radial Overcut.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31152020 A Performance Analysis of Different Scheduling Schemes in WiMAX
Authors: A. Youseef
Abstract:
IEEE 802.16 (WiMAX) aims to present high speed wireless access to cover wide range coverage. The base station (BS) and the subscriber station (SS) are the main parts of WiMAX. WiMAX uses either Point-to-Multipoint (PMP) or mesh topologies. In the PMP mode, the SSs connect to the BS to gain access to the network. However, in the mesh mode, the SSs connect to each other to gain access to the BS. The main components of QoS management in the 802.16 standard are the admission control, buffer management and packet scheduling. In this paper, we use QualNet 5.0.2 to study the performance of different scheduling schemes, such as WFQ, SCFQ, RR and SP when the numbers of SSs increase. We find that when the number of SSs increases, the average jitter and average end-to-end delay is increased and the throughput is reduced.
Keywords: WiMAX, Scheduling Scheme, QoS, QualNet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20812019 A Novel SVM-Based OOK Detector in Low SNR Infrared Channels
Authors: J. P. Dubois, O. M. Abdul-Latif
Abstract:
Support Vector Machine (SVM) is a recent class of statistical classification and regression techniques playing an increasing role in applications to detection problems in various engineering problems, notably in statistical signal processing, pattern recognition, image analysis, and communication systems. In this paper, SVM is applied to an infrared (IR) binary communication system with different types of channel models including Ricean multipath fading and partially developed scattering channel with additive white Gaussian noise (AWGN) at the receiver. The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these channel stochastic models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to classical binary signal maximum likelihood detection using a matched filter driven by On-Off keying (OOK) modulation. We found that the performance of SVM is superior to that of the traditional optimal detection schemes used in statistical communication, especially for very low signal-to-noise ratio (SNR) ranges. For large SNR, the performance of the SVM is similar to that of the classical detectors. The implication of these results is that SVM can prove very beneficial to IR communication systems that notoriously suffer from low SNR at the cost of increased computational complexity.
Keywords: Least square-support vector machine, on-off keying, matched filter, maximum likelihood detector, wireless infrared communication.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19532018 Study of Effect of Gear Tooth Accuracy on Transmission Mount Vibration
Authors: Kalyan Deepak Kolla, Ketan Paua, Rajkumar Bhagate
Abstract:
Transmission dynamics occupy major role in customer perception of the product in both senses of touch and quality of sound. The quantity and quality of sound perceived is more concerned with the whine noise of the gears engaged. Whine noise is tonal in nature and tonal noises cause fatigue and irritation to customers, which in turn affect the quality of the product. Transmission error is the usual suspect for whine noise, which can be caused due to misalignments, tolerances, manufacturing variabilities. In-cabin noise is also more sensitive to the gear design. As the details of the gear tooth design and manufacturing are in microns, anything out of the tolerance zone, either in design or manufacturing, will cause a whine noise. This will also cause high variation in stress and deformation due to change in the load and leads to the fatigue failure of the gears. Hence gear design and development take priority in the transmission development process. This paper aims to study such variability by considering five pairs of helical spur gears and their effect on the transmission error, contact pattern and vibration level on the transmission.Keywords: Gears, whine noise, manufacturing variability, mount vibration variability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6462017 Structural and Electrical Properties of BNT-BT0.08 Ceramics Processed by Spark Plasma Sintering
Authors: Ciceron Berbecaru, Marin Cernea, Gheorghe Virgil Aldica, Roxana Trusca
Abstract:
(Bi0.5Na0.5)TiO3 doped with 8 mol % BaTiO3 powder (BNT-BT0.08), prepared by sol-gel method was compacted and sintered by Spark Plasma Sintering (SPS) process. The influence of SPS temperature on the densification of BNT-BT0.08 ceramic was investigated. Starting from sol-gel nanopowder of BNT-BT containing 8 mol % BaTiO3 with an average particles size of about 30 nm, were obtained ceramics with density around 98 % of the theoretical density value when the SPS temperature used was about 850 °C. The average grain size of the resulting ceramics was 80 nm. The BNT-BT0.08 ceramic sample obtained by SPS method has shown good electric properties at various frequencies.Keywords: (Bi0.5Na0.5)TiO3 doped with BaTiO3, Spark PlasmaSintering (SPS), dielectric properties
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25452016 Anaerobic Treatment of Produced Water
Authors: F. C. Khong, M. H. Isa, S. R. M. Kutty, S. A. Farhan
Abstract:
An experimental study of anaerobic treatment was performed by hybrid upflow anaerobic sludge blanket (HUASB) reactor to treat produced water (PW) of an onshore crude oil terminal (COD: 1597 mg/L, NH3-N: 14.7 mg/L, phenol: 13.8 mg/L, BOD5: 862 mg/L, sodium: 6240 mg/L and chloride 9530 mg/L). The produced water with high salinity and other toxic substances will inhibit the methanogens performance if there is no adaptation on biomass before anaerobic digestion. COD removal from produced water was investigated at five different dilutions of produced water and tap water (TW) without any nutrient addition and pre-treatment. The dilution ratios were 1PW:4TW, 2PW:3TW, 3PW:2TW, 4PW:1TW and 5PW:0TW. The reactor was evaluated at mesophilic operating condition (35 ± 2 °C) at 5 days of HRT for 250 days continuous feed. The average COD removals for 1PW:4TW, 2PW:3TW, 3PW:2TW, 4PW:1TW and 5PW:0TW were found to be approximately 76.1%, 73.8%, 70.3%, 46.3% and 61.82% respectively, with final average effluent COD of 123.7 mg/L, 240 mg/L, 294 mg/L, 589 mg/L and 738 mg/L, respectively.
Keywords: Anaerobic, fixed film, hybrid UASB, produced water, inhibitor
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24992015 Optimal Channel Equalization for MIMO Time-Varying Channels
Authors: Ehab F. Badran, Guoxiang Gu
Abstract:
We consider optimal channel equalization for MIMO (multi-input/multi-output) time-varying channels in the sense of MMSE (minimum mean-squared-error), where the observation noise can be non-stationary. We show that all ZF (zero-forcing) receivers can be parameterized in an affine form which eliminates completely the ISI (inter-symbol-interference), and optimal channel equalizers can be designed through minimization of the MSE (mean-squarederror) between the detected signals and the transmitted signals, among all ZF receivers. We demonstrate that the optimal channel equalizer is a modified Kalman filter, and show that under the AWGN (additive white Gaussian noise) assumption, the proposed optimal channel equalizer minimizes the BER (bit error rate) among all possible ZF receivers. Our results are applicable to optimal channel equalization for DWMT (discrete wavelet multitone), multirate transmultiplexers, OFDM (orthogonal frequency division multiplexing), and DS (direct sequence) CDMA (code division multiple access) wireless data communication systems. A design algorithm for optimal channel equalization is developed, and several simulation examples are worked out to illustrate the proposed design algorithm.Keywords: Channel equalization, Kalman filtering, Time-varying systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18342014 Effectiveness of a Malaysian Workplace Intervention Study on Physical Activity Levels
Authors: M. Z. Bin Mohd Ghazali, N. C. Wilson, A. F. Bin Ahmad Fuad, M. A. H. B. Musa, M. U. Mohamad Sani, F. Zulkifli, M. S. Zainal Abidin
Abstract:
Physical activity levels are low in Malaysia and this study was undertaken to determine if a four week work-based intervention program would be effective in changing physical activity levels. The study was conducted in a Malaysian Government Department and had three stages: baseline data collection, four-week intervention and two-month post intervention data collection. During the intervention and two-month post intervention phases, physical activity levels (determined by a pedometer) and basic health profiles (BMI, abdominal obesity, blood pressure) were measured. Staff (58 males, 47 females) with an average age of 33 years completed baseline data collection. Pedometer steps averaged 7,102 steps/day at baseline, although male step counts were significantly higher than females (7,861 vs. 6114). Health profiles were poor: over 50% were overweight/obese (males 66%, females 40%); hypertension (males 23%, females 6%); excess waist circumference (males 52%, females 17%). While 86 staff participated in the intervention, only 49 regularly reported their steps. There was a significant increase (17%) in average daily steps from 8,965 (week 1) to 10,436 (week 4). Unfortunately, participation in the intervention program was avoided by the less healthy staff. Two months after the intervention there was no significant difference in average steps/day, despite the fact that 89% of staff reporting they planned to make long-term changes to their lifestyle. An unexpected average increase of 2kg in body weight occurred in participants, although this was less than the 5.6kg in non-participants. A number of recommendations are made for future interventions, including the conclusion that pedometers were a useful tool and popular with participants.
Keywords: Pedometers, walking, health, intervention.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14782013 Imputing Missing Data in Electronic Health Records: A Comparison of Linear and Non-Linear Imputation Models
Authors: Alireza Vafaei Sadr, Vida Abedi, Jiang Li, Ramin Zand
Abstract:
Missing data is a common challenge in medical research and can lead to biased or incomplete results. When the data bias leaks into models, it further exacerbates health disparities; biased algorithms can lead to misclassification and reduced resource allocation and monitoring as part of prevention strategies for certain minorities and vulnerable segments of patient populations, which in turn further reduce data footprint from the same population – thus, a vicious cycle. This study compares the performance of six imputation techniques grouped into Linear and Non-Linear models, on two different real-world electronic health records (EHRs) datasets, representing 17864 patient records. The mean absolute percentage error (MAPE) and root mean squared error (RMSE) are used as performance metrics, and the results show that the Linear models outperformed the Non-Linear models in terms of both metrics. These results suggest that sometimes Linear models might be an optimal choice for imputation in laboratory variables in terms of imputation efficiency and uncertainty of predicted values.
Keywords: EHR, Machine Learning, imputation, laboratory variables, algorithmic bias.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1772012 An Efficient Burst Errors Combating for Image Transmission over Mobile WPANs
Authors: Mohsen A. M. El-Bendary, Mostafa A. R. El-Tokhy
Abstract:
This paper presents an efficient burst error spreading tool. Also, it studies a vital issue in wireless communications, which is the transmission of images over wireless networks. IEEE ZigBee 802.15.4 is a short-range communication standard that could be used for small distance multimedia transmissions. In fact, the ZigBee network is a Wireless Personal Area Network (WPAN), which needs a strong interleaving mechanism for protection against error bursts. Also, it is low power technology and utilized in the Wireless Sensor Networks (WSN) implementation. This paper presents the chaotic interleaving scheme as a data randomization tool for this purpose. This scheme depends on the chaotic Baker map. The mobility effects on the image transmission are studied with different velocity through utilizing the Jakes’ model. A comparison study between the proposed chaotic interleaving scheme and the traditional block and convolutional interleaving schemes for image transmission over a correlated fading channel is presented. The simulation results show the superiority of the proposed chaotic interleaving scheme over the traditional schemes.
Keywords: WPANs, Burst Errors, Mobility, Interleaving Techniques, Fading channels.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20332011 Design of Parity-Preserving Reversible Logic Signed Array Multipliers
Authors: Mojtaba Valinataj
Abstract:
Reversible logic as a new favorable design domain can be used for various fields especially creating quantum computers because of its speed and intangible power consumption. However, its susceptibility to a variety of environmental effects may lead to yield the incorrect results. In this paper, because of the importance of multiplication operation in various computing systems, some novel reversible logic array multipliers are proposed with error detection capability by incorporating the parity-preserving gates. The new designs are presented for two main parts of array multipliers, partial product generation and multi-operand addition, by exploiting the new arrangements of existing gates, which results in two signed parity-preserving array multipliers. The experimental results reveal that the best proposed 4×4 multiplier in this paper reaches 12%, 24%, and 26% enhancements in the number of constant inputs, number of required gates, and quantum cost, respectively, compared to previous design. Moreover, the best proposed design is generalized for n×n multipliers with general formulations to estimate the main reversible logic criteria as the functions of the multiplier size.Keywords: Array multipliers, Baugh-Wooley method, error detection, parity-preserving gates, quantum computers, reversible logic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10262010 Least Square-SVM Detector for Wireless BPSK in Multi-Environmental Noise
Authors: J. P. Dubois, Omar M. Abdul-Latif
Abstract:
Support Vector Machine (SVM) is a statistical learning tool developed to a more complex concept of structural risk minimization (SRM). In this paper, SVM is applied to signal detection in communication systems in the presence of channel noise in various environments in the form of Rayleigh fading, additive white Gaussian background noise (AWGN), and interference noise generalized as additive color Gaussian noise (ACGN). The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these advanced stochastic noise models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to conventional binary signaling optimal model-based detector driven by binary phase shift keying (BPSK) modulation. We show that the SVM performance is superior to that of conventional matched filter-, innovation filter-, and Wiener filter-driven detectors, even in the presence of random Doppler carrier deviation, especially for low SNR (signal-to-noise ratio) ranges. For large SNR, the performance of the SVM was similar to that of the classical detectors. However, the convergence between SVM and maximum likelihood detection occurred at a higher SNR as the noise environment became more hostile.Keywords: Colour noise, Doppler shift, innovation filter, least square-support vector machine, matched filter, Rayleigh fading, Wiener filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18132009 A Comparative Study of Additive and Nonparametric Regression Estimators and Variable Selection Procedures
Authors: Adriano Z. Zambom, Preethi Ravikumar
Abstract:
One of the biggest challenges in nonparametric regression is the curse of dimensionality. Additive models are known to overcome this problem by estimating only the individual additive effects of each covariate. However, if the model is misspecified, the accuracy of the estimator compared to the fully nonparametric one is unknown. In this work the efficiency of completely nonparametric regression estimators such as the Loess is compared to the estimators that assume additivity in several situations, including additive and non-additive regression scenarios. The comparison is done by computing the oracle mean square error of the estimators with regards to the true nonparametric regression function. Then, a backward elimination selection procedure based on the Akaike Information Criteria is proposed, which is computed from either the additive or the nonparametric model. Simulations show that if the additive model is misspecified, the percentage of time it fails to select important variables can be higher than that of the fully nonparametric approach. A dimension reduction step is included when nonparametric estimator cannot be computed due to the curse of dimensionality. Finally, the Boston housing dataset is analyzed using the proposed backward elimination procedure and the selected variables are identified.Keywords: Additive models, local polynomial regression, residuals, mean square error, variable selection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10112008 On Adaptive Optimization of Filter Performance Based on Markov Representation for Output Prediction Error
Authors: Hong Son Hoang, Remy Baraille
Abstract:
This paper addresses the problem of how one can improve the performance of a non-optimal filter. First the theoretical question on dynamical representation for a given time correlated random process is studied. It will be demonstrated that for a wide class of random processes, having a canonical form, there exists a dynamical system equivalent in the sense that its output has the same covariance function. It is shown that the dynamical approach is more effective for simulating and estimating a Markov and non- Markovian random processes, computationally is less demanding, especially with increasing of the dimension of simulated processes. Numerical examples and estimation problems in low dimensional systems are given to illustrate the advantages of the approach. A very useful application of the proposed approach is shown for the problem of state estimation in very high dimensional systems. Here a modified filter for data assimilation in an oceanic numerical model is presented which is proved to be very efficient due to introducing a simple Markovian structure for the output prediction error process and adaptive tuning some parameters of the Markov equation.Keywords: Statistical simulation, canonical form, dynamical system, Markov and non-Markovian processes, data assimilation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1298