Search results for: ARDL bounds testing
743 Development of the Maturity Sensor Prototype and Method of Its Placement in the Structure
Authors: Ye. B. Utepov, A. S. Tulebekova, A. B. Kazkeyev
Abstract:
Maturity sensors are used to determine concrete strength by the non-destructive method. The method of placement of the maturity sensors determines their number required for a certain frame of a monolithic building. This paper proposes a cheap prototype of an embedded wireless sensor for monitoring concrete structures, as well as an alternative strategy for placing sensors based on the transitional boundaries of the temperature distribution of concrete curing, which were determined by building a heat map of the temperature distribution, where unknown values are calculated by the method of inverse distance weighing. The developed prototype can simultaneously measure temperature and relative humidity over a smartphone-controlled time interval. It implements a maturity method to assess the in-situ strength of concrete, which is considered an alternative to the traditional shock impulse and compression testing method used in Kazakhstan. The prototype was tested in laboratory and field conditions. The tests were aimed at studying the effect of internal and external temperature and relative humidity on concrete's strength gain. Based on an experimentally poured concrete slab with randomly integrated maturity sensors, it the transition boundaries form elliptical forms were determined. Temperature distribution over the largest diameter of the ellipses was plotted, resulting in correct and inverted parabolas. As a result, the distance between the closest opposite crossing points of the parabolas is accepted as the maximum permissible step for setting the maturity sensors. The proposed placement strategy can be applied to sensors that measure various continuous phenomena such as relative humidity. Prototype testing has also revealed Bluetooth inconvenience due to weak signal and inability to access multiple prototypes simultaneously. For this reason, further prototype upgrades are planned in the future work.
Keywords: Heat map, placement strategy, temperature and relative humidity, wireless embedded sensor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 365742 Battery/Supercapacitor Emulator for Chargers Functionality Testing
Authors: S. Farag, A. Kupeman
Abstract:
In this paper, design of solid-state battery/supercapacitor emulator based on dc-dc boost converter is described. The emulator mimics charging behavior of any storage device based on a predefined behavior set by the user. The device is operated by a two-level control structure: high-level emulating controller and low- level input voltage controller. Simulation and experimental results are shown to demonstrate the emulator operation.
Keywords: Battery, Charger, Energy, Storage, Supercapacitor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2860741 Comparative Evaluation of Ice Adhesion Behavior
Authors: T. Strobl, D. Raps, M. Hornung
Abstract:
In this study, the adhesion of ice to solid substrates with different surface properties is compared. Clear ice, similar to atmospheric in-flight icing encounters, is accreted on the different substrates under controlled conditions. The ice adhesion behavior is investigated by means of a dynamic vibration testing technique with an electromagnetic shaker initiating ice de-bonding in the interface between the substrate and the ice. The results of the experiments reveal that the affinity for ice accretion is significantly influenced by the water contact angle of the respective sample.Keywords: Contact angle, dynamic vibration measurement, ice adhesion, interfacial shear stress.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2346740 Reduction of Emissions of Nitrogen Oxides from Traffic
Authors: Frantisek Bozek, Jiri Dvorak, Jaromir Mares, Hana Malachova
Abstract:
The value of emission factor was calculated in the older type of Diesel engine operating on an engine testing bench and then compared with the parameters monitored under similar conditions when the EnviroxTM additive was applied. It has been found out that the additive based on CeO2 nanoparticles reduces emission of NOx. The dependencies of NOx emissions on reduced torque, engine power and revolutions have been observed as well.Keywords: Additive, air, cerium dioxide, emission factor, emissions, nanoparticles, nitrogen oxides
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1720739 Optimizing Usability Testing with Collaborative Method in an E-Commerce Ecosystem
Authors: Markandeya Kunchi
Abstract:
Usability testing (UT) is one of the vital steps in the User-centred design (UCD) process when designing a product. In an e-commerce ecosystem, UT becomes primary as new products, features, and services are launched very frequently. And, there are losses attached to the company if an unusable and inefficient product is put out to market and is rejected by customers. This paper tries to answer why UT is important in the product life-cycle of an E-commerce ecosystem. Secondary user research was conducted to find out work patterns, development methods, type of stakeholders, and technology constraints, etc. of a typical E-commerce company. Qualitative user interviews were conducted with product managers and designers to find out the structure, project planning, product management method and role of the design team in a mid-level company. The paper tries to address the usual apprehensions of the company to inculcate UT within the team. As well, it stresses upon factors like monetary resources, lack of usability expert, narrow timelines, and lack of understanding of higher management as some primary reasons. Outsourcing UT to vendors is also very prevalent with mid-level e-commerce companies, but it has its own severe repercussions like very little team involvement, huge cost, misinterpretation of the findings, elongated timelines, and lack of empathy towards the customer, etc. The shortfalls of the unavailability of a UT process in place within the team and conducting UT through vendors are bad user experiences for customers while interacting with the product, badly designed products which are neither useful and nor utilitarian. As a result, companies see dipping conversions rates in apps and websites, huge bounce rates and increased uninstall rates. Thus, there was a need for a more lean UT system in place which could solve all these issues for the company. This paper highlights on optimizing the UT process with a collaborative method. The degree of optimization and structure of collaborative method is the highlight of this paper. Collaborative method of UT is one in which the centralised design team of the company takes for conducting and analysing the UT. The UT is usually a formative kind where designers take findings into account and uses in the ideation process. The success of collaborative method of UT is due to its ability to sync with the product management method employed by the company or team. The collaborative methods focus on engaging various teams (design, marketing, product, administration, IT, etc.) each with its own defined roles and responsibility in conducting a smooth UT with users In-house. The paper finally highlights the positive results of collaborative UT method after conducting more than 100 In-lab interviews with users across the different lines of businesses. Some of which are the improvement of interaction between stakeholders and the design team, empathy towards users, improved design iteration, better sanity check of design solutions, optimization of time and money, effective and efficient design solution. The future scope of collaborative UT is to make this method leaner, by reducing the number of days to complete the entire project starting from planning between teams to publishing the UT report.
Keywords: Usability testing, collaborative method, e-commerce, product management method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 668738 Determination of Electromagnetic Properties of Human Tissues
Authors: Iliana Marinova, Valentin Mateev
Abstract:
In this paper a computer system for electromagnetic properties measurements is designed. The system employs Agilent 4294A precision impedance analyzer to measure the amplitude and the phase of a signal applied over a tested biological tissue sample. Measured by the developed computer system data could be used for tissue characterization in wide frequency range from 40Hz to 110MHz. The computer system can interface with output devices acquiring flexible testing process.Keywords: Electromagnetic properties, human tissue, bioimpedance, measurement system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2429737 Functionality of Negotiation Agent on Value-based Design Decision
Authors: Arazi Idrus, Christiono Utomo
Abstract:
This paper presents functionality of negotiation agent on value-based design decision. The functionality is based on the characteristics of the system and goal specification. A Prometheus Design Tool model was used for developing the system. Group functionality will be the attribute for negotiation agents, which comprises a coordinator agent and decision- maker agent. The results of the testing of the system to a building system selection on valuebased decision environment are also presented.Keywords: Functionality, negotiation agent, value-baseddecision
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1420736 Contextual SenSe Model: Word Sense Disambiguation Using Sense and Sense Value of Context Surrounding the Target
Authors: Vishal Raj, Noorhan Abbas
Abstract:
Ambiguity in NLP (Natural Language Processing) refers to the ability of a word, phrase, sentence, or text to have multiple meanings. This results in various kinds of ambiguities such as lexical, syntactic, semantic, anaphoric and referential. This study is focused mainly on solving the issue of Lexical ambiguity. Word Sense Disambiguation (WSD) is an NLP technique that aims to resolve lexical ambiguity by determining the correct meaning of a word within a given context. Most WSD solutions rely on words for training and testing, but we have used lemma and Part of Speech (POS) tokens of words for training and testing. Lemma adds generality and POS adds properties of word into token. We have designed a method to create an affinity matrix to calculate the affinity between any pair of lemma_POS (a token where lemma and POS of word are joined by underscore) of given training set. Additionally, we have devised an algorithm to create the sense clusters of tokens using affinity matrix under hierarchy of POS of lemma. Furthermore, three different mechanisms to predict the sense of target word using the affinity/similarity value are devised. Each contextual token contributes to the sense of target word with some value and whichever sense gets higher value becomes the sense of target word. So, contextual tokens play a key role in creating sense clusters and predicting the sense of target word, hence, the model is named Contextual SenSe Model (CSM). CSM exhibits a noteworthy simplicity and explication lucidity in contrast to contemporary deep learning models characterized by intricacy, time-intensive processes, and challenging explication. CSM is trained on SemCor training data and evaluated on SemEval test dataset. The results indicate that despite the naivety of the method, it achieves promising results when compared to the Most Frequent Sense (MFS) model.
Keywords: Word Sense Disambiguation, WSD, Contextual SenSe Model, Most Frequent Sense, part of speech, POS, Natural Language Processing, NLP, OOV, out of vocabulary, ELMo, Embeddings from Language Model, BERT, Bidirectional Encoder Representations from Transformers, Word2Vec, lemma_POS, Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 389735 Bootstrap and MLS Methods-based Individual Bioequivalence Assessment
Authors: Kongsheng Zhang, Li Ge
Abstract:
It is a one-sided hypothesis testing process for assessing bioequivalence. Bootstrap and modified large-sample(MLS) methods are considered to study individual bioequivalence(IBE), type I error and power of hypothesis tests are simulated and compared with FDA(2001). The results show that modified large-sample method is equivalent to the method of FDA(2001) .
Keywords: Individual bioequivalence, bootstrap, Bayesian bootstrap, modified large-sample.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1584734 Complexity of Component-based Development of Embedded Systems
Authors: M. Zheng, V. S. Alagar
Abstract:
The paper discusses complexity of component-based development (CBD) of embedded systems. Although CBD has its merits, it must be augmented with methods to control the complexities that arise due to resource constraints, timeliness, and run-time deployment of components in embedded system development. Software component specification, system-level testing, and run-time reliability measurement are some ways to control the complexity.Keywords: Components, embedded systems, complexity, softwaredevelopment, traffic controller system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1499733 Air Handling Units Power Consumption Using Generalized Additive Model for Anomaly Detection: A Case Study in a Singapore Campus
Authors: Ju Peng Poh, Jun Yu Charles Lee, Jonathan Chew Hoe Khoo
Abstract:
The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.
Keywords: Anomaly detection, digital twin, Generalised Additive Model, Power Consumption Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 502732 Improved Computational Efficiency of Machine Learning Algorithms Based on Evaluation Metrics to Control the Spread of Coronavirus in the UK
Authors: Swathi Ganesan, Nalinda Somasiri, Rebecca Jeyavadhanam, Gayathri Karthick
Abstract:
The COVID-19 crisis presents a substantial and critical hazard to worldwide health. Since the occurrence of the disease in late January 2020 in the UK, the number of infected people confirmed to acquire the illness has increased tremendously across the country, and the number of individuals affected is undoubtedly considerably high. The purpose of this research is to figure out a predictive machine learning (ML) archetypal that could forecast the COVID-19 cases within the UK. This study concentrates on the statistical data collected from 31st January 2020 to 31st March 2021 in the United Kingdom. Information on total COVID-19 cases registered, new cases encountered on a daily basis, total death registered, and patients’ death per day due to Coronavirus is collected from World Health Organization (WHO). Data preprocessing is carried out to identify any missing values, outliers, or anomalies in the dataset. The data are split into 8:2 ratio for training and testing purposes to forecast future new COVID-19 cases. Support Vector Machine (SVM), Random Forest (RF), and linear regression (LR) algorithms are chosen to study the model performance in the prediction of new COVID-19 cases. From the evaluation metrics such as r-squared value and mean squared error, the statistical performance of the model in predicting the new COVID-19 cases is evaluated. RF outperformed the other two ML algorithms with a training accuracy of 99.47% and testing accuracy of 98.26% when n = 30. The mean square error obtained for RF is 4.05e11, which is lesser compared to the other predictive models used for this study. From the experimental analysis, RF algorithm can perform more effectively and efficiently in predicting the new COVID-19 cases, which could help the health sector to take relevant control measures for the spread of the virus.
Keywords: COVID-19, machine learning, supervised learning, unsupervised learning, linear regression, support vector machine, random forest.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 172731 Induction of alpha-Amylase in Wheat Grain Cultivars as an Indicator of Resistance to Pre-harvest Sprouting
Authors: Aidar A. Khakimzhanov, Vladimir A. Kuzovlev, Nurgul S. Mamytova, Dinara A. Shansharova, Oleg V. Fursov
Abstract:
The influence of humidity and low temperature on the α- amylase activity and isoenzyme composition of grains of different wheat varieties have been studied. The identified samples of varieties have significant difference in the level of enzyme induction under the impact of high humidity and low temperature. It is proposed to use this methodological approach for testing genotypes and wheat breeding lines for resistance to pre-harvest sprouting (PHS).
Keywords: α-Amylase, isoenzymes, wheat, pre-harvest sprouting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1852730 Proposal of Solidification/Stabilisation Process of Chosen Hazardous Waste by Cementation
Authors: Bozena Dohnalkova
Abstract:
This paper presents a part of the project solving which is dedicated to the identification of the hazardous waste with the most critical production within the Czech Republic with the aim to study and find the optimal composition of the cement matrix that will ensure maximum content disposal of chosen hazardous waste. In the first stage of project solving – which represents this paper – a specific hazardous waste was chosen, its properties were identified and suitable solidification agents were chosen. Consequently solidification formulas and testing methodology was proposed.
Keywords: Cementation, solidification, waste, binder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1869729 Pilot Scale Production and Compatibility Criteria of New Self-Cleaning Materials
Authors: J. Ranogajec, O. Rudic, S. Pasalic, S. Vucetic, D. Cjepa
Abstract:
The paper involves a chain of activities from synthesis, establishment of the methodology for characterization and testing of novel protective materials through the pilot production and application on model supports. It summarizes the results regarding the development of the pilot production protocol for newly developed self-cleaning materials. The optimization of the production parameters was completed in order to improve the most important functional properties (mineralogy characteristics, particle size, self-cleaning properties and photocatalytic activity) of the newly designed nanocomposite material.
Keywords: Cultural heritage. Materials compatibility. Pilot production. Self-cleaning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2318728 Sensitivity Analysis of Real-Time Systems
Authors: Benjamin Gorry, Andrew Ireland, Peter King
Abstract:
Verification of real-time software systems can be expensive in terms of time and resources. Testing is the main method of proving correctness but has been shown to be a long and time consuming process. Everyday engineers are usually unwilling to adopt formal approaches to correctness because of the overhead associated with developing their knowledge of such techniques. Performance modelling techniques allow systems to be evaluated with respect to timing constraints. This paper describes PARTES, a framework which guides the extraction of performance models from programs written in an annotated subset of C.Keywords: Performance Modelling, Real-time, SensitivityAnalysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513727 Some Factors Affecting the Compressive Behaviour of Structural Masonry at Small Scales
Authors: A. Mohammed, T. G. Hughes
Abstract:
This paper presents part of a research into the small scale modelling of masonry. Small scale testing of masonry has been carried out by many authors, but few have attempted a systematic determination of the parameters that affect masonry at a small scale. The effect of increasing mortar strength and different sand gradings under compression were investigated. The results show masonry strength at small scale is influenced by increasing mortar strength and different sand gradings.Keywords: Compression, masonry, models, mortar, sand gradings
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2545726 A Failure Analysis Tool for HDD Analysis
Authors: C. Kumjeera, T. Unchim, B. Marungsri, A. Oonsivilai
Abstract:
The study of piezoelectric material in the past was in T-Domain form; however, no one has studied piezoelectric material in the S-Domain form. This paper will present the piezoelectric material in the transfer function or S-Domain model. S-Domain is a well known mathematical model, used for analyzing the stability of the material and determining the stability limits. By using S-Domain in testing stability of piezoelectric material, it will provide a new tool for the scientific world to study this material in various forms.
Keywords: Hard disk drive, failure analysis, tool, time
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2751725 Replacement of Power Transformers basis on Diagnostic Results and Load Forecasting
Authors: G. Gavrilovs, O. Borscevskis
Abstract:
This paper describes interconnection between technical and economical making decision. The reason of this dealing could be different: poor technical condition, change of substation (electrical network) regime, power transformer owner budget deficit and increasing of tariff on electricity. Establishing of recommended practice as well as to give general advice and guidance in economical sector, testing, diagnostic power transformers to establish its conditions, identify problems and provide potential remedies.Keywords: Diagnostic results, load forecasting, power supplysystem, replacement of power transformer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2066724 The Benefits of Regional Brand for Companies
Authors: H. Starzyczna, M. Stoklasa, K. Matusinska
Abstract:
This article deals with the benefits of regional brands for companies in the Czech Republic. Research was focused on finding out the expected and actual benefits of regional brands for companies. The data were obtained by questionnaire survey and analysed by IBM SPSS. Representative sample of 204 companies was created. The research analysis disclosed the expected benefits that the regional brand should bring to companies. But the actual benefits are much worse. The statistical testing of hypotheses revealed that the benefits depend on the region of origin, which surprised both us and the regional coordinators.
Keywords: Brand, regional brands, product protective branding programs, brand benefits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1460723 The Influence of Social Network Websites on Level of user Satisfaction
Authors: Pedram Behyar, Maryam Heidari, Zahra Bayat
Abstract:
the purpose of this research is to identify and clarify factors which have positive effect among user satisfaction and their social networking through websites. The examined factors in this research are; innovation, ease of use, trustworthy and customer support which are defined as satisfaction factors. To obtain reliable research approaches and to have better result in this research four hypothesizes used to test. This hypothesis testing has been done by correlation, regression and test of normality by using “SPSS16" also the data which was analyzed by this software. this data was gathered from prepaid questionnaire.Keywords: Customer Satisfaction, Social Network Website
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1858722 Sensory Acceptability of Novel Sorrel/Roselle (Hibiscus sabdariffa L.)
Authors: Tamara Anderson, Neela Badrie
Abstract:
Consumers are demanding novel beverages that are healthier, convenient and have appealing consumer acceptance. The objectives of this study were to investigate the effects of adding grape polyphenols and the influence of presenting health claims on the sensory acceptability of wines. Fresh red sorrel calyces were fermented into wines. The total soluble solids of the pectinase-treated sorrel puree were from 4°Brix to 23.8°Brix. Polyphenol in the form of grape pomace extract was added to sorrel wines (w/v) in specified levels to give 0. 25. 50 and 75 ppm. A focus group comprising of 12 panelists was use to select the level of polyphenol to be added to sorrel wines for sensory preference The sensory attributed of the wines which were evaluated were colour, clarity, aroma, flavor, mouth-feel, sweetness, astringency and overall preference. The sorrel wine which was most preferred from focus group evaluation was presented for hedonic rating. In the first stage of hedonic testing, the sorrel wine was served chilled at 7°C for 24 h prior to sensory evaluation. Each panelist was provided with a questionnaire and was asked to rate the wines on colour, aroma, flavor, mouth-feel, sweetness, astringency and overall acceptability using a 9-point hedonic scale. In the second stage of hedonic testing, the panelist were instructed to read a health abstract on the health benefits of polyphenolic compounds and again to rate sorrel wine with added 25 ppm polyphenol. Paired t-test was used for the analysis of the influence of presenting health information on polyphenols on hedonic scoring of sorrel wines. Focus groups found that the addition of polyphenol addition had no significant effect on sensory color and aroma but affected clarity and flavor. A 25 ppm wine was liked moderately in overall acceptability. The presentation of information on the health benefit of polyphenols in sorrel wines to panelists had no significant influence on the sensory acceptance of wine. More than half of panelists would drink this wine now and then. This wine had color L 19.86±0.68, chroma 2.10±0.12, hue° 16.90 ±3.10 and alcohol content of 13.0%. The sorrel wine was liked moderately in overall acceptability with the added polyphenols.Keywords: Sorrel wines, Roselle Hibiscus sabdariffa L, novel wine, polyphenols, health benefits, physicochemical properties.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2353721 A Study on the Relation of Corporate Governance and Pricing for Initial Public Offerings
Authors: Chei-Chang Chiou, Sen-Wei Wang, Yu-Min Wang
Abstract:
The purpose of this study is to investigate the relationship between corporate governance and pricing for initial public offerings (IPOs). Empirical result finds that the prediction of pricing of IPOs with corporate governance added can have a rather higher degree of predicting accuracy than that of non governance added during the training and testing samples. Therefore, it can be observed that corporate governance mechanism can affect the pricing of IPOsKeywords: Artificial neural networks, corporate governance, initial public offerings.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1807720 Peak-to-Average Power Ratio Reduction in OFDM Systems using Huffman Coding
Authors: Ashraf A. Eltholth, Adel R. Mikhail, A. Elshirbini, Moawad I. Moawad, A. I. Abdelfattah
Abstract:
In this paper we proposed the use of Huffman coding to reduce the PAR of an OFDM system as a distortionless scrambling technique, and we utilize the amount saved in the total bit rate by the Huffman coding to send the encoding table for accurate decoding at the receiver without reducing the effective throughput. We found that the use of Huffman coding reduces the PAR by about 6 dB. Also we have investigated the effect of PAR reduction due to Huffman coding through testing the spectral spreading and the inband distortion due to HPA with different IBO values. We found a complete match of our expectation from the proposed solution with the obtained simulation results.Keywords: HPA, Huffman coding, OFDM, PAR
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2599719 Transfer Function of Piezoelectric Material
Authors: C. Worakitjaroenphon, A. Oonsivilai
Abstract:
The study of piezoelectric material in the past was in T-Domain form; however, no one has studied piezoelectric material in the S-Domain form. This paper will present the piezoelectric material in the transfer function or S-Domain model. S-Domain is a well known mathematical model, used for analyzing the stability of the material and determining the stability limits. By using S-Domain in testing stability of piezoelectric material, it will provide a new tool for the scientific world to study this material in various forms.Keywords: Piezoelectric, Stability, S-Domain, Transfer function
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3866718 Evaluation of Tension Capacity of Pile (Case Study in Sandy Soil)
Authors: Shooshpasha I., Kiakojoori M., Mirzagoltabar R. A.
Abstract:
High building constructions are increasing in south beaches of the Caspian Sea because of tourist attractions and limitation of residential areas. According to saturated alluvial fields transfer of load from high structures to the soil by piles is inevitable. In spite of most of these piles are under compression forces, tension piles are used in special conditions. Few studies have been conducted because of the limited use of these piles. Tension capacity of openended pipe piles in full scale was tested in this study. The length of the bored piles was 420 up to 480 cm and all were in 120 cm diameter. The results of testing 7 piles were compared with the results of relations given by researches.Keywords: piles, tension capacity, sand, shaft friction
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6982717 Off-Line Signature Recognition Based On Angle Features and GRNN Neural Networks
Authors: Laila Y. Fannas, Ahmed Y. Ben Sasi
Abstract:
This research presents a handwritten signature recognition based on angle feature vector using Artificial Neural Network (ANN). Each signature image will be represented by an Angle vector. The feature vector will constitute the input to the ANN. The collection of signature images will be divided into two sets. One set will be used for training the ANN in a supervised fashion. The other set which is never seen by the ANN will be used for testing. After training, the ANN will be tested for recognition of the signature. When the signature is classified correctly, it is considered correct recognition otherwise it is a failure.
Keywords: Signature Recognition, Artificial Neural Network, Angle Features.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2496716 An Intelligent Text Independent Speaker Identification Using VQ-GMM Model Based Multiple Classifier System
Authors: Cheima Ben Soltane, Ittansa Yonas Kelbesa
Abstract:
Speaker Identification (SI) is the task of establishing identity of an individual based on his/her voice characteristics. The SI task is typically achieved by two-stage signal processing: training and testing. The training process calculates speaker specific feature parameters from the speech and generates speaker models accordingly. In the testing phase, speech samples from unknown speakers are compared with the models and classified. Even though performance of speaker identification systems has improved due to recent advances in speech processing techniques, there is still need of improvement. In this paper, a Closed-Set Tex-Independent Speaker Identification System (CISI) based on a Multiple Classifier System (MCS) is proposed, using Mel Frequency Cepstrum Coefficient (MFCC) as feature extraction and suitable combination of vector quantization (VQ) and Gaussian Mixture Model (GMM) together with Expectation Maximization algorithm (EM) for speaker modeling. The use of Voice Activity Detector (VAD) with a hybrid approach based on Short Time Energy (STE) and Statistical Modeling of Background Noise in the pre-processing step of the feature extraction yields a better and more robust automatic speaker identification system. Also investigation of Linde-Buzo-Gray (LBG) clustering algorithm for initialization of GMM, for estimating the underlying parameters, in the EM step improved the convergence rate and systems performance. It also uses relative index as confidence measures in case of contradiction in identification process by GMM and VQ as well. Simulation results carried out on voxforge.org speech database using MATLAB highlight the efficacy of the proposed method compared to earlier work.Keywords: Feature Extraction, Speaker Modeling, Feature Matching, Mel Frequency Cepstrum Coefficient (MFCC), Gaussian mixture model (GMM), Vector Quantization (VQ), Linde-Buzo-Gray (LBG), Expectation Maximization (EM), pre-processing, Voice Activity Detection (VAD), Short Time Energy (STE), Background Noise Statistical Modeling, Closed-Set Tex-Independent Speaker Identification System (CISI).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1889715 Speaker Identification by Atomic Decomposition of Learned Features Using Computational Auditory Scene Analysis Principals in Noisy Environments
Authors: Thomas Bryan, Veton Kepuska, Ivica Kostanic
Abstract:
Speaker recognition is performed in high Additive White Gaussian Noise (AWGN) environments using principals of Computational Auditory Scene Analysis (CASA). CASA methods often classify sounds from images in the time-frequency (T-F) plane using spectrograms or cochleargrams as the image. In this paper atomic decomposition implemented by matching pursuit performs a transform from time series speech signals to the T-F plane. The atomic decomposition creates a sparsely populated T-F vector in “weight space” where each populated T-F position contains an amplitude weight. The weight space vector along with the atomic dictionary represents a denoised, compressed version of the original signal. The arraignment or of the atomic indices in the T-F vector are used for classification. Unsupervised feature learning implemented by a sparse autoencoder learns a single dictionary of basis features from a collection of envelope samples from all speakers. The approach is demonstrated using pairs of speakers from the TIMIT data set. Pairs of speakers are selected randomly from a single district. Each speak has 10 sentences. Two are used for training and 8 for testing. Atomic index probabilities are created for each training sentence and also for each test sentence. Classification is performed by finding the lowest Euclidean distance between then probabilities from the training sentences and the test sentences. Training is done at a 30dB Signal-to-Noise Ratio (SNR). Testing is performed at SNR’s of 0 dB, 5 dB, 10 dB and 30dB. The algorithm has a baseline classification accuracy of ~93% averaged over 10 pairs of speakers from the TIMIT data set. The baseline accuracy is attributable to short sequences of training and test data as well as the overall simplicity of the classification algorithm. The accuracy is not affected by AWGN and produces ~93% accuracy at 0dB SNR.
Keywords: Time-frequency plane, atomic decomposition, envelope sampling, Gabor atoms, matching pursuit, sparse dictionary learning, sparse autoencoder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1570714 A Classification Scheme for Game Input and Output
Authors: P. Prema, B. Ramadoss
Abstract:
Computer game industry has experienced exponential growth in recent years. A game is a recreational activity involving one or more players. Game input is information such as data, commands, etc., which is passed to the game system at run time from an external source. Conversely, game outputs are information which are generated by the game system and passed to an external target, but which is not used internally by the game. This paper identifies a new classification scheme for game input and output, which is based on player-s input and output. Using this, relationship table for game input classifier and output classifier is developed.Keywords: Game Classification, Game Input, Game Output, Game Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1983