Search results for: grammatical error correction
2170 Tests for Zero Inflation in Count Data with Measurement Error in Covariates
Authors: Man-Yu Wong, Siyu Zhou, Zhiqiang Cao
Abstract:
In quality of life, health service utilization is an important determinant of medical resource expenditures on Colorectal cancer (CRC) care, a better understanding of the increased utilization of health services is essential for optimizing the allocation of healthcare resources to services and thus for enhancing the service quality, especially for high expenditure on CRC care like Hong Kong region. In assessing the association between the health-related quality of life (HRQOL) and health service utilization in patients with colorectal neoplasm, count data models can be used, which account for over dispersion or extra zero counts. In our data, the HRQOL evaluation is a self-reported measure obtained from a questionnaire completed by the patients, misreports and variations in the data are inevitable. Besides, there are more zero counts from the observed number of clinical consultations (observed frequency of zero counts = 206) than those from a Poisson distribution with mean equal to 1.33 (expected frequency of zero counts = 156). This suggests that excess of zero counts may exist. Therefore, we study tests for detecting zero-inflation in models with measurement error in covariates. Method: Under classical measurement error model, the approximate likelihood function for zero-inflation Poisson regression model can be obtained, then Approximate Maximum Likelihood Estimation(AMLE) can be derived accordingly, which is consistent and asymptotically normally distributed. By calculating score function and Fisher information based on AMLE, a score test is proposed to detect zero-inflation effect in ZIP model with measurement error. The proposed test follows asymptotically standard normal distribution under H0, and it is consistent with the test proposed for zero-inflation effect when there is no measurement error. Results: Simulation results show that empirical power of our proposed test is the highest among existing tests for zero-inflation in ZIP model with measurement error. In real data analysis, with or without considering measurement error in covariates, existing tests, and our proposed test all imply H0 should be rejected with P-value less than 0.001, i.e., zero-inflation effect is very significant, ZIP model is superior to Poisson model for analyzing this data. However, if measurement error in covariates is not considered, only one covariate is significant; if measurement error in covariates is considered, only another covariate is significant. Moreover, the direction of coefficient estimations for these two covariates is different in ZIP regression model with or without considering measurement error. Conclusion: In our study, compared to Poisson model, ZIP model should be chosen when assessing the association between condition-specific HRQOL and health service utilization in patients with colorectal neoplasm. and models taking measurement error into account will result in statistically more reliable and precise information.Keywords: count data, measurement error, score test, zero inflation
Procedia PDF Downloads 2882169 The Use of Image Processing Responses Tools Applied to Analysing Bouguer Gravity Anomaly Map (Tangier-Tetuan's Area-Morocco)
Authors: Saad Bakkali
Abstract:
Image processing is a powerful tool for the enhancement of edges in images used in the interpretation of geophysical potential field data. Arial and terrestrial gravimetric surveys were carried out in the region of Tangier-Tetuan. From the observed and measured data of gravity Bouguer gravity anomalies map was prepared. This paper reports the results and interpretations of the transformed maps of Bouguer gravity anomaly of the Tangier-Tetuan area using image processing. Filtering analysis based on classical image process was applied. Operator image process like logarithmic and gamma correction are used. This paper also present the results obtained from this image processing analysis of the enhancement edges of the Bouguer gravity anomaly map of the Tangier-Tetuan zone.Keywords: bouguer, tangier, filtering, gamma correction, logarithmic enhancement edges
Procedia PDF Downloads 4222168 Forecast Based on an Empirical Probability Function with an Adjusted Error Using Propagation of Error
Authors: Oscar Javier Herrera, Manuel Angel Camacho
Abstract:
This paper addresses a cutting edge method of business demand forecasting, based on an empirical probability function when the historical behavior of the data is random. Additionally, it presents error determination based on the numerical method technique ‘propagation of errors’. The methodology was conducted characterization and process diagnostics demand planning as part of the production management, then new ways to predict its value through techniques of probability and to calculate their mistake investigated, it was tools used numerical methods. All this based on the behavior of the data. This analysis was determined considering the specific business circumstances of a company in the sector of communications, located in the city of Bogota, Colombia. In conclusion, using this application it was possible to obtain the adequate stock of the products required by the company to provide its services, helping the company reduce its service time, increase the client satisfaction rate, reduce stock which has not been in rotation for a long time, code its inventory, and plan reorder points for the replenishment of stock.Keywords: demand forecasting, empirical distribution, propagation of error, Bogota
Procedia PDF Downloads 6302167 Assessment of Intern Students' Attitudes towards Medical Errors
Authors: Nilgün Katrancı, Pınar Göv
Abstract:
With the acceleration and assessment of quality and patient safety works in healthcare services in the 21st century, activities to reduce errors have gained importance. The prevention and reduction of unintended consequences related to healthcare services and errors made during the delivery of healthcare services can be achieved by understanding the causes of the errors. Communication is the basic reason most frequently seen in such cases. Nurses who communicate with patients more closely and for longer time play a more critical role in ensuring patient safety compared to other healthcare professionals. To reduce the risk of medical errors and increase the quality of care, it is important to raise the awareness of nurses about patient safety in training period. This descriptive study was conducted between February 2017 and May 2017 to assess intern students' attitudes towards and knowledge of patient safety and medical errors. The target population of the study consists of intern students at the Faculty of Nursing in Gaziantep University (N=180). The study did not apply any sample selection method, and the research group consisted of 90 female and 37 male senior students who were available and accepted to take part in the study (N=127). The study used personal information form and medical error attitude scale to collect data. The medical error attitude scale consists of 16 items and 3 sub-dimensions. The most frequently seen medical error in the clinics the interns worked at was found as ‘Failure to comply with asepsis rules’ with a rate of 67,7%. The most frequent case among reasons for not disclosing an error is ‘noticing and correcting the error before affecting the patient’ with the rate of 70,9%. The most frequently expressed implications of disclosing a serious error for the intern students participating in the study are ‘harming patient trust (78%)’ and ‘possibility of overreaction by patient (62,2%)’. According to the results of the study, the awareness of the students about the importance of medical errors and error reporting was found high (3,48 ± 0,49). Consequently, it is important to assess and positively improve the attitudes of nurses and other healthcare professionals towards medical errors for the determination of causes of medical errors and their prevention.Keywords: healthcare service, intern student, medical error, patient safety
Procedia PDF Downloads 2032166 In-Flight Aircraft Performance Model Enhancement Using Adaptive Lookup Tables
Authors: Georges Ghazi, Magali Gelhaye, Ruxandra Botez
Abstract:
Over the years, the Flight Management System (FMS) has experienced a continuous improvement of its many features, to the point of becoming the pilot’s primary interface for flight planning operation on the airplane. With the assistance of the FMS, the concept of distance and time has been completely revolutionized, providing the crew members with the determination of the optimized route (or flight plan) from the departure airport to the arrival airport. To accomplish this function, the FMS needs an accurate Aircraft Performance Model (APM) of the aircraft. In general, APMs that equipped most modern FMSs are established before the entry into service of an individual aircraft, and results from the combination of a set of ordinary differential equations and a set of performance databases. Unfortunately, an aircraft in service is constantly exposed to dynamic loads that degrade its flight characteristics. These degradations endow two main origins: airframe deterioration (control surfaces rigging, seals missing or damaged, etc.) and engine performance degradation (fuel consumption increase for a given thrust). Thus, after several years of service, the performance databases and the APM associated to a specific aircraft are no longer representative enough of the actual aircraft performance. It is important to monitor the trend of the performance deterioration and correct the uncertainties of the aircraft model in order to improve the accuracy the flight management system predictions. The basis of this research lies in the new ability to continuously update an Aircraft Performance Model (APM) during flight using an adaptive lookup table technique. This methodology was developed and applied to the well-known Cessna Citation X business aircraft. For the purpose of this study, a level D Research Aircraft Flight Simulator (RAFS) was used as a test aircraft. According to Federal Aviation Administration the level D is the highest certification level for the flight dynamics modeling. Basically, using data available in the Flight Crew Operating Manual (FCOM), a first APM describing the variation of the engine fan speed and aircraft fuel flow w.r.t flight conditions was derived. This model was next improved using the proposed methodology. To do that, several cruise flights were performed using the RAFS. An algorithm was developed to frequently sample the aircraft sensors measurements during the flight and compare the model prediction with the actual measurements. Based on these comparisons, a correction was performed on the actual APM in order to minimize the error between the predicted data and the measured data. In this way, as the aircraft flies, the APM will be continuously enhanced, making the FMS more and more precise and the prediction of trajectories more realistic and more reliable. The results obtained are very encouraging. Indeed, using the tables initialized with the FCOM data, only a few iterations were needed to reduce the fuel flow prediction error from an average relative error of 12% to 0.3%. Similarly, the FCOM prediction regarding the engine fan speed was reduced from a maximum error deviation of 5.0% to 0.2% after only ten flights.Keywords: aircraft performance, cruise, trajectory optimization, adaptive lookup tables, Cessna Citation X
Procedia PDF Downloads 2642165 Dynamic Compensation for Environmental Temperature Variation in the Coolant Refrigeration Cycle as a Means of Increasing Machine-Tool Precision
Authors: Robbie C. Murchison, Ibrahim Küçükdemiral, Andrew Cowell
Abstract:
Thermal effects are the largest source of dimensional error in precision machining, and a major proportion is caused by ambient temperature variation. The use of coolant is a primary means of mitigating these effects, but there has been limited work on coolant temperature control. This research critically explored whether CNC-machine coolant refrigeration systems adapted to actively compensate for ambient temperature variation could increase machining accuracy. Accuracy data were collected from operators’ checklists for a CNC 5-axis mill and statistically reduced to bias and precision metrics for observations of one day over a sample period of 27 days. Temperature data were collected using three USB dataloggers in ambient air, the chiller inflow, and the chiller outflow. The accuracy and temperature data were analysed using Pearson correlation, then the thermodynamics of the system were described using system identification with MATLAB. It was found that 75% of thermal error is reflected in the hot coolant temperature but that this is negligibly dependent on ambient temperature. The effect of the coolant refrigeration process on hot coolant outflow temperature was also found to be negligible. Therefore, the evidence indicated that it would not be beneficial to adapt coolant chillers to compensate for ambient temperature variation. However, it is concluded that hot coolant outflow temperature is a robust and accessible source of thermal error data which could be used for prevention strategy evaluation or as the basis of other thermal error strategies.Keywords: CNC manufacturing, machine-tool, precision machining, thermal error
Procedia PDF Downloads 892164 Velocity Logs Error Reduction for In-Service Calibration of Vessel Performance Indicators
Authors: Maria Tsompanoglou, Dimitris Armenis
Abstract:
Vessel behavior in different operational and weather conditions constitutes the main area of interest for the ship operator. Ship speed and fuel consumption are the most decisive parameters in this respect, as their correlation provides information about the economic and environmental efficiency of the vessel, becoming the basis of decision making in terms of maintenance and trading. In the analysis of vessel operational profile for the evaluation of fuel consumption and the equivalent CO2 emissions footprint, the indications of Speed Through Water are widely used. The seasonal and regional variations in seawater characteristics, which are available nowadays, can provide the basis for accurate estimation of the errors in Speed Through Water indications at any time. Accuracy in the speed value on a route basis can enable operator identify the ship fuel and propulsion efficiency and proceed with improvements. This paper discusses case studies, where the actual vessel speed was corrected by a post-processing algorithm. The effects of the vessel correction to standard Key Performance Indicators, as well as operational findings not identified earlier, are also discussed.Keywords: data analytics, MATLAB, vessel performance monitoring, speed through water
Procedia PDF Downloads 3002163 Modelling Volatility of Cryptocurrencies: Evidence from GARCH Family of Models with Skewed Error Innovation Distributions
Authors: Timothy Kayode Samson, Adedoyin Isola Lawal
Abstract:
The past five years have shown a sharp increase in public interest in the crypto market, with its market capitalization growing from $100 billion in June 2017 to $2158.42 billion on April 5, 2022. Despite the outrageous nature of the volatility of cryptocurrencies, the use of skewed error innovation distributions in modelling the volatility behaviour of these digital currencies has not been given much research attention. Hence, this study models the volatility of 5 largest cryptocurrencies by market capitalization (Bitcoin, Ethereum, Tether, Binance coin, and USD Coin) using four variants of GARCH models (GJR-GARCH, sGARCH, EGARCH, and APARCH) estimated using three skewed error innovation distributions (skewed normal, skewed student- t and skewed generalized error innovation distributions). Daily closing prices of these currencies were obtained from Yahoo Finance website. Finding reveals that the Binance coin reported higher mean returns compared to other digital currencies, while the skewness indicates that the Binance coin, Tether, and USD coin increased more than they decreased in values within the period of study. For both Bitcoin and Ethereum, negative skewness was obtained, meaning that within the period of study, the returns of these currencies decreased more than they increased in value. Returns from these cryptocurrencies were found to be stationary but not normality distributed with evidence of the ARCH effect. The skewness parameters in all best forecasting models were all significant (p<.05), justifying of use of skewed error innovation distributions with a fatter tail than normal, Student-t, and generalized error innovation distributions. For Binance coin, EGARCH-sstd outperformed other volatility models, while for Bitcoin, Ethereum, Tether, and USD coin, the best forecasting models were EGARCH-sstd, APARCH-sstd, EGARCH-sged, and GJR-GARCH-sstd, respectively. This suggests the superiority of skewed Student t- distribution and skewed generalized error distribution over the skewed normal distribution.Keywords: skewed generalized error distribution, skewed normal distribution, skewed student t- distribution, APARCH, EGARCH, sGARCH, GJR-GARCH
Procedia PDF Downloads 1182162 Hybrid Robust Estimation via Median Filter and Wavelet Thresholding with Automatic Boundary Correction
Authors: Alsaidi M. Altaher, Mohd Tahir Ismail
Abstract:
Wavelet thresholding has been a power tool in curve estimation and data analysis. In the presence of outliers this non parametric estimator can not suppress the outliers involved. This study proposes a new two-stage combined method based on the use of the median filter as primary step before applying wavelet thresholding. After suppressing the outliers in a signal through the median filter, the classical wavelet thresholding is then applied for removing the remaining noise. We use automatic boundary corrections; using a low order polynomial model or local polynomial model as a more realistic rule to correct the bias at the boundary region; instead of using the classical assumptions such periodic or symmetric. A simulation experiment has been conducted to evaluate the numerical performance of the proposed method. Results show strong evidences that the proposed method is extremely effective in terms of correcting the boundary bias and eliminating outlier’s sensitivity.Keywords: boundary correction, median filter, simulation, wavelet thresholding
Procedia PDF Downloads 4282161 On Constructing a Cubically Convergent Numerical Method for Multiple Roots
Authors: Young Hee Geum
Abstract:
We propose the numerical method defined by xn+1 = xn − λ[f(xn − μh(xn))/]f'(xn) , n ∈ N, and determine the control parameter λ and μ to converge cubically. In addition, we derive the asymptotic error constant. Applying this proposed scheme to various test functions, numerical results show a good agreement with the theory analyzed in this paper and are proven using Mathematica with its high-precision computability.Keywords: asymptotic error constant, iterative method, multiple root, root-finding
Procedia PDF Downloads 2202160 Accentuation Moods of Blaming Utterances in Egyptian Arabic: A Pragmatic Study of Prosodic Focus
Authors: Reda A. H. Mahmoud
Abstract:
This paper investigates the pragmatic meaning of prosodic focus through four accentuation moods of blaming utterances in Egyptian Arabic. Prosodic focus results in various pragmatic meanings when the speaker utters the same blaming expression in different emotional moods: the angry, the mocking, the frustrated, and the informative moods. The main objective of this study is to interpret the meanings of these four accentuation moods in relation to their illocutionary forces and pre-locutionary effects, the integrated features of prosodic focus (e.g., tone movement distributions, pitch accents, lengthening of vowels, deaccentuation of certain syllables/words, and tempo), and the consonance between the former prosodic features and certain lexico-grammatical components to communicate the intentions of the speaker. The data on blaming utterances has been collected via elicitation and pre-recorded material, and the selection of blaming utterances is based on the criteria of lexical and prosodic regularity to be processed and verified by three computer programs, Praat, Speech Analyzer, and Spectrogram Freeware. A dual pragmatic approach is established to interpret expressive blaming utterance and their lexico-grammatical distributions into intonational focus structure units. The pragmatic component of this approach explains the variable psychological attitudes through the expressions of blaming and their effects whereas the analysis of prosodic focus structure is used to describe the intonational contours of blaming utterances and other prosodic features. The study concludes that every accentuation mood has its different prosodic configuration which influences the listener’s interpretation of the pragmatic meanings of blaming utterances.Keywords: pragmatics, pragmatic interpretation, prosody, prosodic focus
Procedia PDF Downloads 1522159 Application of a Universal Distortion Correction Method in Stereo-Based Digital Image Correlation Measurement
Authors: Hu Zhenxing, Gao Jianxin
Abstract:
Stereo-based digital image correlation (also referred to as three-dimensional (3D) digital image correlation (DIC)) is a technique for both 3D shape and surface deformation measurement of a component, which has found increasing applications in academia and industries. The accuracy of the reconstructed coordinate depends on many factors such as configuration of the setup, stereo-matching, distortion, etc. Most of these factors have been investigated in literature. For instance, the configuration of a binocular vision system determines the systematic errors. The stereo-matching errors depend on the speckle quality and the matching algorithm, which can only be controlled in a limited range. And the distortion is non-linear particularly in a complex imaging acquisition system. Thus, the distortion correction should be carefully considered. Moreover, the distortion function is difficult to formulate in a complex imaging acquisition system using conventional models in such cases where microscopes and other complex lenses are involved. The errors of the distortion correction will propagate to the reconstructed 3D coordinates. To address the problem, an accurate mapping method based on 2D B-spline functions is proposed in this study. The mapping functions are used to convert the distorted coordinates into an ideal plane without distortions. This approach is suitable for any image acquisition distortion models. It is used as a prior process to convert the distorted coordinate to an ideal position, which enables the camera to conform to the pin-hole model. A procedure of this approach is presented for stereo-based DIC. Using 3D speckle image generation, numerical simulations were carried out to compare the accuracy of both the conventional method and the proposed approach.Keywords: distortion, stereo-based digital image correlation, b-spline, 3D, 2D
Procedia PDF Downloads 4982158 Channel Estimation for LTE Downlink
Authors: Rashi Jain
Abstract:
The LTE systems employ Orthogonal Frequency Division Multiplexing (OFDM) as the multiple access technology for the Downlink channels. For enhanced performance, accurate channel estimation is required. Various algorithms such as Least Squares (LS), Minimum Mean Square Error (MMSE) and Recursive Least Squares (RLS) can be employed for the purpose. The paper proposes channel estimation algorithm based on Kalman Filter for LTE-Downlink system. Using the frequency domain pilots, the initial channel response is obtained using the LS criterion. Then Kalman Filter is employed to track the channel variations in time-domain. To suppress the noise within a symbol, threshold processing is employed. The paper draws comparison between the LS, MMSE, RLS and Kalman filter for channel estimation. The parameters for evaluation are Bit Error Rate (BER), Mean Square Error (MSE) and run-time.Keywords: LTE, channel estimation, OFDM, RLS, Kalman filter, threshold
Procedia PDF Downloads 3552157 Developing Soil Accumulation Effect Correction Factor for Solar Photovoltaic Module
Authors: Kelebaone Tsamaase, Rapelang Kemoabe, Japhet Sakala, Edward Rakgati, Ishmael Zibani
Abstract:
Increasing demand for energy, depletion of non-renewable energy, effects of climate change, the abundance of renewable energy such as solar energy have increased the interest in investing in renewable energies, in particular solar photovoltaic (PV) energy. Solar photovoltaic energy systems as part of clean technology are considered to be environmentally friendly, freely available, offer clean production systems, long term costs benefits as opposed to conventional sources, and are the attractive power source for a wide range of applications in remote areas where there is no easy access to the national grid. To get maximum electrical power, maximum solar power should penetrate the module and be converted accordingly. However, some environmental and other geographical related factors reduce the electrical power. One of them is dust which accumulates on the surface of the module and forming a dust layer and in the process obstructing the solar power from penetrating PV module. This study intends to improve the performance of solar photovoltaic (PV) energy modules by establishing soil accumulation effects correction factor from dust characteristics and properties, and also from dust accumulation and retention pattern on PV module surface. The non-urban dry deposition flux model was adapted to determine monthly and yearly dust accumulation pattern. Consideration was done on prevailing environmental and other geographical conditions. Preliminary results showed that cumulative dust settlement increased during the months of July to October leading to a higher drop in module electrical output power.Keywords: dust, electrical power output, PV module, soil correction factor
Procedia PDF Downloads 1332156 Turkish University Level EFL Learners’ Collocational Knowledge at Receptive and Productive Levels
Authors: Nazife Duygu Bagci
Abstract:
Collocations are an important part of vocabulary knowledge, and it is a subject that has recently attracted attention, while still in need of more research. The aim of this study is to answer three research questions related to the collocational knowledge of Turkish university level EFL learners at different proficiency levels of English. The first research question aims to compare the pre-intermediate (PIN) and the advanced (ADV) level learners’ collocational knowledge at receptive and productive levels. The second one is to analyze the performance of the PIN and the ADV students in two main collocation categories; lexical and grammatical. Lastly, the performance of both groups are focused on to find the collocation type (among verb-noun, adjective- noun, adjective-preposition, noun-preposition collocation types) they show the best performance in. Two offline tests were used to answer these questions. The results show that there is a significant difference between the PIN and the ADV groups at both receptive and productive levels. It can be concluded that proficiency is an important criterion in collocational knowledge, and learners do not necessarily know the collocates of the vocabulary items that they know. Although there is no significant difference between the PIN group’s performance in lexical and grammatical collocations, the ADV group showed a better performance in lexical collocations. Lastly, the PIN group at receptive and the ADV group at both receptive and productive levels showed the best performance in verb-noun collocations, which is in line with the previous research focusing on different collocation types.Keywords: collocational knowledge, EFL, language proficiency, testing
Procedia PDF Downloads 3892155 Data Gathering and Analysis for Arabic Historical Documents
Authors: Ali Dulla
Abstract:
This paper introduces a new dataset (and the methodology used to generate it) based on a wide range of historical Arabic documents containing clean data simple and homogeneous-page layouts. The experiments are implemented on printed and handwritten documents obtained respectively from some important libraries such as Qatar Digital Library, the British Library and the Library of Congress. We have gathered and commented on 150 archival document images from different locations and time periods. It is based on different documents from the 17th-19th century. The dataset comprises differing page layouts and degradations that challenge text line segmentation methods. Ground truth is produced using the Aletheia tool by PRImA and stored in an XML representation, in the PAGE (Page Analysis and Ground truth Elements) format. The dataset presented will be easily available to researchers world-wide for research into the obstacles facing various historical Arabic documents such as geometric correction of historical Arabic documents.Keywords: dataset production, ground truth production, historical documents, arbitrary warping, geometric correction
Procedia PDF Downloads 1682154 A Novel Approach to Design of EDDR Architecture for High Speed Motion Estimation Testing Applications
Authors: T. Gangadhararao, K. Krishna Kishore
Abstract:
Motion Estimation (ME) plays a critical role in a video coder, testing such a module is of priority concern. While focusing on the testing of ME in a video coding system, this work presents an error detection and data recovery (EDDR) design, based on the residue-and-quotient (RQ) code, to embed into ME for video coding testing applications. An error in processing Elements (PEs), i.e. key components of a ME, can be detected and recovered effectively by using the proposed EDDR design. The proposed EDDR design for ME testing can detect errors and recover data with an acceptable area overhead and timing penalty.Keywords: area overhead, data recovery, error detection, motion estimation, reliability, residue-and-quotient (RQ) code
Procedia PDF Downloads 4312153 Uncertainty Quantification of Corrosion Anomaly Length of Oil and Gas Steel Pipelines Based on Inline Inspection and Field Data
Authors: Tammeen Siraj, Wenxing Zhou, Terry Huang, Mohammad Al-Amin
Abstract:
The high resolution inline inspection (ILI) tool is used extensively in the pipeline industry to identify, locate, and measure metal-loss corrosion anomalies on buried oil and gas steel pipelines. Corrosion anomalies may occur singly (i.e. individual anomalies) or as clusters (i.e. a colony of corrosion anomalies). Although the ILI technology has advanced immensely, there are measurement errors associated with the sizes of corrosion anomalies reported by ILI tools due limitations of the tools and associated sizing algorithms, and detection threshold of the tools (i.e. the minimum detectable feature dimension). Quantifying the measurement error in the ILI data is crucial for corrosion management and developing maintenance strategies that satisfy the safety and economic constraints. Studies on the measurement error associated with the length of the corrosion anomalies (in the longitudinal direction of the pipeline) has been scarcely reported in the literature and will be investigated in the present study. Limitations in the ILI tool and clustering process can sometimes cause clustering error, which is defined as the error introduced during the clustering process by including or excluding a single or group of anomalies in or from a cluster. Clustering error has been found to be one of the biggest contributory factors for relatively high uncertainties associated with ILI reported anomaly length. As such, this study focuses on developing a consistent and comprehensive framework to quantify the measurement errors in the ILI-reported anomaly length by comparing the ILI data and corresponding field measurements for individual and clustered corrosion anomalies. The analysis carried out in this study is based on the ILI and field measurement data for a set of anomalies collected from two segments of a buried natural gas pipeline currently in service in Alberta, Canada. Data analyses showed that the measurement error associated with the ILI-reported length of the anomalies without clustering error, denoted as Type I anomalies is markedly less than that for anomalies with clustering error, denoted as Type II anomalies. A methodology employing data mining techniques is further proposed to classify the Type I and Type II anomalies based on the ILI-reported corrosion anomaly information.Keywords: clustered corrosion anomaly, corrosion anomaly assessment, corrosion anomaly length, individual corrosion anomaly, metal-loss corrosion, oil and gas steel pipeline
Procedia PDF Downloads 3092152 Reproducibility of Dopamine Transporter Density Measured with I-123-N-ω-Fluoropropyl-2β-Carbomethoxy-3β-(4-Iodophenyl)Nortropane SPECT in Phantom Studies and Parkinson’s Disease Patients
Authors: Yasuyuki Takahashi, Genta Hoshi, Kyoko Saito
Abstract:
Objectives: The objective of this study was to evaluate the reproducibility of I-123-N-ω-fluoropropyl-2β-carbomethoxy-3β-(4- iodophenyl) nortropane (I-123 FP-CIT) SPECT by using specific binding ratio (SBR) in phantom studies and Parkinson’s Disease (PD) patients. Methods: We made striatum phantom originally and confirmed reproducibility. The phantom studies changed head position and accumulation of FP-CIT, each. And image processing confirms influence on SBR by 30 cases. 30 PD received a SPECT for 3 hours post injection of I-123 FP-CIT 167MBq. Results: SBR decreased in rotatory direction by the patient position by the phantom studies. And, SBR improved the influence after the attenuation and the scatter correction in the cases (y=0.99x+0.57 r2=0.83). However, Stage II recognized dispersion in SBR by low accumulation. Conclusion: Than the phantom studies that assumed the normal cases, the SPECT image after the attenuation and scatter correction had better reproducibility.Keywords: 123I-FP-CIT, specific binding ratio, Parkinson’s disease
Procedia PDF Downloads 4292151 A Study Investigating Word Association Behaviour in People with Acquired Language and Communication Disorders
Authors: Angela Maria Fenu
Abstract:
The aim of this study was to better characterize the nature of word association responses in people with aphasia. The participants selected for the experimental group were 4 individuals with mild Broca’s aphasia. The control group consisted of 51 cognitively intact age- and gender-matched individuals. The participants were asked to perform a word association task in which they had to say the first word they thought of when hearing each cue. The cue words (n= 16) were the translation in Italian of the set of English cue words of a published study. The participants from the experimental group were administered the word association test every two weeks for a period of two months when they received speech-language therapy A combination of analytical approaches to measure the data was used. To analyse different patterns of word association responses in both groups, the nature of the relationship between the cue and the response was examined: responses were divided into five categories of association. To investigate the similarity between aphasic and non-aphasic subjects, the stereotypy of responses was examined.While certain stimulus words (nouns, adjectives) elicited responses from Broca’s aphasics that tended to resemble those made by non-aphasic subjects; others (adverbs, verbs) showed the tendency to elicit responses different from the ones given by normal subjects. This suggests that some mechanisms underlying certain types of associations are degraded in aphasics individuals, while others display little evidence of disruption. The high number of paradigmatic associations given in response to a noun or an adjective might imply that the mechanisms, largely semantic, underlying paradigmatic associations are relatively preserved in Broca’s aphasia, but it might also mean that some words are more easily processed depending on their grammatical class (nouns, adjectives). The most significant variation was noticed when the grammatical class of the cue word was an adverb. Unlike the normal individuals, the experimental subjects gave the most idiosyncratic associations, which are often produced when the attempt to give a paradigmatic response fails. In turn, the failure to retrieve paradigmatic responses when the cue is an adverb might suggest that Broca’s aphasics are more sensitive to this grammatical class.The findings from this study suggest that, from research on word associations in people with aphasia, important data can arise concerning the specific lexical retrieval impairments that characterize the different types of aphasia and the various treatments that might positively influence the kinds of word association responses affected by language disruption.Keywords: aphasia therapy, clinical linguistics, word-association behaviour, mental lexicon
Procedia PDF Downloads 882150 Sectoral Energy Consumption in South Africa and Its Implication for Economic Growth
Authors: Kehinde Damilola Ilesanmi, Dev Datt Tewari
Abstract:
South Africa is in its post-industrial era moving from the primary and secondary sector to the tertiary sector. The study investigated the impact of the disaggregated energy consumption (coal, oil, and electricity) on the primary, secondary and tertiary sectors of the economy between 1980 and 2012 in South Africa. Using vector error correction model, it was established that South Africa is an energy dependent economy, and that energy (especially electricity and oil) is a limiting factor of growth. This implies that implementation of energy conservation policies may hamper economic growth. Output growth is significantly outpacing energy supply, which has necessitated load shedding. To meet up the excess energy demand, there is a need to increase the generating capacity which will necessitate increased investment in the electricity sector as well as strategic steps to increase oil production. There is also need to explore more renewable energy sources, in order to meet the growing energy demand without compromising growth and environmental sustainability. Policy makers should also pursue energy efficiency policies especially at sectoral level of the economy.Keywords: causality, economic growth, energy consumption, hypothesis, sectoral output
Procedia PDF Downloads 4702149 DIAL Measurements of Vertical Distribution of Ozone at the Siberian Lidar Station in Tomsk
Authors: Oleg A. Romanovskii, Vladimir D. Burlakov, Sergey I. Dolgii, Olga V. Kharchenko, Alexey A. Nevzorov, Alexey V. Nevzorov
Abstract:
The paper presents the results of DIAL measurements of the vertical ozone distribution. The ozone lidar operate as part of the measurement complex at Siberian Lidar Station (SLS) of V.E. Zuev Institute of Atmospheric Optics SB RAS, Tomsk (56.5ºN; 85.0ºE) and designed for study of the vertical ozone distribution in the upper troposphere–lower stratosphere. Most suitable wavelengths for measurements of ozone profiles are selected. We present an algorithm for retrieval of vertical distribution of ozone with temperature and aerosol correction during DIAL lidar sounding of the atmosphere. The temperature correction of ozone absorption coefficients is introduced in the software to reduce the retrieval errors. Results of lidar measurement at wavelengths of 299 and 341 nm agree with model estimates, which point to acceptable accuracy of ozone sounding in the 6–18 km altitude range.Keywords: lidar, ozone distribution, atmosphere, DIAL
Procedia PDF Downloads 4972148 Macroeconomic Impact of Economic Growth on Unemployment: A Case of South Africa
Authors: Ashika Govender
Abstract:
This study seeks to determine whether Okun’s Law is valid for the South African economy, using time series data for the period 2004 to 2014. The data were accessed from the South African Reserve Bank and Stats SA. The stationarity of the variables was analysed by applying unit root tests via the Augmented Dickey-Fuller test (ADF), the Phillips-Perron (PP) test, and the Kwiatkowski–Phillips–Schmidt–Shin test (KPSS) test. The study used an ordinary least square (OLS) model in analysing the dynamic version of Okun’s law. The Error Correction Model (ECM) was used to analyse the short-run impact of GDP growth on unemployment, as well as the speed of adjustment. The results indicate a short run and long run relationship between unemployment rate and GDP growth rate in period 2004q1-2014q4, suggesting that Okun’s law is valid for the South African economy. With a 1 percent increase in GDP, unemployment can decrease by 0.13 percent, ceteris paribus. The research culminates in important policy recommendations, highlighting the relationship between unemployment and economic growth in the spirit of the National Development Plan.Keywords: unemployment, economic growth, Okun's law, South Africa
Procedia PDF Downloads 2722147 Knowledge Required for Avoiding Lexical Errors at Machine Translation
Authors: Yukiko Sasaki Alam
Abstract:
This research aims at finding out the causes that led to wrong lexical selections in machine translation (MT) rather than categorizing lexical errors, which has been a main practice in error analysis. By manually examining and analyzing lexical errors outputted by a MT system, it suggests what knowledge would help the system reduce lexical errors.Keywords: machine translation, error analysis, lexical errors, evaluation
Procedia PDF Downloads 3372146 Ant Lion Optimization in a Fuzzy System for Benchmark Control Problem
Authors: Leticia Cervantes, Edith Garcia, Oscar Castillo
Abstract:
At today, there are several control problems where the main objective is to obtain the best control in the study to decrease the error in the application. Many techniques can use to control these problems such as Neural Networks, PID control, Fuzzy Logic, Optimization techniques and many more. In this case, fuzzy logic with fuzzy system and an optimization technique are used to control the case of study. In this case, Ant Lion Optimization is used to optimize a fuzzy system to control the velocity of a simple treadmill. The main objective is to achieve the control of the velocity in the control problem using the ALO optimization. First, a simple fuzzy system was used to control the velocity of the treadmill it has two inputs (error and error change) and one output (desired speed), then results were obtained but to decrease the error the ALO optimization was developed to optimize the fuzzy system of the treadmill. Having the optimization, the simulation was performed, and results can prove that using the ALO optimization the control of the velocity was better than a conventional fuzzy system. This paper describes some basic concepts to help to understand the idea in this work, the methodology of the investigation (control problem, fuzzy system design, optimization), the results are presented and the optimization is used for the fuzzy system. A comparison between the simple fuzzy system and the optimized fuzzy systems are presented where it can be proving the optimization improved the control with good results the major findings of the study is that ALO optimization is a good alternative to improve the control because it helped to decrease the error in control applications even using any control technique to optimized, As a final statement is important to mentioned that the selected methodology was good because the control of the treadmill was improve using the optimization technique.Keywords: ant lion optimization, control problem, fuzzy control, fuzzy system
Procedia PDF Downloads 3992145 Development of Quasi Real-Time Comprehensive System for Earthquake Disaster
Authors: Zhi Liu, Hui Jiang, Jin Li, Kunhao Chen, Langfang Zhang
Abstract:
Fast acquisition of the seismic information and accurate assessment of the earthquake disaster is the key problem for emergency rescue after a destructive earthquake. In order to meet the requirements of the earthquake emergency response and rescue for the cities and counties, a quasi real-time comprehensive evaluation system for earthquake disaster is developed. Based on monitoring data of Micro-Electro-Mechanical Systems (MEMS) strong motion network, structure database of a county area and the real-time disaster information by the mobile terminal after an earthquake, fragility analysis method and dynamic correction algorithm are synthetically obtained in the developed system. Real-time evaluation of the seismic disaster in the county region is finally realized to provide scientific basis for seismic emergency command, rescue and assistant decision.Keywords: quasi real-time, earthquake disaster data collection, MEMS accelerometer, dynamic correction, comprehensive evaluation
Procedia PDF Downloads 2132144 Random Walks and Option Pricing for European and American Options
Authors: Guillaume Leduc
Abstract:
In this paper, we describe a broad setting under which the error of the approximation can be quantified, controlled, and for which convergence occurs at a speed of n⁻¹ for European and American options. We describe how knowledge of the error allows for arbitrarily fast acceleration of the convergence.Keywords: random walk approximation, European and American options, rate of convergence, option pricing
Procedia PDF Downloads 4632143 Cointegration Dynamics in Asian Stock Markets: Implications for Long-Term Portfolio Management
Authors: Xinyi Xu
Abstract:
This study conducts a detailed examination of Asian stock markets over the period from 2008 to 2023, with a focus on the dynamics of cointegration and their relevance for long-term investment strategies. Specifically, we assess the co-movement and potential for pairs trading—a strategy where investors take opposing positions on two stocks, indices, or financial instruments that historically move together. For example, we explore the relationship between the Nikkei 225 (N225), Japan’s benchmark stock index, and the Straits Times Index (STI) of Singapore, as well as the relationship between the Korea Composite Stock Price Index (KS11) and the STI. The methodology includes tests for normality, stationarity, cointegration, and the application of Vector Error Correction Modeling (VECM). Our findings reveal significant long-term relationships between these pairs, indicating opportunities for pairs trading strategies. Furthermore, the research underscores the challenges posed by model instability and the influence of major global incidents, which are identified as structural breaks. These findings pave the way for further exploration into the intricacies of financial market dynamics.Keywords: normality tests, stationarity, cointegration, VECM, pairs trading
Procedia PDF Downloads 562142 Lookup Table Reduction and Its Error Analysis of Hall Sensor-Based Rotation Angle Measurement
Authors: Young-San Shin, Seongsoo Lee
Abstract:
Hall sensor is widely used to measure rotation angle. When the Hall voltage is measured for linear displacement, it is converted to angular displacement using arctangent function, which requires a large lookup table. In this paper, a lookup table reduction technique is presented for angle measurement. When the input of the lookup table is small within a certain threshold, the change of the outputs with respect to the change of the inputs is relatively small. Thus, several inputs can share same output, which significantly reduce the lookup table size. Its error analysis was also performed, and the threshold was determined so as to maintain the error less than 1°. When the Hall voltage has 11-bit resolution, the lookup table size is reduced from 1,024 samples to 279 samples.Keywords: hall sensor, angle measurement, lookup table, arctangent
Procedia PDF Downloads 3362141 The Effectiveness of Orthogonal Frequency Division Multiplexing as Modulation Technique
Authors: Mohamed O. Babana
Abstract:
In wireless channel multipath is the propagation phenomena where the transmitted signal arrive at the receiver side with many of paths, the signal at these paths arrive with different time delay the results is random signal fading due to intersymbols interference(ISI). This paper deals with identification of orthogonal frequency division multiplexing (OFDM) technology, and how it is used to overcome intersymbol interference due to multipath. Also investigates the effect of Additive White Gaussian Noise Channel (AWGN) on OFDM using multi-level modulation of Phase Shift Keying (PSK), computer simulation to calculate the bit error rate (BER) under AWGN channel is applied. A comparison study is carried out to obtain the Bit Error Rate performance for OFDM to identify the best multi-level modulation of PSK.Keywords: intersymbol interference(ISI), bit error rate(BER), modulation, multiplexing, simulation
Procedia PDF Downloads 422