Search results for: measuring precision
245 Readiness of Intellectual Capital Measurement: A Review of the Property Development and Investment Industry
Authors: Edward C. W. Chan, Benny C. F. Cheung
Abstract:
In the knowledge economy, the financial indicator is not the unique instrument to gauge the performance of a company. The role of intellectual capital contributing to the company performance is increasing. To measure the company performance due to intellectual capital, the value-added intellectual capital (VAIC) model is adopted to measure the intellectual capital utilization efficiency of the subject companies. The purpose of this study is to review the readiness of measuring intellectual capital for the Hong Kong listed companies in the property development and property investment industry by using VAIC model. This study covers the financial reports from the representative Hong Kong listed property development companies and property investment companies in the period 2014-2019. The findings from this study indicated the industry is ready for IC measurement employing VAIC framework but not yet ready for using the extended VAIC model.
Keywords: Intellectual capital, intellectual capital measurement, property development, property investment, Skandia Navigator, VAIC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 558244 Do C-Test and Cloze Procedure Measure what they Purport to be Measuring? A Case of Criterion-Related Validity
Authors: Masoud Saeedi, Mansour Tavakoli, Shirin Rahimi Kazerooni, Vahid Parvaresh
Abstract:
This article investigated the validity of C-test and Cloze test which purport to measure general English proficiency. To provide empirical evidence pertaining to the validity of the interpretations based on the results of these integrative language tests, their criterion-related validity was investigated. In doing so, the test of English as a foreign language (TOEFL) which is an established, standardized, and internationally administered test of general English proficiency was used as the criterion measure. Some 90 Iranian English majors participated in this study. They were seniors studying English at a university in Tehran, Iran. The results of analyses showed that there is a statistically significant correlation among participants- scores on Cloze test, C-test, and the TOEFL. Building on the findings of the study and considering criterion-related validity as the evidential basis of the validity argument, it was cautiously deducted that these tests measure the same underlying trait. However, considering the limitations of using criterion measures to validate tests, no absolute claims can be made as to the construct validity of these integrative tests.
Keywords: Integrative testing, C-test, Cloze test, theTOEFL, Validity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3329243 X-Ray Intensity Measurement Using Frequency Output Sensor for Computed Tomography
Authors: R. M. Siddiqui, D. Z. Moghaddam, T. R. Turlapati, S. H. Khan, I. Ul Ahad
Abstract:
Quality of 2D and 3D cross-sectional images produce by Computed Tomography primarily depend upon the degree of precision of primary and secondary X-Ray intensity detection. Traditional method of primary intensity detection is apt to errors. Recently the X-Ray intensity measurement system along with smart X-Ray sensors is developed by our group which is able to detect primary X-Ray intensity unerringly. In this study a new smart X-Ray sensor is developed using Light-to-Frequency converter TSL230 from Texas Instruments which has numerous advantages in terms of noiseless data acquisition and transmission. TSL230 construction is based on a silicon photodiode which converts incoming X-Ray radiation into the proportional current signal. A current to frequency converter is attached to this photodiode on a single monolithic CMOS integrated circuit which provides proportional frequency count to incoming current signal in the form of the pulse train. The frequency count is delivered to the center of PICDEM FS USB board with PIC18F4550 microcontroller mounted on it. With highly compact electronic hardware, this Demo Board efficiently read the smart sensor output data. The frequency output approaches overcome nonlinear behavior of sensors with analog output thus un-attenuated X-Ray intensities could be measured precisely and better normalization could be acquired in order to attain high resolution.Keywords: Computed tomography, detector technology, X-Ray intensity measurement
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2609242 A Short Form of the Taiwan Health Literacy Scale (THLS) for Chinese-Speaking Adults
Authors: Frank C. Pan
Abstract:
The Taiwan Health Literacy Scale (THLS) was developed to cope with the need of measuring heath literacy of Chinese-speaking adults in Taiwan. Although the scale was proven having good reliability and validity, it was not popularly adopted by the practitioners due to the length, and the time required completing. Based on the THLS, this research further invited healthcare professionals to review the original scale for a possible shorten work. Under the logic of THLS, the research adopted an analytic hierarchy process technique to consolidate the healthcare experts- assessments to shorten the original scale. There are fifteen items out of the original 66 items were identified having higher loadings. Confirmed by the experts and passed a pilot test with 40 undergraduate students, a short form of THLS is then introduced. This research then used 839 samples from the major cities of the Hua-lien county in the eastern part of Taiwan to test the reliability and validity of this new scale. The reliability of the scale is high and acceptable. The current scale is also highly correlated with the original, of which provide evidence for the validity of the scale.
Keywords: Health literacy, THLS, health education, STHLS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1919241 Temperature Dependence of Relative Permittivity: A Measurement Technique Using Split Ring Resonators
Authors: Sreedevi P. Chakyar, Jolly Andrews, V. P. Joseph
Abstract:
A compact method for measuring the relative permittivity of a dielectric material at different temperatures using a single circular Split Ring Resonator (SRR) metamaterial unit working as a test probe is presented in this paper. The dielectric constant of a material is dependent upon its temperature and the LC resonance of the SRR depends on its dielectric environment. Hence, the temperature of the dielectric material in contact with the resonator influences its resonant frequency. A single SRR placed between transmitting and receiving probes connected to a Vector Network Analyser (VNA) is used as a test probe. The dependence of temperature between 30 oC and 60 oC on resonant frequency of SRR is analysed. Relative permittivities ‘ε’ of test samples for different temperatures are extracted from a calibration graph drawn between the relative permittivity of samples of known dielectric constant and their corresponding resonant frequencies. This method is found to be an easy and efficient technique for analysing the temperature dependent permittivity of different materials.
Keywords: Metamaterials, negative permeability, permittivity measurement techniques, split ring resonators, temperature dependent dielectric constant.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2585240 An Algorithm Proposed for FIR Filter Coefficients Representation
Authors: Mohamed Al Mahdi Eshtawie, Masuri Bin Othman
Abstract:
Finite impulse response (FIR) filters have the advantage of linear phase, guaranteed stability, fewer finite precision errors, and efficient implementation. In contrast, they have a major disadvantage of high order need (more coefficients) than IIR counterpart with comparable performance. The high order demand imposes more hardware requirements, arithmetic operations, area usage, and power consumption when designing and fabricating the filter. Therefore, minimizing or reducing these parameters, is a major goal or target in digital filter design task. This paper presents an algorithm proposed for modifying values and the number of non-zero coefficients used to represent the FIR digital pulse shaping filter response. With this algorithm, the FIR filter frequency and phase response can be represented with a minimum number of non-zero coefficients. Therefore, reducing the arithmetic complexity needed to get the filter output. Consequently, the system characteristic i.e. power consumption, area usage, and processing time are also reduced. The proposed algorithm is more powerful when integrated with multiplierless algorithms such as distributed arithmetic (DA) in designing high order digital FIR filters. Here the DA usage eliminates the need for multipliers when implementing the multiply and accumulate unit (MAC) and the proposed algorithm will reduce the number of adders and addition operations needed through the minimization of the non-zero values coefficients to get the filter output.
Keywords: Pulse shaping Filter, Distributed Arithmetic, Optimization algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3175239 Fabrication, Testing and Machinability Evaluation of Glass Fiber Reinforced Epoxy Composites
Authors: S. S. Panda, Arkesh Chouhan, Yogesh Deshpande
Abstract:
The present paper deals with designing and fabricating an apparatus for the speedy and accurate manufacturing of fiber reinforced composite lamina of different orientation, thickness and stacking sequences for testing. Properties derived through an analytical approach are verified through measuring the elastic modulus, ultimate tensile strength, flexural modulus and flexural strength of the samples. The 00 orientation ply looks stiffer compared to the 900 ply. Similarly, the flexural strength of 00 ply is higher than to the 900 ply. Sample machinability has been studied by conducting numbers of drilling based on Taguchi Design experiments. Multi Responses (Delamination and Damage grading) is obtained using the desirability approach and optimum cutting condition (spindle speed, feed and drill diameter), at which responses are minimized is obtained thereafter. Delamination increases nonlinearly with the increase in spindle speed. Similarly, the influence of the drill diameter on delamination is higher than the spindle speed and feed rate.
Keywords: Delamination, FRP composite, multi response optimization, Taguchi design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1233238 Affective Adaptation Design for Better Gaming Experiences
Authors: Ollie Hall, Salma ElSayed
Abstract:
Affective adaptation is a creative way for game designers to add an extra layer of engagement to their productions. When player’s emotions are an explicit factor in mechanics design, endless possibilities for imaginative gameplay emerge. Whilst gaining popularity, existing affective game research mostly runs controlled experiments in restrictive settings and rely on one or more specialist devices for measuring player’s emotional state. These conditions albeit effective, are not necessarily realistic. Moreover, the simplified narrative and intrusive wearables may not be suitable for players. This exploratory study investigates delivering an immersive affective experience in the wild with minimal requirements, in an attempt for the average developer to reach the average player. A puzzle game is created with rich narrative and creative mechanics. It employs both explicit and implicit adaptation and only requires a web camera. Participants played the game on their own machines in various settings. Whilst it was rated feasible, very engaging and enjoyable, it remains questionable whether a fully immersive experience was delivered due to the limited sample size.
Keywords: affective games, dynamic adaptation, emotion recognition, game design
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 846237 Recycling of Aggregates from Construction Demolition Wastes in Concrete: Study of Physical and Mechanical Properties
Authors: M. Saidi, F. Ait-Medjber, B. Safi, M. Samar
Abstract:
This work is focused on the study of valuation of recycled concrete aggregates, by measuring certain properties of concrete in the fresh and hardened state. In this study, rheological tests and physic-mechanical characterization on concretes and mortars were conducted with recycled concrete whose geometric properties were identified aggregates. Mortars were elaborated with recycled fine aggregate (0/5mm) and concretes were manufactured using recycled coarse aggregates (5/12.5 mm and 12.5/20 mm). First, a study of the mortars was conducted to determine the effectiveness of polycarboxylate superplasticizer on the workability of these and their action deflocculating of the recycled sand. The rheological behavior of mortars based on fine aggregate recycled was characterized. The results confirm that the mortars composed of different fractions of recycled sand (0 /5) have a better mechanical properties (compressive and flexural strength) compared to normal mortar. Also, the mechanical strengths of concretes made with recycled aggregates (5/12.5 mm and 12.5/20 mm), are comparable to those of conventional concrete with conventional aggregates, provided that the implementation can be improved by the addition of a superplasticizer.
Keywords: Demolition wastes, recycled coarse aggregate, concrete, workability, mechanical strength, porosity/water absorption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3389236 Performance Evaluation of Conventional and Wiper Carbide Tools When Turning 6060 Aluminium Alloy: Analysis of Surface Roughness
Authors: Salah Gariani, Taher Dao, Khaled Jegandi
Abstract:
Wiper inserts are widely used nowadays, particularly in turning and milling operations, due to their unique geometric characteristics that generate superb surface finish and improve productivity. Wiper inserts can produce double the feed rate while preserving comparable surface roughness compared to that produced by conventional cutting tools. This paper reports an experimental investigation of surface quality generated in the precision dry turning of 6060 Aluminium alloy using conventional and wiper inserts at different cutting conditions. The Taguchi L9 array, Analysis of Means (AOM) and variance (ANOVA) were employed in the development of the experimental design and to optimise the process parameter identified: average surface roughness (Ra). The experimental results show that the wiper inserts substantially improved the surface quality of the machined samples by a factor of two compared to those for the conventional insert under all cutting conditions. The ANOVA and AOM analysis showed that the type of insert is the most significant factor affecting surface roughness, with a Percentage Contribution Ratio (PCR) value of 67.41%. Feed rate also significantly affected surface roughness but contributed less to its variation. No significant difference was found between values of Ra using wiper inserts under dry and wet cooling modes when turning 6060 Aluminium alloy.
Keywords: 6060 Aluminium alloy, conventional and wiper carbide tools, dry turning, average surface roughness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 323235 Measuring the Relationship between Customers- Satisfaction and Cognitions: A Case of Janfusun Fancyworld in Taiwan
Authors: Wan-Yu Liu, Yen-Hsiang Liu, Shing-Yi Huang, Hao-Zhi Wen
Abstract:
The private theme parks are gradually surpassing public-owned scenic areas after many years of development and have become a mainstream choice for domestic tourists. Previous studies show that visitors from different backgrounds differ in consumer behavior and satisfaction factors. An understanding of visitor satisfaction is therefore of extreme importance to operators of privately-owned theme parks. Importance-Performance Analysis (IPA) is used to measure consumer's potential satisfaction with services and has become a widely used management tool for strength and weakness analysis for brands, products, services and point of sales. As IPA has so far not been used to evaluate the visitor satisfaction with privately-owned theme parks, in this study the IPA method is used to analyze visitor satisfaction with Janfusun Fancyworld (one of the most popular private theme parks in Taiwan) and to rank visitor focus and satisfaction on/in theme park facilities and services. Results of the analysis provide private theme park operators with an understanding of user or consumer demands as well as an assessment of the quality of services currently offered.Keywords: Satisfaction, Importance-Performance Analysis, Theme Parks, Janfusun Fancyworld.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2448234 Decoy-pulse Protocol for Frequency-coded Quantum Key Distribution
Authors: Sudeshna Bhattacharya, Pratyush Pandey, Pradeep Kumar K
Abstract:
We propose a decoy-pulse protocol for frequency-coded implementation of B92 quantum key distribution protocol. A direct extension of decoy-pulse method to frequency-coding scheme results in security loss as an eavesdropper can distinguish between signal and decoy pulses by measuring the carrier photon number without affecting other statistics. We overcome this problem by optimizing the ratio of carrier photon number of decoy-to-signal pulse to be as close to unity as possible. In our method the switching between signal and decoy pulses is achieved by changing the amplitude of RF signal as opposed to modulating the intensity of optical signal thus reducing system cost. We find an improvement by a factor of 100 approximately in the key generation rate using decoy-state protocol. We also study the effect of source fluctuation on key rate. Our simulation results show a key generation rate of 1.5×10-4/pulse for link lengths up to 70km. Finally, we discuss the optimum value of average photon number of signal pulse for a given key rate while also optimizing the carrier ratio.
Keywords: B92, decoy-pulse, frequency-coding, quantum key distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1727233 Urban Big Data: An Experimental Approach to Building-Value Estimation Using Web-Based Data
Authors: Sun-Young Jang, Sung-Ah Kim, Dongyoun Shin
Abstract:
Current real-estate value estimation, difficult for laymen, usually is performed by specialists. This paper presents an automated estimation process based on big data and machine-learning technology that calculates influences of building conditions on real-estate price measurement. The present study analyzed actual building sales sample data for Nonhyeon-dong, Gangnam-gu, Seoul, Korea, measuring the major influencing factors among the various building conditions. Further to that analysis, a prediction model was established and applied using RapidMiner Studio, a graphical user interface (GUI)-based tool for derivation of machine-learning prototypes. The prediction model is formulated by reference to previous examples. When new examples are applied, it analyses and predicts accordingly. The analysis process discerns the crucial factors effecting price increases by calculation of weighted values. The model was verified, and its accuracy determined, by comparing its predicted values with actual price increases.Keywords: Big data, building-value analysis, machine learning, price prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1164232 Attribute Weighted Class Complexity: A New Metric for Measuring Cognitive Complexity of OO Systems
Authors: Dr. L. Arockiam, A. Aloysius
Abstract:
In general, class complexity is measured based on any one of these factors such as Line of Codes (LOC), Functional points (FP), Number of Methods (NOM), Number of Attributes (NOA) and so on. There are several new techniques, methods and metrics with the different factors that are to be developed by the researchers for calculating the complexity of the class in Object Oriented (OO) software. Earlier, Arockiam et.al has proposed a new complexity measure namely Extended Weighted Class Complexity (EWCC) which is an extension of Weighted Class Complexity which is proposed by Mishra et.al. EWCC is the sum of cognitive weights of attributes and methods of the class and that of the classes derived. In EWCC, a cognitive weight of each attribute is considered to be 1. The main problem in EWCC metric is that, every attribute holds the same value but in general, cognitive load in understanding the different types of attributes cannot be the same. So here, we are proposing a new metric namely Attribute Weighted Class Complexity (AWCC). In AWCC, the cognitive weights have to be assigned for the attributes which are derived from the effort needed to understand their data types. The proposed metric has been proved to be a better measure of complexity of class with attributes through the case studies and experimentsKeywords: Software Complexity, Attribute Weighted Class Complexity, Weighted Class Complexity, Data Type
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121231 An Activity Based Trajectory Search Approach
Authors: Mohamed Mahmoud Hasan, Hoda M. O. Mokhtar
Abstract:
With the gigantic increment in portable applications use and the spread of positioning and location-aware technologies that we are seeing today, new procedures and methodologies for location-based strategies are required. Location recommendation is one of the highly demanded location-aware applications uniquely with the wide accessibility of social network applications that are location-aware including Facebook check-ins, Foursquare, and others. In this paper, we aim to present a new methodology for location recommendation. The proposed approach coordinates customary spatial traits alongside other essential components including shortest distance, and user interests. We also present another idea namely, "activity trajectory" that represents trajectory that fulfills the set of activities that the user is intrigued to do. The approach dispatched acquaints the related distance value to select trajectory(ies) with minimum cost value (distance) and spatial-area to prune unneeded directions. The proposed calculation utilizes the idea of movement direction to prescribe most comparable N-trajectory(ies) that matches the client's required action design with least voyaging separation. To upgrade the execution of the proposed approach, parallel handling is applied through the employment of a MapReduce based approach. Experiments taking into account genuine information sets were built up and tested for assessing the proposed approach. The exhibited tests indicate how the proposed approach beets different strategies giving better precision and run time.
Keywords: Location-based recommendation, map-reduce, recommendation system, trajectory search.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 980230 Measuring the Effect of Intercollegiate Athletic Success on Private Giving and Enrollment
Authors: Jamie L. Stangel
Abstract:
Increased popularity and visibility of college athletics has contributed to an environment in which institutions—most of which lack self-sufficient athletics department budgets—reallocate monies from the university general fund and seek additional funding sources to keep up with increasing levels of spending on athletics. Given the prevalence of debates on student debt levels, coach salaries, and athlete pay, empirical evidence on whether this spending yields expected return on investment is necessary. This study considered the relationship between the independent variable of winning percentage of the men’s basketball team at a mid-major university, moderated by National Collegiate Athletic Association (NCAA) tournament appearance, and number of applicants, number of enrollments, average SAT score of students, and donor giving to the university general and athletic funds. The results indicate that, other than a small correlation between athletic success and number of applicants, only when NCAA tournament appearance is used as a moderating variable, these purported benefits are not supported, suggesting the need for a reevaluation of athletic department spending and perceptions on tangible and intangible benefits for universities.
Keywords: Athletic success, enrollment, NCAA, private giving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 369229 Improving Production Capacity through Efficient PPC System: Lesson from Leather Manufacturing
Authors: Mengist Hailemariam, Silma Yoseph
Abstract:
A well designed and executed Production Planning and Control (PPC) system is one of the key levers for superior performance in the current manufacturing set-up. Hence, measuring the PPC system performance has become a necessity for long term success. The present study examined PPC related issues which impact the production capacity and productivity of leather companies with special focus on Kombolcha Tannery Share Company (KTSC), Ethiopia. Physical observation, interview, and questionnaire were used to generate necessary information from the respondents and reach valid conclusions. Company annual reports were referred and analyzed to triangulate primary data. Consequently, the study revealed that KTSC runs below its capacity due to its inefficient PPC system being in use for which the root causes were identified. The study thereby conceptualizes a PPC system improvement framework comprising three pillars viz., management culture, internal capability and performance measurement together with key considerations in each case. The study findings enable the company to recognize the importance of efficient PPC system as a source of competitive advantage. It also aid managers in evaluating various PPC execution schemes to enhance productivity.
Keywords: Ethiopia, Leather manufacturing, Production planning and control, PPC improvement framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3734228 Stability Analysis and Controller Design of Further Development of MIMOS II for Space Applications with Focus on the Extended Lyapunov Method: Part I
Authors: Mohammad Beyki, Justus Pawlak, Robert Patzke, Franz Renz
Abstract:
In the context of planetary exploration, the MIMOS II (miniaturized M¨ossbauer spectrometer) serves as a proven and reliable measuring instrument. The transmission behaviour of the electronics in the M¨ossbauer spectroscopy is further developed and optimized. For this purpose, the overall electronics is split into three parts. This elaboration deals exclusively with the first part of the signal chain for the evaluation of photons in experiments with gamma radiation. Parallel to the analysis of the electronics, an additional method for analysing the stability of linear and non-linear systems is presented: The extended method of Lyapunov’s stability criteria. The design helps to weigh advantages and disadvantages against other simulated circuits in order to optimize the MIMOS II for the terestric and extraterestric measurment. Finally, after stability analysis, the controller design according to Ackermann is performed, achieving the best possible optimization of the output variable through a skillful pole assignment.
Keywords: Controller design for MIMOS II, stability analysis, M¨ossbauer spectroscopy, electronic signal amplifier, light processing technology, photocurrent, transimpedance amplifier, extended Lyapunov method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52227 Theory of Mind and Its Brain Distribution in Patients with Temporal Lobe Epilepsy
Authors: Wei-Han Wang, Hsiang-Yu Yu, Mau-Sun Hua
Abstract:
Theory of Mind (ToM) refers to the ability to infer another’s mental state. With appropriate ToM, one can behave well in social interactions. A growing body of evidence has demonstrated that patients with temporal lobe epilepsy (TLE) may damage ToM by affecting on regions of the underlying neural network of ToM. However, the question of whether there is cerebral laterality for ToM functions remains open. This study aimed to examine whether there is cerebral lateralization for ToM abilities in TLE patients. Sixty-seven adult TLE patients and 30 matched healthy controls (HC) were recruited. Patients were classified into right (RTLE), left (LTLE), and bilateral (BTLE) TLE groups on the basis of a consensus panel review of their seizure semiology, EEG findings, and brain imaging results. All participants completed an intellectual test and four tasks measuring basic and advanced ToM. The results showed that, on all ToM tasks, (1) each patient group performed worse than HC; (2) there were no significant differences between LTLE and RTLE groups; and (3) the BTLE group performed the worst. It appears that the neural network responsible for ToM is distributed evenly between the cerebral hemispheres.Keywords: Cerebral lateralization, social cognition, temporal lobe epilepsy, theory of mind.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2040226 An Efficient Algorithm for Computing all Program Forward Static Slices
Authors: Jehad Al Dallal
Abstract:
Program slicing is the task of finding all statements in a program that directly or indirectly influence the value of a variable occurrence. The set of statements that can affect the value of a variable at some point in a program is called a program backward slice. In several software engineering applications, such as program debugging and measuring program cohesion and parallelism, several slices are computed at different program points. The existing algorithms for computing program slices are introduced to compute a slice at a program point. In these algorithms, the program, or the model that represents the program, is traversed completely or partially once. To compute more than one slice, the same algorithm is applied for every point of interest in the program. Thus, the same program, or program representation, is traversed several times. In this paper, an algorithm is introduced to compute all forward static slices of a computer program by traversing the program representation graph once. Therefore, the introduced algorithm is useful for software engineering applications that require computing program slices at different points of a program. The program representation graph used in this paper is called Program Dependence Graph (PDG).Keywords: Program slicing, static slicing, forward slicing, program dependence graph (PDG).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1465225 A 3D Approach for Extraction of the Coronaryartery and Quantification of the Stenosis
Authors: Mahdi Mazinani, S. D. Qanadli, Rahil Hosseini, Tim Ellis, Jamshid Dehmeshki
Abstract:
Segmentation and quantification of stenosis is an important task in assessing coronary artery disease. One of the main challenges is measuring the real diameter of curved vessels. Moreover, uncertainty in segmentation of different tissues in the narrow vessel is an important issue that affects accuracy. This paper proposes an algorithm to extract coronary arteries and measure the degree of stenosis. Markovian fuzzy clustering method is applied to model uncertainty arises from partial volume effect problem. The algorithm employs: segmentation, centreline extraction, estimation of orthogonal plane to centreline, measurement of the degree of stenosis. To evaluate the accuracy and reproducibility, the approach has been applied to a vascular phantom and the results are compared with real diameter. The results of 10 patient datasets have been visually judged by a qualified radiologist. The results reveal the superiority of the proposed method compared to the Conventional thresholding Method (CTM) on both datasets.Keywords: 3D coronary artery tree extraction, segmentation, quantification, fuzzy clustering, and Markov random field
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1582224 A Neurofuzzy Learning and its Application to Control System
Authors: Seema Chopra, R. Mitra, Vijay Kumar
Abstract:
A neurofuzzy approach for a given set of input-output training data is proposed in two phases. Firstly, the data set is partitioned automatically into a set of clusters. Then a fuzzy if-then rule is extracted from each cluster to form a fuzzy rule base. Secondly, a fuzzy neural network is constructed accordingly and parameters are tuned to increase the precision of the fuzzy rule base. This network is able to learn and optimize the rule base of a Sugeno like Fuzzy inference system using Hybrid learning algorithm, which combines gradient descent, and least mean square algorithm. This proposed neurofuzzy system has the advantage of determining the number of rules automatically and also reduce the number of rules, decrease computational time, learns faster and consumes less memory. The authors also investigate that how neurofuzzy techniques can be applied in the area of control theory to design a fuzzy controller for linear and nonlinear dynamic systems modelling from a set of input/output data. The simulation analysis on a wide range of processes, to identify nonlinear components on-linely in a control system and a benchmark problem involving the prediction of a chaotic time series is carried out. Furthermore, the well-known examples of linear and nonlinear systems are also simulated under the Matlab/Simulink environment. The above combination is also illustrated in modeling the relationship between automobile trips and demographic factors.
Keywords: Fuzzy control, neuro-fuzzy techniques, fuzzy subtractive clustering, extraction of rules, and optimization of membership functions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2593223 Measuring the Cognitive Abilities of Teenage Basketball Players in Singapore
Authors: Stella Y. Ng, John B. Peacock, Kay Chuan Tan
Abstract:
This paper discusses the use of a computerized test to measure the decision-making abilities of teenage basketball players in Singapore. There are five sections in this test – Competitive state anxiety inventory-2 (CSAI-2) questionnaire (measures player’s cognitive anxiety, somatic anxiety and self-confidence), Corsi block-tapping task (measures player’s short-term spatial memory), situation awareness global assessment technique (SAGAT) (measures players’ situation awareness in a basketball game), multiple choice questions on basketball knowledge (measures players’ knowledge of basketball rules and concepts), and lastly, a learning test that requires participants to recall and recognize basketball set plays (measures player’s ability to learn and recognize set plays). A total of 25 basketball players, aged 14 to 16 years old, from three secondary school teams participated in this experiment. The results that these basketball players obtained from this cognitive test were then used to compare with their physical fitness and basketball performance.
Keywords: Basketball, cognitive abilities, computerized test, decision-making.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2402222 Measuring Risk Levels and Efficacy of Risk Management Strategies in Vietnamese Catfish Farming
Authors: Tru C. Le, France Cheong
Abstract:
Although the Vietnamese catfish farming has grown at very high rates in recent years, the industry has also faced many problems affecting its sustainability. This paper studies the perceptions of catfish farmers regarding risk and risk management strategies in their production activities. Specifically, the study aims to measure the consequences, likelihoods, and levels of risks as well as the efficacy of risk management in Vietnamese catfish farming. Data for the study were collected through a sample of 261 catfish farmers in the Mekong Delta, Vietnam using a questionnaire survey in 2008. Results show that, in general, price and production risks were perceived as the most important risks. Farm management and technical measures were perceived more effective than other kinds of risk management strategies in risk reduction. Although price risks were rated as important risks, price risk management strategies were not perceived as important measures for risk mitigation. The results of the study are discussed to provide implications for various industry stakeholders, including policy makers, processors, advisors, and developers of new risk management strategies.Keywords: Aquaculture, catfish farming, sources of risk, riskmanagement, risk strategies, risk mitigation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1986221 Observations of Conformity in the Health Professions
Authors: Tanya N. Beran, Michelle A. Drefs, Ghazwan Altabbaa, Nouf Al Harbi, Noof Al Baz, Elizabeth Oddone Paolucci
Abstract:
Although interprofessional practice is a collaborative approach for problem solving among health professionals, its implementation can present challenges to its team members. In particular, they may feel pressured to agree with or conform to other members who share information that is contrary to their own understanding. Obtaining evidence of this phenomenon is challenging, as team members may underreport their conformity behaviors due to reasons such as social desirability. In this paper, a series of studies are reviewed in which several approaches to assessing conformity in the health care professions are tested. Simulations, questionnaires, and behavior checklists can be used to measure conformity behaviors. Insights from these studies show that a significant proportion of people conform either in the presence or absence of others, express a variety of verbal and nonverbal behaviors when considering whether to conform to others, may shift between conforming and moments later not conforming (and vice versa), and may not accurately report whether they conformed. A method of measuring conformity using the implicit bias test is also discussed. People at all levels in the healthcare system are encouraged to develop both formal and informal strategies to manage the conformity pressures that people face.
Keywords: Conformity, decision-making, interprofessional teams, medical simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 434220 Perceptual JPEG Compliant Coding by Using DCT-Based Visibility Thresholds of Color Images
Authors: Kuo-Cheng Liu
Abstract:
Effective estimation of just noticeable distortion (JND) for images is helpful to increase the efficiency of a compression algorithm in which both the statistical redundancy and the perceptual redundancy should be accurately removed. In this paper, we design a DCT-based model for estimating JND profiles of color images. Based on a mathematical model of measuring the base detection threshold for each DCT coefficient in the color component of color images, the luminance masking adjustment, the contrast masking adjustment, and the cross masking adjustment are utilized for luminance component, and the variance-based masking adjustment based on the coefficient variation in the block is proposed for chrominance components. In order to verify the proposed model, the JND estimator is incorporated into the conventional JPEG coder to improve the compression performance. A subjective and fair viewing test is designed to evaluate the visual quality of the coding image under the specified viewing condition. The simulation results show that the JPEG coder integrated with the proposed DCT-based JND model gives better coding bit rates at visually lossless quality for a variety of color images.
Keywords: Just-noticeable distortion (JND), discrete cosine transform (DCT), JPEG.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2581219 Rank-Based Chain-Mode Ensemble for Binary Classification
Authors: Chongya Song, Kang Yen, Alexander Pons, Jin Liu
Abstract:
In the field of machine learning, the ensemble has been employed as a common methodology to improve the performance upon multiple base classifiers. However, the true predictions are often canceled out by the false ones during consensus due to a phenomenon called “curse of correlation” which is represented as the strong interferences among the predictions produced by the base classifiers. In addition, the existing practices are still not able to effectively mitigate the problem of imbalanced classification. Based on the analysis on our experiment results, we conclude that the two problems are caused by some inherent deficiencies in the approach of consensus. Therefore, we create an enhanced ensemble algorithm which adopts a designed rank-based chain-mode consensus to overcome the two problems. In order to evaluate the proposed ensemble algorithm, we employ a well-known benchmark data set NSL-KDD (the improved version of dataset KDDCup99 produced by University of New Brunswick) to make comparisons between the proposed and 8 common ensemble algorithms. Particularly, each compared ensemble classifier uses the same 22 base classifiers, so that the differences in terms of the improvements toward the accuracy and reliability upon the base classifiers can be truly revealed. As a result, the proposed rank-based chain-mode consensus is proved to be a more effective ensemble solution than the traditional consensus approach, which outperforms the 8 ensemble algorithms by 20% on almost all compared metrices which include accuracy, precision, recall, F1-score and area under receiver operating characteristic curve.
Keywords: Consensus, curse of correlation, imbalanced classification, rank-based chain-mode ensemble.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 734218 Context Aware Lightweight Energy Efficient Framework
Authors: D. Sathan, A. Meetoo, R. K. Subramaniam
Abstract:
Context awareness is a capability whereby mobile computing devices can sense their physical environment and adapt their behavior accordingly. The term context-awareness, in ubiquitous computing, was introduced by Schilit in 1994 and has become one of the most exciting concepts in early 21st-century computing, fueled by recent developments in pervasive computing (i.e. mobile and ubiquitous computing). These include computing devices worn by users, embedded devices, smart appliances, sensors surrounding users and a variety of wireless networking technologies. Context-aware applications use context information to adapt interfaces, tailor the set of application-relevant data, increase the precision of information retrieval, discover services, make the user interaction implicit, or build smart environments. For example: A context aware mobile phone will know that the user is currently in a meeting room, and reject any unimportant calls. One of the major challenges in providing users with context-aware services lies in continuously monitoring their contexts based on numerous sensors connected to the context aware system through wireless communication. A number of context aware frameworks based on sensors have been proposed, but many of them have neglected the fact that monitoring with sensors imposes heavy workloads on ubiquitous devices with limited computing power and battery. In this paper, we present CALEEF, a lightweight and energy efficient context aware framework for resource limited ubiquitous devices.Keywords: Context-Aware, Energy-Efficient, Lightweight, Ubiquitous Devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1947217 Non-parametric Linear Technique for Measuring the Efficiency of Winter Road Maintenance in the Arctic Area
Authors: Mahshid Hatamzad, Geanette Polanco
Abstract:
Improving the performance of Winter Road Maintenance (WRM) can increase the traffic safety and reduce the cost as well as environmental impacts. This study evaluates the efficiency of WRM technique, named salting, in the Arctic area by using Data Envelopment Analysis (DEA), which is a non-parametric linear method to measure the efficiencies of decision-making units (DMUs) based on handling multiple inputs and multiple outputs at the same time that their associated weights are not known. Here, roads are considered as DMUs for which the efficiency must be determined. The three input variables considered are traffic flow, road area and WRM cost. In addition, the two output variables included are level of safety in the roads and environment impacts resulted from WRM, which is also considered as an uncontrollable factor in the second scenario. The results show the performance of DMUs from the most efficient WRM to the inefficient/least efficient one and this information provides decision makers with technical support and the required suggested improvements for inefficient WRM, in order to achieve a cost-effective WRM and a safe road transportation during wintertime in the Arctic areas.
Keywords: DEA, environmental impacts, risk and safety, WRM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 555216 A Model-Reference Sliding Mode for Dual-Stage Actuator Servo Control in HDD
Authors: S. Sonkham, U. Pinsopon, W. Chatlatanagulchai
Abstract:
This paper presents a method of sliding mode control (SMC) designing and developing for the servo system in a dual-stage actuator (DSA) hard disk drive. Mathematical modeling of hard disk drive actuators is obtained, extracted from measuring frequency response of the voice-coil motor (VCM) and PZT micro-actuator separately. Matlab software tools are used for mathematical model estimation and also for controller design and simulation. A model-reference approach for tracking requirement is selected as a proposed technique. The simulation results show that performance of a model-reference SMC controller design in DSA servo control can be satisfied in the tracking error, as well as keeping the positioning of the head within the boundary of +/-5% of track width under the presence of internal and external disturbance. The overall results of model-reference SMC design in DSA are met per requirement specifications and significant reduction in %off track is found when compared to the single-state actuator (SSA).
Keywords: Hard Disk Drive, Dual-Stage Actuator, Track Following, HDD Servo Control, Sliding Mode Control, Model-Reference, Tracking Control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1962