Search results for: scale invariant feature
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7287

Search results for: scale invariant feature

7107 An Automatic Feature Extraction Technique for 2D Punch Shapes

Authors: Awais Ahmad Khan, Emad Abouel Nasr, H. M. A. Hussein, Abdulrahman Al-Ahmari

Abstract:

Sheet-metal parts have been widely applied in electronics, communication and mechanical industries in recent decades; but the advancement in sheet-metal part design and manufacturing is still behind in comparison with the increasing importance of sheet-metal parts in modern industry. This paper presents a methodology for automatic extraction of some common 2D internal sheet metal features. The features used in this study are taken from Unipunch ™ catalogue. The extraction process starts with the data extraction from STEP file using an object oriented approach and with the application of suitable algorithms and rules, all features contained in the catalogue are automatically extracted. Since the extracted features include geometry and engineering information, they will be effective for downstream application such as feature rebuilding and process planning.

Keywords: feature extraction, internal features, punch shapes, sheet metal

Procedia PDF Downloads 587
7106 Improving Fake News Detection Using K-means and Support Vector Machine Approaches

Authors: Kasra Majbouri Yazdi, Adel Majbouri Yazdi, Saeid Khodayi, Jingyu Hou, Wanlei Zhou, Saeed Saedy

Abstract:

Fake news and false information are big challenges of all types of media, especially social media. There is a lot of false information, fake likes, views and duplicated accounts as big social networks such as Facebook and Twitter admitted. Most information appearing on social media is doubtful and in some cases misleading. They need to be detected as soon as possible to avoid a negative impact on society. The dimensions of the fake news datasets are growing rapidly, so to obtain a better result of detecting false information with less computation time and complexity, the dimensions need to be reduced. One of the best techniques of reducing data size is using feature selection method. The aim of this technique is to choose a feature subset from the original set to improve the classification performance. In this paper, a feature selection method is proposed with the integration of K-means clustering and Support Vector Machine (SVM) approaches which work in four steps. First, the similarities between all features are calculated. Then, features are divided into several clusters. Next, the final feature set is selected from all clusters, and finally, fake news is classified based on the final feature subset using the SVM method. The proposed method was evaluated by comparing its performance with other state-of-the-art methods on several specific benchmark datasets and the outcome showed a better classification of false information for our work. The detection performance was improved in two aspects. On the one hand, the detection runtime process decreased, and on the other hand, the classification accuracy increased because of the elimination of redundant features and the reduction of datasets dimensions.

Keywords: clustering, fake news detection, feature selection, machine learning, social media, support vector machine

Procedia PDF Downloads 145
7105 The Examination of Prospective ICT Teachers’ Attitudes towards Application of Computer Assisted Instruction

Authors: Agâh Tuğrul Korucu, Ismail Fatih Yavuzaslan, Lale Toraman

Abstract:

Nowadays, thanks to development of technology, integration of technology into teaching and learning activities is spreading. Increasing technological literacy which is one of the expected competencies for individuals of 21st century is associated with the effective use of technology in education. The most important factor in effective use of technology in education institutions is ICT teachers. The concept of computer assisted instruction (CAI) refers to the utilization of information and communication technology as a tool aided teachers in order to make education more efficient and improve its quality in the process of educational. Teachers can use computers in different places and times according to owned hardware and software facilities and characteristics of the subject and student in CAI. Analyzing teachers’ use of computers in education is significant because teachers are the ones who manage the course and they are the most important element in comprehending the topic by students. To accomplish computer-assisted instruction efficiently is possible through having positive attitude of teachers. Determination the level of knowledge, attitude and behavior of teachers who get the professional knowledge from educational faculties and elimination of deficiencies if any are crucial when teachers are at the faculty. Therefore, the aim of this paper is to identify ICT teachers' attitudes toward computer-assisted instruction in terms of different variables. Research group consists of 200 prospective ICT teachers studying at Necmettin Erbakan University Ahmet Keleşoğlu Faculty of Education CEIT department. As data collection tool of the study; “personal information form” developed by the researchers and used to collect demographic data and "the attitude scale related to computer-assisted instruction" are used. The scale consists of 20 items. 10 of these items show positive feature, while 10 of them show negative feature. The Kaiser-Meyer-Olkin (KMO) coefficient of the scale is found 0.88 and Barlett test significance value is found 0.000. The Cronbach’s alpha reliability coefficient of the scale is found 0.93. In order to analyze the data collected by data collection tools computer-based statistical software package used; statistical techniques such as descriptive statistics, t-test, and analysis of variance are utilized. It is determined that the attitudes of prospective instructors towards computers do not differ according to their educational branches. On the other hand, the attitudes of prospective instructors who own computers towards computer-supported education are determined higher than those of the prospective instructors who do not own computers. It is established that the departments of students who previously received computer lessons do not affect this situation so much. The result is that; the computer experience affects the attitude point regarding the computer-supported education positively.

Keywords: computer based instruction, teacher candidate, attitude, technology based instruction, information and communication technologies

Procedia PDF Downloads 262
7104 Using Self Organizing Feature Maps for Classification in RGB Images

Authors: Hassan Masoumi, Ahad Salimi, Nazanin Barhemmat, Babak Gholami

Abstract:

Artificial neural networks have gained a lot of interest as empirical models for their powerful representational capacity, multi input and output mapping characteristics. In fact, most feed-forward networks with nonlinear nodal functions have been proved to be universal approximates. In this paper, we propose a new supervised method for color image classification based on self organizing feature maps (SOFM). This algorithm is based on competitive learning. The method partitions the input space using self-organizing feature maps to introduce the concept of local neighborhoods. Our image classification system entered into RGB image. Experiments with simulated data showed that separability of classes increased when increasing training time. In additional, the result shows proposed algorithms are effective for color image classification.

Keywords: classification, SOFM algorithm, neural network, neighborhood, RGB image

Procedia PDF Downloads 445
7103 Joint Modeling of Longitudinal and Time-To-Event Data with Latent Variable

Authors: Xinyuan Y. Song, Kai Kang

Abstract:

Joint models for analyzing longitudinal and survival data are widely used to investigate the relationship between a failure time process and time-variant predictors. A common assumption in conventional joint models in the survival analysis literature is that all predictors are observable. However, this assumption may not always be supported because unobservable traits, namely, latent variables, which are indirectly observable and should be measured through multiple observed variables, are commonly encountered in the medical, behavioral, and financial research settings. In this study, a joint modeling approach to deal with this feature is proposed. The proposed model comprises three parts. The first part is a dynamic factor analysis model for characterizing latent variables through multiple observed indicators over time. The second part is a random coefficient trajectory model for describing the individual trajectories of latent variables. The third part is a proportional hazard model for examining the effects of time-invariant predictors and the longitudinal trajectories of time-variant latent risk factors on hazards of interest. A Bayesian approach coupled with a Markov chain Monte Carlo algorithm to perform statistical inference. An application of the proposed joint model to a study on the Alzheimer's disease neuroimaging Initiative is presented.

Keywords: Bayesian analysis, joint model, longitudinal data, time-to-event data

Procedia PDF Downloads 112
7102 Graph Codes - 2D Projections of Multimedia Feature Graphs for Fast and Effective Retrieval

Authors: Stefan Wagenpfeil, Felix Engel, Paul McKevitt, Matthias Hemmje

Abstract:

Multimedia Indexing and Retrieval is generally designed and implemented by employing feature graphs. These graphs typically contain a significant number of nodes and edges to reflect the level of detail in feature detection. A higher level of detail increases the effectiveness of the results but also leads to more complex graph structures. However, graph-traversal-based algorithms for similarity are quite inefficient and computation intensive, especially for large data structures. To deliver fast and effective retrieval, an efficient similarity algorithm, particularly for large graphs, is mandatory. Hence, in this paper, we define a graph-projection into a 2D space (Graph Code) as well as the corresponding algorithms for indexing and retrieval. We show that calculations in this space can be performed more efficiently than graph-traversals due to a simpler processing model and a high level of parallelization. In consequence, we prove that the effectiveness of retrieval also increases substantially, as Graph Codes facilitate more levels of detail in feature fusion. Thus, Graph Codes provide a significant increase in efficiency and effectiveness (especially for Multimedia indexing and retrieval) and can be applied to images, videos, audio, and text information.

Keywords: indexing, retrieval, multimedia, graph algorithm, graph code

Procedia PDF Downloads 129
7101 Correlation of Spirometry with Six Minute Walk Test and Grading of Dyspnoea in COPD Patients

Authors: Anand K. Patel

Abstract:

Background: Patients with COPD have decreased pulmonary functions, which in turn reflect on their day-to-day activities. Objectives: To assess the correlation between functional vital capacity (FVC) and forced expiratory volume in one second (FEV1) with 6 minutes walk test (6MWT). To correlate the Borg rating for perceived exertion scale (Borg scale) and Modified medical research council (MMRC) dyspnea scale with the 6MWT, FVC and FEV1. Method: In this prospective study total 72 patients with COPD diagnosed by the GOLD guidelines were enrolled after taking written consent. They were first asked to rate physical exertion on the Borg scale as well as the modified medical research council dyspnea scale and then were subjected to perform pre and post bronchodilator spirometry followed by 6 minute walk test. The findings were correlated by calculating the Pearson coefficient for each set and obtaining the p-values, with a p < 0.05 being clinically significant. Result: There was a significant correlation between spirometry and 6MWT suggesting that patients with lower measurements were unable to walk for longer distances. However, FVC had the stronger correlation than FEV1. MMRC scale had a stronger correlation with 6MWT as compared to the Borg scale. Conclusion: The study suggests that 6MWT is a better test for monitoring the patients of COPD. In spirometry, FVC should be used in monitoring patients with COPD, instead of FEV1. MMRC scale shows a stronger correlation than the Borg scale, and we should use it more often.

Keywords: spirometry, 6 minute walk test, MMRC, Borg scale

Procedia PDF Downloads 168
7100 Pantograph-Catenary Contact Force: Features Evaluation for Catenary Diagnostics

Authors: Mehdi Brahimi, Kamal Medjaher, Noureddine Zerhouni, Mohammed Leouatni

Abstract:

The Prognostics and Health Management is a system engineering discipline which provides solutions and models to the implantation of a predictive maintenance. The approach is based on extracting useful information from monitoring data to assess the “health” state of an industrial equipment or an asset. In this paper, we examine multiple extracted features from Pantograph-Catenary contact force in order to select the most relevant ones to achieve a diagnostics function. The feature extraction methodology is based on simulation data generated thanks to a Pantograph-Catenary simulation software called INPAC and measurement data. The feature extraction method is based on both statistical and signal processing analyses. The feature selection method is based on statistical criteria.

Keywords: catenary/pantograph interaction, diagnostics, Prognostics and Health Management (PHM), quality of current collection

Procedia PDF Downloads 260
7099 Development of an Attitude Scale Towards Social Networking Sites

Authors: Münevver Başman, Deniz Gülleroğlu

Abstract:

The purpose of this study is to develop a scale to determine the attitudes towards social networking sites. 45 tryout items, prepared for this aim, were applied to 342 students studying at Marmara University, Faculty of Education. The reliability and the validity of the scale were conducted with the help of these students. As a result of exploratory factor analysis with Varimax rotation, 41 items grouped according to the structure with three factors (interest, reality and negative effects) is obtained. While alpha reliability of the scale is obtained as .899; the reliability of factors is obtained as .899, .799, .775, respectively.

Keywords: Attitude, reliability, social networking sites, validity.

Procedia PDF Downloads 349
7098 Prediction of the Torsional Vibration Characteristics of a Rotor-Shaft System Using Its Scale Model and Scaling Laws

Authors: Jia-Jang Wu

Abstract:

This paper presents the scaling laws that provide the criteria of geometry and dynamic similitude between the full-size rotor-shaft system and its scale model, and can be used to predict the torsional vibration characteristics of the full-size rotor-shaft system by manipulating the corresponding data of its scale model. The scaling factors, which play fundamental roles in predicting the geometry and dynamic relationships between the full-size rotor-shaft system and its scale model, for torsional free vibration problems between scale and full-size rotor-shaft systems are firstly obtained from the equation of motion of torsional free vibration. Then, the scaling factor of external force (i.e., torque) required for the torsional forced vibration problems is determined based on the Newton’s second law. Numerical results show that the torsional free and forced vibration characteristics of a full-size rotor-shaft system can be accurately predicted from those of its scale models by using the foregoing scaling factors. For this reason, it is believed that the presented approach will be significant for investigating the relevant phenomenon in the scale model tests.

Keywords: torsional vibration, full-size model, scale model, scaling laws

Procedia PDF Downloads 369
7097 A Nonlinear Feature Selection Method for Hyperspectral Image Classification

Authors: Pei-Jyun Hsieh, Cheng-Hsuan Li, Bor-Chen Kuo

Abstract:

For hyperspectral image classification, feature reduction is an important pre-processing for avoiding the Hughes phenomena due to the difficulty for collecting training samples. Hence, lots of researches developed feature selection methods such as F-score, HSIC (Hilbert-Schmidt Independence Criterion), and etc., to improve hyperspectral image classification. However, most of them only consider the class separability in the original space, i.e., a linear class separability. In this study, we proposed a nonlinear class separability measure based on kernel trick for selecting an appropriate feature subset. The proposed nonlinear class separability was formed by a generalized RBF kernel with different bandwidths with respect to different features. Moreover, it considered the within-class separability and the between-class separability. A genetic algorithm was applied to tune these bandwidths such that the smallest with-class separability and the largest between-class separability simultaneously. This indicates the corresponding feature space is more suitable for classification. In addition, the corresponding nonlinear classification boundary can separate classes very well. These optimal bandwidths also show the importance of bands for hyperspectral image classification. The reciprocals of these bandwidths can be viewed as weights of bands. The smaller bandwidth, the larger weight of the band, and the more importance for classification. Hence, the descending order of the reciprocals of the bands gives an order for selecting the appropriate feature subsets. In the experiments, three hyperspectral image data sets, the Indian Pine Site data set, the PAVIA data set, and the Salinas A data set, were used to demonstrate the selected feature subsets by the proposed nonlinear feature selection method are more appropriate for hyperspectral image classification. Only ten percent of samples were randomly selected to form the training dataset. All non-background samples were used to form the testing dataset. The support vector machine was applied to classify these testing samples based on selected feature subsets. According to the experiments on the Indian Pine Site data set with 220 bands, the highest accuracies by applying the proposed method, F-score, and HSIC are 0.8795, 0.8795, and 0.87404, respectively. However, the proposed method selects 158 features. F-score and HSIC select 168 features and 217 features, respectively. Moreover, the classification accuracies increase dramatically only using first few features. The classification accuracies with respect to feature subsets of 10 features, 20 features, 50 features, and 110 features are 0.69587, 0.7348, 0.79217, and 0.84164, respectively. Furthermore, only using half selected features (110 features) of the proposed method, the corresponding classification accuracy (0.84168) is approximate to the highest classification accuracy, 0.8795. For other two hyperspectral image data sets, the PAVIA data set and Salinas A data set, we can obtain the similar results. These results illustrate our proposed method can efficiently find feature subsets to improve hyperspectral image classification. One can apply the proposed method to determine the suitable feature subset first according to specific purposes. Then researchers can only use the corresponding sensors to obtain the hyperspectral image and classify the samples. This can not only improve the classification performance but also reduce the cost for obtaining hyperspectral images.

Keywords: hyperspectral image classification, nonlinear feature selection, kernel trick, support vector machine

Procedia PDF Downloads 241
7096 A Study on the Development of Social Participation Activity Scale for the Elderly

Authors: Young-Kwang Lee, Eun-Gu Ji, Min-Joo Kim, Seung-Jae Oh

Abstract:

The purpose of this study is to develop a social participation activity scale for the elderly. As a result of exploratory factor analysis, confirmatory factor analysis was conducted using maximum likelihood method using bundled items. In conclusion, thirteen items of social participation activity scale seemed appropriate. Finally, convergent validity and discriminant validity were verified on the scale with the fit. The convergent validity was based on the variance extracted value. In other words, the hypothesis that the variables are the same is rejected and the validity is confirmed. This study extensively considered the measurement items of the social participation activity scale used to measure social participation activities of the elderly. In the future, it will be meaningful that it can be used as a tool to verify the effectiveness of services in organizations that provide social welfare services to elderly people such as comprehensive social welfare centers and the elderly comprehensive social welfare centers.

Keywords: elderly, social participation, scale development, validity

Procedia PDF Downloads 153
7095 Efficient Human Motion Detection Feature Set by Using Local Phase Quantization Method

Authors: Arwa Alzughaibi

Abstract:

Human Motion detection is a challenging task due to a number of factors including variable appearance, posture and a wide range of illumination conditions and background. So, the first need of such a model is a reliable feature set that can discriminate between a human and a non-human form with a fair amount of confidence even under difficult conditions. By having richer representations, the classification task becomes easier and improved results can be achieved. The Aim of this paper is to investigate the reliable and accurate human motion detection models that are able to detect the human motions accurately under varying illumination levels and backgrounds. Different sets of features are tried and tested including Histogram of Oriented Gradients (HOG), Deformable Parts Model (DPM), Local Decorrelated Channel Feature (LDCF) and Aggregate Channel Feature (ACF). However, we propose an efficient and reliable human motion detection approach by combining Histogram of oriented gradients (HOG) and local phase quantization (LPQ) as the feature set, and implementing search pruning algorithm based on optical flow to reduce the number of false positive. Experimental results show the effectiveness of combining local phase quantization descriptor and the histogram of gradient to perform perfectly well for a large range of illumination conditions and backgrounds than the state-of-the-art human detectors. Areaunder th ROC Curve (AUC) of the proposed method achieved 0.781 for UCF dataset and 0.826 for CDW dataset which indicates that it performs comparably better than HOG, DPM, LDCF and ACF methods.

Keywords: human motion detection, histograms of oriented gradient, local phase quantization, local phase quantization

Procedia PDF Downloads 229
7094 Human Gait Recognition Using Moment with Fuzzy

Authors: Jyoti Bharti, Navneet Manjhi, M. K.Gupta, Bimi Jain

Abstract:

A reliable gait features are required to extract the gait sequences from an images. In this paper suggested a simple method for gait identification which is based on moments. Moment values are extracted on different number of frames of gray scale and silhouette images of CASIA database. These moment values are considered as feature values. Fuzzy logic and nearest neighbour classifier are used for classification. Both achieved higher recognition.

Keywords: gait, fuzzy logic, nearest neighbour, recognition rate, moments

Procedia PDF Downloads 722
7093 Modeling the Effect of Scale Deposition on Heat Transfer in Desalination Multi-Effect Distillation Evaporators

Authors: K. Bourouni, M. Chacha, T. Jaber, A. Tchantchane

Abstract:

In Multi-Effect Distillation (MED) desalination evaporators, the scale deposit outside the tubes presents a barrier to heat transfers reducing the global heat transfer coefficient and causing a decrease in water production; hence a loss of efficiency and an increase in operating and maintenance costs. Scale removal (by acid cleaning) is the main maintenance operation and constitutes the major reason for periodic plant shutdowns. A better understanding of scale deposition mechanisms will lead to an accurate determination of the variation of scale thickness around the tubes and an improved accuracy of the overall heat transfer coefficient calculation. In this paper, a coupled heat transfer-calcium carbonate scale deposition model on a horizontal tube bundle is presented. The developed tool is used to determine precisely the heat transfer area leading to a significant cost reduction for a given water production capacity. Simulations are carried to investigate the influence of different parameters such as water salinity, temperature, etc. on the heat transfer.

Keywords: multi-effect-evaporator, scale deposition, water desalination, heat transfer coefficient

Procedia PDF Downloads 119
7092 Video Text Information Detection and Localization in Lecture Videos Using Moments

Authors: Belkacem Soundes, Guezouli Larbi

Abstract:

This paper presents a robust and accurate method for text detection and localization over lecture videos. Frame regions are classified into text or background based on visual feature analysis. However, lecture video shows significant degradation mainly related to acquisition conditions, camera motion and environmental changes resulting in low quality videos. Hence, affecting feature extraction and description efficiency. Moreover, traditional text detection methods cannot be directly applied to lecture videos. Therefore, robust feature extraction methods dedicated to this specific video genre are required for robust and accurate text detection and extraction. Method consists of a three-step process: Slide region detection and segmentation; Feature extraction and non-text filtering. For robust and effective features extraction moment functions are used. Two distinct types of moments are used: orthogonal and non-orthogonal. For orthogonal Zernike Moments, both Pseudo Zernike moments are used, whereas for non-orthogonal ones Hu moments are used. Expressivity and description efficiency are given and discussed. Proposed approach shows that in general, orthogonal moments show high accuracy in comparison to the non-orthogonal one. Pseudo Zernike moments are more effective than Zernike with better computation time.

Keywords: text detection, text localization, lecture videos, pseudo zernike moments

Procedia PDF Downloads 122
7091 Experimental Investigation of Heat Transfer and Scale Growth Characteristics of Crystallisation Scale in Agitation Tank

Authors: Prasanjit Das, M .M. K. Khan, M. G. Rasul, Jie Wu, I. Youn

Abstract:

Crystallisation scale occurs when dissolved minerals precipitate from an aqueous solution. To investigate the crystallisation scale growth of normal solubility salt, a lab-scale agitation tank with and without baffles were used as a benchmark using potassium nitrate as the test fluid. Potassium nitrate (KNO3) solution in this test leads to crystallisation scale on heat transfer surfaces. This experimental investigation has focused on the effect of surface crystallisation of potassium nitrate on the low-temperature heat exchange surfaces on the wall of the agitation tank. The impeller agitation rate affects the scaling rate at the low-temperature agitation wall and it shows a decreasing scaling rate with an increasing agitation rate. It was observed that there was a significant variation of heat transfer coefficients and scaling resistance coefficients with different agitation rate as well as with varying impeller size, tank with and without baffles and solution concentration.

Keywords: crystallisation, heat transfer coefficient, scale, resistance

Procedia PDF Downloads 153
7090 Development and Validation of the 'Short Form BASIC Scale' Psychotic Tendencies Subscale

Authors: Chia-Chun Wu, Ying-Yao Cheng

Abstract:

The purpose of this study was developing the 'short-form BASIC scale' psychotic tendencies subscale so as to provide a more efficient, economical and effective way to assess the mental health of recruits. 1749 students from Naval Recruit Training Center participated in this study. The multidimensional constructs of psychotic tendencies subscale include four dimensions: schizophrenic tendencies, manic tendencies, depression tendencies, and suicidal ideation. We cut down the 36-item psychotic tendencies subscale to 25 items by using multidimension Rasch techniques. They were applied to assess model-data fit and to provide the validity evidence of the short form BASIC scale of psychotic tendencies subscale. The person separation reliabilities of the measures from four dimensions were .70, .67, .74 and .57, respectively. In addition, there is a notable correlation between the length version and short version of schizophrenic tendencies (scaled .89), manic tendencies (.96), depression tendencies (.97) and suicidal ideation (.97). The results have indicated that the development of the study of short-form scale sufficient to replace the original scale. Therefore, it is suggested that short-form basic scale is used to assess the mental health with participants being more willing to answer questions to ensure the validation of assessments.

Keywords: BASIC scale, military, Rasch analysis, short-form scale

Procedia PDF Downloads 333
7089 Methods of Livable Goal-Oriented Master Urban Design: A Case Study on Zibo City

Authors: Xiaoping Zhang, Fengying Yan

Abstract:

The implementation of the 'Urban Design Management Measures' requires that the master urban design should aim at creating a livable urban space. However, to our best knowledge, the existing researches and practices of master urban design not only focus less on the livable space but also face a number of problems such as paying more attention to the image of the city, ignoring the people-oriented and lacking dynamic continuity. In order to make the master urban design can better guide the construction of city. Firstly, the paper proposes the livable city hierarchy system to meet the needs of different groups of people and then constructs the framework of livable goal-oriented master urban design based on the theory of livable content and the ideological origin of people-oriented. Secondly, the paper takes the master urban design practice of Zibo as a sample and puts forward the design strategy of strengthening the pattern, improve the quality of space, shape the feature, and establish a series of action plans based on the strategy of urban space development. Finally, the paper explores the method system of livable goal-oriented master urban design from the aspects of safety pattern, morphology pattern, neighborhood scale, open space, street space, public interface, style feature, public participation and action plans.

Keywords: livable, master urban design, public participation, zibo city

Procedia PDF Downloads 274
7088 Musical Instruments Classification Using Machine Learning Techniques

Authors: Bhalke D. G., Bormane D. S., Kharate G. K.

Abstract:

This paper presents classification of musical instrument using machine learning techniques. The classification has been carried out using temporal, spectral, cepstral and wavelet features. Detail feature analysis is carried out using separate and combined features. Further, instrument model has been developed using K-Nearest Neighbor and Support Vector Machine (SVM). Benchmarked McGill university database has been used to test the performance of the system. Experimental result shows that SVM performs better as compared to KNN classifier.

Keywords: feature extraction, SVM, KNN, musical instruments

Procedia PDF Downloads 454
7087 Fast and Scale-Adaptive Target Tracking via PCA-SIFT

Authors: Yawen Wang, Hongchang Chen, Shaomei Li, Chao Gao, Jiangpeng Zhang

Abstract:

As the main challenge for target tracking is accounting for target scale change and real-time, we combine Mean-Shift and PCA-SIFT algorithm together to solve the problem. We introduce similarity comparison method to determine how the target scale changes, and taking different strategies according to different situation. For target scale getting larger will cause location error, we employ backward tracking to reduce the error. Mean-Shift algorithm has poor performance when tracking scale-changing target due to the fixed bandwidth of its kernel function. In order to overcome this problem, we introduce PCA-SIFT matching. Through key point matching between target and template, that adjusting the scale of tracking window adaptively can be achieved. Because this algorithm is sensitive to wrong match, we introduce RANSAC to reduce mismatch as far as possible. Furthermore target relocating will trigger when number of match is too small. In addition we take comprehensive consideration about target deformation and error accumulation to put forward a new template update method. Experiments on five image sequences and comparison with 6 kinds of other algorithm demonstrate favorable performance of the proposed tracking algorithm.

Keywords: target tracking, PCA-SIFT, mean-shift, scale-adaptive

Procedia PDF Downloads 405
7086 Development and Psychometric Properties of the Relational Mobility Scale for the Indonesian Population

Authors: Sukaesi Marianti

Abstract:

This study aims to develop the Relational Mobility Scale for the Indonesian population and to investigate its psychometric properties. New items of the scale were created taking into account the Indonesian population which consists of two parallel forms (A and A’). This study uses 30 newly orchestrated items while keeping in mind the characteristics of the targeted population. The scale was administered to 433 public high school students in Malang, Indonesia. Construct validity of its factor structure was demonstrated using exploratory factor analysis and confirmatory factor analysis. The result exhibits that he model fits the data, and that the delayed alternate form method shows acceptable result. Results yielded that 21 items of the three-dimensional Relational Mobility Scale is suitable for measuring relational mobility in high school students of Indonesian population.

Keywords: confirmatory factor analysis, delayed alternate form, Indonesian population, relational mobility scale

Procedia PDF Downloads 227
7085 Dynamic Distribution Calibration for Improved Few-Shot Image Classification

Authors: Majid Habib Khan, Jinwei Zhao, Xinhong Hei, Liu Jiedong, Rana Shahzad Noor, Muhammad Imran

Abstract:

Deep learning is increasingly employed in image classification, yet the scarcity and high cost of labeled data for training remain a challenge. Limited samples often lead to overfitting due to biased sample distribution. This paper introduces a dynamic distribution calibration method for few-shot learning. Initially, base and new class samples undergo normalization to mitigate disparate feature magnitudes. A pre-trained model then extracts feature vectors from both classes. The method dynamically selects distribution characteristics from base classes (both adjacent and remote) in the embedding space, using a threshold value approach for new class samples. Given the propensity of similar classes to share feature distributions like mean and variance, this research assumes a Gaussian distribution for feature vectors. Subsequently, distributional features of new class samples are calibrated using a corrected hyperparameter, derived from the distribution features of both adjacent and distant base classes. This calibration augments the new class sample set. The technique demonstrates significant improvements, with up to 4% accuracy gains in few-shot classification challenges, as evidenced by tests on miniImagenet and CUB datasets.

Keywords: deep learning, computer vision, image classification, few-shot learning, threshold

Procedia PDF Downloads 32
7084 Use of a Symptom Scale Based on Degree of Functional Impairment for Acute Concussion

Authors: Matthew T. McCarthy, Sarah Janse, Natalie M. Pizzimenti, Anthony K. Savino, Brian Crosser, Sean C. Rose

Abstract:

Concussion is diagnosed clinically using a comprehensive history and exam, supported by ancillary testing. Frequently, symptom checklists are used as part of the evaluation of concussion. Existing symptom scales are based on a subjective Likert scale, without relation of symptoms to clinical or functional impairment. This is a retrospective review of 133 patients under age 30 seen in an outpatient neurology practice within 30 days of a probable or definite concussion. Each patient completed 2 symptom checklists at the initial visit – the SCAT-3 symptom evaluation (22 symptoms, 0-6 scale) and a scale based on the degree of clinical impairment for each symptom (22 symptoms, 0-3 scale related to functional impact of the symptom). Final clearance date was determined by the treating physician. 60.9% of patients were male with mean age 15.7 years (SD 2.3). Mean time from concussion to first visit was 6.9 days (SD 6.2), and 101 patients had definite concussions (75.9%), while 32 were diagnosed as probable (24.1%). 94 patients had a known clearance date (70.7%) with mean clearance time of 20.6 days (SD 18.6) and median clearance time of 19 days (95% CI 16-21). Mean total symptom score was 27.2 (SD 22.9) on the SCAT-3 and 14.7 (SD 11.9) for the functional impairment scale. Pearson’s correlation between the two scales was 0.98 (p < 0.001). After adjusting for patient and injury characteristics, an equivalent increase in score on each scale was associated with longer time to clearance (SCAT-3 hazard ratio 0.885, 95%CI 0.835-0.938, p < 0.001; functional impairment scale hazard ratio 0.851, 95%CI 0.802-0.902, p < 0.001). A concussion symptom scale based on degree of functional impairment correlates strongly with the SCAT-3 scale and demonstrates a similar association with time to clearance. By assessing the degree of impact on clinical functioning, this symptom scale reflects a more intuitive approach to rating symptoms and can be used in the management of concussion.

Keywords: checklist, concussion, neurology, scale, sports, symptoms

Procedia PDF Downloads 118
7083 A Preliminary Conceptual Scale to Discretize the Distributed Manufacturing Continuum

Authors: Ijaz Ul Haq, Fiorenzo Franceschini

Abstract:

The distributed manufacturing methodology brings a new concept of decentralized manufacturing operations close to the proximity of end users. A preliminary scale, to measure distributed capacity and evaluate positioning of firms, is developed in this research. In the first part of the paper, a literature review has been performed which highlights the explorative nature of the studies conducted to present definitions and classifications due to novelty of this topic. From literature, five dimensions of distributed manufacturing development stages have been identified: localization, manufacturing technologies, customization and personalization, digitalization and democratization of design. Based on these determinants a conceptual scale is proposed to measure the status of distributed manufacturing of a generic firm. A multiple case study is then conducted in two steps to test the conceptual scale and to identify the corresponding level of distributed potential in each case study firm.

Keywords: distributed manufacturing, distributed capacity, localized production, ordinal scale

Procedia PDF Downloads 139
7082 PYTHEIA: A Scale for Assessing Rehabilitation and Assistive Robotics

Authors: Yiannis Koumpouros, Effie Papageorgiou, Alexandra Karavasili, Foteini Koureta

Abstract:

The objective of the present study was to develop a scale called PYTHEIA. The PYTHEIA is a self-reported measure for the assessment of rehabilitation and assistive robotics and other assistive technology devices. The development of PYTHEIA faced the absence of a valid instrument that can be used to evaluate the assistive robotic devices both as a whole, as well as any of their individual components or functionalities implemented. According to the results presented, PYTHEIA is a valid and reliable scale able to be applied to different target groups for the subjective evaluation of various assistive technology devices.

Keywords: rehabilitation, assistive technology, assistive robots, rehabilitation robots, scale, psychometric test, assessment, validation, user satisfaction

Procedia PDF Downloads 282
7081 Detection of Curvilinear Structure via Recursive Anisotropic Diffusion

Authors: Sardorbek Numonov, Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Dongeun Choi, Byung-Woo Hong

Abstract:

The detection of curvilinear structures often plays an important role in the analysis of images. In particular, it is considered as a crucial step for the diagnosis of chronic respiratory diseases to localize the fissures in chest CT imagery where the lung is divided into five lobes by the fissures that are characterized by linear features in appearance. However, the characteristic linear features for the fissures are often shown to be subtle due to the high intensity variability, pathological deformation or image noise involved in the imaging procedure, which leads to the uncertainty in the quantification of anatomical or functional properties of the lung. Thus, it is desired to enhance the linear features present in the chest CT images so that the distinctiveness in the delineation of the lobe is improved. We propose a recursive diffusion process that prefers coherent features based on the analysis of structure tensor in an anisotropic manner. The local image features associated with certain scales and directions can be characterized by the eigenanalysis of the structure tensor that is often regularized via isotropic diffusion filters. However, the isotropic diffusion filters involved in the computation of the structure tensor generally blur geometrically significant structure of the features leading to the degradation of the characteristic power in the feature space. Thus, it is required to take into consideration of local structure of the feature in scale and direction when computing the structure tensor. We apply an anisotropic diffusion in consideration of scale and direction of the features in the computation of the structure tensor that subsequently provides the geometrical structure of the features by its eigenanalysis that determines the shape of the anisotropic diffusion kernel. The recursive application of the anisotropic diffusion with the kernel the shape of which is derived from the structure tensor leading to the anisotropic scale-space where the geometrical features are preserved via the eigenanalysis of the structure tensor computed from the diffused image. The recursive interaction between the anisotropic diffusion based on the geometry-driven kernels and the computation of the structure tensor that determines the shape of the diffusion kernels yields a scale-space where geometrical properties of the image structure are effectively characterized. We apply our recursive anisotropic diffusion algorithm to the detection of curvilinear structure in the chest CT imagery where the fissures present curvilinear features and define the boundary of lobes. It is shown that our algorithm yields precise detection of the fissures while overcoming the subtlety in defining the characteristic linear features. The quantitative evaluation demonstrates the robustness and effectiveness of the proposed algorithm for the detection of fissures in the chest CT in terms of the false positive and the true positive measures. The receiver operating characteristic curves indicate the potential of our algorithm as a segmentation tool in the clinical environment. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: anisotropic diffusion, chest CT imagery, chronic respiratory disease, curvilinear structure, fissure detection, structure tensor

Procedia PDF Downloads 208
7080 Evidence on Scale Economies in National Bank of Pakistan

Authors: Sohail Zafar, Sardar Javaid Iqbal Khan

Abstract:

We use a parametric approach within a translog cost function framework to estimate the economies of scale in National Bank of Pakistan from 1997 to 2013. The results indicate significant economies of scale throughout the sample at aggregates and disaggregates taking in account size subject to stipulation ownership. The factor markets often produce scale inefficiencies in the banking of developing countries like Pakistan such inefficiencies are common due to distortion in factor markets leading to the use of inappropriate factor proportions. The findings suggest that National Bank of Pakistan diversify their asset portfolios that it has cost advantage, therefore, expansion in size should be encouraged under current technology because it appears to be cost effective. In addition, our findings support the implementation of universal banking model in Pakistan.

Keywords: scale economies, cost function, disaggregates, aggregates

Procedia PDF Downloads 299
7079 Bayesian Network and Feature Selection for Rank Deficient Inverse Problem

Authors: Kyugneun Lee, Ikjin Lee

Abstract:

Parameter estimation with inverse problem often suffers from unfavorable conditions in the real world. Useless data and many input parameters make the problem complicated or insoluble. Data refinement and reformulation of the problem can solve that kind of difficulties. In this research, a method to solve the rank deficient inverse problem is suggested. A multi-physics system which has rank deficiency caused by response correlation is treated. Impeditive information is removed and the problem is reformulated to sequential estimations using Bayesian network (BN) and subset groups. At first, subset grouping of the responses is performed. Feature selection with singular value decomposition (SVD) is used for the grouping. Next, BN inference is used for sequential conditional estimation according to the group hierarchy. Directed acyclic graph (DAG) structure is organized to maximize the estimation ability. Variance ratio of response to noise is used to pairing the estimable parameters by each response.

Keywords: Bayesian network, feature selection, rank deficiency, statistical inverse analysis

Procedia PDF Downloads 284
7078 Towards Computational Fluid Dynamics Based Methodology to Accelerate Bioprocess Scale Up and Scale Down

Authors: Vishal Kumar Singh

Abstract:

Bioprocess development is a time-constrained activity aimed at harnessing the full potential of culture performance in an ambience that is not natural to cells. Even with the use of chemically defined media and feeds, a significant amount of time is devoted in identifying the apt operating parameters. In addition, the scale-up of these processes is often accompanied by loss of antibody titer and product quality, which further delays the commercialization of the drug product. In such a scenario, the investigation of this disparity of culture performance is done by further experimentation at a smaller scale that is representative of at-scale production bioreactors. These scale-down model developments are also time-intensive. In this study, a computation fluid dynamics-based multi-objective scaling approach has been illustrated to speed up the process transfer. For the implementation of this approach, a transient multiphase water-air system has been studied in Ansys CFX to visualize the air bubble distribution and volumetric mass transfer coefficient (kLa) profiles, followed by the design of experiment based parametric optimization approach to define the operational space. The proposed approach is completely in silico and requires minimum experimentation, thereby rendering a high throughput to the overall process development.

Keywords: bioprocess development, scale up, scale down, computation fluid dynamics, multi-objective, Ansys CFX, design of experiment

Procedia PDF Downloads 50