Search results for: test data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9422

Search results for: test data

8372 Predictions Using Data Mining and Case-based Reasoning: A Case Study for Retinopathy

Authors: Vimala Balakrishnan, Mohammad R. Shakouri, Hooman Hoodeh, Loo, Huck-Soo

Abstract:

Diabetes is one of the high prevalence diseases worldwide with increased number of complications, with retinopathy as one of the most common one. This paper describes how data mining and case-based reasoning were integrated to predict retinopathy prevalence among diabetes patients in Malaysia. The knowledge base required was built after literature reviews and interviews with medical experts. A total of 140 diabetes patients- data were used to train the prediction system. A voting mechanism selects the best prediction results from the two techniques used. It has been successfully proven that both data mining and case-based reasoning can be used for retinopathy prediction with an improved accuracy of 85%.

Keywords: Case-Based Reasoning, Data Mining, Prediction, Retinopathy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3020
8371 Research on the Layout of Ground Control Points in Plain area 1:10000 DLG Production Using POS Technique

Authors: Dong Ming, Chen Haipeng

Abstract:

POS (also been called DGPS/IMU) technique can obtain the Exterior Orientation Elements of aerial photo, so the triangulation and DLG production using POS can save large numbers of ground control points (GCP), and this will improve the produce efficiency of DLG and reduce the cost of collecting GCP. This paper mainly research on POS technique in production of 1:10 000 scale DLG on GCP distribution. We designed 23 kinds of ground control points distribution schemes, using integrated sensor direction method to do the triangulation experiments, based on the results of triangulation, we produce a map with the scale of 1:10 000 and test its accuracy. This paper put forward appropriate GCP distributing schemes by experiments and research above, and made preparations for the application of POS technique on photogrammetry 4D data production.

Keywords: POS, IMU, DGPS, DLG, ground control point, triangulation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1704
8370 Zero Truncated Strict Arcsine Model

Authors: Y. N. Phang, E. F. Loh

Abstract:

The zero truncated model is usually used in modeling count data without zero. It is the opposite of zero inflated model. Zero truncated Poisson and zero truncated negative binomial models are discussed and used by some researchers in analyzing the abundance of rare species and hospital stay. Zero truncated models are used as the base in developing hurdle models. In this study, we developed a new model, the zero truncated strict arcsine model, which can be used as an alternative model in modeling count data without zero and with extra variation. Two simulated and one real life data sets are used and fitted into this developed model. The results show that the model provides a good fit to the data. Maximum likelihood estimation method is used in estimating the parameters.

Keywords: Hurdle models, maximum likelihood estimation method, positive count data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1856
8369 Li-Fi Technology: Data Transmission through Visible Light

Authors: Shahzad Hassan, Kamran Saeed

Abstract:

People are always in search of Wi-Fi hotspots because Internet is a major demand nowadays. But like all other technologies, there is still room for improvement in the Wi-Fi technology with regards to the speed and quality of connectivity. In order to address these aspects, Harald Haas, a professor at the University of Edinburgh, proposed what we know as the Li-Fi (Light Fidelity). Li-Fi is a new technology in the field of wireless communication to provide connectivity within a network environment. It is a two-way mode of wireless communication using light. Basically, the data is transmitted through Light Emitting Diodes which can vary the intensity of light very fast, even faster than the blink of an eye. From the research and experiments conducted so far, it can be said that Li-Fi can increase the speed and reliability of the transfer of data. This paper pays particular attention on the assessment of the performance of this technology. In other words, it is a 5G technology which uses LED as the medium of data transfer. For coverage within the buildings, Wi-Fi is good but Li-Fi can be considered favorable in situations where large amounts of data are to be transferred in areas with electromagnetic interferences. It brings a lot of data related qualities such as efficiency, security as well as large throughputs to the table of wireless communication. All in all, it can be said that Li-Fi is going to be a future phenomenon where the presence of light will mean access to the Internet as well as speedy data transfer.

Keywords: Communication, LED, Li-Fi, Wi-Fi.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2165
8368 Business Rules for Data Warehouse

Authors: Rajeev Kaula

Abstract:

Business rules and data warehouse are concepts and technologies that impact a wide variety of organizational tasks. In general, each area has evolved independently, impacting application development and decision-making. Generating knowledge from data warehouse is a complex process. This paper outlines an approach to ease import of information and knowledge from a data warehouse star schema through an inference class of business rules. The paper utilizes the Oracle database for illustrating the working of the concepts. The star schema structure and the business rules are stored within a relational database. The approach is explained through a prototype in Oracle-s PL/SQL Server Pages.

Keywords: Business Rules, Data warehouse, PL/SQL ServerPages, Relational model, Web Application.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2983
8367 Moderation in Temperature Dependence on Counter Frictional Coefficient and Prevention of Wear of C/C Composites by Synthesizing SiC around Surface and Internal Vacancies

Authors: Noboru Wakamoto, Kiyotaka Obunai, Kazuya Okubo, Toru Fujii

Abstract:

The aim of this study is to moderate the dependence of counter frictional coefficient on temperature between counter surfaces and to reduce the wear of C/C composites at low temperature. To modify the C/C composites, Silica (SiO2) powders were added into phenolic resin for carbon precursor. The preform plate of the precursor of C/C composites was prepared by conventional filament winding method. The C/C composites plates were obtained by carbonizing preform plate at 2200 °C under an argon atmosphere. At that time, the silicon carbides (SiC) were synthesized around the surfaces and the internal vacancies of the C/C composites. The frictional coefficient on the counter surfaces and specific wear volumes of the C/C composites were measured by our developed frictional test machine like pin-on disk type. The XRD indicated that SiC was synthesized in the body of C/C composite fabricated by current method. The results of friction test showed that coefficient of friction of unmodified C/C composites have temperature dependence when the test condition was changed. In contrast, frictional coefficient of the C/C composite modified with SiO2 powders was almost constant at about 0.27 when the temperature condition was changed from Room Temperature (RT) to 300 °C. The specific wear rate decreased from 25×10-6 mm2/N to 0.1×10-6 mm2/N. The observations of the surfaces after friction tests showed that the frictional surface of the modified C/C composites was covered with a film produced by the friction. This study found that synthesizing SiC around surface and internal vacancies of C/C composites was effective to moderate the dependence on the frictional coefficient and reduce to the abrasion of C/C composites.

Keywords: C/C composites, frictional coefficient, SiC, wear.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 827
8366 Authorization of Commercial Communication Satellite Grounds for Promoting Turkish Data Relay System

Authors: Celal Dudak, Aslı Utku, Burak Yağlioğlu

Abstract:

Uninterrupted and continuous satellite communication through the whole orbit time is becoming more indispensable every day. Data relay systems are developed and built for various high/low data rate information exchanges like TDRSS of USA and EDRSS of Europe. In these missions, a couple of task-dedicated communication satellites exist. In this regard, for Turkey a data relay system is attempted to be defined exchanging low data rate information (i.e. TTC) for Earth-observing LEO satellites appointing commercial GEO communication satellites all over the world. First, justification of this attempt is given, demonstrating duration enhancements in the link. Discussion of preference of RF communication is, also, given instead of laser communication. Then, preferred communication GEOs – including TURKSAT4A already belonging to Turkey- are given, together with the coverage enhancements through STK simulations and the corresponding link budget. Also, a block diagram of the communication system is given on the LEO satellite.

Keywords: Communication, satellite, data relay system, coverage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1415
8365 An Efficient Approach to Mining Frequent Itemsets on Data Streams

Authors: Sara Ansari, Mohammad Hadi Sadreddini

Abstract:

The increasing importance of data stream arising in a wide range of advanced applications has led to the extensive study of mining frequent patterns. Mining data streams poses many new challenges amongst which are the one-scan nature, the unbounded memory requirement and the high arrival rate of data streams. In this paper, we propose a new approach for mining itemsets on data stream. Our approach SFIDS has been developed based on FIDS algorithm. The main attempts were to keep some advantages of the previous approach and resolve some of its drawbacks, and consequently to improve run time and memory consumption. Our approach has the following advantages: using a data structure similar to lattice for keeping frequent itemsets, separating regions from each other with deleting common nodes that results in a decrease in search space, memory consumption and run time; and Finally, considering CPU constraint, with increasing arrival rate of data that result in overloading system, SFIDS automatically detect this situation and discard some of unprocessing data. We guarantee that error of results is bounded to user pre-specified threshold, based on a probability technique. Final results show that SFIDS algorithm could attain about 50% run time improvement than FIDS approach.

Keywords: Data stream, frequent itemset, stream mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1418
8364 Studying the Effect of Climate Change on the Conditions of Isfahan-s Province Tourism

Authors: A.Gandomkar, F. Khorasanizadeh

Abstract:

Tourism is a phenomenon respected by the human communities since a long time ago. It has been evoloving continually based on a variety of social and economic needs and with respect to increasingly development of communication and considerable increase of tourist-s number and resulted exchange income has attained much out come such as employment for the communities. or the purpose of tourism development in this zone suitable times and locations need to be specified in the zone for the tourist-s attendance. One of the most important needs of the tourists is the knowledge of climate conditions and suitable times for sightseeing. In this survey, the climate trend condition has been identified for attending the tourists in Isfahan province using the modified tourism climate index (TCI) as well as SPSS, GIS, excel, surfer softwares. This index evoluates systematically the climate conditions for tourism affairs and activities using the monthly maximum mean parameters of daily temperature, daily mean temperature, minimum relative humidity, daily mean relative humidity, precipitation (mm), total sunny hours, wind speed and dust. The results obtaind using kendal-s correlation test show that the months January, February, March, April, May, June, July, August, September, October, November and December are significant and have an increasing trend that indicates the best condition for attending the tourists. S, P, T mean , T max and dust are estimated from 1976-2005 and do kendal-s correlation test again to see which parameter has been effective. Based on the test, we also observed on the effective parameters that the rate of dust in February, March, April, May, June, July, August, October and November is decreasing and precipitation in September and January is increasing and also the radiation rate in May and August is increasing that indicate a better condition of convenience. Maximum temperature in June is also decreasing. Isfahan province has two spring and fall peaks and the best places for tourism are in the north and western areas.

Keywords: Climate, Tourism, Correlation Test, Tourism Climate Index, Isfahan Province

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1698
8363 Transmission Pricing based on Voltage Angle Decomposition

Authors: M. Oloomi-Buygi, M. Reza Salehizadeh

Abstract:

In this paper a new approach for transmission pricing is presented. The main idea is voltage angle allocation, i.e. determining the contribution of each contract on the voltage angle of each bus. DC power flow is used to compute a primary solution for angle decomposition. To consider the impacts of system non-linearity on angle decomposition, the primary solution is corrected in different iterations of decoupled Newton-Raphson power flow. Then, the contribution of each contract on power flow of each transmission line is computed based on angle decomposition. Contract-related flows are used as a measure for “extent of use" of transmission network capacity and consequently transmission pricing. The presented approach is applied to a 4-bus test system and IEEE 30-bus test system.

Keywords: Deregulation, Power electric markets, Transmission pricing methodologies, decoupled Newton-Raphson power flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1661
8362 Study Habits and Level of Difficulty Encountered by Maltese Students Studying Biology Advanced Level Topics

Authors: Marthese Azzopardi, Liberato Camilleri

Abstract:

This research was performed to investigate the study habits and level of difficulty perceived by post-secondary students in Biology at Advanced-level topics after completing their first year of study. At the end of a two-year ‘sixth form’ course, Maltese students sit for the Matriculation and Secondary Education Certificate (MATSEC) Advanced-level biology exam as a requirement to pursue science-related studies at the University of Malta. The sample was composed of 23 students (16 taking Chemistry and seven taking some ‘Other’ subject at the Advanced Level). The cohort comprised seven males and 16 females. A questionnaire constructed by the authors, was answered anonymously during the last lecture at the end of the first year of study, in May 2016. The Chi square test revealed that gender plays no effect on the various study habits (c2 (6) = 5.873, p = 0.438). ‘Reading both notes and textbooks’ was the most common method adopted by males (71.4%), whereas ‘Writing notes on each topic’ was that mostly used by females (81.3%). The Mann-Whitney U test showed no significant difference in the study habits of students and the mean assessment mark obtained at the end of the first year course (p = 0.231). Statistical difference was found with the One-ANOVA test when comparing the mean assessment mark obtained at the end of the first year course when students are clustered by their Secondary Education Certificate (SEC) grade (p < 0.001). Those obtaining a SEC grade of 2 and 3 got the highest mean assessment of 68.33% and 66.9%, respectively [SEC grading is 1-7, where 1 is the highest]. The Friedman test was used to compare the mean difficulty rating scores provided for the difficulty of each topic. The mean difficulty rating score ranges from 1 to 4, where the larger the mean rating score, the higher the difficulty. When considering the whole group of students, nine topics out of 21 were perceived as significantly more difficult than the other topics. Protein synthesis, DNA Replication and Biomolecules were the most difficult, in that order. The Mann-Whitney U test revealed that the perceived level of difficulty in comprehending Biomolecules is significantly lower for students taking Chemistry compared to those not choosing the subject (p = 0.018). Protein Synthesis was claimed as the most difficult by Chemistry students and Biomolecules by those not studying Chemistry. DNA Replication was the second most difficult topic perceived by both groups. The Mann-Whitney U test was used to examine the effect of gender on the perceived level of difficulty in comprehending various topics. It was found that females have significantly more difficulty in comprehending Biomolecules than males (p=0.039). Protein synthesis was perceived as the most difficult topic by males (mean difficulty rating score = 3.14), while Biomolecules, DNA Replication and Protein synthesis were of equal difficulty for females (mean difficulty rating score = 3.00). Males and females perceived DNA Replication as equally difficult (mean difficulty rating score = 3.00). Discovering the students’ study habits and perceived level of difficulty of specific topics is vital for the lecturer to offer guidance that leads to higher academic achievement.

Keywords: Biology, Perceived difficulty, Post-secondary, Study habits.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1366
8361 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model

Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin

Abstract:

Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.

Keywords: Anomaly detection, autoencoder, data centers, deep learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 741
8360 A Multi-Level GA Search with Application to the Resource-Constrained Re-Entrant Flow Shop Scheduling Problem

Authors: Danping Lin, C.K.M. Lee

Abstract:

Re-entrant scheduling is an important search problem with many constraints in the flow shop. In the literature, a number of approaches have been investigated from exact methods to meta-heuristics. This paper presents a genetic algorithm that encodes the problem as multi-level chromosomes to reflect the dependent relationship of the re-entrant possibility and resource consumption. The novel encoding way conserves the intact information of the data and fastens the convergence to the near optimal solutions. To test the effectiveness of the method, it has been applied to the resource-constrained re-entrant flow shop scheduling problem. Computational results show that the proposed GA performs better than the simulated annealing algorithm in the measure of the makespan

Keywords: Resource-constrained, re-entrant, genetic algorithm (GA), multi-level encoding

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1791
8359 Carotenoid Potential to Protect Cow-s Milk Fat Against Oxidative Deterioration

Authors: U. Antone, V. Sterna, J. Zagorska

Abstract:

Milk from differently fed cows (supplemented with carotenoids from carrots or palm oil product Carotino CAF 100) was obtained in a conventional dairy farm to assess the carotenoid potential to protect milk fat against oxidation. The extracted anhydrous milk fat (AMF) was tested by peroxide value, and Rancimat tests. Temperature, and light stimulation for reaction acceleration was used. The oxidative stability enhancement by carotenoids was detected in peroxide value test – the strongest effect was observed in palm oil, following by carrot supplemented group, compared to control group, whose feed was unchanged. Rancimat accelerated oxidation test results did not show any superiority of the oxidative stability of the AMF samples from milk of the carotenoidsupplemented cow groups. The average oxidation stability of AMF dark-stored samples was 12.59 ± 0.294 h, and it was significantly (p < 0.05) higher than that of AMF light-affected samples, i.e. 2.60 ± 0.191 h.

Keywords: antioxidants, dairy products, forages, lipid aging, peroxide

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2119
8358 AnQL: A Query Language for Annotation Documents

Authors: Neerja Bhatnagar, Ben A. Juliano, Renee S. Renner

Abstract:

This paper presents data annotation models at five levels of granularity (database, relation, column, tuple, and cell) of relational data to address the problem of unsuitability of most relational databases to express annotations. These models do not require any structural and schematic changes to the underlying database. These models are also flexible, extensible, customizable, database-neutral, and platform-independent. This paper also presents an SQL-like query language, named Annotation Query Language (AnQL), to query annotation documents. AnQL is simple to understand and exploits the already-existent wide knowledge and skill set of SQL.

Keywords: Annotation query language, data annotations, data annotation models, semantic data annotations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1841
8357 Machine Learning-Enabled Classification of Climbing Using Small Data

Authors: Nicholas Milburn, Yu Liang, Dalei Wu

Abstract:

Athlete performance scoring within the climbing domain presents interesting challenges as the sport does not have an objective way to assign skill. Assessing skill levels within any sport is valuable as it can be used to mark progress while training, and it can help an athlete choose appropriate climbs to attempt. Machine learning-based methods are popular for complex problems like this. The dataset available was composed of dynamic force data recorded during climbing; however, this dataset came with challenges such as data scarcity, imbalance, and it was temporally heterogeneous. Investigated solutions to these challenges include data augmentation, temporal normalization, conversion of time series to the spectral domain, and cross validation strategies. The investigated solutions to the classification problem included light weight machine classifiers KNN and SVM as well as the deep learning with CNN. The best performing model had an 80% accuracy. In conclusion, there seems to be enough information within climbing force data to accurately categorize climbers by skill.

Keywords: Classification, climbing, data imbalance, data scarcity, machine learning, time sequence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 565
8356 Assessment of Influence of Short-Lasting Whole-Body Vibration on Joint Position Sense and Body Balance–A Randomised Masked Study

Authors: Anna Słupik, Anna Mosiołek, Sebastian Wójtowicz, Dariusz Białoszewski

Abstract:

Introduction: Whole-Body Vibration (WBV) uses high frequency mechanical stimuli generated by a vibration plate and transmitted through bone, muscle and connective tissues to the whole body. Research has shown that long-term vibration-plate training improves neuromuscular facilitation, especially in afferent neural pathways, responsible for the conduction of vibration and proprioceptive stimuli, muscle function, balance and proprioception. Some researchers suggest that the vibration stimulus briefly inhibits the conduction of afferent signals from proprioceptors and can interfere with the maintenance of body balance. The aim of this study was to evaluate the influence of a single set of exercises associated with whole-body vibration on the joint position sense and body balance. Material and methods: The study enrolled 55 people aged 19-24 years. These individuals were randomly divided into a test group (30 persons) and a control group (25 persons). Both groups performed the same set of exercises on a vibration plate. The following vibration parameters: frequency of 20Hz and amplitude of 3mm, were used in the test group. The control group performed exercises on the vibration plate while it was off. All participants were instructed to perform six dynamic exercises lasting 30 seconds each with a 60-second period of rest between them. The exercises involved large muscle groups of the trunk, pelvis and lower limbs. Measurements were carried out before and immediately after exercise. Joint position sense (JPS) was measured in the knee joint for the starting position at 45° in an open kinematic chain. JPS error was measured using a digital inclinometer. Balance was assessed in a standing position with both feet on the ground with the eyes open and closed (each test lasting 30 sec). Balance was assessed using Matscan with FootMat 7.0 SAM software. The surface of the ellipse of confidence and front-back as well as right-left swing were measured to assess balance. Statistical analysis was performed using Statistica 10.0 PL software. Results: There were no significant differences between the groups, both before and after the exercise (p> 0.05). JPS did not change in both the test (10.7° vs. 8.4°) and control groups (9.0° vs. 8.4°). No significant differences were shown in any of the test parameters during balance tests with the eyes open or closed in both the test and control groups (p> 0.05). Conclusions: 1. Deterioration in proprioception or balance was not observed immediately after the vibration stimulus. This suggests that vibrationinduced blockage of proprioceptive stimuli conduction can have only a short-lasting effect that occurs only as long as a vibration stimulus is present. 2. Short-term use of vibration in treatment does not impair proprioception and seems to be safe for patients with proprioceptive impairment. 3. These results need to be supplemented with an assessment of proprioception during the application of vibration stimuli. Additionally, the impact of vibration parameters used in the exercises should be evaluated.

Keywords: Balance, joint position sense, proprioception, whole body vibration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1608
8355 Cardiopulmonary Exercise Testing in Young Asthmatic Children Ages 6-10 Years Old

Authors: Yen-Ting Wang, Kenny Wen-Chyuan Chen, I-Tsun Chiang, Lung-Ching Liang, Alex J.Y. Lee

Abstract:

The aim of this study was to establish the feasibility of a minute incremental exercise testing protocol in young asthma children. Twenty-two children with clinically diagnosed mild to moderate asthma volunteered to participate. The maximum incremental exercise test was performed using a cycle ergometer with an electromagnetic braking. A warm-up unloaded for 2 minutes then the workload was started at 40 watts for 2 minutes, and then stepwise increments of 8 watts per 2 minutes were applied. The pedaling frequency was set at 50 rpm. Ventilation and gas exchange were measured with a breath-by-breath automatic metabolic measurement system. Results showed that this test was well tolerated by all asthmatic children. Most of the children reached the VO2 plateau and satisfied the criteria for maximal respiratory exchange ratio of ≥ 1. This Study demonstrated that this testing protocol was suitable for young asthmatic children.

Keywords: Asthma, Child, Exercise, Pediatrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2519
8354 Wavelet - Based Classification of Outdoor Natural Scenes by Resilient Neural Network

Authors: Amitabh Wahi, Sundaramurthy S.

Abstract:

Natural outdoor scene classification is active and promising research area around the globe. In this study, the classification is carried out in two phases. In the first phase, the features are extracted from the images by wavelet decomposition method and stored in a database as feature vectors. In the second phase, the neural classifiers such as back-propagation neural network (BPNN) and resilient back-propagation neural network (RPNN) are employed for the classification of scenes. Four hundred color images are considered from MIT database of two classes as forest and street. A comparative study has been carried out on the performance of the two neural classifiers BPNN and RPNN on the increasing number of test samples. RPNN showed better classification results compared to BPNN on the large test samples.

Keywords: BPNN, Classification, Feature extraction, RPNN, Wavelet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1941
8353 One-Class Support Vector Machine for Sentiment Analysis of Movie Review Documents

Authors: Chothmal, Basant Agarwal

Abstract:

Sentiment analysis means to classify a given review document into positive or negative polar document. Sentiment analysis research has been increased tremendously in recent times due to its large number of applications in the industry and academia. Sentiment analysis models can be used to determine the opinion of the user towards any entity or product. E-commerce companies can use sentiment analysis model to improve their products on the basis of users’ opinion. In this paper, we propose a new One-class Support Vector Machine (One-class SVM) based sentiment analysis model for movie review documents. In the proposed approach, we initially extract features from one class of documents, and further test the given documents with the one-class SVM model if a given new test document lies in the model or it is an outlier. Experimental results show the effectiveness of the proposed sentiment analysis model.

Keywords: Feature selection methods, Machine learning, NB, One-class SVM, Sentiment Analysis, Support Vector Machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3303
8352 FCNN-MR: A Parallel Instance Selection Method Based on Fast Condensed Nearest Neighbor Rule

Authors: Lu Si, Jie Yu, Shasha Li, Jun Ma, Lei Luo, Qingbo Wu, Yongqi Ma, Zhengji Liu

Abstract:

Instance selection (IS) technique is used to reduce the data size to improve the performance of data mining methods. Recently, to process very large data set, several proposed methods divide the training set into some disjoint subsets and apply IS algorithms independently to each subset. In this paper, we analyze the limitation of these methods and give our viewpoint about how to divide and conquer in IS procedure. Then, based on fast condensed nearest neighbor (FCNN) rule, we propose a large data sets instance selection method with MapReduce framework. Besides ensuring the prediction accuracy and reduction rate, it has two desirable properties: First, it reduces the work load in the aggregation node; Second and most important, it produces the same result with the sequential version, which other parallel methods cannot achieve. We evaluate the performance of FCNN-MR on one small data set and two large data sets. The experimental results show that it is effective and practical.

Keywords: Instance selection, data reduction, MapReduce, kNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1015
8351 Optimization of Real Time Measured Data Transmission, Given the Amount of Data Transmitted

Authors: Michal Kopcek, Tomas Skulavik, Michal Kebisek, Gabriela Krizanova

Abstract:

The operation of nuclear power plants involves continuous monitoring of the environment in their area. This monitoring is performed using a complex data acquisition system, which collects status information about the system itself and values of many important physical variables e.g. temperature, humidity, dose rate etc. This paper describes a proposal and optimization of communication that takes place in teledosimetric system between the central control server responsible for the data processing and storing and the decentralized measuring stations, which are measuring the physical variables. Analyzes of ongoing communication were performed and consequently the optimization of the system architecture and communication was done.

Keywords: Communication protocol, transmission optimization, data acquisition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1819
8350 Integrating Bioremediation and Phytoremediation to Clean up Polychlorinated Biphenyls Contaminated Soils

Authors: Petruzzelli G., Pedron F., Rosellini I., Tassi E., Gorini F., Barbafieri M.

Abstract:

This work involved the use of phytoremediation to remediate an aged soil contaminated with polychlorinated biphenyls (PCBs). At microcosm scale, tests were prepared using soil samples that have been collected in an industrial area with a total PCBs concentration of about 250 μg kg-1. Medicago sativa and Lolium italicum were the species selected in this study that is used as “feasibility test" for full scale remediation. The experiment was carried out with the addition of a mixture of randomly methylatedbeta- cyclodextrins (RAMEB). At the end of the experiment analysis of soil samples showed that in general the presence of plants has led to a higher degradation of most congeners with respect to not vegetated soil. The two plant species efficiencies were comparable and improved by RAMEB addition with a final reduction of total PCBs near to 50%. With increasing the chlorination of the congeners the removal percentage of PCBs progressively decreased.

Keywords: contaminated soil, feasibility test, phytoremediation, polychlorinated biphenyls

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1799
8349 Creative Mapping Landuse and Human Activities: From the Inventories of Factories to the History of the City and Citizens

Authors: R. Tamborrino, F. Rinaudo

Abstract:

Digital technologies offer possibilities to effectively convert historical archives into instruments of knowledge able to provide a guide for the interpretation of historical phenomena. Digital conversion and management of those documents allow the possibility to add other sources in a unique and coherent model that permits the intersection of different data able to open new interpretations and understandings. Urban history uses, among other sources, the inventories that register human activities in a specific space (e.g. cadastres, censuses, etc.). The geographic localisation of that information inside cartographic supports allows for the comprehension and visualisation of specific relationships between different historical realities registering both the urban space and the peoples living there. These links that merge the different nature of data and documentation through a new organisation of the information can suggest a new interpretation of other related events. In all these kinds of analysis, the use of GIS platforms today represents the most appropriate answer. The design of the related databases is the key to realise the ad-hoc instrument to facilitate the analysis and the intersection of data of different origins. Moreover, GIS has become the digital platform where it is possible to add other kinds of data visualisation. This research deals with the industrial development of Turin at the beginning of the 20th century. A census of factories realized just prior to WWI provides the opportunity to test the potentialities of GIS platforms for the analysis of urban landscape modifications during the first industrial development of the town. The inventory includes data about location, activities, and people. GIS is shaped in a creative way linking different sources and digital systems aiming to create a new type of platform conceived as an interface integrating different kinds of data visualisation. The data processing allows linking this information to an urban space, and also visualising the growth of the city at that time. The sources, related to the urban landscape development in that period, are of a different nature. The emerging necessity to build, enlarge, modify and join different buildings to boost the industrial activities, according to their fast development, is recorded by different official permissions delivered by the municipality and now stored in the Historical Archive of the Municipality of Turin. Those documents, which are reports and drawings, contain numerous data on the buildings themselves, including the block where the plot is located, the district, and the people involved such as the owner, the investor, and the engineer or architect designing the industrial building. All these collected data offer the possibility to firstly re-build the process of change of the urban landscape by using GIS and 3D modelling technologies thanks to the access to the drawings (2D plans, sections and elevations) that show the previous and the planned situation. Furthermore, they access information for different queries of the linked dataset that could be useful for different research and targets such as economics, biographical, architectural, or demographical. By superimposing a layer of the present city, the past meets to the present-industrial heritage, and people meet urban history.

Keywords: Digital urban history, census, digitalisation, GIS, modelling, digital humanities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1226
8348 System Reliability by Prediction of Generator Output and Losses in a Competitive Energy Market

Authors: Perumal Nallagownden, Ravindra N. Mukerjee, Syafrudin Masri

Abstract:

In a competitive energy market, system reliability should be maintained at all times. Power system operation being of online in nature, the energy balance requirements must be satisfied to ensure reliable operation the system. To achieve this, information regarding the expected status of the system, the scheduled transactions and the relevant inputs necessary to make either a transaction contract or a transmission contract operational, have to be made available in real time. The real time procedure proposed, facilitates this. This paper proposes a quadratic curve learning procedure, which enables a generator-s contribution to the retailer demand, power loss of transaction in a line at the retail end and its associated losses for an oncoming operating scenario to be predicted. Matlab program was used to test in on a 24-bus IEE Reliability Test System, and the results are found to be acceptable.

Keywords: Deregulation, learning coefficients, reliability, prediction, competitive energy market.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1477
8347 Empirical Process Monitoring Via Chemometric Analysis of Partially Unbalanced Data

Authors: Hyun-Woo Cho

Abstract:

Real-time or in-line process monitoring frameworks are designed to give early warnings for a fault along with meaningful identification of its assignable causes. In artificial intelligence and machine learning fields of pattern recognition various promising approaches have been proposed such as kernel-based nonlinear machine learning techniques. This work presents a kernel-based empirical monitoring scheme for batch type production processes with small sample size problem of partially unbalanced data. Measurement data of normal operations are easy to collect whilst special events or faults data are difficult to collect. In such situations, noise filtering techniques can be helpful in enhancing process monitoring performance. Furthermore, preprocessing of raw process data is used to get rid of unwanted variation of data. The performance of the monitoring scheme was demonstrated using three-dimensional batch data. The results showed that the monitoring performance was improved significantly in terms of detection success rate of process fault.

Keywords: Process Monitoring, kernel methods, multivariate filtering, data-driven techniques, quality improvement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1745
8346 Calcium Silicate Bricks – Ultrasonic Pulse Method: Effects of Natural Frequency of Transducers on Measurement Results

Authors: Jiri Brozovsky

Abstract:

Modulus of elasticity is one of the important parameters of construction materials, which considerably influence their deformation properties and which can also be determined by means of non-destructive test methods like ultrasonic pulse method. However, measurement results of ultrasonic pulse methods are influenced by various factors, one of which is the natural frequency of the transducers. The paper states knowledge about influence of natural frequency of the transducers (54; 82 and 150kHz) on ultrasonic pulse velocity and dynamic modulus of elasticity (Young's Dynamic modulus of elasticity). Differences between ultrasonic pulse velocity and dynamic modulus of elasticity were found with the same smallest dimension of test specimen in the direction of sounding and density their value decreases as the natural frequency of transducers grew.

Keywords: Calcium silicate brick, ultrasonic pulse method, ultrasonic pulse velocity, dynamic modulus of elasticity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2244
8345 Owner/Managers’ External Financing Used and Preference towards Islamic Banking

Authors: Khalid Hassan Abdesamed, Kalsom Abd Wahab

Abstract:

Economic development and growth are significantly linked to the consistent and sustainable sector of small and medium enterprises (SMEs). Banks are the frontrunners in financing and advising SMEs. The main objective of the study is to assess the tendency of SMEs to use the Islamic bank. Model was developed using quantitative method with a hypothetical-deductive testing approach. Model (N = 364) used primary data on the tendency of SMEs to use Islamic banks gathered from questionnaire. It is found by Mann-Whitney test that the tendency to use Islamic bank varies between those firms which consider formal financing with the ones relying on informal financing with the latter tends more to use Islamic bank. This study can serve academic researchers, policy makers, and developing countries as a model of SMEs’ desirability to Islamic banking.

Keywords: Formal financing, informal financing, Islamic bank, SMEs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1665
8344 Applying Theory of Perceived Risk and Technology Acceptance Model in the Online Shopping Channel

Authors: Yong-Hui Li, Jing-Wen Huang

Abstract:

As the advancement of technology, online shopping channel develops rapidly in recent years. According to the report of Taiwan Network Information Center, there are almost eighty percents of internet population shopping in online channel. Synthesizing insights from the previous research, this study develops the conceptual model to integrate Theory of Perceived Risk (TPR) and Technology Acceptance Model (TAM) to apply in online shopping. Using data collected from 637 respondents from online survey website, we use structural equation modeling to test measurement and structural models. The results suggest the need for consideration of perceived risk as an antecedent in the Technology Acceptance Model. The limitations and implications are discussed.

Keywords: perceived risk, perceived usefulness, perceived ease of use, behavioral intention, actual purchase behavior

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6433
8343 A Comparison of Image Data Representations for Local Stereo Matching

Authors: André Smith, Amr Abdel-Dayem

Abstract:

The stereo matching problem, while having been present for several decades, continues to be an active area of research. The goal of this research is to find correspondences between elements found in a set of stereoscopic images. With these pairings, it is possible to infer the distance of objects within a scene, relative to the observer. Advancements in this field have led to experimentations with various techniques, from graph-cut energy minimization to artificial neural networks. At the basis of these techniques is a cost function, which is used to evaluate the likelihood of a particular match between points in each image. While at its core, the cost is based on comparing the image pixel data; there is a general lack of consistency as to what image data representation to use. This paper presents an experimental analysis to compare the effectiveness of more common image data representations. The goal is to determine the effectiveness of these data representations to reduce the cost for the correct correspondence relative to other possible matches.

Keywords: Colour data, local stereo matching, stereo correspondence, disparity map.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 915