Search results for: teaching-learning based optimization
9055 A Performance Evaluation of Oscillation Based Test in Continuous Time Filters
Authors: Eduardo Romero, Marcelo Costamagna, Gabriela Peretti, Carlos Marqués
Abstract:
This work evaluates the ability of OBT for detecting parametric faults in continuous-time filters. To this end, we adopt two filters with quite different topologies as cases of study and a previously reported statistical fault model. In addition, we explore the behavior of the test schemes when a particular test condition is changed. The new data reported here, obtained from a fault simulation process, reveal a lower performance of OBT not observed in previous work using single-deviation faults, even under the change in the test condition.
Keywords: Testing, analog fault simulation, analog filter test, oscillation based test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14449054 Miller’s Model for Developing Critical Thinking Skill of Pre-Service Teachers at Suan Sunandha Rajabhat University
Authors: Suttipong Boonphadung, Thassanant Unnanantn
Abstract:
This research focused on comparing the critical thinking of the teacher students before and after using Miller’s Model learning activities and investigating their opinions. The sampling groups were (1) fourth year 33 student teachers majoring in Early Childhood Education and enrolling in semester 1 of academic year 2013 (2) third year 28 student teachers majoring in English and enrolling in semester 2 of academic year 2013 and (3) third year 22 student teachers majoring in Thai and enrolling in semester 2 of academic year 2013. The research instruments were (1) lesson plans where the learning activities were settled based on Miller’s Model (2) critical thinking assessment criteria and (3) a questionnaire on opinions towards Miller’s Model based learning activities. The statistical treatment was mean, deviation, different scores and T-test. The result unfolded that (1) the critical thinking of the students after the assigned activities was better than before and (2) the students’ opinions towards the critical thinking improvement activities based on Miller’s Model ranged from the level of high to highest.
Keywords: Critical thinking, Miller’s model, Opinions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20709053 Advanced Image Analysis Tools Development for the Early Stage Bronchial Cancer Detection
Authors: P. Bountris, E. Farantatos, N. Apostolou
Abstract:
Autofluorescence (AF) bronchoscopy is an established method to detect dysplasia and carcinoma in situ (CIS). For this reason the “Sotiria" Hospital uses the Karl Storz D-light system. However, in early tumor stages the visualization is not that obvious. With the help of a PC, we analyzed the color images we captured by developing certain tools in Matlab®. We used statistical methods based on texture analysis, signal processing methods based on Gabor models and conversion algorithms between devicedependent color spaces. Our belief is that we reduced the error made by the naked eye. The tools we implemented improve the quality of patients' life.Keywords: Bronchoscopy, digital image processing, lung cancer, texture analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14339052 The Consumer Private Space: What is and How it can be Approached without Affecting the Consumer's Privacy
Authors: Calin Veghes
Abstract:
The concept of privacy, seen in connection to the consumer's private space and personalization, has recently gained a higher importance as a consequence of the increasing marketing efforts of the organizations based on the capturing, processing and usage of consumer-s personal data.Paper intends to provide a definition of the consumer-s private space based on the types of personal data the consumer is willing to disclose, to assess the attitude toward personalization and to identify the means preferred by consumers to control their personal data and defend their private space. Several implications generated through the definition of the consumer-s private space are identified and weighted from both the consumers- and organizations- perspectives.
Keywords: Consumer private space, personalization, privacy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15669051 Spike Sorting Method Using Exponential Autoregressive Modeling of Action Potentials
Authors: Sajjad Farashi
Abstract:
Neurons in the nervous system communicate with each other by producing electrical signals called spikes. To investigate the physiological function of nervous system it is essential to study the activity of neurons by detecting and sorting spikes in the recorded signal. In this paper a method is proposed for considering the spike sorting problem which is based on the nonlinear modeling of spikes using exponential autoregressive model. The genetic algorithm is utilized for model parameter estimation. In this regard some selected model coefficients are used as features for sorting purposes. For optimal selection of model coefficients, self-organizing feature map is used. The results show that modeling of spikes with nonlinear autoregressive model outperforms its linear counterpart. Also the extracted features based on the coefficients of exponential autoregressive model are better than wavelet based extracted features and get more compact and well-separated clusters. In the case of spikes different in small-scale structures where principal component analysis fails to get separated clouds in the feature space, the proposed method can obtain well-separated cluster which removes the necessity of applying complex classifiers.
Keywords: Exponential autoregressive model, Neural data, spike sorting, time series modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17709050 Development of Coronal Field and Solar Wind Components for MHD Interplanetary Simulations
Authors: Ljubomir Nikolic, Larisa Trichtchenko
Abstract:
The connection between solar activity and adverse phenomena in the Earth’s environment that can affect space and ground based technologies has spurred interest in Space Weather (SW) research. A great effort has been put on the development of suitable models that can provide advanced forecast of SW events. With the progress in computational technology, it is becoming possible to develop operational large scale physics based models which can incorporate the most important physical processes and domains of the Sun-Earth system. In order to enhance our SW prediction capabilities we are developing advanced numerical tools. With operational requirements in mind, our goal is to develop a modular simulation framework of propagation of the disturbances from the Sun through interplanetary space to the Earth. Here, we report and discuss on the development of coronal field and solar wind components for a large scale MHD code. The model for these components is based on a potential field source surface model and an empirical Wang-Sheeley-Arge solar wind relation.
Keywords: Space weather, numerical modeling, coronal field, solar wind.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21399049 Statistical Properties and Performance of Ecological Indices Based On Relative Abundances
Authors: Gebriel M. Shamia
Abstract:
The Improved Generalized Diversity Index (IGDI) has been proposed as a tool that can be used to identify areas that have high conservation value and measure the ecological condition of an area. IGDI is based on the species relative abundances. This paper is concerned with particular attention is given to comparisons involving the MacArthur model of species abundances. The properties and performance of various species indices were assessed. Both IGDI and species richness increased with sampling area according to a power function. IGDI were also found to be acceptable ecological indicators of conditions and consistently outperformed coefficient of conservatism indices.Keywords: Statistical ecology, MacArthur model, Functional Diversity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20289048 Disinfection of Water by Adsorption with Electrochemical Regeneration
Authors: S. N. Hussain, H. M. A. Asghar, E. P. L. Roberts, N. W. Brown
Abstract:
Arvia®, a spin-out company of University of Manchester, UK is commercialising a water treatment technology for the removal of low concentrations of organics from water. This technology is based on the adsorption of organics onto graphite based adsorbents coupled with their electrochemical regeneration in a simple electrochemical cell. In this paper, the potential of the process to adsorb microorganisms and electrochemically disinfect them present in water has been demonstrated. Bench scale experiments have indicated that the process of adsorption using graphite adsorbents with electrochemical regeneration can be used for water disinfection effectively. The most likely mechanisms of disinfection of water through this process include direct electrochemical oxidation and electrochemical chlorination.Keywords: Arvia, Adsorption, Electrochemical Regeneration, Nyex
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21599047 Decolorization and COD Reduction Efficiency of Magnesium over Iron based Salt for the Treatment of Textile Wastewater Containing Diazo and Anthraquinone Dyes
Authors: Akshaya Kumar Verma, Puspendu Bhunia*, Rajesh Roshan Dash
Abstract:
Magnesium chloride, though cost wise roughly same as of ferrous sulphate, is less commonly used coagulant in comparison to the ferrous sulphate for the treatment of wastewater. The present study was conducted to investigate the comparative effectiveness of ferrous sulphate (FeSO4.7H2O) as iron based salt and magnesium chloride (MgCl2) as magnesium based salt in terms of decolorization and chemical oxygen demand (COD) reduction efficiency of textile wastewater. The coagulants were evaluated for synthetic textile wastewater containing two diazo dyes namely Reactive Black 5 (RB5) and Congo Red (CR) and one anthraquinone dye as Disperse Blue 3 (DB3), in seven possible equi-ratio combinations. Other chemical constituents that are normally released from different textile processing units were also added to replicate a practical scenario. From this study, MgCl2/Lime was found to be a superior coagulant system as compared to FeSO4.7H2O/Lime, FeSO4.7H2O/NaOH and MgCl2/NaOH.
Keywords: Coagulation, Color removal, Magnesium chloride, Textile wastewater
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34309046 Decentralised Edge Authentication in the Industrial Enterprise IoT Space
Authors: C. P. Autry, A.W. Roscoe
Abstract:
Authentication protocols based on public key infrastructure (PKI) and trusted third party (TTP) are no longer adequate for industrial scale IoT networks thanks to issues such as low compute and power availability, the use of widely distributed and commercial off-the-shelf (COTS) systems, and the increasingly sophisticated attackers and attacks we now have to counter. For example, there is increasing concern about nation-state-based interference and future quantum computing capability. We have examined this space from first principles and have developed several approaches to group and point-to-point authentication for IoT that do not depend on the use of a centralised client-server model. We emphasise the use of quantum resistant primitives such as strong cryptographic hashing and the use multi-factor authentication.
Keywords: Authentication, enterprise IoT cybersecurity, public key infrastructure, trusted third party.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4729045 Response of the Residential Building Structureon Load Technical Seismicity due to Mining Activities
Authors: V. Salajka, Z. Kaláb, J. Kala, P. Hradil
Abstract:
In the territories where high-intensity earthquakes are frequent is paid attention to the solving of the seismic problems. In the paper are described two computational model variants based on finite element method of the construction with different subsoil simulation (rigid or elastic subsoil) is used. For simulation and calculations program system based on method final elements ANSYS was used. Seismic responses calculations of residential building structure were effected on loading characterized by accelerogram for comparing with the responses spectra method.Keywords: Accelerogram, ANSYS, mining induced seismic, residential building structure, spectra, subsoil.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15389044 ANN-Based Classification of Indirect Immuno Fluorescence Images
Authors: P. Soda, G.Iannello
Abstract:
In this paper we address the issue of classifying the fluorescent intensity of a sample in Indirect Immuno-Fluorescence (IIF). Since IIF is a subjective, semi-quantitative test in its very nature, we discuss a strategy to reliably label the image data set by using the diagnoses performed by different physicians. Then, we discuss image pre-processing, feature extraction and selection. Finally, we propose two ANN-based classifiers that can separate intrinsically dubious samples and whose error tolerance can be flexibly set. Measured performance shows error rates less than 1%, which candidates the method to be used in daily medical practice either to perform pre-selection of cases to be examined, or to act as a second reader.
Keywords: Artificial neural networks, computer aided diagnosis, image classification, indirect immuno-fluorescence, pattern recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15699043 Enhancing the Performance of Wireless Sensor Networks Using Low Power Design
Authors: N. Mahendran, R. Madhuranthi
Abstract:
Wireless sensor networks (WSNs), are constantly in demand to process information more rapidly with less energy and area cost. Presently, processor based solutions have difficult to achieve high processing speed with low-power consumption. This paper presents a simple and accurate data processing scheme for low power wireless sensor node, based on reduced number of processing element (PE). The presented model provides a simple recursive structure (SRS) to process the sampled data in the wireless sensor environment and to reduce the power consumption in wireless sensor node. Based on this model, to process the incoming samples and produce a smaller amount of data sufficient to reconstruct the original signal. The ModelSim simulator used to simulate SRS structure. Functional simulation is carried out for the validation of the presented architecture. Xilinx Power Estimator (XPE) tool is used to measure the power consumption. The experimental results show the average power consumption of 91 mW; this is 42% improvement compared to the folded tree architecture.Keywords: Power consumption, energy efficiency, low power WSN node, recursive structure, sleep/wake scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10149042 Analyzing the Factors that Cause Parallel Performance Degradation in Parallel Graph-Based Computations Using Graph500
Authors: Mustafa Elfituri, Jonathan Cook
Abstract:
Recently, graph-based computations have become more important in large-scale scientific computing as they can provide a methodology to model many types of relations between independent objects. They are being actively used in fields as varied as biology, social networks, cybersecurity, and computer networks. At the same time, graph problems have some properties such as irregularity and poor locality that make their performance different than regular applications performance. Therefore, parallelizing graph algorithms is a hard and challenging task. Initial evidence is that standard computer architectures do not perform very well on graph algorithms. Little is known exactly what causes this. The Graph500 benchmark is a representative application for parallel graph-based computations, which have highly irregular data access and are driven more by traversing connected data than by computation. In this paper, we present results from analyzing the performance of various example implementations of Graph500, including a shared memory (OpenMP) version, a distributed (MPI) version, and a hybrid version. We measured and analyzed all the factors that affect its performance in order to identify possible changes that would improve its performance. Results are discussed in relation to what factors contribute to performance degradation.
Keywords: Graph computation, Graph500 benchmark, parallel architectures, parallel programming, workload characterization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5489041 PZ: A Z-based Formalism for Modeling Probabilistic Behavior
Authors: Hassan Haghighi
Abstract:
Probabilistic techniques in computer programs are becoming more and more widely used. Therefore, there is a big interest in the formal specification, verification, and development of probabilistic programs. In our work-in-progress project, we are attempting to make a constructive framework for developing probabilistic programs formally. The main contribution of this paper is to introduce an intermediate artifact of our work, a Z-based formalism called PZ, by which one can build set theoretical models of probabilistic programs. We propose to use a constructive set theory, called CZ set theory, to interpret the specifications written in PZ. Since CZ has an interpretation in Martin-L¨of-s theory of types, this idea enables us to derive probabilistic programs from correctness proofs of their PZ specifications.Keywords: formal specification, formal program development, probabilistic programs, CZ set theory, type theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12039040 PhilSHORE: Development of a WebGIS-Based Marine Spatial Planning Tool for Tidal Current Energy Resource Assessment and Site Suitability Analysis
Authors: Ma. Rosario Concepcion O. Ang, Luis Caezar Ian K. Panganiban, Charmyne B. Mamador, Oliver Dan G. De Luna, Michael D. Bausas, Joselito P. Cruz
Abstract:
PhilSHORE is a multi-site, multi-device and multicriteria decision support tool designed to support the development of tidal current energy in the Philippines. Its platform is based on Geographic Information Systems (GIS) which allows for the collection, storage, processing, analyses and display of geospatial data. Combining GIS tools with open source web development applications, PhilSHORE becomes a webGIS-based marine spatial planning tool. To date, PhilSHORE displays output maps and graphs of power and energy density, site suitability and site-device analysis. It enables stakeholders and the public easy access to the results of tidal current energy resource assessments and site suitability analyses. Results of the initial development show that PhilSHORE is a promising decision support tool for ORE project developments.Keywords: GIS, Site Suitability Analysis, Tidal Current Energy Resource Assessment, WebGIS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27149039 Fuzzy Logic Based Determination of Battery Charging Efficiency Applied to Hybrid Power System
Authors: Priyanka Paliwal, N. P. Patidar, R. K. Nema
Abstract:
Battery storage system is emerging as an essential component of hybrid power system based on renewable energy resources such as solar and wind in order to make these sources dispatchable. Accurate modeling of battery storage system is ssential in order to ensure optimal planning of hybrid power systems incorporating battery storage. Majority of the system planning studies involving battery storage assume battery charging efficiency to be constant. However a strong correlation exists between battery charging efficiency and battery state of charge. In this work a Fuzzy logic based model has been presented for determining battery charging efficiency relative to a particular SOC. In order to demonstrate the efficacy of proposed approach, reliability evaluation studies are carried out for a hypothetical autonomous hybrid power system located in Jaisalmer, Rajasthan, India. The impact of considering battery charging efficiency as a function of state of charge is compared against the assumption of fixed battery charging efficiency for three different configurations comprising of wind-storage, solar-storage and wind-solar-storage.
Keywords: Battery Storage, Charging efficiency, Fuzzy Logic, Hybrid Power System, Reliability
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20939038 Application of the Piloting Law Based on Adaptive Differentiators via Second Order Sliding Mode for a Fixed Wing Aircraft
Authors: Zaouche Mohammed, Amini Mohammed, Foughali Khaled, Hamissi Aicha, Aktouf Mohand Arezki, Boureghda Ilyes
Abstract:
In this paper, we present a piloting law based on the adaptive differentiators via high order sliding mode controller, by using an aircraft in virtual simulated environment. To deal with the design of an autopilot controller, we propose a framework based on Software in the Loop (SIL) methodology and we use MicrosoftTM Flight Simulator (FS-2004) as the environment for plane simulation. The aircraft dynamic model is nonlinear, Multi-Input Multi-Output (MIMO) and tightly coupled. The nonlinearity resides in the dynamic equations and also in the aerodynamic coefficients' variability. In our case, two (02) aircrafts are used in the flight tests, the Zlin-142 and MQ-1 Predator. For both aircrafts and in a very low altitude flight, we send the piloting control inputs to the aircraft which has stalled due to a command disconnection. Then, we present the aircraft’s dynamic behavior analysis while reestablishing the command transmission. Finally, a comparative study between the two aircraft’s dynamic behaviors is presented.
Keywords: Adaptive differentiators, Microsoft Flight Simulator, MQ-1 predator, second order sliding modes, Zlin-142.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12179037 Scale-Space Volume Descriptors for Automatic 3D Facial Feature Extraction
Authors: Daniel Chen, George Mamic, Clinton Fookes, Sridha Sridharan
Abstract:
An automatic method for the extraction of feature points for face based applications is proposed. The system is based upon volumetric feature descriptors, which in this paper has been extended to incorporate scale space. The method is robust to noise and has the ability to extract local and holistic features simultaneously from faces stored in a database. Extracted features are stable over a range of faces, with results indicating that in terms of intra-ID variability, the technique has the ability to outperform manual landmarking.
Keywords: Scale space volume descriptor, feature extraction, 3D facial landmarking
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15089036 Feasibility Study of MongoDB and Radio Frequency Identification Technology in Asset Tracking System
Authors: Mohd Noah A. Rahman, Afzaal H. Seyal, Sharul T. Tajuddin, Hartiny Md Azmi
Abstract:
Taking into consideration the real time situation specifically the higher academic institutions, small, medium to large companies, public to private sectors and the remaining sectors, do experience the inventory or asset shrinkages due to theft, loss or even inventory tracking errors. This happening is due to a zero or poor security systems and measures being taken and implemented in their organizations. Henceforth, implementing the Radio Frequency Identification (RFID) technology into any manual or existing web-based system or web application can simply deter and will eventually solve certain major issues to serve better data retrieval and data access. Having said, this manual or existing system can be enhanced into a mobile-based system or application. In addition to that, the availability of internet connections can aid better services of the system. Such involvement of various technologies resulting various privileges to individuals or organizations in terms of accessibility, availability, mobility, efficiency, effectiveness, real-time information and also security. This paper will look deeper into the integration of mobile devices with RFID technologies with the purpose of asset tracking and control. Next, it is to be followed by the development and utilization of MongoDB as the main database to store data and its association with RFID technology. Finally, the development of a web based system which can be viewed in a mobile based formation with the aid of Hypertext Preprocessor (PHP), MongoDB, Hyper-Text Markup Language 5 (HTML5), Android, JavaScript and AJAX programming language.
Keywords: RFID, asset tracking system, MongoDB, NoSQL.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16499035 Recognition-based Segmentation in Persian Character Recognition
Authors: Mohsen Zand, Ahmadreza Naghsh Nilchi, S. Amirhassan Monadjemi
Abstract:
Optical character recognition of cursive scripts presents a number of challenging problems in both segmentation and recognition processes in different languages, including Persian. In order to overcome these problems, we use a newly developed Persian word segmentation method and a recognition-based segmentation technique to overcome its segmentation problems. This method is robust as well as flexible. It also increases the system-s tolerances to font variations. The implementation results of this method on a comprehensive database show a high degree of accuracy which meets the requirements for commercial use. Extended with a suitable pre and post-processing, the method offers a simple and fast framework to develop a full OCR system.Keywords: OCR, Persian, Recognition, Segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18409034 Using PFA in Feature Analysis and Selection for H.264 Adaptation
Authors: Nora A. Naguib, Ahmed E. Hussein, Hesham A. Keshk, Mohamed I. El-Adawy
Abstract:
Classification of video sequences based on their contents is a vital process for adaptation techniques. It helps decide which adaptation technique best fits the resource reduction requested by the client. In this paper we used the principal feature analysis algorithm to select a reduced subset of video features. The main idea is to select only one feature from each class based on the similarities between the features within that class. Our results showed that using this feature reduction technique the source video features can be completely omitted from future classification of video sequences.
Keywords: Adaptation, feature selection, H.264, Principal Feature Analysis (PFA)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16079033 Improved Feature Processing for Iris Biometric Authentication System
Authors: Somnath Dey, Debasis Samanta
Abstract:
Iris-based biometric authentication is gaining importance in recent times. Iris biometric processing however, is a complex process and computationally very expensive. In the overall processing of iris biometric in an iris-based biometric authentication system, feature processing is an important task. In feature processing, we extract iris features, which are ultimately used in matching. Since there is a large number of iris features and computational time increases as the number of features increases, it is therefore a challenge to develop an iris processing system with as few as possible number of features and at the same time without compromising the correctness. In this paper, we address this issue and present an approach to feature extraction and feature matching process. We apply Daubechies D4 wavelet with 4 levels to extract features from iris images. These features are encoded with 2 bits by quantizing into 4 quantization levels. With our proposed approach it is possible to represent an iris template with only 304 bits, whereas existing approaches require as many as 1024 bits. In addition, we assign different weights to different iris region to compare two iris templates which significantly increases the accuracy. Further, we match the iris template based on a weighted similarity measure. Experimental results on several iris databases substantiate the efficacy of our approach.Keywords: Iris recognition, biometric, feature processing, patternrecognition, pattern matching.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21409032 Visual Search Based Indoor Localization in Low Light via RGB-D Camera
Authors: Yali Zheng, Peipei Luo, Shinan Chen, Jiasheng Hao, Hong Cheng
Abstract:
Most of traditional visual indoor navigation algorithms and methods only consider the localization in ordinary daytime, while we focus on the indoor re-localization in low light in the paper. As RGB images are degraded in low light, less discriminative infrared and depth image pairs are taken, as the input, by RGB-D cameras, the most similar candidates, as the output, are searched from databases which is built in the bag-of-word framework. Epipolar constraints can be used to relocalize the query infrared and depth image sequence. We evaluate our method in two datasets captured by Kinect2. The results demonstrate very promising re-localization results for indoor navigation system in low light environments.Keywords: Indoor navigation, low light, RGB-D camera, vision based.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16769031 A Reusability Evaluation Model for OO-Based Software Components
Authors: Parvinder S. Sandhu, Hardeep Singh
Abstract:
The requirement to improve software productivity has promoted the research on software metric technology. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. CK metric suit is most widely used metrics for the objectoriented (OO) software; we critically analyzed the CK metrics, tried to remove the inconsistencies and devised the framework of metrics to obtain the structural analysis of OO-based software components. Neural network can learn new relationships with new input data and can be used to refine fuzzy rules to create fuzzy adaptive system. Hence, Neuro-fuzzy inference engine can be used to evaluate the reusability of OO-based component using its structural attributes as inputs. In this paper, an algorithm has been proposed in which the inputs can be given to Neuro-fuzzy system in form of tuned WMC, DIT, NOC, CBO , LCOM values of the OO software component and output can be obtained in terms of reusability. The developed reusability model has produced high precision results as expected by the human experts.Keywords: CK-Metric, ID3, Neuro-fuzzy, Reusability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18199030 The Impact of Video Games in Children-s Learning of Mathematics
Authors: Muhammad Ridhuan Tony Lim Abdullah, Zulqarnain Abu Bakar, Razol Mahari Ali, Ibrahima Faye, Hilmi Hasan
Abstract:
This paper describes a research project on Year 3 primary school students in Malaysia in their use of computer-based video game to enhance learning of multiplication facts (tables) in the Mathematics subject. This study attempts to investigate whether video games could actually contribute to positive effect on children-s learning or otherwise. In conducting this study, the researchers assume a neutral stand in the investigation as an unbiased outcome of the study would render reliable response to the impact of video games in education which would contribute to the literature of technology-based education as well as impact to the pedagogical aspect of formal education. In order to conduct the study, a subject (Mathematics) with a specific topic area in the subject (multiplication facts) is chosen. The study adopts a causal-comparative research to investigate the impact of the inclusion of a computer-based video game designed to teach multiplication facts to primary level students. Sample size is 100 students divided into two i.e., A: conventional group and B conventional group aided by video games. The conventional group (A) would be taught multiplication facts (timetables) and skills conventionally. The other group (B) underwent the same lessons but with supplementary activity: a computer-based video game on multiplication which is called Timez-Attack. Analysis of marks accrued from pre-test will be compared to post- test using comparisons of means, t tests, and ANOVA tests to investigate the impact of computer games as an added learning activity. The findings revealed that video games as a supplementary activity to classroom learning brings significant and positive effect on students- retention and mastery of multiplication tables as compared to students who rely only upon formal classroom instructions.
Keywords: Technology for education, Gaming for education, Computer-based video games, Cognitive learning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42609029 A Heuristic Statistical Model for Lifetime Distribution Analysis of Complicated Systems in the Reliability Centered Maintenance
Authors: Mojtaba Mahdavi, Mohamad Mahdavi, Maryam Yazdani
Abstract:
A heuristic conceptual model for to develop the Reliability Centered Maintenance (RCM), especially in preventive strategy, has been explored during this paper. In most real cases which complicity of system obligates high degree of reliability, this model proposes a more appropriate reliability function between life time distribution based and another which is based on relevant Extreme Value (EV) distribution. A statistical and mathematical approach is used to estimate and verify these two distribution functions. Then best one is chosen just among them, whichever is more reliable. A numeric Industrial case study will be reviewed to represent the concepts of this paper, more clearly.Keywords: Lifetime distribution, Reliability, Estimation, Extreme value, Improving model, Series, Parallel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14809028 A New Method of Combined Classifier Design Based on Fuzzy Neural Network
Abstract:
To make the modulation classification system more suitable for signals in a wide range of signal to noise rate (SNR), a novel method of designing combined classifier based on fuzzy neural network (FNN) is presented in this paper. The method employs fuzzy neural network classifiers and interclass distance (ICD) to improve recognition reliability. Experimental results show that the proposed combined classifier has high recognition rate with large variation range of SNR (success rates are over 99.9% when SNR is not lower than 5dB).Keywords: Modulation classification, combined classifier, fuzzy neural network, interclass distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12249027 Manual to Automated Testing: An Effort-Based Approach for Determining the Priority of Software Test Automation
Authors: Peter Sabev, Katalina Grigorova
Abstract:
Test automation allows performing difficult and time consuming manual software testing tasks efficiently, quickly and repeatedly. However, development and maintenance of automated tests is expensive, so it needs a proper prioritization what to automate first. This paper describes a simple yet efficient approach for such prioritization of test cases based on the effort needed for both manual execution and software test automation. The suggested approach is very flexible because it allows working with a variety of assessment methods, and adding or removing new candidates at any time. The theoretical ideas presented in this article have been successfully applied in real world situations in several software companies by the authors and their colleagues including testing of real estate websites, cryptographic and authentication solutions, OSGi-based middleware framework that has been applied in various systems for smart homes, connected cars, production plants, sensors, home appliances, car head units and engine control units (ECU), vending machines, medical devices, industry equipment and other devices that either contain or are connected to an embedded service gateway.Keywords: Automated Testing, Manual Testing, Test Automation, Software testing, Test Prioritization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33859026 Text Mining Technique for Data Mining Application
Authors: M. Govindarajan
Abstract:
Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In decision tree approach is most useful in classification problem. With this technique, tree is constructed to model the classification process. There are two basic steps in the technique: building the tree and applying the tree to the database. This paper describes a proposed C5.0 classifier that performs rulesets, cross validation and boosting for original C5.0 in order to reduce the optimization of error ratio. The feasibility and the benefits of the proposed approach are demonstrated by means of medial data set like hypothyroid. It is shown that, the performance of a classifier on the training cases from which it was constructed gives a poor estimate by sampling or using a separate test file, either way, the classifier is evaluated on cases that were not used to build and evaluate the classifier are both are large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772 case training set and a 1000 case test set, C5.0 might construct a different classifier with a lower or higher error rate on the test cases. An important feature of see5 is its ability to classifiers called rulesets. The ruleset has an error rate 0.5 % on the test cases. The standard errors of the means provide an estimate of the variability of results. One way to get a more reliable estimate of predictive is by f-fold –cross- validation. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. The Boost option with x trials instructs See5 to construct up to x classifiers in this manner. Trials over numerous datasets, large and small, show that on average 10-classifier boosting reduces the error rate for test cases by about 25%.Keywords: C5.0, Error Ratio, text mining, training data, test data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2489