Search results for: test data.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9349

Search results for: test data.

7999 A Novel Web Metric for the Evaluation of Internet Trends

Authors: Radek Malinský, Ivan Jelínek

Abstract:

Web 2.0 (social networking, blogging and online forums) can serve as a data source for social science research because it contains vast amount of information from many different users. The volume of that information has been growing at a very high rate and becoming a network of heterogeneous data; this makes things difficult to find and is therefore not almost useful. We have proposed a novel theoretical model for gathering and processing data from Web 2.0, which would reflect semantic content of web pages in better way. This article deals with the analysis part of the model and its usage for content analysis of blogs. The introductory part of the article describes methodology for the gathering and processing data from blogs. The next part of the article is focused on the evaluation and content analysis of blogs, which write about specific trend.

Keywords: Blog, Sentiment Analysis, Web 2.0, Webometrics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3517
7998 The Feasibility of Using Milled Glass Wastes in Concrete to Resist Freezing-Thawing Action

Authors: Raed Abendeh, Mousa Bani Baker, Zaydoun Abu Salem, Heham Ahmad

Abstract:

The using of waste materials in the construction industry can reduce the dependence on the natural aggregates which are going at the end to deplete. The glass waste is generated in a huge amount which can make one of its disposals in concrete industry effective not only as a green solution but also as an advantage to enhance the performance of mechanical properties and durability of concrete. This article reports the performance of concrete specimens containing different percentages of milled glass waste as a partial replacement of cement (Powder), when they are subject to cycles of freezing and thawing. The tests were conducted on 75-mm cubes and 75 x 75 x 300-mm prisms. Compressive strength based on laboratory testing and non-destructive ultrasonic pulse velocity test were performed during the action of freezing-thawing cycles (F/T). The results revealed that the incorporation of glass waste in concrete mixtures is not only feasible but also showed generally better strength and durability performance than control concrete mixture. It may be said that the recycling of waste glass in concrete mixes is not only a disposal way, but also it can be an exploitation in concrete industry.

Keywords: Durability, glass waste, freeze-thaw cycles, nondestructive test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2637
7997 Combined Effect of Moving and Open Boundary Conditions in the Simulation of Inland Inundation Due to Far Field Tsunami

Authors: M. Ashaque Meah, Md. Fazlul Karim, M. Shah Noor, Nazmun Nahar Papri, M. Khalid Hossen, M. Ismoen

Abstract:

Tsunami and inundation modelling due to far field tsunami propagation in a limited area is a very challenging numerical task because it involves many aspects such as the formation of various types of waves and the irregularities of coastal boundaries. To compute the effect of far field tsunami and extent of inland inundation due to far field tsunami along the coastal belts of west coast of Malaysia and Southern Thailand, a formulated boundary condition and a moving boundary condition are simultaneously used. In this study, a boundary fitted curvilinear grid system is used in order to incorporate the coastal and island boundaries accurately as the boundaries of the model domain are curvilinear in nature and the bending is high. The tsunami response of the event 26 December 2004 along the west open boundary of the model domain is computed to simulate the effect of far field tsunami. Based on the data of the tsunami source at the west open boundary of the model domain, a boundary condition is formulated and applied to simulate the tsunami response along the coastal and island boundaries. During the simulation process, a moving boundary condition is initiated instead of fixed vertical seaside wall. The extent of inland inundation and tsunami propagation pattern are computed. Some comparisons are carried out to test the validation of the simultaneous use of the two boundary conditions. All simulations show excellent agreement with the data of observation.

Keywords: Open boundary condition, moving boundary condition, boundary-fitted curvilinear grids, far field tsunami, Shallow Water Equations, tsunami source, Indonesian tsunami of 2004.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2024
7996 The Effect of e-learning on the Promotion of Optoelectronics Technology and Daily Livings Literacy among Students in Universities of Technology

Authors: Chin-Pin Chen, David W.S. Tai, Wen-Jong Chen, Hui-Min Lai

Abstract:

This study aims to analyze the effect of e-learning on photonics technology and daily livings among college students. The course contents of photonics technology and daily livings are first drafted based on research discussions and expert interviews. Having expert questionnaires with Delphi Technique for three times, the knowledge units and items for the course of photonics technology and daily livings are established. The e-learning materials and the drafts of instructional strategies, academic achievement, and learning attitude scales are then developed. With expert inspection, reliability and validity test, and experimental instructions, the scales and the material are further revised. Finally, the formal instructions are implemented to test the effect of different instructional methods on the academic achievement of photonics technology and daily livings among students in universities of technology. The research results show that e-learning could effectively promote academic achievement and learning attitude, and the students with e-learning obviously outperform the ones with trandition instructions.

Keywords: E-learning, Photonics Technology and Daily Livings, Academic Achievement

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1676
7995 Simulated Annealing Algorithm for Data Aggregation Trees in Wireless Sensor Networks and Comparison with Genetic Algorithm

Authors: Ladan Darougaran, Hossein Shahinzadeh, Hajar Ghotb, Leila Ramezanpour

Abstract:

In ad hoc networks, the main issue about designing of protocols is quality of service, so that in wireless sensor networks the main constraint in designing protocols is limited energy of sensors. In fact, protocols which minimize the power consumption in sensors are more considered in wireless sensor networks. One approach of reducing energy consumption in wireless sensor networks is to reduce the number of packages that are transmitted in network. The technique of collecting data that combines related data and prevent transmission of additional packages in network can be effective in the reducing of transmitted packages- number. According to this fact that information processing consumes less power than information transmitting, Data Aggregation has great importance and because of this fact this technique is used in many protocols [5]. One of the Data Aggregation techniques is to use Data Aggregation tree. But finding one optimum Data Aggregation tree to collect data in networks with one sink is a NP-hard problem. In the Data Aggregation technique, related information packages are combined in intermediate nodes and form one package. So the number of packages which are transmitted in network reduces and therefore, less energy will be consumed that at last results in improvement of longevity of network. Heuristic methods are used in order to solve the NP-hard problem that one of these optimization methods is to solve Simulated Annealing problems. In this article, we will propose new method in order to build data collection tree in wireless sensor networks by using Simulated Annealing algorithm and we will evaluate its efficiency whit Genetic Algorithm.

Keywords: Data aggregation, wireless sensor networks, energy efficiency, simulated annealing algorithm, genetic algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1664
7994 Encoding and Compressing Data for Decreasing Number of Switches in Baseline Networks

Authors: Mohammad Ali Jabraeil Jamali, Ahmad Khademzadeh, Hasan Asil, Amir Asil

Abstract:

This method decrease usage power (expenditure) in networks on chips (NOC). This method data coding for data transferring in order to reduces expenditure. This method uses data compression reduces the size. Expenditure calculation in NOC occurs inside of NOC based on grown models and transitive activities in entry ports. The goal of simulating is to weigh expenditure for encoding, decoding and compressing in Baseline networks and reduction of switches in this type of networks. KeywordsNetworks on chip, Compression, Encoding, Baseline networks, Banyan networks.

Keywords: Networks on chip, Compression, Encoding, Baseline networks, Banyan networks

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1958
7993 Sampled-Data Control for Fuel Cell Systems

Authors: H. Y. Jung, Ju H. Park, S. M. Lee

Abstract:

Sampled-data controller is presented for solid oxide fuel cell systems which is expressed by a sector bounded nonlinear model. The proposed control law is obtained by solving a convex problem satisfying several linear matrix inequalities. Simulation results are given to show the effectiveness of the proposed design method.

Keywords: Sampled-data control, Sector bound, Solid oxide fuel cell, Time-delay.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1699
7992 Automatic Detection and Spatio-temporal Analysis of Commercial Accumulations Using Digital Yellow Page Data

Authors: Yuki. Akiyama, Hiroaki. Sengoku, Ryosuke. Shibasaki

Abstract:

In this study, the locations and areas of commercial accumulations were detected by using digital yellow page data. An original buffering method that can accurately create polygons of commercial accumulations is proposed in this paper.; by using this method, distribution of commercial accumulations can be easily created and monitored over a wide area. The locations, areas, and time-series changes of commercial accumulations in the South Kanto region can be monitored by integrating polygons of commercial accumulations with the time-series data of digital yellow page data. The circumstances of commercial accumulations were shown to vary according to areas, that is, highly- urbanized regions such as the city center of Tokyo and prefectural capitals, suburban areas near large cities, and suburban and rural areas.

Keywords: Commercial accumulations, Spatio-temporal analysis, Urban monitoring, Yellow page data

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1241
7991 Characterization of a Pure Diamond-Like Carbon Film Deposited by Nanosecond Pulsed Laser Deposition

Authors: Camilla G. Goncalves, Benedito Christ, Walter Miyakawa, Antonio J. Abdalla

Abstract:

This work aims to investigate the properties and microstructure of diamond-like carbon film deposited by pulsed laser deposition by ablation of a graphite target in a vacuum chamber on a steel substrate. The equipment was mounted to provide one laser beam. The target of high purity graphite and the steel substrate were polished. The mechanical and tribological properties of the film were characterized using Raman spectroscopy, nanoindentation test, scratch test, roughness profile, tribometer, optical microscopy and SEM images. It was concluded that the pulsed laser deposition (PLD) technique associated with the low-pressure chamber and a graphite target provides a good fraction of sp3 bonding, that the process variable as surface polishing and laser parameter have great influence in tribological properties and in adherence tests performance. The optical microscopy images are efficient to identify the metallurgical bond.

Keywords: Characterization, diamond-like carbon, DLC, mechanical properties, pulsed laser deposition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 674
7990 Performance Analysis of Search Medical Imaging Service on Cloud Storage Using Decision Trees

Authors: González A. Julio, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

Telemedicine services use a large amount of data, most of which are diagnostic images in Digital Imaging and Communications in Medicine (DICOM) and Health Level Seven (HL7) formats. Metadata is generated from each related image to support their identification. This study presents the use of decision trees for the optimization of information search processes for diagnostic images, hosted on the cloud server. To analyze the performance in the server, the following quality of service (QoS) metrics are evaluated: delay, bandwidth, jitter, latency and throughput in five test scenarios for a total of 26 experiments during the loading and downloading of DICOM images, hosted by the telemedicine group server of the Universidad Militar Nueva Granada, Bogotá, Colombia. By applying decision trees as a data mining technique and comparing it with the sequential search, it was possible to evaluate the search times of diagnostic images in the server. The results show that by using the metadata in decision trees, the search times are substantially improved, the computational resources are optimized and the request management of the telemedicine image service is improved. Based on the experiments carried out, search efficiency increased by 45% in relation to the sequential search, given that, when downloading a diagnostic image, false positives are avoided in management and acquisition processes of said information. It is concluded that, for the diagnostic images services in telemedicine, the technique of decision trees guarantees the accessibility and robustness in the acquisition and manipulation of medical images, in improvement of the diagnoses and medical procedures in patients.

Keywords: Cloud storage, decision trees, diagnostic image, search, telemedicine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 912
7989 EEG Waves Classifier using Wavelet Transform and Fourier Transform

Authors: Maan M. Shaker

Abstract:

The electroencephalograph (EEG) signal is one of the most widely signal used in the bioinformatics field due to its rich information about human tasks. In this work EEG waves classification is achieved using the Discrete Wavelet Transform DWT with Fast Fourier Transform (FFT) by adopting the normalized EEG data. The DWT is used as a classifier of the EEG wave's frequencies, while FFT is implemented to visualize the EEG waves in multi-resolution of DWT. Several real EEG data sets (real EEG data for both normal and abnormal persons) have been tested and the results improve the validity of the proposed technique.

Keywords: Bioinformatics, DWT, EEG waves, FFT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5525
7988 Obstacle Classification Method Based On 2D LIDAR Database

Authors: Moohyun Lee, Soojung Hur, Yongwan Park

Abstract:

We propose obstacle classification method based on 2D LIDAR Database. The existing obstacle classification method based on 2D LIDAR, has an advantage in terms of accuracy and shorter calculation time. However, it was difficult to classifier the type of obstacle and therefore accurate path planning was not possible. In order to overcome this problem, a method of classifying obstacle type based on width data of obstacle was proposed. However, width data was not sufficient to improve accuracy. In this paper, database was established by width and intensity data; the first classification was processed by the width data; the second classification was processed by the intensity data; classification was processed by comparing to database; result of obstacle classification was determined by finding the one with highest similarity values. An experiment using an actual autonomous vehicle under real environment shows that calculation time declined in comparison to 3D LIDAR and it was possible to classify obstacle using single 2D LIDAR.

Keywords: Obstacle, Classification, LIDAR, Segmentation, Width, Intensity, Database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3421
7987 An Empirical Mode Decomposition Based Method for Action Potential Detection in Neural Raw Data

Authors: Sajjad Farashi, Mohammadjavad Abolhassani, Mostafa Taghavi Kani

Abstract:

Information in the nervous system is coded as firing patterns of electrical signals called action potential or spike so an essential step in analysis of neural mechanism is detection of action potentials embedded in the neural data. There are several methods proposed in the literature for such a purpose. In this paper a novel method based on empirical mode decomposition (EMD) has been developed. EMD is a decomposition method that extracts oscillations with different frequency range in a waveform. The method is adaptive and no a-priori knowledge about data or parameter adjusting is needed in it. The results for simulated data indicate that proposed method is comparable with wavelet based methods for spike detection. For neural signals with signal-to-noise ratio near 3 proposed methods is capable to detect more than 95% of action potentials accurately.

Keywords: EMD, neural data processing, spike detection, wavelet decomposition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2357
7986 Classification Control for Discrimination between Interictal Epileptic and Non – Epileptic Pathological EEG Events

Authors: Sozon H. Papavlasopoulos, Marios S. Poulos, George D. Bokos, Angelos M. Evangelou

Abstract:

In this study, the problem of discriminating between interictal epileptic and non- epileptic pathological EEG cases, which present episodic loss of consciousness, investigated. We verify the accuracy of the feature extraction method of autocross-correlated coefficients which extracted and studied in previous study. For this purpose we used in one hand a suitable constructed artificial supervised LVQ1 neural network and in other a cross-correlation technique. To enforce the above verification we used a statistical procedure which based on a chi- square control. The classification and the statistical results showed that the proposed feature extraction is a significant accurate method for diagnostic discrimination cases between interictal and non-interictal EEG events and specifically the classification procedure showed that the LVQ neural method is superior than the cross-correlation one.

Keywords: Cross-Correlation Methods, Diagnostic Test, Interictal Epileptic, LVQ1 neural network, Auto-Cross-Correlation Methods, chi-square test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1499
7985 Platform-as-a-Service Sticky Policies for Privacy Classification in the Cloud

Authors: Maha Shamseddine, Amjad Nusayr, Wassim Itani

Abstract:

In this paper, we present a Platform-as-a-Service (PaaS) model for controlling the privacy enforcement mechanisms applied on user data when stored and processed in Cloud data centers. The proposed architecture consists of establishing user configurable ‘sticky’ policies on the Graphical User Interface (GUI) data-bound components during the application development phase to specify the details of privacy enforcement on the contents of these components. Various privacy classification classes on the data components are formally defined to give the user full control on the degree and scope of privacy enforcement including the type of execution containers to process the data in the Cloud. This not only enhances the privacy-awareness of the developed Cloud services, but also results in major savings in performance and energy efficiency due to the fact that the privacy mechanisms are solely applied on sensitive data units and not on all the user content. The proposed design is implemented in a real PaaS cloud computing environment on the Microsoft Azure platform.

Keywords: Privacy enforcement, Platform-as-a-Service privacy awareness, cloud computing privacy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 728
7984 DIFFER: A Propositionalization approach for Learning from Structured Data

Authors: Thashmee Karunaratne, Henrik Böstrom

Abstract:

Logic based methods for learning from structured data is limited w.r.t. handling large search spaces, preventing large-sized substructures from being considered by the resulting classifiers. A novel approach to learning from structured data is introduced that employs a structure transformation method, called finger printing, for addressing these limitations. The method, which generates features corresponding to arbitrarily complex substructures, is implemented in a system, called DIFFER. The method is demonstrated to perform comparably to an existing state-of-art method on some benchmark data sets without requiring restrictions on the search space. Furthermore, learning from the union of features generated by finger printing and the previous method outperforms learning from each individual set of features on all benchmark data sets, demonstrating the benefit of developing complementary, rather than competing, methods for structure classification.

Keywords: Machine learning, Structure classification, Propositionalization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1201
7983 Improving the Performance of Proxy Server by Using Data Mining Technique

Authors: P. Jomsri

Abstract:

Currently, web usage make a huge data from a lot of user attention. In general, proxy server is a system to support web usage from user and can manage system by using hit rates. This research tries to improve hit rates in proxy system by applying data mining technique. The data set are collected from proxy servers in the university and are investigated relationship based on several features. The model is used to predict the future access websites. Association rule technique is applied to get the relation among Date, Time, Main Group web, Sub Group web, and Domain name for created model. The results showed that this technique can predict web content for the next day, moreover the future accesses of websites increased from 38.15% to 85.57 %. This model can predict web page access which tends to increase the efficient of proxy servers as a result. In additional, the performance of internet access will be improved and help to reduce traffic in networks.

Keywords: Association rule, proxy server, data mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3035
7982 Appling Eyring-s Accelerated Life Testing Model to “Times to Breakdown“ of Insulating Fluid: A Combined Approach of an Accelerated and a Sequential Life Testing

Authors: D. I. De Souza, D. R. Fonseca, D. Kipper

Abstract:

In this paper, the test purpose will be to assess whether or not the accelerated model proposed by Eyring will be able to translate results for the shape and scale parameters of an underlying Weibull model, obtained under two accelerating using conditions, to expected normal using condition results for these parameters. The product being analyzed is a new type of insulate fluid, and the accelerating factor is the voltage stresses applied to the fluid at two different levels (30KV and 40KV). The normal operating voltage is 25KV. In this case, it was possible to test the insulate fluid at normal voltage using condition. Both results for the two parameters of the Weibull model, obtained under normal using condition and translated from accelerated using conditions to normal conditions, will be compared to each other to assess the accuracy of the Eyring model when the accelerating factor is only the voltage stress.

Keywords: Eyring Accelerated Model, Sequential Life Testing, Two-Parameter Weibull Distribution, Voltage Stresses.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2262
7981 Piezoelectric Power Output Predictions Using Single-Phase Flow to Power Flow Meters

Authors: Umar Alhaji Mukhtar, Abubakar Mohammed El-jummah

Abstract:

This research involved the utilization of fluid flow energy to predict power output using Lead Zirconate Titanate (PZT) piezoelectric stacks. The aim of this work is to extract energy from a controlled level of pressure fluctuation in single-phase flow which forms a part of the energy harvesting technology that powers flow meters. A device- Perspex box was developed and fixed to 50.8 mm rig to induce pressure fluctuation in the flow. An experimental test was carried out using the single-phase water flow in the developed rig in order to measure the power output generation from the piezoelectric stacks. 16 sets of experimental tests were conducted to ensure the maximum output result. The acquired signal of the pressure fluctuation was used to simulate the expected electrical output from the piezoelectric material. The results showed a maximum output voltage of 12 V with an instantaneous output power of 1 µW generated, when the pressure amplitude is 2.6 kPa at a frequency of 2.4 Hz.

Keywords: Energy harvesting, experimental test, perspex rig, pressure fluctuation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 672
7980 Effect of Physical Contact (Hand-Holding) on Heart Rate Variability

Authors: T. Pishbin, S.M.P. Firoozabadi, N. Jafarnia Dabanloo, F. Mohammadi, S. Koozehgari

Abstract:

Heart-s electric field can be measured anywhere on the surface of the body (ECG). When individuals touch, one person-s ECG signal can be registered in other person-s EEG and elsewhere on his body. Now, the aim of this study was to test the hypothesis that physical contact (hand-holding) of two persons changes their heart rate variability. Subjects were sixteen healthy female (age: 20- 26) which divided into eight sets. In each sets, we had two friends that they passed intimacy test of J.sternberg. ECG of two subjects (each set) acquired for 5 minutes before hand-holding (as control group) and 5 minutes during they held their hands (as experimental group). Then heart rate variability signals were extracted from subjects' ECG and analyzed in linear feature space (time and frequency domain) and nonlinear feature space. Considering the results, we conclude that physical contact (hand-holding of two friends) increases parasympathetic activity, as indicate by increase SD1, SD1/SD2, HF and MF power (p<0.05) and decreases sympathetic activity, as indicate by decrease LF power (p<0.01) and LF/HF ratio (p<0.05).

Keywords: Autonomic nervous system (ANS), Hand- holding, Heart rate variability (HRV), Power spectral density analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3071
7979 Performance Analysis of the Subgroup Method for Collective I/O

Authors: Kwangho Cha, Hyeyoung Cho, Sungho Kim

Abstract:

As many scientific applications require large data processing, the importance of parallel I/O has been increasingly recognized. Collective I/O is one of the considerable features of parallel I/O and enables application programmers to easily handle their large data volume. In this paper we measured and analyzed the performance of original collective I/O and the subgroup method, the way of using collective I/O of MPI effectively. From the experimental results, we found that the subgroup method showed good performance with small data size.

Keywords: Collective I/O, MPI, parallel file system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1549
7978 Statistical Analysis for Overdispersed Medical Count Data

Authors: Y. N. Phang, E. F. Loh

Abstract:

Many researchers have suggested the use of zero inflated Poisson (ZIP) and zero inflated negative binomial (ZINB) models in modeling overdispersed medical count data with extra variations caused by extra zeros and unobserved heterogeneity. The studies indicate that ZIP and ZINB always provide better fit than using the normal Poisson and negative binomial models in modeling overdispersed medical count data. In this study, we proposed the use of Zero Inflated Inverse Trinomial (ZIIT), Zero Inflated Poisson Inverse Gaussian (ZIPIG) and zero inflated strict arcsine models in modeling overdispered medical count data. These proposed models are not widely used by many researchers especially in the medical field. The results show that these three suggested models can serve as alternative models in modeling overdispersed medical count data. This is supported by the application of these suggested models to a real life medical data set. Inverse trinomial, Poisson inverse Gaussian and strict arcsine are discrete distributions with cubic variance function of mean. Therefore, ZIIT, ZIPIG and ZISA are able to accommodate data with excess zeros and very heavy tailed. They are recommended to be used in modeling overdispersed medical count data when ZIP and ZINB are inadequate.

Keywords: Zero inflated, inverse trinomial distribution, Poisson inverse Gaussian distribution, strict arcsine distribution, Pearson’s goodness of fit.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3284
7977 Effect of Load Ratio on Probability Distribution of Fatigue Crack Propagation Life in Magnesium Alloys

Authors: Seon Soon Choi

Abstract:

It is necessary to predict a fatigue crack propagation life for estimation of structural integrity. Because of an uncertainty and a randomness of a structural behavior, it is also required to analyze stochastic characteristics of the fatigue crack propagation life at a specified fatigue crack size. The essential purpose of this study is to find the effect of load ratio on probability distribution of the fatigue crack propagation life at a specified grown crack size and to confirm the good probability distribution in magnesium alloys under various fatigue load ratio conditions. To investigate a stochastic crack growth behavior, fatigue crack propagation experiments are performed in laboratory air under several conditions of fatigue load ratio using AZ31. By Anderson-Darling test, a goodness-of-fit test for probability distribution of the fatigue crack propagation life is performed. The effect of load ratio on variability of fatigue crack propagation life is also investigated.

Keywords: Load ratio, fatigue crack propagation life, Magnesium alloys, probability distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703
7976 NaCl Erosion-Corrosion of Mild Steel under Submerged Impingement Jet

Authors: M. Sadique, S. Ainane, Y. F. Yap, P. Rostron, E. Al Hajri

Abstract:

The presence of sand in production lines in the oil and gas industries causes material degradation due to erosion-corrosion. The material degradation caused by erosion-corrosion in pipelines can result in a high cost of monitoring and maintenance and in major accidents. The process of erosion-corrosion consists of erosion, corrosion, and their interactions. Investigating and understanding how the erosion-corrosion process affects the degradation process in certain materials will allow for a reduction in economic loss and help prevent accidents. In this study, material loss due to erosion-corrosion of mild steel under impingement of sand-laden water at 90˚ impingement angle is investigated using a submerged impingement jet (SIJ) test. In particular, effects of jet velocity and sand loading on TWL due to erosion-corrosion, weight loss due to pure erosion and erosion-corrosion interactions, at a temperature of 29-33 °C in sea water environment (3.5% NaCl), are analyzed. The results show that the velocity and sand loading have a great influence on the removal of materials, and erosion is more dominant under all conditions studied. Changes in the surface characteristics of the specimen after impingement test are also discussed.

Keywords: Erosion-corrosion, flow velocity, jet impingement, sand loading.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1584
7975 Student Satisfaction Data for Work Based Learners

Authors: Rosie Borup, Hanifa Shah

Abstract:

This paper aims to describe how student satisfaction is measured for work-based learners as these are non-traditional learners, conducting academic learning in the workplace, typically their curricula have a high degree of negotiation, and whose motivations are directly related to their employers- needs, as well as their own career ambitions. We argue that while increasing WBL participation, and use of SSD are both accepted as being of strategic importance to the HE agenda, the use of WBL SSD is rarely examined, and lessons can be learned from the comparison of SSD from a range of WBL programmes, and increased visibility of this type of data will provide insight into ways to improve and develop this type of delivery. The key themes that emerged from the analysis of the interview data were: learners profiles and needs, employers drivers, academic staff drivers, organizational approach, tools for collecting data and visibility of findings. The paper concludes with observations on best practice in the collection, analysis and use of WBL SSD, thus offering recommendations for both academic managers and practitioners.

Keywords: Student satisfaction data, work based learning, employer engagement, NSS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1468
7974 A Consistency Protocol Multi-Layer for Replicas Management in Large Scale Systems

Authors: Ghalem Belalem, Yahya Slimani

Abstract:

Large scale systems such as computational Grid is a distributed computing infrastructure that can provide globally available network resources. The evolution of information processing systems in Data Grid is characterized by a strong decentralization of data in several fields whose objective is to ensure the availability and the reliability of the data in the reason to provide a fault tolerance and scalability, which cannot be possible only with the use of the techniques of replication. Unfortunately the use of these techniques has a height cost, because it is necessary to maintain consistency between the distributed data. Nevertheless, to agree to live with certain imperfections can improve the performance of the system by improving competition. In this paper, we propose a multi-layer protocol combining the pessimistic and optimistic approaches conceived for the data consistency maintenance in large scale systems. Our approach is based on a hierarchical representation model with tree layers, whose objective is with double vocation, because it initially makes it possible to reduce response times compared to completely pessimistic approach and it the second time to improve the quality of service compared to an optimistic approach.

Keywords: Data Grid, replication, consistency, optimistic approach, pessimistic approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1549
7973 Seismic Assessment of Old Existing RC Buildings with Masonry Infill in Madinah as per ASCE

Authors: Tarek M. Alguhane, Ayman H. Khalil, M. N. Fayed, Ayman M. Ismail

Abstract:

An existing RC building in Madinah is seismically evaluated with and without infill wall. Four model systems have been considered i.e. model I (no infill), model IIA (strut infill-update from field test), model IIB (strut infill- ASCE/SEI 41) and model IIC (strut infill-Soft storey- ASCE/SEI 41). Three dimensional pushover analyses have been carried out using SAP2000 software incorporating inelastic material behavior for concrete, steel and infill walls. Infill wall has been modeled as equivalent strut according to suggested equation matching field test measurements and to the ASCE/SEI 41 equation. The effect of building modeling on the performance point as well as capacity and demand spectra due to EQ design spectrum function in Madinah area has been investigated. The response modification factor (R) for the 5 story RC building is evaluated from capacity and demand spectra (ATC-40) for the studied models. The results are summarized and discussed.

Keywords: Infill wall, Pushover Analysis, Response Modification Factor, Seismic Assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3227
7972 A Force Measurement Evaluation Tool for Telerobotic Cutting Applications: Development of an Effective Characterization Platform

Authors: Dean J. Callaghan, Mark M. McGrath

Abstract:

Sensorized instruments that accurately measure the interaction forces (between biological tissue and instrument endeffector) during surgical procedures offer surgeons a greater sense of immersion during minimally invasive robotic surgery. Although there is ongoing research into force measurement involving surgical graspers little corresponding effort has been carried out on the measurement of forces between scissor blades and tissue. This paper presents the design and development of a force measurement test apparatus, which will serve as a sensor characterization and evaluation platform. The primary aim of the experiments is to ascertain whether the system can differentiate between tissue samples with differing mechanical properties in a reliable, repeatable manner. Force-angular displacement curves highlight trends in the cutting process as well the forces generated along the blade during a cutting procedure. Future applications of the test equipment will involve the assessment of new direct force sensing technologies for telerobotic surgery.

Keywords: Force measurement, minimally invasive surgery, scissor blades, tissue cutting.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1809
7971 Discovery of Time Series Event Patterns based on Time Constraints from Textual Data

Authors: Shigeaki Sakurai, Ken Ueno, Ryohei Orihara

Abstract:

This paper proposes a method that discovers time series event patterns from textual data with time information. The patterns are composed of sequences of events and each event is extracted from the textual data, where an event is characteristic content included in the textual data such as a company name, an action, and an impression of a customer. The method introduces 7 types of time constraints based on the analysis of the textual data. The method also evaluates these constraints when the frequency of a time series event pattern is calculated. We can flexibly define the time constraints for interesting combinations of events and can discover valid time series event patterns which satisfy these conditions. The paper applies the method to daily business reports collected by a sales force automation system and verifies its effectiveness through numerical experiments.

Keywords: Text mining, sequential mining, time constraints, daily business reports.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1465
7970 Different Multimedia Presentation Types and Students' Interpretation Achievement

Authors: Cenk Akbiyik, Gonul Altin Akbiyik

Abstract:

The main purpose of the study was to determine whether students- interpretation achievement differed with the use of various multimedia presentation types. Four groups of students, text only (T), audio only (A), text and audio (TA), text and image (TI), were arranged and they were presented the same story via different types of multimedia presentations. Inference achievement was measured by a critical thinking inference test. Higher mean scores for the TA group compared to the other three groups were found. Also when compared pairwise, interpretation achievement of the TA group differed significantly from scores of the T and TI groups. These differences were interpreted with the increased cognitive load. Increased cognitive load for the TA group may have invited students to put more effort into comprehending the text, thus resulting in better test scores. Findings of the study can be seen as a sign of the importance of learning situations and learning outcomes in multimedia-supported learning environments and may have practical benefits for instructional designers.

Keywords: Multimedia, cognitive multimedia, dual coding, cognitive load, critical thinking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3422