Search results for: measured data.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8344

Search results for: measured data.

6244 Congo Red Photocatalytic Decolourization using Modified Titanium

Authors: A. López–Vásquez, D. Santamaría, M. Tibatá, C. Gómez

Abstract:

A study concerning the photocatalytic decolourization of Congo red (CR) dye, over artificial UV irradiation is presented. Photocatalysts based on a commercial titanium dioxide (TiO2) modified with transition metals (Ni, Cu and Zn) were used. The dopage method used was wet impregnation. A TiO2 sample without salt was subjected to the same hydrothermal treatment to be used as reference. Congo red solutions to several pH conditions (natural and basic) were used to evaluate photocatalytic performance of each doped catalysts. Photodecolourization percentage was measured spectrofotrometically after 3 h of treatment to 499 nm as response variable. Kinetics investigations of photodegradation indicated that reactions obey to Langmuir-Hinshelwood model and pseudo–first order law. The rate constant studies of photocatalytic decolourization reactions for Zn–TiO2 and Cu–TiO2 photocatalysts indicated that in all cases the rate constant of the reaction was higher than that of TiO2 undoped. These results show that nature of the metal modifying the TiO2 influence on the efficiency of the photocatalyst evaluated in process. Ni does not present an additional effect compared with TiO2, while Zn enhances the photoactivity due to its electronic properties.

Keywords: Congo red, Dopage, Photodecolourization, Titanium dioxide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2133
6243 Development of a Fiber based Interferometric Sensor for Non-contact Displacement Measurement

Authors: S. Pullteap

Abstract:

In this paper, a fiber based Fabry-Perot interferometer is proposed and demonstrated for a non-contact displacement measurement. A piece of micro-prism which attached to the mechanical vibrator is served as the target reflector. Interference signal is generated from the superposition between the sensing beam and the reference beam within the sensing arm of the fiber sensor. This signal is then converted to the displacement value by using a developed program written in visual Cµ programming with a resolution of λ/8. A classical function generator is operated for controlling the vibrator. By fixing an excitation frequency of 100 Hz and varying the excitation amplitude range of 0.1 – 3 Volts, the output displacements measured by the fiber sensor are obtained from 1.55 μm to 30.225 μm. A reference displacement sensor with a sensitivity of ~0.4 μm is also employed for comparing the displacement errors between both sensors. We found that over the entire displacement range, a maximum and average measurement error are obtained of 0.977% and 0.44% respectively.

Keywords: Non-contact displacement measurement, extrinsicfiber based Fabry-Perot interferometer, interference signal, zerocrossingfringe counting technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1996
6242 Customer Churn Prediction: A Cognitive Approach

Authors: Damith Senanayake, Lakmal Muthugama, Laksheen Mendis, Tiroshan Madushanka

Abstract:

Customer churn prediction is one of the most useful areas of study in customer analytics. Due to the enormous amount of data available for such predictions, machine learning and data mining have been heavily used in this domain. There exist many machine learning algorithms directly applicable for the problem of customer churn prediction, and here, we attempt to experiment on a novel approach by using a cognitive learning based technique in an attempt to improve the results obtained by using a combination of supervised learning methods, with cognitive unsupervised learning methods.

Keywords: Growing Self Organizing Maps, Kernel Methods, Churn Prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2529
6241 Web Content Mining: A Solution to Consumer's Product Hunt

Authors: Syed Salman Ahmed, Zahid Halim, Rauf Baig, Shariq Bashir

Abstract:

With the rapid growth in business size, today's businesses orient towards electronic technologies. Amazon.com and e-bay.com are some of the major stakeholders in this regard. Unfortunately the enormous size and hugely unstructured data on the web, even for a single commodity, has become a cause of ambiguity for consumers. Extracting valuable information from such an everincreasing data is an extremely tedious task and is fast becoming critical towards the success of businesses. Web content mining can play a major role in solving these issues. It involves using efficient algorithmic techniques to search and retrieve the desired information from a seemingly impossible to search unstructured data on the Internet. Application of web content mining can be very encouraging in the areas of Customer Relations Modeling, billing records, logistics investigations, product cataloguing and quality management. In this paper we present a review of some very interesting, efficient yet implementable techniques from the field of web content mining and study their impact in the area specific to business user needs focusing both on the customer as well as the producer. The techniques we would be reviewing include, mining by developing a knowledge-base repository of the domain, iterative refinement of user queries for personalized search, using a graphbased approach for the development of a web-crawler and filtering information for personalized search using website captions. These techniques have been analyzed and compared on the basis of their execution time and relevance of the result they produced against a particular search.

Keywords: Data mining, web mining, search engines, knowledge discovery.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2027
6240 Spatial Analysis and Statistics for Zoning of Urban Areas

Authors: Benedetto Manganelli, Beniamino Murgante

Abstract:

The use of statistical data and of the neural networks, capable of elaborate a series of data and territorial info, have allowed the making of a model useful in the subdivision of urban places into homogeneous zone under the profile of a social, real estate, environmental and urbanist background of a city. The development of homogeneous zone has fiscal and urbanist advantages. The tools in the model proposed, able to be adapted to the dynamic changes of the city, allow the application of the zoning fast and dynamic.

Keywords: Homogeneous Urban Areas, Multidimensional Scaling, Neural Network, Real Estate Market, Urban Planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1905
6239 A Comparison of SVM-based Criteria in Evolutionary Method for Gene Selection and Classification of Microarray Data

Authors: Rameswar Debnath, Haruhisa Takahashi

Abstract:

An evolutionary method whose selection and recombination operations are based on generalization error-bounds of support vector machine (SVM) can select a subset of potentially informative genes for SVM classifier very efficiently [7]. In this paper, we will use the derivative of error-bound (first-order criteria) to select and recombine gene features in the evolutionary process, and compare the performance of the derivative of error-bound with the error-bound itself (zero-order) in the evolutionary process. We also investigate several error-bounds and their derivatives to compare the performance, and find the best criteria for gene selection and classification. We use 7 cancer-related human gene expression datasets to evaluate the performance of the zero-order and first-order criteria of error-bounds. Though both criteria have the same strategy in theoretically, experimental results demonstrate the best criterion for microarray gene expression data.

Keywords: support vector machine, generalization error-bound, feature selection, evolutionary algorithm, microarray data

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1512
6238 A Grounded Theory on Marist Spirituality/Charism from the Perspective of the Lay Marists in the Philippines

Authors: Nino M. Pizarro

Abstract:

To the author’s knowledge, despite the written documents about Marist spirituality/charism, nothing has been done concerning a clear theoretical framework that highlights Marist spirituality/charism from the perspective or lived experience of the lay Marists of St. Marcellin Champagnat. The participants of the study are the lay Marist - educators who are from Marist Schools in the Philippines. Since the study would like to find out the respondents’ own concepts and meanings about Marist spirituality/charism, qualitative methodology is considered the approach to be used in the study. In particular, the study will use the qualitative methods of Barney Glaser. The theory will be generated systematically from data collection, coding and analyzing through memoing, theoretical sampling, sorting and writing and using the constant comparative method. The data collection method that will be employed in this grounded theory research is the in-depth interview that is semi-structured and participant driven. Data collection will be done through snowball sampling that is purposive. The study is considering to come up with a theoretical framework that will help the lay Marists to deepen their understanding of the Marist spirituality/charism and their vocation as lay partners of the Marist Brothers of the Schools.

Keywords: Grounded Theory, Lay Marists, Lived Experience, Marist Spirituality/Charism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 949
6237 Information Delivery and Advanced Traffic Information Systems in Istanbul

Authors: Kevser Simsek, Rahime Gunay

Abstract:

In this paper, we focused primarily on Istanbul data that is gathered by using intelligent transportation systems (ITS), and considered the developments in traffic information delivery and future applications that are being planned for implementation. Since traffic congestion is increasing and travel times are becoming less consistent and less predictable, traffic information delivery has become a critical issue. Considering the fuel consumption and wasted time in traffic, advanced traffic information systems are becoming increasingly valuable which enables travelers to plan their trips more accurately and easily.

Keywords: Data Fusion, Istanbul, ITS, Real Time Information, Traffic Information, Travel Time, Urban Mobility

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2016
6236 Modulation of Lipopolysaccharide Induced Interleukin-17F and Cyclooxygenase-2 Gene Expression by Echinacea purpurea in Broiler Chickens

Authors: Ali Asghar Saki, Sayed Ali Hosseini Siyar, Abbass Ashoori

Abstract:

This study was conducted to evaluate the effect of Echinacea purpurea on the expression of cyclooxygenase-2 (COX-2), interleukin-17F (IL-17F) in seven-day-old broiler chickens. Four groups were fed with concentration of 0 g/kg, 5 g/kg, 10 g/kg and 20 g/kg from the root of E. purpurea in the basal diet and two other groups were only fed with the basal diet for 21 days. At the 28th day, lipopolysaccharide (LPS, 2 mg/kg diet) was injected in four groups and the basal diet group was injected by saline as control. The chickens’ spleen RNA expression was measured for the COX-2 and IL-17F genes by Real-Time PCR. The results have shown that chickens which were fed E. purpurea had a lower COX-2 and IL-17F mRNA expression. The chickens who have received LPS only, lymphocyte was lower than other treatments. Vital organ weights were not significantly different, but body weight loss was recovered by dietary herbs inclusion. The results of this study have shown the positive effect of an anti-inflammatory herb to prevent the undesirable effect of inflammation.

Keywords: Echinacea purpurea, broiler chickens, gene expression, lipopolysaccharide.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 954
6235 MLOps Scaling Machine Learning Lifecycle in an Industrial Setting

Authors: Yizhen Zhao, Adam S. Z. Belloum, Gonc¸alo Maia da Costa, Zhiming Zhao

Abstract:

Machine learning has evolved from an area of academic research to a real-world applied field. This change comes with challenges, gaps and differences exist between common practices in academic environments and the ones in production environments. Following continuous integration, development and delivery practices in software engineering, similar trends have happened in machine learning (ML) systems, called MLOps. In this paper we propose a framework that helps to streamline and introduce best practices that facilitate the ML lifecycle in an industrial setting. This framework can be used as a template that can be customized to implement various machine learning experiments. The proposed framework is modular and can be recomposed to be adapted to various use cases (e.g. data versioning, remote training on Cloud). The framework inherits practices from DevOps and introduces other practices that are unique to the machine learning system (e.g.data versioning). Our MLOps practices automate the entire machine learning lifecycle, bridge the gap between development and operation.

Keywords: Cloud computing, continuous development, data versioning, DevOps, industrial setting, MLOps, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1006
6234 Dynamic Bayesian Networks Modeling for Inferring Genetic Regulatory Networks by Search Strategy: Comparison between Greedy Hill Climbing and MCMC Methods

Authors: Huihai Wu, Xiaohui Liu

Abstract:

Using Dynamic Bayesian Networks (DBN) to model genetic regulatory networks from gene expression data is one of the major paradigms for inferring the interactions among genes. Averaging a collection of models for predicting network is desired, rather than relying on a single high scoring model. In this paper, two kinds of model searching approaches are compared, which are Greedy hill-climbing Search with Restarts (GSR) and Markov Chain Monte Carlo (MCMC) methods. The GSR is preferred in many papers, but there is no such comparison study about which one is better for DBN models. Different types of experiments have been carried out to try to give a benchmark test to these approaches. Our experimental results demonstrated that on average the MCMC methods outperform the GSR in accuracy of predicted network, and having the comparable performance in time efficiency. By proposing the different variations of MCMC and employing simulated annealing strategy, the MCMC methods become more efficient and stable. Apart from comparisons between these approaches, another objective of this study is to investigate the feasibility of using DBN modeling approaches for inferring gene networks from few snapshots of high dimensional gene profiles. Through synthetic data experiments as well as systematic data experiments, the experimental results revealed how the performances of these approaches can be influenced as the target gene network varies in the network size, data size, as well as system complexity.

Keywords: Genetic regulatory network, Dynamic Bayesian network, GSR, MCMC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1862
6233 Design and Implementation of Secure Electronic Payment System (Client)

Authors: Pyae Pyae Hun

Abstract:

Secure electronic payment system is presented in this paper. This electronic payment system is to be secure for clients such as customers and shop owners. The security architecture of the system is designed by RC5 encryption / decryption algorithm. This eliminates the fraud that occurs today with stolen credit card numbers. The symmetric key cryptosystem RC5 can protect conventional transaction data such as account numbers, amount and other information. This process can be done electronically using RC5 encryption / decryption program written by Microsoft Visual Basic 6.0. There is no danger of any data sent within the system being intercepted, and replaced. The alternative is to use the existing network, and to encrypt all data transmissions. The system with encryption is acceptably secure, but that the level of encryption has to be stepped up, as computing power increases. Results In order to be secure the system the communication between modules is encrypted using symmetric key cryptosystem RC5. The system will use simple user name, password, user ID, user type and cipher authentication mechanism for identification, when the user first enters the system. It is the most common method of authentication in most computer system.

Keywords: A 128-bit block cipher, Microsoft visual basic 6.0, RC5 encryption /decryption algorithm and TCP/IP protocol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2341
6232 Equivalent Field Calculation to Irregular Symmetric and Asymmetric Photon Fields

Authors: N. Chegeni, M. J. Tahmasebi Birgani

Abstract:

Equivalent fields are frequently used for central axis depth-dose calculations of rectangular and irregular shaped photon beam. Since most of the proposed models to calculate the equivalent square field, are dosimetry-based, a simple physical-based method to calculate the equivalent square field size was used as the basis of this study. The table of the sides of the equivalent square for rectangular fields was constructed and then compared with the well-known tables of BJR and Venselaar with the average relative error percentage of 2.5±2.5 % and 1.5±1.5 % respectively. To evaluate the accuracy of this method, the PDDs were measured for some special irregular symmetric and asymmetric treatment fields and their equivalent squares for Siemens Primus Plus linear accelerator for both energies 6 and 18MV. The mean relative differences of PDDs measurement for these fields and their equivalent square was approximately 1% or less. As a result, this method can be employed to calculate equivalent field not only for rectangular fields but also for any irregular symmetric or asymmetric field.

Keywords: Equivalent field, asymmetric field, irregular field, multi leaf collimators.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5521
6231 Bootstrap Confidence Intervals and Parameter Estimation for Zero Inflated Strict Arcsine Model

Authors: Y. N. Phang, E. F. Loh

Abstract:

Zero inflated Strict Arcsine model is a newly developed model which is found to be appropriate in modeling overdispersed count data. In this study, maximum likelihood estimation method is used in estimating the parameters for zero inflated strict arcsine model. Bootstrapping is then employed to compute the confidence intervals for the estimated parameters.

Keywords: overdispersed count data, maximum likelihood estimation, simulated annealing, BCa confidence intervals.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2259
6230 Impact of Implementing VPN to Secure Wireless LAN

Authors: H. Bourdoucen, A. Al Naamany, A. Al Kalbani

Abstract:

Many corporations are seriously concerned about security of networks and therefore, their network supervisors are still reluctant to install WLANs. In this regards, the IEEE802.11i standard was developed to address the security problems, even though the mistrust of the wireless LAN technology is still existing. The thought was that the best security solutions could be found in open standards based technologies that can be delivered by Virtual Private Networking (VPN) being used for long time without addressing any security holes for the past few years. This work, addresses this issue and presents a simulated wireless LAN of IEEE802.11g protocol, and analyzes impact of integrating Virtual Private Network technology to secure the flow of traffic between the client and the server within the LAN, using OPNET WLAN utility. Two Wireless LAN scenarios have been introduced and simulated. These are based on normal extension to a wired network and VPN over extension to a wired network. The results of the two scenarios are compared and indicate the impact of improving performance, measured by response time and load, of Virtual Private Network over wireless LAN.

Keywords: IEEE802.11, VPN, Networking, Secure Wireless, WLAN, Opnet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2967
6229 Energy Loss at Drops using Neuro Solutions

Authors: Farzin Salmasi

Abstract:

Energy dissipation in drops has been investigated by physical models. After determination of effective parameters on the phenomenon, three drops with different heights have been constructed from Plexiglas. They have been installed in two existing flumes in the hydraulic laboratory. Several runs of physical models have been undertaken to measured required parameters for determination of the energy dissipation. Results showed that the energy dissipation in drops depend on the drop height and discharge. Predicted relative energy dissipations varied from 10.0% to 94.3%. This work has also indicated that the energy loss at drop is mainly due to the mixing of the jet with the pool behind the jet that causes air bubble entrainment in the flow. Statistical model has been developed to predict the energy dissipation in vertical drops denotes nonlinear correlation between effective parameters. Further an artificial neural networks (ANNs) approach was used in this paper to develop an explicit procedure for calculating energy loss at drops using NeuroSolutions. Trained network was able to predict the response with R2 and RMSE 0.977 and 0.0085 respectively. The performance of ANN was found effective when compared to regression equations in predicting the energy loss.

Keywords: Air bubble, drop, energy loss, hydraulic jump, NeuroSolutions

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1615
6228 One Hour Ahead Load Forecasting Using Artificial Neural Network for the Western Area of Saudi Arabia

Authors: A. J. Al-Shareef, E. A. Mohamed, E. Al-Judaibi

Abstract:

Load forecasting has become in recent years one of the major areas of research in electrical engineering. Most traditional forecasting models and artificial intelligence neural network techniques have been tried out in this task. Artificial neural networks (ANN) have lately received much attention, and a great number of papers have reported successful experiments and practical tests. This article presents the development of an ANN-based short-term load forecasting model with improved generalization technique for the Regional Power Control Center of Saudi Electricity Company, Western Operation Area (SEC-WOA). The proposed ANN is trained with weather-related data and historical electric load-related data using the data from the calendar years 2001, 2002, 2003, and 2004 for training. The model tested for one week at five different seasons, typically, winter, spring, summer, Ramadan and fall seasons, and the mean absolute average error for one hour-ahead load forecasting found 1.12%.

Keywords: Artificial neural networks, short-term load forecasting, back propagation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2086
6227 Theoretical Considerations for Software Component Metrics

Authors: V. Lakshmi Narasimhan, Bayu Hendradjaya

Abstract:

We have defined two suites of metrics, which cover static and dynamic aspects of component assembly. The static metrics measure complexity and criticality of component assembly, wherein complexity is measured using Component Packing Density and Component Interaction Density metrics. Further, four criticality conditions namely, Link, Bridge, Inheritance and Size criticalities have been identified and quantified. The complexity and criticality metrics are combined to form a Triangular Metric, which can be used to classify the type and nature of applications. Dynamic metrics are collected during the runtime of a complete application. Dynamic metrics are useful to identify super-component and to evaluate the degree of utilisation of various components. In this paper both static and dynamic metrics are evaluated using Weyuker-s set of properties. The result shows that the metrics provide a valid means to measure issues in component assembly. We relate our metrics suite with McCall-s Quality Model and illustrate their impact on product quality and to the management of component-based product development.

Keywords: Component Assembly, Component Based SoftwareEngineering, CORBA Component Model, Software ComponentMetrics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2258
6226 A Proposal for a Secure and Interoperable Data Framework for Energy Digitalization

Authors: Hebberly Ahatlan

Abstract:

The process of digitizing energy systems involves transforming traditional energy infrastructure into interconnected, data-driven systems that enhance efficiency, sustainability, and responsiveness. As smart grids become increasingly integral to the efficient distribution and management of electricity from both fossil and renewable energy sources, the energy industry faces strategic challenges associated with digitalization and interoperability — particularly in the context of modern energy business models, such as virtual power plants (VPPs). The critical challenge in modern smart grids is to seamlessly integrate diverse technologies and systems, including virtualization, grid computing and service-oriented architecture (SOA), across the entire energy ecosystem. Achieving this requires addressing issues like semantic interoperability, Information Technology (IT) and Operational Technology (OT) convergence, and digital asset scalability, all while ensuring security and risk management. This paper proposes a four-layer digitalization framework to tackle these challenges, encompassing persistent data protection, trusted key management, secure messaging, and authentication of IoT resources. Data assets generated through this framework enable AI systems to derive insights for improving smart grid operations, security, and revenue generation. Furthermore, this paper also proposes a Trusted Energy Interoperability Alliance as a universal guiding standard in the development of this digitalization framework to support more dynamic and interoperable energy markets.

Keywords: Digitalization, IT/OT convergence, semantic interoperability, TEIA alliance, VPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 64
6225 Needs of Omani Children in First Grade during Their Transition from Kindergarten to Primary School: An Ethnographic Study

Authors: Zainab Algharibi, Julie McAdam, Catherine Fagan

Abstract:

The purpose of this paper is to shed light on how Omani children in the first grade experience their needs during their transition to primary school. Theoretically, the paper was built on two perspectives: Dewey's concept of continuity of experience and the boundary objects introduced by Vygotsky (CHAT). The methodology of the study is based on the crucial role of children’s agency which is a very important activity as an educational tool to enhance the child’s participation in the learning process and develop their ability to face various issues in their life. Thus, the data were obtained from 45 children in grade one from four different primary schools using drawing and visual narrative activities, in addition to researcher observations during the start of the first weeks of the academic year for the first grade. As the study dealt with children, all of the necessary ethical laws were followed. This paper is considered original since it seeks to deal with the issue of children's transition from kindergarten to primary school not only in Oman, but in the Arab region. Therefore, it is expected to fill an important gap in this field and present a proposal that will be a door for researchers to enter this research field later. The analysis of drawing and visual narrative was performed according to the social semiotics approach in two phases. The first is to read out the surface message “denotation,” while the second is to go in-depth via the symbolism obtained from children while they talked and drew letters and signs. This stage is known as “signified”; a video was recorded of each child talking about their drawing and expressing themself. Then, the data were organised and classified according to a cross-data network. Regarding the researcher observation analyses, the collected data were analysed according to the "grounded theory". It is based on comparing the recent data collected from observations with data previously encoded by other methods in which children were drawing alongside the visual narrative in the current study, in order to identify the similarities and differences, and also to clarify the meaning of the accessed categories and to identify sub-categories of them with a description of possible links between them. This is a kind of triangulation in data collection. The study came up with a set of findings, the most vital being that the children's greatest interest goes to their social and psychological needs, such as friends, their teacher, and playing. Also, their biggest fears are a new place, a new teacher, and not having friends, while they showed less concern for their need for educational knowledge and skills.

Keywords: Children’s academic needs, children’s social needs, children transition, primary school.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 138
6224 Location Based Clustering in Wireless Sensor Networks

Authors: Ashok Kumar, Narottam Chand, Vinod Kumar

Abstract:

Due to the limited energy resources, energy efficient operation of sensor node is a key issue in wireless sensor networks. Clustering is an effective method to prolong the lifetime of energy constrained wireless sensor network. However, clustering in wireless sensor network faces several challenges such as selection of an optimal group of sensor nodes as cluster, optimum selection of cluster head, energy balanced optimal strategy for rotating the role of cluster head in a cluster, maintaining intra and inter cluster connectivity and optimal data routing in the network. In this paper, we propose a protocol supporting an energy efficient clustering, cluster head selection/rotation and data routing method to prolong the lifetime of sensor network. Simulation results demonstrate that the proposed protocol prolongs network lifetime due to the use of efficient clustering, cluster head selection/rotation and data routing.

Keywords: Wireless sensor networks, clustering, energy efficient, localization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2662
6223 Investigation of Active Modified Atmosphere and Nanoparticle Packaging on Quality of Tomatoes

Authors: M. Ghasemi-Varnamkhasti, S. H. Yoosefian, A. Mohammad-Razdari

Abstract:

This study investigated the effects of Ag nanoparticle polyethylene film and active modified atmosphere on the postharvest quality of tomatoes stored at 6 ºC. The atmosphere composition used in the packaging was 7% O2 + 7% CO2 + 86% N2, and synthetic air (control). The variables measured were weight loss, firmness, color and respiration rate over 21 days. The results showed that the combination of Ag nanoparticle polyethylene film and modified atmosphere could extend the shelf life of tomatoes to 21 days and could influence the postharvest quality of tomatoes. Also, existence of Ag nanoparticles caused preventing from increasing weight loss, a*, b*, Chroma, Hue angle and reducing firmness and L*. As well as, tomatoes at Ag nanoparticle polyethylene films had lower respiration rate than Polyethylene and paper bags to 13.27% and 23.50%, respectively. The combination of Ag nanoparticle polyethylene film and active modified atmosphere was effective with regard to delaying maturity during the storage period, and preserving the quality of tomatoes.

Keywords: Ag nanoparticles, modified atmosphere, polyethylene film, tomato.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1117
6222 Probability Distribution of Rainfall Depth at Hourly Time-Scale

Authors: S. Dan'azumi, S. Shamsudin, A. A. Rahman

Abstract:

Rainfall data at fine resolution and knowledge of its characteristics plays a major role in the efficient design and operation of agricultural, telecommunication, runoff and erosion control as well as water quality control systems. The paper is aimed to study the statistical distribution of hourly rainfall depth for 12 representative stations spread across Peninsular Malaysia. Hourly rainfall data of 10 to 22 years period were collected and its statistical characteristics were estimated. Three probability distributions namely, Generalized Pareto, Exponential and Gamma distributions were proposed to model the hourly rainfall depth, and three goodness-of-fit tests, namely, Kolmogorov-Sminov, Anderson-Darling and Chi-Squared tests were used to evaluate their fitness. Result indicates that the east cost of the Peninsular receives higher depth of rainfall as compared to west coast. However, the rainfall frequency is found to be irregular. Also result from the goodness-of-fit tests show that all the three models fit the rainfall data at 1% level of significance. However, Generalized Pareto fits better than Exponential and Gamma distributions and is therefore recommended as the best fit.

Keywords: Goodness-of-fit test, Hourly rainfall, Malaysia, Probability distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2887
6221 Classifying and Predicting Efficiencies Using Interval DEA Grid Setting

Authors: Yiannis G. Smirlis

Abstract:

The classification and the prediction of efficiencies in Data Envelopment Analysis (DEA) is an important issue, especially in large scale problems or when new units frequently enter the under-assessment set. In this paper, we contribute to the subject by proposing a grid structure based on interval segmentations of the range of values for the inputs and outputs. Such intervals combined, define hyper-rectangles that partition the space of the problem. This structure, exploited by Interval DEA models and a dominance relation, acts as a DEA pre-processor, enabling the classification and prediction of efficiency scores, without applying any DEA models.

Keywords: Data envelopment analysis, interval DEA, efficiency classification, efficiency prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 911
6220 Application of Gamma Frailty Model in Survival of Liver Cirrhosis Patients

Authors: Elnaz Saeedi, Jamileh Abolaghasemi, Mohsen Nasiri Tousi, Saeedeh Khosravi

Abstract:

Goals and Objectives: A typical analysis of survival data involves the modeling of time-to-event data, such as the time till death. A frailty model is a random effect model for time-to-event data, where the random effect has a multiplicative influence on the baseline hazard function. This article aims to investigate the use of gamma frailty model with concomitant variable in order to individualize the prognostic factors that influence the liver cirrhosis patients’ survival times. Methods: During the one-year study period (May 2008-May 2009), data have been used from the recorded information of patients with liver cirrhosis who were scheduled for liver transplantation and were followed up for at least seven years in Imam Khomeini Hospital in Iran. In order to determine the effective factors for cirrhotic patients’ survival in the presence of latent variables, the gamma frailty distribution has been applied. In this article, it was considering the parametric model, such as Exponential and Weibull distributions for survival time. Data analysis is performed using R software, and the error level of 0.05 was considered for all tests. Results: 305 patients with liver cirrhosis including 180 (59%) men and 125 (41%) women were studied. The age average of patients was 39.8 years. At the end of the study, 82 (26%) patients died, among them 48 (58%) were men and 34 (42%) women. The main cause of liver cirrhosis was found hepatitis 'B' with 23%, followed by cryptogenic with 22.6% were identified as the second factor. Generally, 7-year’s survival was 28.44 months, for dead patients and for censoring was 19.33 and 31.79 months, respectively. Using multi-parametric survival models of progressive and regressive, Exponential and Weibull models with regard to the gamma frailty distribution were fitted to the cirrhosis data. In both models, factors including, age, bilirubin serum, albumin serum, and encephalopathy had a significant effect on survival time of cirrhotic patients. Conclusion: To investigate the effective factors for the time of patients’ death with liver cirrhosis in the presence of latent variables, gamma frailty model with parametric distributions seems desirable.

Keywords: Frailty model, latent variables, liver cirrhosis, parametric distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1034
6219 Removal of Pharmaceutical Compounds by a Sequential Treatment of Ozonation Followed by Fenton Process: Influence of the Water Matrix

Authors: Almudena Aguinaco, Olga Gimeno, Fernando J. Beltrán, Juan José P. Sagasti

Abstract:

A sequential treatment of ozonation followed by a Fenton or photo-Fenton process, using black light lamps (365 nm) in this latter case, has been applied to remove a mixture of pharmaceutical compounds and the generated by-products both in ultrapure and secondary treated wastewater. The scientifictechnological innovation of this study stems from the in situ generation of hydrogen peroxide from the direct ozonation of pharmaceuticals, and can later be used in the application of Fenton and photo-Fenton processes. The compounds selected as models were sulfamethoxazol and acetaminophen. It should be remarked that the use of a second process is necessary as a result of the low mineralization yield reached by the exclusive application of ozone. Therefore, the influence of the water matrix has been studied in terms of hydrogen peroxide concentration, individual compound concentration and total organic carbon removed. Moreover, the concentration of different iron species in solution has been measured.

Keywords: Fenton, photo-Fenton, ozone, pharmaceutical compounds, hydrogen peroxide, water treatment

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1818
6218 Operational Risk – Scenario Analysis

Authors: Milan Rippel, Petr Teply

Abstract:

This paper focuses on operational risk measurement techniques and on economic capital estimation methods. A data sample of operational losses provided by an anonymous Central European bank is analyzed using several approaches. Loss Distribution Approach and scenario analysis method are considered. Custom plausible loss events defined in a particular scenario are merged with the original data sample and their impact on capital estimates and on the financial institution is evaluated. Two main questions are assessed – What is the most appropriate statistical method to measure and model operational loss data distribution? and What is the impact of hypothetical plausible events on the financial institution? The g&h distribution was evaluated to be the most suitable one for operational risk modeling. The method based on the combination of historical loss events modeling and scenario analysis provides reasonable capital estimates and allows for the measurement of the impact of extreme events on banking operations.

Keywords: operational risk, scenario analysis, economic capital, loss distribution approach, extreme value theory, stress testing

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2399
6217 Performance Evaluation of Data Mining Techniques for Predicting Software Reliability

Authors: Pradeep Kumar, Abdul Wahid

Abstract:

Accurate software reliability prediction not only enables developers to improve the quality of software but also provides useful information to help them for planning valuable resources. This paper examines the performance of three well-known data mining techniques (CART, TreeNet and Random Forest) for predicting software reliability. We evaluate and compare the performance of proposed models with Cascade Correlation Neural Network (CCNN) using sixteen empirical databases from the Data and Analysis Center for Software. The goal of our study is to help project managers to concentrate their testing efforts to minimize the software failures in order to improve the reliability of the software systems. Two performance measures, Normalized Root Mean Squared Error (NRMSE) and Mean Absolute Errors (MAE), illustrate that CART model is accurate than the models predicted using Random Forest, TreeNet and CCNN in all datasets used in our study. Finally, we conclude that such methods can help in reliability prediction using real-life failure datasets.

Keywords: Classification, Cascade Correlation Neural Network, Random Forest, Software reliability, TreeNet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1811
6216 Ultrasound Mechanical Index as a Parameter Affecting of the Ability of Proliferation of Cells

Authors: Z. Hormozi Moghaddam, M. Mokhtari-Dizaji, M. Movahedin, M. E. Ravari

Abstract:

Mechanical index (MI) is used for quantifying acoustic cavitation and the relationship between acoustic pressure and the frequency. In this study, modeling of the MI was applied to provide treatment protocol and to understand the effective physical processes on reproducibility of stem cells. The acoustic pressure and MI equations are modeled and solved to estimate optimal MI for 28, 40, 150 kHz and 1 MHz frequencies. Radial and axial acoustic pressure distribution was extracted. To validate the results of the modeling, the acoustic pressure in the water and near field depth was measured by a piston hydrophone. Results of modeling and experiments show that the model is consistent well to experimental results with 0.91 and 0.90 correlation of coefficient (p<0.05) for 1 MHz and 40 kHz. Low intensity ultrasound with 0.40 MI is more effective on the proliferation rate of the spermatogonial stem cells during the seven days of culture, in contrast, high MI has a harmful effect on the spermatogonial stem cells. This model provides proper treatment planning in vitro and in vivo by estimating the cavitation phenomenon.

Keywords: Ultrasound, mechanical index, modeling, stem cell.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 934
6215 Application of Wireless Visual Sensor for Semi- Autonomous Mine Navigation System

Authors: Vinay Kumar Pilania, Debashish Chakravarty

Abstract:

The present paper represent the efforts undertaken for the development of an semi-automatic robot that may be used for various post-disaster rescue operation planning and their subsequent execution using one-way communication of video and data from the robot to the controller and controller to the robot respectively. Wireless communication has been used for the purpose so that the robot may access the unapproachable places easily without any difficulties. It is expected that the information obtained from the robot would be of definite help to the rescue team for better planning and execution of their operations.

Keywords: Mine environment, mine navigation, mine rescue robot, video data transmission.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1703