Search results for: digital forensic processes and procedures
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2922

Search results for: digital forensic processes and procedures

852 Predictors of Academic Achievement of Student ICT Teachers with Different Learning Styles

Authors: Deniz Deryakulu, Şener Büyüköztürk Hüseyin Özçınar

Abstract:

The main purpose of this study was to determine the predictors of academic achievement of student Information and Communications Technologies (ICT) teachers with different learning styles. Participants were 148 student ICT teachers from Ankara University. Participants were asked to fill out a personal information sheet, the Turkish version of Kolb-s Learning Style Inventory, Weinstein-s Learning and Study Strategies Inventory, Schommer's Epistemological Beliefs Questionnaire, and Eysenck-s Personality Questionnaire. Stepwise regression analyses showed that the statistically significant predictors of the academic achievement of the accommodators were attitudes and high school GPAs; of the divergers was anxiety; of the convergers were gender, epistemological beliefs, and motivation; and of the assimilators were gender, personality, and test strategies. Implications for ICT teaching-learning processes and teacher education are discussed.

Keywords: Academic achievement, student ICT teachers, Kolb learning styles, experiential learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2581
851 Integrating Fast Karnough Map and Modular Neural Networks for Simplification and Realization of Complex Boolean Functions

Authors: Hazem M. El-Bakry

Abstract:

In this paper a new fast simplification method is presented. Such method realizes Karnough map with large number of variables. In order to accelerate the operation of the proposed method, a new approach for fast detection of group of ones is presented. Such approach implemented in the frequency domain. The search operation relies on performing cross correlation in the frequency domain rather than time one. It is proved mathematically and practically that the number of computation steps required for the presented method is less than that needed by conventional cross correlation. Simulation results using MATLAB confirm the theoretical computations. Furthermore, a powerful solution for realization of complex functions is given. The simplified functions are implemented by using a new desigen for neural networks. Neural networks are used because they are fault tolerance and as a result they can recognize signals even with noise or distortion. This is very useful for logic functions used in data and computer communications. Moreover, the implemented functions are realized with minimum amount of components. This is done by using modular neural nets (MNNs) that divide the input space into several homogenous regions. Such approach is applied to implement XOR function, 16 logic functions on one bit level, and 2-bit digital multiplier. Compared to previous non- modular designs, a clear reduction in the order of computations and hardware requirements is achieved.

Keywords: Boolean Functions, Simplification, KarnoughMap, Implementation of Logic Functions, Modular NeuralNetworks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1797
850 Chose the Right Mutation Rate for Better Evolve Combinational Logic Circuits

Authors: Emanuele Stomeo, Tatiana Kalganova, Cyrille Lambert

Abstract:

Evolvable hardware (EHW) is a developing field that applies evolutionary algorithm (EA) to automatically design circuits, antennas, robot controllers etc. A lot of research has been done in this area and several different EAs have been introduced to tackle numerous problems, as scalability, evolvability etc. However every time a specific EA is chosen for solving a particular task, all its components, such as population size, initialization, selection mechanism, mutation rate, and genetic operators, should be selected in order to achieve the best results. In the last three decade the selection of the right parameters for the EA-s components for solving different “test-problems" has been investigated. In this paper the behaviour of mutation rate for designing logic circuits, which has not been done before, has been deeply analyzed. The mutation rate for an EHW system modifies the number of inputs of each logic gates, the functionality (for example from AND to NOR) and the connectivity between logic gates. The behaviour of the mutation has been analyzed based on the number of generations, genotype redundancy and number of logic gates for the evolved circuits. The experimental results found provide the behaviour of the mutation rate during evolution for the design and optimization of simple logic circuits. The experimental results propose the best mutation rate to be used for designing combinational logic circuits. The research presented is particular important for those who would like to implement a dynamic mutation rate inside the evolutionary algorithm for evolving digital circuits. The researches on the mutation rate during the last 40 years are also summarized.

Keywords: Design of logic circuit, evolutionary computation, evolvable hardware, mutation rate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1676
849 On the use of Ionic Liquids for CO2 Capturing

Authors: Emad Ali, Inas Alnashef, Abdelhamid Ajbar, Mohamed HadjKali, Sarwono Mulyono

Abstract:

In this work, ionic liquids (ILs) for CO2 capturing in typical absorption/stripper process are considered. The use of ionic liquids is considered to be cost-effective because it requires less energy for solvent recovery compared to other conventional processes. A mathematical model is developed for the process based on Peng-Robinson (PR) equation of state (EoS) which is validated with experimental data for various solutions involving CO2. The model is utilized to study the sorbent and energy demand for three types of ILs at specific CO2 capturing rates. The energy demand is manifested by the vapor-liquid equilibrium temperature necessary to remove the captured CO2 from the used solvent in the regeneration step. It is found that higher recovery temperature is required for solvents with higher solubility coefficient. For all ILs, the temperature requirement is less than that required by the typical monoethanolamine (MEA) solvent. The effect of the CO2 loading in the sorbent stream on the process performance is also examined.

Keywords: Ionic liquid, CO2 capturing, CO2 solubility, Vaporliquid equilibrium.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2696
848 Determination of the Specific Activity of Soil and Fertilizers in Sergipe - Brazil

Authors: Leandro X. Cardoso, Susana O. Souza, Fernanda C. L. Ferreira, Orlando C. Ferreira, Elenilson Barboza, Carlos E. Alhanati

Abstract:

Measurements of radioactivity in the environment is of great importance to monitor and control the levels of radiation to which man is exposed directly or indirectly. It is necessary to show that regardless of working or being close to nuclear power plants, people are daily in contact with some amount of radiation from the actual environment and food that are ingested, contradicting the view of most of them. The aim of this study was to analyze the rate of natural and artificial radiation from radionuclides present in cement, soil and fertilizers used in Sergipe State – Brazil. The radionuclide activitiesmeasured all samples arebelow the Brazilian limit of the exclusion and exemption criteria from the requirement of radiation protection.It was detected Be-7 in organic fertilizers that means a short interval between the brewing processes for use in agriculture. It was also detected an unexpected Cs-137 in some samples; however its activities does not represent risk for the population. Th-231 was also found in samples of soil and cement in the state of Sergipe that is an unprecedented result.

Keywords: Cs-137, Be-7, Th-231 radiation dose, radio isotopes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1991
847 Distances over Incomplete Diabetes and Breast Cancer Data Based on Bhattacharyya Distance

Authors: Loai AbdAllah, Mahmoud Kaiyal

Abstract:

Missing values in real-world datasets are a common problem. Many algorithms were developed to deal with this problem, most of them replace the missing values with a fixed value that was computed based on the observed values. In our work, we used a distance function based on Bhattacharyya distance to measure the distance between objects with missing values. Bhattacharyya distance, which measures the similarity of two probability distributions. The proposed distance distinguishes between known and unknown values. Where the distance between two known values is the Mahalanobis distance. When, on the other hand, one of them is missing the distance is computed based on the distribution of the known values, for the coordinate that contains the missing value. This method was integrated with Wikaya, a digital health company developing a platform that helps to improve prevention of chronic diseases such as diabetes and cancer. In order for Wikaya’s recommendation system to work distance between users need to be measured. Since there are missing values in the collected data, there is a need to develop a distance function distances between incomplete users profiles. To evaluate the accuracy of the proposed distance function in reflecting the actual similarity between different objects, when some of them contain missing values, we integrated it within the framework of k nearest neighbors (kNN) classifier, since its computation is based only on the similarity between objects. To validate this, we ran the algorithm over diabetes and breast cancer datasets, standard benchmark datasets from the UCI repository. Our experiments show that kNN classifier using our proposed distance function outperforms the kNN using other existing methods.

Keywords: Missing values, distance metric, Bhattacharyya distance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 765
846 Integrating Fast Karnough Map and Modular Neural Networks for Simplification and Realization of Complex Boolean Functions

Authors: Hazem M. El-Bakry

Abstract:

In this paper a new fast simplification method is presented. Such method realizes Karnough map with large number of variables. In order to accelerate the operation of the proposed method, a new approach for fast detection of group of ones is presented. Such approach implemented in the frequency domain. The search operation relies on performing cross correlation in the frequency domain rather than time one. It is proved mathematically and practically that the number of computation steps required for the presented method is less than that needed by conventional cross correlation. Simulation results using MATLAB confirm the theoretical computations. Furthermore, a powerful solution for realization of complex functions is given. The simplified functions are implemented by using a new desigen for neural networks. Neural networks are used because they are fault tolerance and as a result they can recognize signals even with noise or distortion. This is very useful for logic functions used in data and computer communications. Moreover, the implemented functions are realized with minimum amount of components. This is done by using modular neural nets (MNNs) that divide the input space into several homogenous regions. Such approach is applied to implement XOR function, 16 logic functions on one bit level, and 2-bit digital multiplier. Compared to previous non- modular designs, a clear reduction in the order of computations and hardware requirements is achieved.

Keywords: Boolean functions, simplification, Karnough map, implementation of logic functions, modular neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2055
845 A Proposed Hybrid Color Image Compression Based on Fractal Coding with Quadtree and Discrete Cosine Transform

Authors: Shimal Das, Dibyendu Ghoshal

Abstract:

Fractal based digital image compression is a specific technique in the field of color image. The method is best suited for irregular shape of image like snow bobs, clouds, flame of fire; tree leaves images, depending on the fact that parts of an image often resemble with other parts of the same image. This technique has drawn much attention in recent years because of very high compression ratio that can be achieved. Hybrid scheme incorporating fractal compression and speedup techniques have achieved high compression ratio compared to pure fractal compression. Fractal image compression is a lossy compression method in which selfsimilarity nature of an image is used. This technique provides high compression ratio, less encoding time and fart decoding process. In this paper, fractal compression with quad tree and DCT is proposed to compress the color image. The proposed hybrid schemes require four phases to compress the color image. First: the image is segmented and Discrete Cosine Transform is applied to each block of the segmented image. Second: the block values are scanned in a zigzag manner to prevent zero co-efficient. Third: the resulting image is partitioned as fractals by quadtree approach. Fourth: the image is compressed using Run length encoding technique.

Keywords: Fractal coding, Discrete Cosine Transform, Iterated Function System (IFS), Affine Transformation, Run length encoding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1553
844 A New Approach to Workforce Planning

Authors: M. Othman, N. Bhuiyan, G. J. Gouw

Abstract:

In today-s global and competitive market, manufacturing companies are working hard towards improving their production system performance. Most companies develop production systems that can help in cost reduction. Manufacturing systems consist of different elements including production methods, machines, processes, control and information systems. Human issues are an important part of manufacturing systems, yet most companies do not pay sufficient attention to them. In this paper, a workforce planning (WP) model is presented. A non-linear programming model is developed in order to minimize the hiring, firing, training and overtime costs. The purpose is to determine the number of workers for each worker type, the number of workers trained, and the number of overtime hours. Moreover, a decision support system (DSS) based on the proposed model is introduced using the Excel-Lingo software interfacing feature. This model will help to improve the interaction between the workers, managers and the technical systems in manufacturing.

Keywords: Decision Support System, Human Factors, Manufacturing System, Workforce Planning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2530
843 A Survey on Performance Tools for OpenMP

Authors: Mubarak S. Mohsen, Rosni Abdullah, Yong M. Teo

Abstract:

Advances in processors architecture, such as multicore, increase the size of complexity of parallel computer systems. With multi-core architecture there are different parallel languages that can be used to run parallel programs. One of these languages is OpenMP which embedded in C/Cµ or FORTRAN. Because of this new architecture and the complexity, it is very important to evaluate the performance of OpenMP constructs, kernels, and application program on multi-core systems. Performance is the activity of collecting the information about the execution characteristics of a program. Performance tools consists of at least three interfacing software layers, including instrumentation, measurement, and analysis. The instrumentation layer defines the measured performance events. The measurement layer determines what performance event is actually captured and how it is measured by the tool. The analysis layer processes the performance data and summarizes it into a form that can be displayed in performance tools. In this paper, a number of OpenMP performance tools are surveyed, explaining how each is used to collect, analyse, and display data collection.

Keywords: Parallel performance tools, OpenMP, multi-core.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1899
842 Hydrological Characterization of a Watershed for Streamflow Prediction

Authors: Oseni Taiwo Amoo, Bloodless Dzwairo

Abstract:

In this paper, we extend the versatility and usefulness of GIS as a methodology for any river basin hydrologic characteristics analysis (HCA). The Gurara River basin located in North-Central Nigeria is presented in this study. It is an on-going research using spatial Digital Elevation Model (DEM) and Arc-Hydro tools to take inventory of the basin characteristics in order to predict water abstraction quantification on streamflow regime. One of the main concerns of hydrological modelling is the quantification of runoff from rainstorm events. In practice, the soil conservation service curve (SCS) method and the Conventional procedure called rational technique are still generally used these traditional hydrological lumped models convert statistical properties of rainfall in river basin to observed runoff and hydrograph. However, the models give little or no information about spatially dispersed information on rainfall and basin physical characteristics. Therefore, this paper synthesizes morphometric parameters in generating runoff. The expected results of the basin characteristics such as size, area, shape, slope of the watershed and stream distribution network analysis could be useful in estimating streamflow discharge. Water resources managers and irrigation farmers could utilize the tool for determining net return from available scarce water resources, where past data records are sparse for the aspect of land and climate.

Keywords: Hydrological characteristic, land and climate, runoff discharge, streamflow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1439
841 Secure Mobile E-Business Applications

Authors: Hala A. Alrumaih

Abstract:

It is widely believed that mobile device is a promising technology for lending the opportunity for the third wave of electronic commerce. Mobile devices have changed the way companies do business. Many applications are under development or being incorporated into business processes. In this day, mobile applications are a vital component of any industry strategy.One of the greatest benefits of selling merchandise and providing services on a mobile application is that it widens a company’s customer base significantly.Mobile applications are accessible to interested customers across regional and international borders in different electronic business (e-business) area. But there is a dark side to this success story. The security risks associated with mobile devices and applications are very significant. This paper introduces a broad risk analysis for the various threats, vulnerabilities, and risks in mobile e-business applications and presents some important risk mitigation approaches. It reviews and compares two different frameworks for security assurance in mobile e-business applications. Based on the comparison, the paper suggests some recommendations for applications developers and business owners in mobile e-business application development process.

Keywords: E-business, Mobile Applications, Risk mitigations, Security assurance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2480
840 Analysis of a TBM Tunneling Effect on Surface Subsidence: A Case Study from Tehran, Iran

Authors: A. R. Salimi, M. Esmaeili, B. Salehi

Abstract:

The development and extension of large cities induced a need for shallow tunnel in soft ground of building areas. Estimation of ground settlement caused by the tunnel excavation is important engineering point. In this paper, prediction of surface subsidence caused by tunneling in one section of seventh line of Tehran subway is considered. On the basis of studied geotechnical conditions of the region, tunnel with the length of 26.9km has been excavated applying a mechanized method using an EPB-TBM with a diameter of 9.14m. In this regard, settlement is estimated utilizing both analytical and numerical finite element method. The numerical method shows that the value of settlement in this section is 5cm. Besides, the analytical consequences (Bobet and Loganathan-Polous) are 5.29 and 12.36cm, respectively. According to results of this study, due tosaturation of this section, there are good agreement between Bobet and numerical methods. Therefore, tunneling processes in this section needs a special consolidation measurement and support system before the passage of tunnel boring machine.

Keywords: TBM, Subsidence, Numerical Method, Analytical Method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5395
839 Multidimensional Visualization Tools for Analysis of Expression Data

Authors: Urska Cvek, Marjan Trutschl, Randolph Stone II, Zanobia Syed, John L. Clifford, Anita L. Sabichi

Abstract:

Expression data analysis is based mostly on the statistical approaches that are indispensable for the study of biological systems. Large amounts of multidimensional data resulting from the high-throughput technologies are not completely served by biostatistical techniques and are usually complemented with visual, knowledge discovery and other computational tools. In many cases, in biological systems we only speculate on the processes that are causing the changes, and it is the visual explorative analysis of data during which a hypothesis is formed. We would like to show the usability of multidimensional visualization tools and promote their use in life sciences. We survey and show some of the multidimensional visualization tools in the process of data exploration, such as parallel coordinates and radviz and we extend them by combining them with the self-organizing map algorithm. We use a time course data set of transitional cell carcinoma of the bladder in our examples. Analysis of data with these tools has the potential to uncover additional relationships and non-trivial structures.

Keywords: microarrays, visualization, parallel coordinates, radviz, self-organizing maps.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2494
838 All-Silicon Raman Laser with Quasi-Phase-Matched Structures and Resonators

Authors: Isao Tomita

Abstract:

The principle of all-silicon Raman lasers for an output wavelength of 1.3 μm is presented, which employs quasi-phase-matched structures and resonators to enhance the output power. 1.3-μm laser beams for GE-PONs in FTTH systems generated from a silicon device are very important because such a silicon device can be monolithically integrated with the silicon planar lightwave circuits (Si PLCs) used in the GE-PONs. This reduces the device fabrication processes and time and also optical losses at the junctions between optical waveguides of the Si PLCs and Si laser devices when compared with 1.3-μm III-V semiconductor lasers set on the Si PLCs employed at present. We show that the quasi-phase-matched Si Raman laser with resonators can produce about 174 times larger laser power at 1.3 μm (at maximum) than that without resonators for a Si waveguide of Raman gain 20 cm/GW and optical loss 1.2 dB/cm, pumped at power 10 mW, where the length of the waveguide is 3 mm and its cross-section is (1.5 μm)2.

Keywords: All-silicon raman laser, FTTH, GE-PON, quasi-phase-matched structure, resonator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 874
837 Investigation into Heterotrophic Activities and Algal Biomass in Surface Flow Stormwater Wetlands

Authors: Wendong Tao

Abstract:

Stormwater wetlands have been mainly designed in an empirical approach for water quality improvement, with little quantitative understanding of the internal microbial processes. This study investigated into heterotrophic bacterial production rate, heterotrophic bacterial mineralization percentage, and algal biomass in hypertrophic and eutrophic surface flow stormwater wetlands. Compared to a nearby wood leachate treatment wetland, the stormwater wetlands had much higher chlorophyll-a concentrations. The eutrophic stormwater wetland had improved water quality, whereas the hypertrophic stormwater wetland had degraded water quality. Heterotrophic bacterial activities in water were limited in the stormwater wetlands due to competition of algal growth for nutrients. The relative contribution of biofilms to the overall heterotrophic activities was higher in the stormwater wetlands than that in the wood leachate treatment wetland.

Keywords: chlorophyll-a, constructed wetland, heterotrophicproduction, mineralization, stormwater

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1474
836 Statistical Modeling of Mobile Fading Channels Based on Triply Stochastic Filtered Marked Poisson Point Processes

Authors: Jihad S. Daba, J. P. Dubois

Abstract:

Understanding the statistics of non-isotropic scattering multipath channels that fade randomly with respect to time, frequency, and space in a mobile environment is very crucial for the accurate detection of received signals in wireless and cellular communication systems. In this paper, we derive stochastic models for the probability density function (PDF) of the shift in the carrier frequency caused by the Doppler Effect on the received illuminating signal in the presence of a dominant line of sight. Our derivation is based on a generalized Clarke’s and a two-wave partially developed scattering models, where the statistical distribution of the frequency shift is shown to be consistent with the power spectral density of the Doppler shifted signal.

Keywords: Doppler shift, filtered Poisson process, generalized Clark’s model, non-isotropic scattering, partially developed scattering, Rician distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 818
835 DEA ANN Approach in Supplier Evaluation System

Authors: Dilek Özdemir, Gül Tekin Temur

Abstract:

In Supply Chain Management (SCM), strengthening partnerships with suppliers is a significant factor for enhancing competitiveness. Hence, firms increasingly emphasize supplier evaluation processes. Supplier evaluation systems are basically developed in terms of criteria such as quality, cost, delivery, and flexibility. Because there are many variables to be analyzed, this process becomes hard to execute and needs expertise. On this account, this study aims to develop an expert system on supplier evaluation process by designing Artificial Neural Network (ANN) that is supported with Data Envelopment Analysis (DEA). The methods are applied on the data of 24 suppliers, which have longterm relationships with a medium sized company from German Iron and Steel Industry. The data of suppliers consists of variables such as material quality (MQ), discount of amount (DOA), discount of cash (DOC), payment term (PT), delivery time (DT) and annual revenue (AR). Meanwhile, the efficiency that is generated by using DEA is added to the supplier evaluation system in order to use them as system outputs.

Keywords: Artificial Neural Network (ANN), DataEnvelopment Analysis (DEA), Supplier Evaluation System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2136
834 Routing Medical Images with Tabu Search and Simulated Annealing: A Study on Quality of Service

Authors: Mejía M. Paula, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

In telemedicine, the image repository service is important to increase the accuracy of diagnostic support of medical personnel. This study makes comparison between two routing algorithms regarding the quality of service (QoS), to be able to analyze the optimal performance at the time of loading and/or downloading of medical images. This study focused on comparing the performance of Tabu Search with other heuristic and metaheuristic algorithms that improve QoS in telemedicine services in Colombia. For this, Tabu Search and Simulated Annealing heuristic algorithms are chosen for their high usability in this type of applications; the QoS is measured taking into account the following metrics: Delay, Throughput, Jitter and Latency. In addition, routing tests were carried out on ten images in digital image and communication in medicine (DICOM) format of 40 MB. These tests were carried out for ten minutes with different traffic conditions, reaching a total of 25 tests, from a server of Universidad Militar Nueva Granada (UMNG) in Bogotá-Colombia to a remote user in Universidad de Santiago de Chile (USACH) - Chile. The results show that Tabu search presents a better QoS performance compared to Simulated Annealing, managing to optimize the routing of medical images, a basic requirement to offer diagnostic images services in telemedicine.

Keywords: Medical image, QoS, simulated annealing, Tabu search, telemedicine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 934
833 A Fitted Random Sampling Scheme for Load Distribution in Grid Networks

Authors: O. A. Rahmeh, P. Johnson, S. Lehmann

Abstract:

Grid networks provide the ability to perform higher throughput computing by taking advantage of many networked computer-s resources to solve large-scale computation problems. As the popularity of the Grid networks has increased, there is a need to efficiently distribute the load among the resources accessible on the network. In this paper, we present a stochastic network system that gives a distributed load-balancing scheme by generating almost regular networks. This network system is self-organized and depends only on local information for load distribution and resource discovery. The in-degree of each node is refers to its free resources, and job assignment and resource discovery processes required for load balancing is accomplished by using fitted random sampling. Simulation results show that the generated network system provides an effective, scalable, and reliable load-balancing scheme for the distributed resources accessible on Grid networks.

Keywords: Complex networks, grid networks, load-balancing, random sampling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1767
832 Adaptive Naïve Bayesian Anti-Spam Engine

Authors: Wojciech P. Gajewski

Abstract:

The problem of spam has been seriously troubling the Internet community during the last few years and currently reached an alarming scale. Observations made at CERN (European Organization for Nuclear Research located in Geneva, Switzerland) show that spam mails can constitute up to 75% of daily SMTP traffic. A naïve Bayesian classifier based on a Bag Of Words representation of an email is widely used to stop this unwanted flood as it combines good performance with simplicity of the training and classification processes. However, facing the constantly changing patterns of spam, it is necessary to assure online adaptability of the classifier. This work proposes combining such a classifier with another NBC (naïve Bayesian classifier) based on pairs of adjacent words. Only the latter will be retrained with examples of spam reported by users. Tests are performed on considerable sets of mails both from public spam archives and CERN mailboxes. They suggest that this architecture can increase spam recall without affecting the classifier precision as it happens when only the NBC based on single words is retrained.

Keywords: Text classification, naïve Bayesian classification, spam, email.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4402
831 Reducing Humic Acid and Disinfection By-products in Raw Water using a Bio-activated Carbon Filter

Authors: Wei-Pin Tseng, Jie-Chung Lou, Ming-Ching Wu, Huang-Ming Fang

Abstract:

For stricter drinking water regulations in the future, reducing the humic acid and disinfection byproducts in raw water, namely, trihalomethanes (THMs) and haloacetic acids (HAAs) is worthy for research. To investigate the removal of waterborne organic material using a lab-scale of bio-activated carbon filter under different EBCT, the concentrations of humic acid prepared were 0.01, 0.03, 0.06, 0.12, 0.17, 0.23, and 0.29 mg/L. Then we conducted experiments using a pilot plant with in-field of the serially connected bio-activated carbon filters and hollow fiber membrane processes employed in traditional water purification plants. Results showed under low TOC conditions of humic acid in influent (0.69 to 1.03 mg TOC/L) with an EBCT of 30 min, 40 min, and 50 min, TOC removal rates increases with greater EBCT, attaining about 39 % removal rate. The removal rate of THMs and HAAs by BACF was 54.8 % and 89.0 %, respectively.

Keywords: Bio-activated carbon filter, hollow fiber membrane, humic acid, THMs, HAAs, Water Treatment

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2147
830 Automation of Web-Portal Construction Processes with SQL Server for the Black Sea Ecosystem Monitoring

Authors: Gia Surguladze, Nino Topuria, Ana Gavardashvili, Tsatsa Namchevadze

Abstract:

The present article discusses design and development of Information System for monitoring ecology within the Black Sea basin of Georgia. Sea parameters, river, estuary, vulnerable district, water sample, etc. were considered as the major parameters of the sea ecosystem. A conceptual schema has been developed for the Black Sea ecosystem based on object-role model. The experimental database for the Black Sea ecosystem has been constructed using Ms SQL Server, while the object-role model NORMA has been developed using graphical instrument Ms Visual Studio within the integrated environment of .NET Framework 4.5. Web portal has been designed based on Ms SharePoint Server. The server database connection with web-portal has been carried out by means of External List of Ms SharePoint Server Designer.

Keywords: Web-application, service-oriented architecture, database, object-role modelling, SharePoint, Black sea, river, estuary, ecology, monitoring system, automation of data processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1285
829 The Effect of Clamping Restrain on the Prediction of Drape Simulation Software Tool

Authors: T.A. Adegbola, IEA Aghachi, E.R. Sadiku

Abstract:

To investigates the effect of fiberglass clamping process improvement on drape simulation prediction. This has great effect on the mould and the fiber during manufacturing process. This also, improves the fiber strain, the quality of the fiber orientation in the area of folding and wrinkles formation during the press-forming process. Drape simulation software tool was used to digitalize the process, noting the formation problems on the contour sensitive part. This was compared with the real life clamping processes using single and double frame set-ups to observe the effects. Also, restrains are introduced by using clips, and the G-clamps with predetermine revolution to; restrain the fabric deformation during the forming process.The incorporation of clamping and fabric restrain deformation improved on the prediction of the simulation tool. Therefore, for effective forming process, incorporation of clamping process into the drape simulation process will assist in the development of fiberglass application in manufacturing process.

Keywords: clamping, fiberglass, drape simulation, pressforming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1541
828 Engineered Cement Composite Materials Characterization for Tunneling Applications

Authors: S. Boughanem, D. A. Jesson, M. J. Mulheron, P.A. Smith C. Eddie, S. Psomas, M. Rimes

Abstract:

Cements, which are intrinsically brittle materials, can exhibit a degree of pseudo-ductility when reinforced with a sufficient volume fraction of a fibrous phase. This class of materials, called Engineered Cement Composites (ECC) has the potential to be used in future tunneling applications where a level of pseudo-ductility is required to avoid brittle failures. However uncertainties remain regarding mechanical performance. Previous work has focused on comparatively thin specimens; however for future civil engineering applications, it is imperative that the behavior in tension of thicker specimens is understood. In the present work, specimens containing cement powder and admixtures have been manufactured following two different processes and tested in tension. Multiple matrix cracking has been observed during tensile testing, leading to a “strain-hardening" behavior, confirming the possible suitability of ECC material when used as thick sections (greater than 50mm) in tunneling applications.

Keywords: Cement composite, polymeric fibers, pseudoductility, test-geometry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450
827 Processes Simulation Study of Coal to Methanol Based on Gasification Technology

Authors: Po-Chuang Chen, Hsiu-Mei Chiu, Yau-Pin Chyou, Chiou-Shia Yu

Abstract:

This study presents a simulation model for converting coal to methanol, based on gasification technology with the commercial chemical process simulator, Pro/II® V8.1.1. The methanol plant consists of air separation unit (ASU), gasification unit, gas clean-up unit, and methanol synthetic unit. The clean syngas is produced with the first three operating units, and the model has been verified with the reference data from United States Environment Protection Agency. The liquid phase methanol (LPMEOHTM) process is adopted in the methanol synthetic unit. Clean syngas goes through gas handing section to reach the reaction requirement, reactor loop/catalyst to generate methanol, and methanol distillation to get desired purity over 99.9 wt%. The ratio of the total energy combined with methanol and dimethyl ether to that of feed coal is 78.5% (gross efficiency). The net efficiency is 64.2% with the internal power consumption taken into account, based on the assumption that the efficiency of electricity generation is 40%.

Keywords: Gasification, Methanol, LPMEOH, System-levelsimulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5334
826 Variogram Fitting Based on the Wilcoxon Norm

Authors: Hazem Al-Mofleh, John Daniels, Joseph McKean

Abstract:

Within geostatistics research, effective estimation of the variogram points has been examined, particularly in developing robust alternatives. The parametric fit of these variogram points which eventually defines the kriging weights, however, has not received the same attention from a robust perspective. This paper proposes the use of the non-linear Wilcoxon norm over weighted non-linear least squares as a robust variogram fitting alternative. First, we introduce the concept of variogram estimation and fitting. Then, as an alternative to non-linear weighted least squares, we discuss the non-linear Wilcoxon estimator. Next, the robustness properties of the non-linear Wilcoxon are demonstrated using a contaminated spatial data set. Finally, under simulated conditions, increasing levels of contaminated spatial processes have their variograms points estimated and fit. In the fitting of these variogram points, both non-linear Weighted Least Squares and non-linear Wilcoxon fits are examined for efficiency. At all levels of contamination (including 0%), using a robust estimation and robust fitting procedure, the non-weighted Wilcoxon outperforms weighted Least Squares.

Keywords: Non-Linear Wilcoxon, robust estimation, Variogram estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 957
825 The Influence of Organisational Culture on the Implementation of Enterprise Resource Planning

Authors: Redha M. Elhuni

Abstract:

The critical key success factors, which have to be targeted with appropriate change management, are the user acceptance and support of a new Enterprise Resource Planning (ERP) system at the early implementation stages. This becomes even more important in Arab context where national and organisational culture with a different value and belief system, resulting in different management styles, might not complement with Western business culture embedded in the predefined standard business processes of existing ERP packages. This study explains and critically evaluates research into national and organizational culture and the influence of different national cultures on the implementation and reengineering process of ERP packages in an Arab context. Using a case study, realized through a quantitative survey testing five of Martinsons’s and Davison’s propositions in a Libyan sample company, confirmed the expected results from the literature review that culture has an impact on the implementation process and that employee empowerment is an unavoidable consequence of an ERP implementation.

Keywords: Enterprise resource planning, ERP systems, organisational culture, Arab context.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1588
824 Transient Three Dimensional FE Modeling for Thermal Analysis of Pulsed Current Gas Tungsten Arc Welding of Aluminum Alloy

Authors: N. Karunakaran, V. Balasubramanian

Abstract:

This paper presents the results of a study aimed at establishing the temperature distribution during the welding of aluminum alloy plates by Pulsed Current Gas Tungsten Arc Welding (PCGTAW) and Constant Current Gas Tungsten Arc Welding (CCGTAW) processes. Pulsing of the GTA welding current influences the dimensions and solidification rate of the fused zone, it also reduces the weld pool volume hence a narrower bead. In this investigation, the base material considered was aluminum alloy AA 6351 T6, which is finding use in aircraft, automobile and high-speed train components. A finite element analysis was carried out using ANSYS, and the results of the FEA were compared with the experimental results. It is evident from the study that the finite element analysis using ANSYS can be effectively used to model PCGTAW process for finding temperature distribution.

Keywords: Gas tungsten arc welding, pulsed current, finite element analysis, thermal analysis, aluminum alloy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2265
823 Displacement Fields in Footing-Sand Interactions under Cyclic Loading

Authors: S. Joseph Antony, Z. K. Jahanger

Abstract:

Soils are subjected to cyclic loading in situ in situations such as during earthquakes and in the compaction of pavements. Investigations on the local scale measurement of the displacements of the grain and failure patterns within the soil bed under the cyclic loading conditions are rather limited. In this paper, using the digital particle image velocimetry (DPIV), local scale displacement fields of a dense sand medium interacting with a rigid footing are measured under the plane-strain condition for two commonly used types of cyclic loading, and the quasi-static loading condition for the purposes of comparison. From the displacement measurements of the grains, the failure envelopes of the sand media are also presented. The results show that, the ultimate cyclic bearing capacity (qultcyc) occurred corresponding to a relatively higher settlement value when compared with that of under the quasi-static loading. For the sand media under the cyclic loading conditions considered here, the displacement fields in the soil media occurred more widely in the horizontal direction and less deeper along the vertical direction when compared with that of under the quasi-static loading. The 'dead zone' in the sand grains beneath the footing is identified for all types of the loading conditions studied here. These grain-scale characteristics have implications on the resulting bulk bearing capacity of the sand media in footing-sand interaction problems.

Keywords: Cyclic loading, DPIV, settlement, soil-structure interactions, strip footing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 859