Search results for: Boussinesq's approximation and Mean Nusselt number.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3935

Search results for: Boussinesq's approximation and Mean Nusselt number.

2795 Attribute Weighted Class Complexity: A New Metric for Measuring Cognitive Complexity of OO Systems

Authors: Dr. L. Arockiam, A. Aloysius

Abstract:

In general, class complexity is measured based on any one of these factors such as Line of Codes (LOC), Functional points (FP), Number of Methods (NOM), Number of Attributes (NOA) and so on. There are several new techniques, methods and metrics with the different factors that are to be developed by the researchers for calculating the complexity of the class in Object Oriented (OO) software. Earlier, Arockiam et.al has proposed a new complexity measure namely Extended Weighted Class Complexity (EWCC) which is an extension of Weighted Class Complexity which is proposed by Mishra et.al. EWCC is the sum of cognitive weights of attributes and methods of the class and that of the classes derived. In EWCC, a cognitive weight of each attribute is considered to be 1. The main problem in EWCC metric is that, every attribute holds the same value but in general, cognitive load in understanding the different types of attributes cannot be the same. So here, we are proposing a new metric namely Attribute Weighted Class Complexity (AWCC). In AWCC, the cognitive weights have to be assigned for the attributes which are derived from the effort needed to understand their data types. The proposed metric has been proved to be a better measure of complexity of class with attributes through the case studies and experiments

Keywords: Software Complexity, Attribute Weighted Class Complexity, Weighted Class Complexity, Data Type

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121
2794 Partially Knowing of Least Support Orthogonal Matching Pursuit (PKLS-OMP) for Recovering Signal

Authors: Israa Sh. Tawfic, Sema Koc Kayhan

Abstract:

Given a large sparse signal, great wishes are to reconstruct the signal precisely and accurately from lease number of measurements as possible as it could. Although this seems possible by theory, the difficulty is in built an algorithm to perform the accuracy and efficiency of reconstructing. This paper proposes a new proved method to reconstruct sparse signal depend on using new method called Least Support Matching Pursuit (LS-OMP) merge it with the theory of Partial Knowing Support (PSK) given new method called Partially Knowing of Least Support Orthogonal Matching Pursuit (PKLS-OMP). The new methods depend on the greedy algorithm to compute the support which depends on the number of iterations. So to make it faster, the PKLS-OMP adds the idea of partial knowing support of its algorithm. It shows the efficiency, simplicity, and accuracy to get back the original signal if the sampling matrix satisfies the Restricted Isometry Property (RIP). Simulation results also show that it outperforms many algorithms especially for compressible signals.

Keywords: Compressed sensing, Lest Support Orthogonal Matching Pursuit, Partial Knowing Support, Restricted isometry property, signal reconstruction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2232
2793 A Review on Comparative Analysis of Path Planning and Collision Avoidance Algorithms

Authors: Divya Agarwal, Pushpendra S. Bharti

Abstract:

Autonomous mobile robots (AMR) are expected as smart tools for operations in every automation industry. Path planning and obstacle avoidance is the backbone of AMR as robots have to reach their goal location avoiding obstacles while traversing through optimized path defined according to some criteria such as distance, time or energy. Path planning can be classified into global and local path planning where environmental information is known and unknown/partially known, respectively. A number of sensors are used for data collection. A number of algorithms such as artificial potential field (APF), rapidly exploring random trees (RRT), bidirectional RRT, Fuzzy approach, Purepursuit, A* algorithm, vector field histogram (VFH) and modified local path planning algorithm, etc. have been used in the last three decades for path planning and obstacle avoidance for AMR. This paper makes an attempt to review some of the path planning and obstacle avoidance algorithms used in the field of AMR. The review includes comparative analysis of simulation and mathematical computations of path planning and obstacle avoidance algorithms using MATLAB 2018a. From the review, it could be concluded that different algorithms may complete the same task (i.e. with a different set of instructions) in less or more time, space, effort, etc.

Keywords: Autonomous mobile robots, obstacle avoidance, path planning, and processing time.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1694
2792 Selection of an Optimum Configuration of Solar PV Array under Partial Shaded Condition Using Particle Swarm Optimization

Authors: R. Ramaprabha

Abstract:

This paper presents an extraction of maximum energy from Solar Photovoltaic Array (SPVA) under partial shaded conditions by optimum selection of array size using Particle Swarm Optimization (PSO) technique. In this paper a detailed study on the output reduction of different SPVA configurations under partial shaded conditions have been carried out. A generalized MATLAB M-code based software model has been used for any required array size, configuration, shading patterns and number of bypass diodes. Comparative study has been carried out on different configurations by testing several shading scenarios. While the number of shading patterns and the rate of change are very low for stationary SPVA but these may be quite large for SPVA mounted on a mobile platforms. This paper presents the suitability of PSO technique to select optimum configuration for mobile arrays by calculating the global peak (GP) of different configurations and to transfer maximum power to the load.

Keywords: Global peak, Mobile PV arrays, Partial shading, optimization, PSO.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4235
2791 Computing Continuous Skyline Queries without Discriminating between Static and Dynamic Attributes

Authors: Ibrahim Gomaa, Hoda M. O. Mokhtar

Abstract:

Although most of the existing skyline queries algorithms focused basically on querying static points through static databases; with the expanding number of sensors, wireless communications and mobile applications, the demand for continuous skyline queries has increased. Unlike traditional skyline queries which only consider static attributes, continuous skyline queries include dynamic attributes, as well as the static ones. However, as skyline queries computation is based on checking the domination of skyline points over all dimensions, considering both the static and dynamic attributes without separation is required. In this paper, we present an efficient algorithm for computing continuous skyline queries without discriminating between static and dynamic attributes. Our algorithm in brief proceeds as follows: First, it excludes the points which will not be in the initial skyline result; this pruning phase reduces the required number of comparisons. Second, the association between the spatial positions of data points is examined; this phase gives an idea of where changes in the result might occur and consequently enables us to efficiently update the skyline result (continuous update) rather than computing the skyline from scratch. Finally, experimental evaluation is provided which demonstrates the accuracy, performance and efficiency of our algorithm over other existing approaches.

Keywords: Continuous query processing, dynamic database, moving object, skyline queries.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1244
2790 Competitiveness and Value Creation of Tourism Sector: In the Case of 10 ASEAN Economies

Authors: Apirada Chinprateep

Abstract:

The ASEAN Economic Community (AEC) is the goal of regional economic integration by 2015. In the region, tourism is an activity that is important, especially as a source of foreign currency, a source of employment creation and a source of income bringing to the region. Given the complexity of the issues entailing the concept of sustainable tourism, this paper tries to assess tourism sustainability with the ASEAN, based on a number of quantitative indicators for all the ten economies, Thailand, Myanmar, Laos, Vietnam, Malaysia, Singapore, Indonesia, Philippines, Cambodia, and Brunei. The methodological framework will provide a number of benchmarks of tourism activities in these countries. They include identification of the dimensions; for example, economic, socio-ecologic, infrastructure and indicators, method of scaling, chart representation and evaluation on Asian countries. This specification shows that a similar level of tourism activity might introduce different implementation in the tourism activity and might have different consequences for the socioecological environment and sustainability. The heterogeneity of developing countries exposed briefly here would be useful to detect and prepare for coping with the main problems of each country in their tourism activities, as well as competitiveness and value creation of tourism for ASEAN economic community, and will compare with other parts of the world.

Keywords: AEC, ASEAN, sustainable, tourism, competitiveness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1762
2789 Rating and Generating Sudoku Puzzles Based On Constraint Satisfaction Problems

Authors: Bahare Fatemi, Seyed Mehran Kazemi, Nazanin Mehrasa

Abstract:

Sudoku is a logic-based combinatorial puzzle game which people in different ages enjoy playing it. The challenging and addictive nature of this game has made it a ubiquitous game. Most magazines, newspapers, puzzle books, etc. publish lots of Sudoku puzzles every day. These puzzles often come in different levels of difficulty so that all people, from beginner to expert, can play the game and enjoy it. Generating puzzles with different levels of difficulty is a major concern of Sudoku designers. There are several works in the literature which propose ways of generating puzzles having a desirable level of difficulty. In this paper, we propose a method based on constraint satisfaction problems to evaluate the difficulty of the Sudoku puzzles. Then we propose a hill climbing method to generate puzzles with different levels of difficulty. Whereas other methods are usually capable of generating puzzles with only few number of difficulty levels, our method can be used to generate puzzles with arbitrary number of different difficulty levels. We test our method by generating puzzles with different levels of difficulty and having a group of 15 people solve all the puzzles and recording the time they spend for each puzzle.

Keywords: Constraint satisfaction problem, generating Sudoku puzzles, hill climbing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3203
2788 Topology of Reverse Von-Kármán Vortex Street in the Wake of a Swimming Whale Shark

Authors: Arash Taheri

Abstract:

In this paper, effects of the ventral body planform of a swimming whale shark on the formation of ‘reverse von-Kármán vortex street’ behind the aquatic animal are studied using Fluid-Structure Interaction (FSI) approach. In this regard, incompressible Navier-Stokes equations around the whale shark’s body with a prescribed deflection dynamics are solved with the aid of Boundary Data Immersion Method (BDIM) and Implicit Large Eddy Simulation (ILES) turbulence treatment by WaterLily.jl solver; fully-written in Julia programming language. The whale shark flow simulations here are performed at high Reynolds number, i.e. 1.4 107 corresponding to the swimming of a 10 meter-whale shark at an average speed of 5 km/h. For comparison purposes, vortical flow generation behind a silky shark with a streamlined forehead eidonomy is also simulated at high Reynolds number, Re = 2 106, corresponding to the swimming of a 2 meter-silky shark at an average speed of 3.6 km/h. The results depict formation of distinct wake topologies behind the swimming sharks depending on the travelling wave oscillating amplitudes.

Keywords: Whale shark, vortex street, BDIM, FSI, functional eidonomy, bionics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1282
2787 Assessment of the Administration and Services of Public Access Computers in Academic Libraries in Kaduna State, Nigeria

Authors: Usman Ahmed Adam, Umar Ibrahim, Ezra S. Gbaje

Abstract:

This study is posed to explore the practice of Public Access Computers (PACs) in academic libraries in Kaduna State, Nigeria. The study aimed to determine the computers and other tools available, their services and challenges of the practices. Three questions were framed to identify number of public computers and tools available, their services and problems faced during the practice. The study used qualitative research design along with semi-constructed interview and observation as tools for data collection. Descriptive analysis was employed to analyze the data. The sample size of the study comprises 52 librarian and IT staff from the seven academic institutions in Kaduna State. The findings revealed that, PACs were provided for access to the Internet, digital resources, library catalogue and training services. The study further explored that, despite the limit number of the computers, users were not allowed to enjoy many services. The study recommends that libraries in Kaduna state should provide more public computers to be able to cover the population of their users; libraries should allow users to use the computers without limitations and restrictions.

Keywords: Academic libraries, computers in the library, digital libraries, public computers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 615
2786 The Trend of Injuries in Building Fire in Tehran from 2002 to 2012

Authors: Mohammadreza Ashouri, Majid Bayatian

Abstract:

Analysis of fire data is a way for the implementation of any plan to improve the level of safety in cities. Such an analysis is able to reveal signs of changes in a given period and can be used as a measure of safety. The information of about 66,341 fires (from 2002 to 2012) released by Tehran Safety Services and Fire-Fighting Organization and data on the population and the number of households provided by Tehran Municipality and the Statistical Yearbook of Iran were extracted. Using the data, the fire changes, the rate of injuries, and mortality rate were determined and analyzed. The rate of injuries and mortality rate of fires per one million population of Tehran were 59.58% and 86.12%, respectively. During the study period, the number of fires and fire stations increased by 104.38% and 102.63%, respectively. Most fires (9.21%) happened in the 4th District of Tehran. The results showed that the recorded fire data have not been systematically planned for fire prevention since one of the ways to reduce injuries caused by fires is to develop a systematic plan for necessary actions in emergency situations. To determine a reliable source for fire prevention, the stages, definitions of working processes and the cause and effect chains should be considered. Therefore, a comprehensive statistical system should be developed for reported and recorded fire data.

Keywords: Fire statistics, fire analysis, accident prevention, Tehran.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 769
2785 Enhancing the Connectedness in Ad–hoc Mesh Networks using the Terranet Technology

Authors: Obeidat I., Bsoul M., Khasawneh A., Kilani Y.

Abstract:

This paper simulates the ad-hoc mesh network in rural areas, where such networks receive great attention due to their cost, since installing the infrastructure for regular networks in these areas is not possible due to the high cost. The distance between the communicating nodes is the most obstacles that the ad-hoc mesh network will face. For example, in Terranet technology, two nodes can communicate if they are only one kilometer far from each other. However, if the distance between them is more than one kilometer, then each node in the ad-hoc mesh networks has to act as a router that forwards the data it receives to other nodes. In this paper, we try to find the critical number of nodes which makes the network fully connected in a particular area, and then propose a method to enhance the intermediate node to accept to be a router to forward the data from the sender to the receiver. Much work was done on technological changes on peer to peer networks, but the focus of this paper will be on another feature which is to find the minimum number of nodes needed for a particular area to be fully connected and then to enhance the users to switch on their phones and accept to work as a router for other nodes. Our method raises the successful calls to 81.5% out of 100% attempt calls.

Keywords: Adjacency matrix, Ad-hoc mesh network, Connectedness, Terranet technology

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1619
2784 Meta Model Based EA for Complex Optimization

Authors: Maumita Bhattacharya

Abstract:

Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, many real life optimization problems often require finding optimal solution to complex high dimensional, multimodal problems involving computationally very expensive fitness function evaluations. Use of evolutionary algorithms in such problem domains is thus practically prohibitive. An attractive alternative is to build meta models or use an approximation of the actual fitness functions to be evaluated. These meta models are order of magnitude cheaper to evaluate compared to the actual function evaluation. Many regression and interpolation tools are available to build such meta models. This paper briefly discusses the architectures and use of such meta-modeling tools in an evolutionary optimization context. We further present two evolutionary algorithm frameworks which involve use of meta models for fitness function evaluation. The first framework, namely the Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model [14] reduces computation time by controlled use of meta-models (in this case approximate model generated by Support Vector Machine regression) to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the metamodel are generated from a single uniform model. This does not take into account uncertain scenarios involving noisy fitness functions. The second model, DAFHEA-II, an enhanced version of the original DAFHEA framework, incorporates a multiple-model based learning approach for the support vector machine approximator to handle noisy functions [15]. Empirical results obtained by evaluating the frameworks using several benchmark functions demonstrate their efficiency

Keywords: Meta model, Evolutionary algorithm, Stochastictechnique, Fitness function, Optimization, Support vector machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2067
2783 Treatment of Inorganic Filler Surface by Silane-Coupling Agent: Investigation of Treatment Condition and Analysis of Bonding State of Reacted Agent

Authors: Hiroshi Hirano, Joji Kadota, Toshiyuki Yamashita, Yasuyuki Agari

Abstract:

It is well known that enhancing interfacial adhesion between inorganic filler and matrix resin in a composite lead to favorable properties such as excellent mechanical properties, high thermal resistance, prominent electric insulation, low expansion coefficient, and so on. But it should be avoided that much excess of coupling agent is reacted due to a negative impact of their final composite-s properties. There is no report to achieve classification of the bonding state excepting investigation of coating layer thickness. Therefore, the analysis of the bonding state of the coupling agent reacted with the filler surface such as BN particles with less functional group and silica particles having much functional group was performed by thermal gravimetric analysis and pyrolysis GC/MS. The reacted number of functional groups on the silane-coupling agent was classified as a result of the analysis. Thus, we succeeded in classifying the reacted number of the functional groups as a result of this study.

Keywords: Inorganic filler, boron nitride, surface treatment, coupling agent, analysis of bonding state

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5043
2782 Capturing an Unknown Moving Target in Unknown Territory using Vision and Coordination

Authors: Kiran Ijaz, Umar Manzoor, Arshad Ali Shahid

Abstract:

In this paper we present an extension to Vision Based LRTA* (VLRTA*) known as Vision Based Moving Target Search (VMTS) for capturing unknown moving target in unknown territory with randomly generated obstacles. Target position is unknown to the agents and they cannot predict its position using any probability method. Agents have omni directional vision but can see in one direction at some point in time. Agent-s vision will be blocked by the obstacles in the search space so agent can not see through the obstacles. Proposed algorithm is evaluated on large number of scenarios. Scenarios include grids of sizes from 10x10 to 100x100. Grids had obstacles randomly placed, occupying 0% to 50%, in increments of 10%, of the search space. Experiments used 2 to 9 agents for each randomly generated maze with same obstacle ratio. Observed results suggests that VMTS is effective in locate target time, solution quality and virtual target. In addition, VMTS becomes more efficient if the number of agents is increased with proportion to obstacle ratio.

Keywords: Vision, MTS, Unknown Target, Coordination, VMTS, Multi-Agent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1464
2781 Tsunami Inundation Modeling in a Boundary Fitted Curvilinear Grid Model Using the Method of Lines Technique

Authors: M. Ashaque Meah, M. Shah Noor, M Asif Arefin, Md. Fazlul Karim

Abstract:

A numerical technique in a boundary-fitted curvilinear grid model is developed to simulate the extent of inland inundation along the coastal belts of Peninsular Malaysia and Southern Thailand due to 2004 Indian ocean tsunami. Tsunami propagation and run-up are also studied in this paper. The vertically integrated shallow water equations are solved by using the method of lines (MOL). For this purpose the boundary-fitted grids are generated along the coastal and island boundaries and the other open boundaries of the model domain. A transformation is used to the governing equations so that the transformed physical domain is converted into a rectangular one. The MOL technique is applied to the transformed shallow water equations and the boundary conditions so that the equations are converted into ordinary differential equations initial value problem. Finally the 4th order Runge-Kutta method is used to solve these ordinary differential equations. The moving boundary technique is applied instead of fixed sea side wall or fixed coastal boundary to ensure the movement of the coastal boundary. The extent of intrusion of water and associated tsunami propagation are simulated for the 2004 Indian Ocean tsunami along the west coast of Peninsular Malaysia and southern Thailand. The simulated results are compared with the results obtained from a finite difference model and the data available in the USGS website. All simulations show better approximation than earlier research and also show excellent agreement with the observed data.

Keywords: Open boundary condition, moving boundary condition, boundary-fitted curvilinear grids, far field tsunami, Shallow Water Equations, tsunami source, Indonesian tsunami of 2004.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 865
2780 Transient Hydrodynamic and Thermal Behaviors of Fluid Flow in a Vertical Porous Microchannel under the Effect of Hyperbolic Heat Conduction Model

Authors: A. F. Khadrawi

Abstract:

The transient hydrodynamics and thermal behaviors of fluid flow in open-ended vertical parallel-plate porous microchannel are investigated semi-analytically under the effect of the hyperbolic heat conduction model. The model that combines both the continuum approach and the possibility of slip at the boundary is adopted in the study. The Effects of Knudsen number , Darcy number , and thermal relaxation time  on the microchannel hydrodynamics and thermal behaviors are investigated using the hyperbolic heat conduction models. It is found that as  increases the slip in the hydrodynamic and thermal boundary condition increases. This slip in the hydrodynamic boundary condition increases as  increases. Also, the slip in the thermal boundary condition increases as  decreases especially the early stage of time.

Keywords: free convection, hyperbolic heat conduction, macroscopic heat conduction models in microchannel, porous media, vertical microchannel, microchannel thermal, hydrodynamic behavior.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1927
2779 Information System Management Factors Related to Behavioral Trend of Online Accommodation Services

Authors: Supattra Kanchanopast

Abstract:

The purpose of this research was to study the customers’ behavioral trend for online accommodation system at Bangkonthi District, Samutsongkhram province. The research collected data from 400 online users. A questionnaire was utilized as the tool in collecting information. Descriptive statistics included frequency, percentage, mean and standard deviation. Independent- sample t- test, analysis of variance and Pearson Correlation were also used. The findings of this research revealed that the majority of the respondents were male, 25-32 years old, and graduated a bachelor degree. The respondents mostly worked in private sectors and had monthly income between 10,001-15,000 baht. The regular online users, visiting this system between 3-4 times/month, spending 1-2 hours/time, searched for online accommodation information. This result showed that the users had good and high attitude towards the system. According to the hypothesis testing, the number of online usage had positive related to the behavioral trends: accommodation purchasing intention and recommend the accommodation to others. Furthermore, both the number of online usage and overall attitude had a significant correlation to accommodation purchase intention and recommend the accommodation to others.

Keywords: Customer Behavior, Information System Management, Online Accommodation Services.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1721
2778 Minimal Spanning Tree based Fuzzy Clustering

Authors: Ágnes Vathy-Fogarassy, Balázs Feil, János Abonyi

Abstract:

Most of fuzzy clustering algorithms have some discrepancies, e.g. they are not able to detect clusters with convex shapes, the number of the clusters should be a priori known, they suffer from numerical problems, like sensitiveness to the initialization, etc. This paper studies the synergistic combination of the hierarchical and graph theoretic minimal spanning tree based clustering algorithm with the partitional Gath-Geva fuzzy clustering algorithm. The aim of this hybridization is to increase the robustness and consistency of the clustering results and to decrease the number of the heuristically defined parameters of these algorithms to decrease the influence of the user on the clustering results. For the analysis of the resulted fuzzy clusters a new fuzzy similarity measure based tool has been presented. The calculated similarities of the clusters can be used for the hierarchical clustering of the resulted fuzzy clusters, which information is useful for cluster merging and for the visualization of the clustering results. As the examples used for the illustration of the operation of the new algorithm will show, the proposed algorithm can detect clusters from data with arbitrary shape and does not suffer from the numerical problems of the classical Gath-Geva fuzzy clustering algorithm.

Keywords: Clustering, fuzzy clustering, minimal spanning tree, cluster validity, fuzzy similarity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2406
2777 Performance Analysis of HSDPA Systems using Low-Density Parity-Check (LDPC)Coding as Compared to Turbo Coding

Authors: K. Anitha Sheela, J. Tarun Kumar

Abstract:

HSDPA is a new feature which is introduced in Release-5 specifications of the 3GPP WCDMA/UTRA standard to realize higher speed data rate together with lower round-trip times. Moreover, the HSDPA concept offers outstanding improvement of packet throughput and also significantly reduces the packet call transfer delay as compared to Release -99 DSCH. Till now the HSDPA system uses turbo coding which is the best coding technique to achieve the Shannon limit. However, the main drawbacks of turbo coding are high decoding complexity and high latency which makes it unsuitable for some applications like satellite communications, since the transmission distance itself introduces latency due to limited speed of light. Hence in this paper it is proposed to use LDPC coding in place of Turbo coding for HSDPA system which decreases the latency and decoding complexity. But LDPC coding increases the Encoding complexity. Though the complexity of transmitter increases at NodeB, the End user is at an advantage in terms of receiver complexity and Bit- error rate. In this paper LDPC Encoder is implemented using “sparse parity check matrix" H to generate a codeword at Encoder and “Belief Propagation algorithm "for LDPC decoding .Simulation results shows that in LDPC coding the BER suddenly drops as the number of iterations increase with a small increase in Eb/No. Which is not possible in Turbo coding. Also same BER was achieved using less number of iterations and hence the latency and receiver complexity has decreased for LDPC coding. HSDPA increases the downlink data rate within a cell to a theoretical maximum of 14Mbps, with 2Mbps on the uplink. The changes that HSDPA enables includes better quality, more reliable and more robust data services. In other words, while realistic data rates are only a few Mbps, the actual quality and number of users achieved will improve significantly.

Keywords: AMC, HSDPA, LDPC, WCDMA, 3GPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048
2776 The Effect of Cyclic Speed on the Wear Properties of Molybdenum Disulfide Greases under Extreme Pressure Loading Using 4 Balls Wear Tests

Authors: Gabi Nehme

Abstract:

The relationship between different types of Molybdenum disulfide greases under extreme pressure loading and different speed situations have been studied using Design of Experiment (DOE) under 1200rpm steady state rotational speed and cyclic frequencies between 2400 and 1200rpm using a Plint machine software to set up the different rotational speed situations.  Research described here is aimed at providing good friction and wear performance while optimizing cyclic frequencies and MoS2 concentration due to the recent concern about grease behavior in extreme pressure applications. Extreme load of 785 Newton was used in conjunction with different cyclic frequencies (2400rpm -3.75min, 1200rpm -7.5min, 2400rpm -3.75min, 1200rpm -7.5min), to examine lithium based grease with and without MoS2 for equal number of revolutions, and a total run of 36000 revolutions; then compared to 1200rpm steady speed for the same total number of revolutions. 4 Ball wear tester was utilized to run large number of experiments randomly selected by the DOE software. The grease was combined with fine grade MoS2 or technical grade then heated to 750C and the wear scar width was collected at the end of each test. DOE model validation results verify that the data were very significant and can be applied to a wide range of extreme pressure applications. Based on simulation results and Scanning Electron images (SEM), it has been found that wear was largely dependent on the cyclic frequency condition. It is believed that technical grade MoS2 greases under faster cyclic speeds perform better and provides antiwear film that can resist extreme pressure loadings. Figures showed reduced wear scars width and improved frictional values.

 

Keywords: MoS2 grease, wear, friction, extreme load, cyclic frequencies, aircraft grade bearing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1898
2775 Calibration of 2D and 3D Optical Measuring Instruments in Industrial Environments at Submillimeter Range

Authors: A. Mínguez-Martínez, J. de Vicente

Abstract:

Modern manufacturing processes have led to the miniaturization of systems and, as a result, parts at the micro and nanoscale are produced. This trend seems to become increasingly important in the near future. Besides, as a requirement of Industry 4.0, the digitalization of the models of production and processes makes it very important to ensure that the dimensions of newly manufactured parts meet the specifications of the models. Therefore, it is possible to reduce the scrap and the cost of non-conformities, ensuring the stability of the production at the same time. To ensure the quality of manufactured parts, it becomes necessary to carry out traceable measurements at scales lower than one millimeter. Providing adequate traceability to the SI unit of length (the meter) to 2D and 3D measurements at this scale is a problem that does not have a unique solution in industrial environments. Researchers in the field of dimensional metrology all around the world are working on this issue. A solution for industrial environments, even if it is not complete, will enable working with some traceability. At this point, we believe that the study of the surfaces could provide us with a first approximation to a solution. In this paper, we propose a calibration procedure for the scales of optical measuring instruments, particularizing for a confocal microscope, using material standards easy to find and calibrate in metrology and quality laboratories in industrial environments. Confocal microscopes are measuring instruments capable of filtering the out-of-focus reflected light so that when it reaches the detector, it is possible to take pictures of the part of the surface that is focused. Varying and taking pictures at different Z levels of the focus, a specialized software interpolates between the different planes, and it could reconstruct the surface geometry into a 3D model. As it is easy to deduce, it is necessary to give traceability to each axis. As a complementary result, the roughness Ra parameter will be traced to the reference. Although the solution is designed for a confocal microscope, it may be used for the calibration of other optical measuring instruments, by applying minor changes.

Keywords: Industrial environment, confocal microscope, optical measuring instrument, traceability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 410
2774 Reversible Watermarking on Stereo Image Sequences

Authors: John N. Ellinas

Abstract:

In this paper, a new reversible watermarking method is presented that reduces the size of a stereoscopic image sequence while keeping its content visible. The proposed technique embeds the residuals of the right frames to the corresponding frames of the left sequence, halving the total capacity. The residual frames may result in after a disparity compensated procedure between the two video streams or by a joint motion and disparity compensation. The residuals are usually lossy compressed before embedding because of the limited embedding capacity of the left frames. The watermarked frames are visible at a high quality and at any instant the stereoscopic video may be recovered by an inverse process. In fact, the left frames may be exactly recovered whereas the right ones are slightly distorted as the residuals are not embedded intact. The employed embedding method reorders the left frame into an array of consecutive pixel pairs and embeds a number of bits according to their intensity difference. In this way, it hides a number of bits in intensity smooth areas and most of the data in textured areas where resulting distortions are less visible. The experimental evaluation demonstrates that the proposed scheme is quite effective.

Keywords: Stereoscopic video, Reversible watermarking, Disparity compensation, Joint compensation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1426
2773 Similitude for Thermal Scale-up of a Multiphase Thermolysis Reactor in the Cu-Cl Cycle of a Hydrogen Production

Authors: Mohammed W. Abdulrahman

Abstract:

The thermochemical copper-chlorine (Cu-Cl) cycle is considered as a sustainable and efficient technology for a hydrogen production, when linked with clean-energy systems such as nuclear reactors or solar thermal plants. In the Cu-Cl cycle, water is decomposed thermally into hydrogen and oxygen through a series of intermediate reactions. This paper investigates the thermal scale up analysis of the three phase oxygen production reactor in the Cu-Cl cycle, where the reaction is endothermic and the temperature is about 530 oC. The paper focuses on examining the size and number of oxygen reactors required to provide enough heat input for different rates of hydrogen production. The type of the multiphase reactor used in this paper is the continuous stirred tank reactor (CSTR) that is heated by a half pipe jacket. The thermal resistance of each section in the jacketed reactor system is studied to examine its effect on the heat balance of the reactor. It is found that the dominant contribution to the system thermal resistance is from the reactor wall. In the analysis, the Cu-Cl cycle is assumed to be driven by a nuclear reactor where two types of nuclear reactors are examined as the heat source to the oxygen reactor. These types are the CANDU Super Critical Water Reactor (CANDU-SCWR) and High Temperature Gas Reactor (HTGR). It is concluded that a better heat transfer rate has to be provided for CANDU-SCWR by 3-4 times than HTGR. The effect of the reactor aspect ratio is also examined in this paper and is found that increasing the aspect ratio decreases the number of reactors and the rate of decrease in the number of reactors decreases by increasing the aspect ratio. Finally, a comparison between the results of heat balance and existing results of mass balance is performed and is found that the size of the oxygen reactor is dominated by the heat balance rather than the material balance.

Keywords: Clean energy, Cu-Cl cycle, heat transfer, sustainable energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1657
2772 Exploration of Least Significant Bit Based Watermarking and Its Robustness against Salt and Pepper Noise

Authors: Kamaldeep Joshi, Rajkumar Yadav, Sachin Allwadhi

Abstract:

Image steganography is the best aspect of information hiding. In this, the information is hidden within an image and the image travels openly on the Internet. The Least Significant Bit (LSB) is one of the most popular methods of image steganography. In this method, the information bit is hidden at the LSB of the image pixel. In one bit LSB steganography method, the total numbers of the pixels and the total number of message bits are equal to each other. In this paper, the LSB method of image steganography is used for watermarking. The watermarking is an application of the steganography. The watermark contains 80*88 pixels and each pixel requirs 8 bits for its binary equivalent form so, the total number of bits required to hide the watermark are 80*88*8(56320). The experiment was performed on standard 256*256 and 512*512 size images. After the watermark insertion, histogram analysis was performed. A noise factor (salt and pepper) of 0.02 was added to the stego image in order to evaluate the robustness of the method. The watermark was successfully retrieved after insertion of noise. An experiment was performed in order to know the imperceptibility of stego and the retrieved watermark. It is clear that the LSB watermarking scheme is robust to the salt and pepper noise.

Keywords: LSB, watermarking, salt and pepper, PSNR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1053
2771 Extraction of Symbolic Rules from Artificial Neural Networks

Authors: S. M. Kamruzzaman, Md. Monirul Islam

Abstract:

Although backpropagation ANNs generally predict better than decision trees do for pattern classification problems, they are often regarded as black boxes, i.e., their predictions cannot be explained as those of decision trees. In many applications, it is desirable to extract knowledge from trained ANNs for the users to gain a better understanding of how the networks solve the problems. A new rule extraction algorithm, called rule extraction from artificial neural networks (REANN) is proposed and implemented to extract symbolic rules from ANNs. A standard three-layer feedforward ANN is the basis of the algorithm. A four-phase training algorithm is proposed for backpropagation learning. Explicitness of the extracted rules is supported by comparing them to the symbolic rules generated by other methods. Extracted rules are comparable with other methods in terms of number of rules, average number of conditions for a rule, and predictive accuracy. Extensive experimental studies on several benchmarks classification problems, such as breast cancer, iris, diabetes, and season classification problems, demonstrate the effectiveness of the proposed approach with good generalization ability.

Keywords: Backpropagation, clustering algorithm, constructivealgorithm, continuous activation function, pruning algorithm, ruleextraction algorithm, symbolic rules.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1616
2770 Meta-Learning for Hierarchical Classification and Applications in Bioinformatics

Authors: Fabio Fabris, Alex A. Freitas

Abstract:

Hierarchical classification is a special type of classification task where the class labels are organised into a hierarchy, with more generic class labels being ancestors of more specific ones. Meta-learning for classification-algorithm recommendation consists of recommending to the user a classification algorithm, from a pool of candidate algorithms, for a dataset, based on the past performance of the candidate algorithms in other datasets. Meta-learning is normally used in conventional, non-hierarchical classification. By contrast, this paper proposes a meta-learning approach for more challenging task of hierarchical classification, and evaluates it in a large number of bioinformatics datasets. Hierarchical classification is especially relevant for bioinformatics problems, as protein and gene functions tend to be organised into a hierarchy of class labels. This work proposes meta-learning approach for recommending the best hierarchical classification algorithm to a hierarchical classification dataset. This work’s contributions are: 1) proposing an algorithm for splitting hierarchical datasets into new datasets to increase the number of meta-instances, 2) proposing meta-features for hierarchical classification, and 3) interpreting decision-tree meta-models for hierarchical classification algorithm recommendation.

Keywords: Algorithm recommendation, meta-learning, bioinformatics, hierarchical classification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1372
2769 Artificial Neural Network with Steepest Descent Backpropagation Training Algorithm for Modeling Inverse Kinematics of Manipulator

Authors: Thiang, Handry Khoswanto, Rendy Pangaldus

Abstract:

Inverse kinematics analysis plays an important role in developing a robot manipulator. But it is not too easy to derive the inverse kinematic equation of a robot manipulator especially robot manipulator which has numerous degree of freedom. This paper describes an application of Artificial Neural Network for modeling the inverse kinematics equation of a robot manipulator. In this case, the robot has three degree of freedoms and the robot was implemented for drilling a printed circuit board. The artificial neural network architecture used for modeling is a multilayer perceptron networks with steepest descent backpropagation training algorithm. The designed artificial neural network has 2 inputs, 2 outputs and varies in number of hidden layer. Experiments were done in variation of number of hidden layer and learning rate. Experimental results show that the best architecture of artificial neural network used for modeling inverse kinematics of is multilayer perceptron with 1 hidden layer and 38 neurons per hidden layer. This network resulted a RMSE value of 0.01474.

Keywords: Artificial neural network, back propagation, inverse kinematics, manipulator, robot.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2288
2768 A Genetic Algorithm Approach Considering Zero Injection Bus Constraint Modeling for Optimal Phasor Measurement Unit Placement

Authors: G. Chandana Sushma, T. R. Jyothsna

Abstract:

This paper presents optimal Phasor Measurement Unit (PMU) Placement in network using a genetic algorithm approach as it is infeasible and require high installation cost to place PMUs at every bus in network. This paper proposes optimal PMU allocation considering observability and redundancy utilizing Genetic Algorithm (GA) approach. The nonlinear constraints of buses are modeled to give accurate results. Constraints associated with Zero Injection (ZI) buses and radial buses are modeled to optimize number of locations for PMU placement. GA is modeled with ZI bus constraints to minimize number of locations without losing complete observability. Redundancy of every bus in network is computed to show optimum redundancy of complete system network. The performance of method is measured by Bus Observability Index (BOI) and Complete System Observability Performance Index (CSOPI). MATLAB simulations are carried out on IEEE -14, -30 and -57 bus-systems and compared with other methods in literature survey to show the effectiveness of the proposed approach.

Keywords: Constraints, genetic algorithm, observability, phasor measurement units, redundancy, synchrophasors, zero injection bus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 784
2767 A Preliminary Study on Factors Determining the Success of High Conservation Value Area in Oil Palm Plantations

Authors: Yanto Santosa, Rozza Tri Kwatrina

Abstract:

High Conservation Value (HCV) is an area with conservation function within oil palm plantation. Despite the important role of HCV area in biodiversity conservation and various studies on HCV, there was a lack of research studying the factors determining its success. A preliminary study was conducted to identify the determinant factor of HCV that affected the diversity. Line transect method was used to calculate the species diversity of butterfly, birds, mammals, and herpetofauna species as well as their richness. Specifically for mammals, camera traps were also used. The research sites comprised of 12 HCV areas in 3 provinces of Indonesia (Central Kalimantan, Riau, and Palembang). The relationship between the HCV biophysical factor with the species number and species diversity for each wildlife class was identified using Chi-Square analysis with Cross tab (contingency table). Results of the study revealed that species diversity varied by research locations. Four factors determining the success of HCV area in relations to the number and diversity of wildlife species are land cover types for mammals, the width of area and distance to rivers for birds, and distance to settlements for butterflies.

Keywords: Ecological factors, high conservation value area, oil palm plantation, wildlife diversity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1509
2766 Performance of Partially Covered N Number of Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) Series Connected Water Heating System

Authors: Rohit Tripathi, Sumit Tiwari, G. N. Tiwari

Abstract:

In present study, an approach is adopted where photovoltaic thermal flat plate collector is integrated with compound parabolic concentrator. Analytical expression of temperature dependent electrical efficiency of N number of partially covered Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) water collector connected in series has been derived with the help of basic thermal energy balance equations. Analysis has been carried for winter weather condition at Delhi location, India. Energy and exergy performance of N - partially covered Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) Water collector system has been compared for two cases: (i) 25% area of water collector covered by PV module, (ii) 75% area of water collector covered by PV module. It is observed that case (i) has been best suited for thermal performance and case (ii) for electrical energy as well as overall exergy.

Keywords: Compound parabolic concentrator, Energy, Photovoltaic thermal, Temperature dependent electrical efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1620