Search results for: Residue Number System
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 11151

Search results for: Residue Number System

4851 Synthesis of Dispersion-Compensating Triangular Lattice Index-Guiding Photonic Crystal Fibers Using the Directed Tabu Search Method

Authors: F. Karim

Abstract:

In this paper, triangular lattice index-guiding photonic crystal fibers (PCFs) are synthesized to compensate the chromatic dispersion of a single mode fiber (SMF-28) for an 80 km optical link operating at 1.55 µm, by using the directed tabu search algorithm. Hole-to-hole distance, circular air-hole diameter, solid-core diameter, ring number and PCF length parameters are optimized for this purpose. Three Synthesized PCFs with different physical parameters are compared in terms of their objective functions values, residual dispersions and compensation ratios.

Keywords: Triangular lattice index-guiding photonic crystal fiber, dispersion compensation, directed tabu search, synthesis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1291
4850 Innovation Development of Food Market of Kazakhstan

Authors: G.B. Nurlikhina, G.N. Azhimetova, R. M. Ashimova

Abstract:

Currently, one of the main directions is developing of development based on the clustering of economic operations of Kazakhstan, providing for the organization and concentration of production capacity in one region or the most optimal system. In the modern economic literature clustering is regarded as one of the most effective tools to ensure competitive businesses, and improve their business itself.

Keywords: A cluster, food market, innovation cluster.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2226
4849 Comparative Analysis of the Third Generation of Research Data for Evaluation of Solar Energy Potential

Authors: Claudineia Brazil, Elison Eduardo Jardim Bierhals, Luciane Teresa Salvi, Rafael Haag

Abstract:

Renewable energy sources are dependent on climatic variability, so for adequate energy planning, observations of the meteorological variables are required, preferably representing long-period series. Despite the scientific and technological advances that meteorological measurement systems have undergone in the last decades, there is still a considerable lack of meteorological observations that form series of long periods. The reanalysis is a system of assimilation of data prepared using general atmospheric circulation models, based on the combination of data collected at surface stations, ocean buoys, satellites and radiosondes, allowing the production of long period data, for a wide gamma. The third generation of reanalysis data emerged in 2010, among them is the Climate Forecast System Reanalysis (CFSR) developed by the National Centers for Environmental Prediction (NCEP), these data have a spatial resolution of 0.50 x 0.50. In order to overcome these difficulties, it aims to evaluate the performance of solar radiation estimation through alternative data bases, such as data from Reanalysis and from meteorological satellites that satisfactorily meet the absence of observations of solar radiation at global and/or regional level. The results of the analysis of the solar radiation data indicated that the reanalysis data of the CFSR model presented a good performance in relation to the observed data, with determination coefficient around 0.90. Therefore, it is concluded that these data have the potential to be used as an alternative source in locations with no seasons or long series of solar radiation, important for the evaluation of solar energy potential.

Keywords: Climate, reanalysis, renewable energy, solar radiation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 875
4848 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation

Authors: H. Khanfari, M. Johari Fard

Abstract:

Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.

Keywords: Carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1762
4847 FPGA Hardware Implementation and Evaluation of a Micro-Network Architecture for Multi-Core Systems

Authors: Yahia Salah, Med Lassaad Kaddachi, Rached Tourki

Abstract:

This paper presents the design, implementation and evaluation of a micro-network, or Network-on-Chip (NoC), based on a generic pipeline router architecture. The router is designed to efficiently support traffic generated by multimedia applications on embedded multi-core systems. It employs a simplest routing mechanism and implements the round-robin scheduling strategy to resolve output port contentions and minimize latency. A virtual channel flow control is applied to avoid the head-of-line blocking problem and enhance performance in the NoC. The hardware design of the router architecture has been implemented at the register transfer level; its functionality is evaluated in the case of the two dimensional Mesh/Torus topology, and performance results are derived from ModelSim simulator and Xilinx ISE 9.2i synthesis tool. An example of a multi-core image processing system utilizing the NoC structure has been implemented and validated to demonstrate the capability of the proposed micro-network architecture. To reduce complexity of the image compression and decompression architecture, the system use image processing algorithm based on classical discrete cosine transform with an efficient zonal processing approach. The experimental results have confirmed that both the proposed image compression scheme and NoC architecture can achieve a reasonable image quality with lower processing time.

Keywords: Generic Pipeline Network-on-Chip Router Architecture, JPEG Image Compression, FPGA Hardware Implementation, Performance Evaluation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3074
4846 In vivo Introduced Extracellular Ubiquitin Regulates Intracellular Processes

Authors: Rusudan Sujashvili, Ekaterine Bakuradze, Irina Modebadze, Davit Dekanoidze

Abstract:

Extracellular ubiquitin in vivo effect on regenerative liver cells and liver histoarchitectonics has been studied. Experiments were performed on mature female white rats. Partial hepatectomy was made using the modified method of Higgins and Anderson. Standard histopathological assessment of liver tissue was used. Proliferative activity of hepatocytes was analyzed by colchicine mitotic index and immunohistochemical staining on ki67. We have found that regardless of number of injections and dose of extracellular ubiquitin liver histology has not been changed, so at tissue level no effect was observed. In vivo double injection of ubiquitin significantly decreases the mitotic activity at 32 hour point after partial hepatectomy. Thus, we can conclude that in vivo injected extracellular ubiquitin inhibits proliferative activity of hepatocytes in partially hepatectomyzed rats.

Keywords: Liver, regeneration, proliferation, ubiquitin.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1531
4845 Defect Detection of Tiles Using 2D-Wavelet Transform and Statistical Features

Authors: M.Ghazvini, S. A. Monadjemi, N. Movahhedinia, K. Jamshidi

Abstract:

In this article, a method has been offered to classify normal and defective tiles using wavelet transform and artificial neural networks. The proposed algorithm calculates max and min medians as well as the standard deviation and average of detail images obtained from wavelet filters, then comes by feature vectors and attempts to classify the given tile using a Perceptron neural network with a single hidden layer. In this study along with the proposal of using median of optimum points as the basic feature and its comparison with the rest of the statistical features in the wavelet field, the relational advantages of Haar wavelet is investigated. This method has been experimented on a number of various tile designs and in average, it has been valid for over 90% of the cases. Amongst the other advantages, high speed and low calculating load are prominent.

Keywords: Defect detection, tile and ceramic quality inspection, wavelet transform, classification, neural networks, statistical features.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2344
4844 Laboratory Evaluation of Geogrids Used for Stabilizing Soft Subgrades

Authors: Magdi M. E. Zumrawi, Nehla Mansour

Abstract:

This paper aims to assess the efficiency of using geogrid reinforcement for subgrade stabilization. The literature of applying geogrid reinforcement technique for pavements built on soft subgrades and the previous experiences were reviewed. Laboratory tests were conducted on soil reinforced with geogrids in one or several layers. The soil specimens were compacted in four layers with or without geogrid sheets. The California Bearing Ratio (CBR) test, in soaking condition, was performed on natural soil and soil-geogrid specimens. The test results revealed that the CBR value is much affected by the geogrid sheet location and the number of sheets used in the soil specimen. When a geogrid sheet was placed at the 1st layer of the soil, there was an increment of 26% in the CBR value. Moreover, the CBR value was significantly increased by 62% when geogrid sheets were placed at all four layers. The high CBR value is attributed to interface friction and interlock involved in the geogrid/ soil interactions. It could be concluded that geogrid reinforcement is successful and more economical technique.

Keywords: Geogrid, reinforcement, stabilization, subgrade.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2775
4843 Decision-Making Strategies on Smart Dairy Farms: A Review

Authors: L. Krpalkova, N. O' Mahony, A. Carvalho, S. Campbell, G. Corkery, E. Broderick, J. Walsh

Abstract:

Farm management and operations will drastically change due to access to real-time data, real-time forecasting and tracking of physical items in combination with Internet of Things (IoT) developments to further automate farm operations. Dairy farms have embraced technological innovations and procured vast amounts of permanent data streams during the past decade; however, the integration of this information to improve the whole farm decision-making process does not exist. It is now imperative to develop a system that can collect, integrate, manage, and analyze on-farm and off-farm data in real-time for practical and relevant environmental and economic actions. The developed systems, based on machine learning and artificial intelligence, need to be connected for useful output, a better understanding of the whole farming issue and environmental impact. Evolutionary Computing (EC) can be very effective in finding the optimal combination of sets of some objects and finally, in strategy determination. The system of the future should be able to manage the dairy farm as well as an experienced dairy farm manager with a team of the best agricultural advisors. All these changes should bring resilience and sustainability to dairy farming as well as improving and maintaining good animal welfare and the quality of dairy products. This review aims to provide an insight into the state-of-the-art of big data applications and EC in relation to smart dairy farming and identify the most important research and development challenges to be addressed in the future. Smart dairy farming influences every area of management and its uptake has become a continuing trend.

Keywords: Big data, evolutionary computing, cloud, precision technologies

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 718
4842 New Methods for E-Commerce Databases Designing in Semantic Web Systems (Modern Systems)

Authors: Karim Heidari, Serajodin Katebi, Ali Reza Mahdavi Far

Abstract:

The purpose of this paper is to study Database Models to use them efficiently in E-commerce websites. In this paper we are going to find a method which can save and retrieve information in Ecommerce websites. Thus, semantic web applications can work with, and we are also going to study different technologies of E-commerce databases and we know that one of the most important deficits in semantic web is the shortage of semantic data, since most of the information is still stored in relational databases, we present an approach to map legacy data stored in relational databases into the Semantic Web using virtually any modern RDF query language, as long as it is closed within RDF. To achieve this goal we study XML structures for relational data bases of old websites and eventually we will come up one level over XML and look for a map from relational model (RDM) to RDF. Noting that a large number of semantic webs get advantage of relational model, opening the ways which can be converted to XML and RDF in modern systems (semantic web) is important.

Keywords: E-Commerce, Semantic Web, Database, XML, RDF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2289
4841 Mathematical Model for Progressive Phase Distribution of Ku-band Reflectarray Antennas

Authors: M. Y. Ismail, M. Inam, A. F. M. Zain, N. Misran

Abstract:

Progressive phase distribution is an important consideration in reflectarray antenna design which is required to form a planar wave in front of the reflectarray aperture. This paper presents a detailed mathematical model in order to determine the required reflection phase values from individual element of a reflectarray designed in Ku-band frequency range. The proposed technique of obtaining reflection phase can be applied for any geometrical design of elements and is independent of number of array elements. Moreover the model also deals with the solution of reflectarray antenna design with both centre and off-set feed configurations. The theoretical modeling has also been implemented for reflectarrays constructed on 0.508mm thickness of different dielectric substrates. The results show an increase in the slope of the phase curve from 4.61°/mm to 22.35°/mm by varying the material properties.

Keywords: Mathematical modeling, Progressive phase distribution, Reflectarray antenna, Reflection phase.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2050
4840 Atmosphere Water Vapour As Main Sweet Water Resource in the Arid Zones of Central Asia

Authors: S.I.Nikolaeva, Yu.V. Petrov, L.Ye.Skipnikova

Abstract:

It has been shown that the solution of water shortage problem in Central Asia closely connected with inclusion of atmosphere water vapour into the system of response and water resources management. Some methods of water extraction from atmosphere have been discussed.

Keywords: potable water, water resources, water problems, water scarcity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1532
4839 Optimal Risk Reduction in the Railway Industry by Using Dynamic Programming

Authors: Michael Todinov, Eberechi Weli

Abstract:

The paper suggests for the first time the use of dynamic programming techniques for optimal risk reduction in the railway industry. It is shown that by using the concept ‘amount of removed risk by a risk reduction option’, the problem related to optimal allocation of a fixed budget to achieve a maximum risk reduction in the railway industry can be reduced to an optimisation problem from dynamic programming. For n risk reduction options and size of the available risk reduction budget B (expressed as integer number), the worst-case running time of the proposed algorithm is O (n x (B+1)), which makes the proposed method a very efficient tool for solving the optimal risk reduction problem in the railway industry. 

Keywords: Optimisation, railway risk reduction, budget constraints, dynamic programming.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2150
4838 A New Method for Contour Approximation Using Basic Ramer Idea

Authors: Ali Abdrhman Ukasha

Abstract:

This paper presented two new efficient algorithms for contour approximation. The proposed algorithm is compared with Ramer (good quality), Triangle (faster) and Trapezoid (fastest) in this work; which are briefly described. Cartesian co-ordinates of an input contour are processed in such a manner that finally contours is presented by a set of selected vertices of the edge of the contour. In the paper the main idea of the analyzed procedures for contour compression is performed. For comparison, the mean square error and signal-to-noise ratio criterions are used. Computational time of analyzed methods is estimated depending on a number of numerical operations. Experimental results are obtained both in terms of image quality, compression ratios, and speed. The main advantages of the analyzed algorithm is small numbers of the arithmetic operations compared to the existing algorithms.

Keywords: Polygonal approximation, Ramer, Triangle and Trapezoid methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1788
4837 Experimental Investigation on the Effect of Bond Thickness on the Interface Behaviour of Fibre Reinforced Polymer Sheet Bonded to Timber

Authors: Abbas Vahedian, Rijun Shrestha, Keith Crews

Abstract:

The bond mechanism between timber and fibre reinforced polymer (FRP) is relatively complex and is influenced by a number of variables including bond thickness, bond width, bond length, material properties, and geometries. This study investigates the influence of bond thickness on the behaviour of interface, failure mode, and bond strength of externally bonded FRP-to-timber interface. In the present study, 106 single shear joint specimens have been investigated. Experiment results showed that higher layers of FRP increase the ultimate load carrying capacity of interface; conversely, such increase led to decrease the slip of interface. Moreover, samples with more layers of FRPs may fail in a brittle manner without noticeable warning that collapse is imminent.

Keywords: FRP, single shear test, bond thickness, bond strength.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 731
4836 Improving Patients Discharge Process in Hospitals by using Six Sigma Approach

Authors: Mahmoud A. El-Banna

Abstract:

The need to increase the efficiency of health care systems is becoming an obligation, and one of area of improvement is the discharge process. The objective of this work is to minimize the patients discharge time (for insured patients) to be less than 50 minutes by using six sigma approach, this improvement will also: lead to an increase in customer satisfaction, increase the number of admissions and turnover on the rooms, increase hospital profitability.Three different departments were considered in this study: Female, Male, and Paediatrics. Six Sigma approach coupled with simulation has been applied to reduce the patients discharge time for pediatrics, female, and male departments at hospital. Upon applying these recommendations at hospital: 60%, 80%, and 22% of insured female, male, and pediatrics patients respectively will have discharge time less than the upper specification time i.e. 50 min.

Keywords: Discharge Time, Healthcare, Hospitals, Patients, Process Improvement, Six Sigma, Simulation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4718
4835 Research on the Aeration Systems’ Efficiency of a Lab-Scale Wastewater Treatment Plant

Authors: Oliver Marunțălu, Elena Elisabeta Manea, Lăcrămioara Diana Robescu, Mihai Necșoiu, Gheorghe Lăzăroiu, Dana Andreya Bondrea

Abstract:

In order to obtain efficient pollutants removal in small-scale wastewater treatment plants, uniform water flow has to be achieved. The experimental setup, designed for treating high-load wastewater (leachate), consists of two aerobic biological reactors and a lamellar settler. Both biological tanks were aerated by using three different types of aeration systems - perforated pipes, membrane air diffusers and tube ceramic diffusers. The possibility of homogenizing the water mass with each of the air diffusion systems was evaluated comparatively. The oxygen concentration was determined by optical sensors with data logging. The experimental data was analyzed comparatively for all three different air dispersion systems aiming to identify the oxygen concentration variation during different operational conditions. The Oxygenation Capacity was calculated for each of the three systems and used as performance and selection parameter. The global mass transfer coefficients were also evaluated as important tools in designing the aeration system. Even though using the tubular porous diffusers leads to higher oxygen concentration compared to the perforated pipe system (which provides medium-sized bubbles in the aqueous solution), it doesn’t achieve the threshold limit of 80% oxygen saturation in less than 30 minutes. The study has shown that the optimal solution for the studied configuration was the radial air diffusers which ensure an oxygen saturation of 80% in 20 minutes. An increment of the values was identified when the air flow was increased.

Keywords: Flow, aeration, bioreactor, oxygen concentration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2434
4834 Integrated Design in Additive Manufacturing Based on Design for Manufacturing

Authors: E. Asadollahi-Yazdi, J. Gardan, P. Lafon

Abstract:

Nowadays, manufactures are encountered with production of different version of products due to quality, cost and time constraints. On the other hand, Additive Manufacturing (AM) as a production method based on CAD model disrupts the design and manufacturing cycle with new parameters. To consider these issues, the researchers utilized Design For Manufacturing (DFM) approach for AM but until now there is no integrated approach for design and manufacturing of product through the AM. So, this paper aims to provide a general methodology for managing the different production issues, as well as, support the interoperability with AM process and different Product Life Cycle Management tools. The problem is that the models of System Engineering which is used for managing complex systems cannot support the product evolution and its impact on the product life cycle. Therefore, it seems necessary to provide a general methodology for managing the product’s diversities which is created by using AM. This methodology must consider manufacture and assembly during product design as early as possible in the design stage. The latest approach of DFM, as a methodology to analyze the system comprehensively, integrates manufacturing constraints in the numerical model in upstream. So, DFM for AM is used to import the characteristics of AM into the design and manufacturing process of a hybrid product to manage the criteria coming from AM. Also, the research presents an integrated design method in order to take into account the knowledge of layers manufacturing technologies. For this purpose, the interface model based on the skin and skeleton concepts is provided, the usage and manufacturing skins are used to show the functional surface of the product. Also, the material flow and link between the skins are demonstrated by usage and manufacturing skeletons. Therefore, this integrated approach is a helpful methodology for designer and manufacturer in different decisions like material and process selection as well as, evaluation of product manufacturability.

Keywords: Additive manufacturing, 3D printing, design for manufacturing, integrated design, interoperability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2232
4833 Fuzzy Metric Approach for Fuzzy Time Series Forecasting based on Frequency Density Based Partitioning

Authors: Tahseen Ahmed Jilani, Syed Muhammad Aqil Burney, C. Ardil

Abstract:

In the last 15 years, a number of methods have been proposed for forecasting based on fuzzy time series. Most of the fuzzy time series methods are presented for forecasting of enrollments at the University of Alabama. However, the forecasting accuracy rates of the existing methods are not good enough. In this paper, we compared our proposed new method of fuzzy time series forecasting with existing methods. Our method is based on frequency density based partitioning of the historical enrollment data. The proposed method belongs to the kth order and time-variant methods. The proposed method can get the best forecasting accuracy rate for forecasting enrollments than the existing methods.

Keywords: Fuzzy logical groups, fuzzified enrollments, fuzzysets, fuzzy time series.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3188
4832 An Improved Method to Compute Sparse Graphs for Traveling Salesman Problem

Authors: Y. Wang

Abstract:

The Traveling salesman problem (TSP) is NP-hard in combinatorial optimization. The research shows the algorithms for TSP on the sparse graphs have the shorter computation time than those for TSP according to the complete graphs. We present an improved iterative algorithm to compute the sparse graphs for TSP by frequency graphs computed with frequency quadrilaterals. The iterative algorithm is enhanced by adjusting two parameters of the algorithm. The computation time of the algorithm is O(CNmaxn2) where C is the iterations, Nmax is the maximum number of frequency quadrilaterals containing each edge and n is the scale of TSP. The experimental results showed the computed sparse graphs generally have less than 5n edges for most of these Euclidean instances. Moreover, the maximum degree and minimum degree of the vertices in the sparse graphs do not have much difference. Thus, the computation time of the methods to resolve the TSP on these sparse graphs will be greatly reduced.

Keywords: Frequency quadrilateral, iterative algorithm, sparse graph, traveling salesman problem.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 988
4831 Multidimensional Data Mining by Means of Randomly Travelling Hyper-Ellipsoids

Authors: Pavel Y. Tabakov, Kevin Duffy

Abstract:

The present study presents a new approach to automatic data clustering and classification problems in large and complex databases and, at the same time, derives specific types of explicit rules describing each cluster. The method works well in both sparse and dense multidimensional data spaces. The members of the data space can be of the same nature or represent different classes. A number of N-dimensional ellipsoids are used for enclosing the data clouds. Due to the geometry of an ellipsoid and its free rotation in space the detection of clusters becomes very efficient. The method is based on genetic algorithms that are used for the optimization of location, orientation and geometric characteristics of the hyper-ellipsoids. The proposed approach can serve as a basis for the development of general knowledge systems for discovering hidden knowledge and unexpected patterns and rules in various large databases.

Keywords: Classification, clustering, data minig, genetic algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1753
4830 Factors Contributing to Building Construction Project’s Cost Overrun in Jordan

Authors: Ghaleb Y. Abbasi, Sufyan Al-Mrayat

Abstract:

This study examined the contribution of 36 factors to building construction project’s cost overrun in Jordan. A questionnaire was distributed to a random sample of 350 stakeholders comprised of owners, consultants, and contractors, of which 285 responded. SPSS analysis was conducted to identify the top five causes of cost overrun, which were a large number of variation orders, inadequate quantities provided in the contract, misunderstanding of the project plan, incomplete bid documents, and choosing the lowest price in the contract bidding. There was an agreement among the study participants in ranking the factors contributing to cost overrun, which indicated that these factors were very commonly encountered in most construction projects in Jordan. Thus, it is crucial to enhance the collaboration among the different project stakeholders to understand the project’s objectives and set a realistic plan that takes into consideration all the factors that might influence the project cost, which might eventually prevent cost overrun.

Keywords: Cost, overrun, building construction projects, Jordan.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 102
4829 A New Center of Motion in Cabling Robots

Authors: A. Abbasi Moshaii, F. Najafi

Abstract:

In this paper a new model for center of motion creating is proposed. This new method uses cables. So, it is very useful in robots because it is light and has easy assembling process. In the robots which need to be in touch with some things this method is so useful. It will be described in the following. The accuracy of the idea is proved by two experiments. This system could be used in the robots which need a fixed point in the contact with some things and make a circular motion.

Keywords: Center of Motion, Robotic cables, permanent touching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1643
4828 The Investigation and Analysis of Village Remains in Jinzhong Prefecture of Shanxi Province, China

Authors: Zhang Yu

Abstract:

Shanxi Province is a province with a long history in China. The historical characteristics of Jinzhong Prefecture in Shaanxi Province are very prominent. This research has done a lot of field research and analysis, and has analyzed a large number of documents. The formation and characteristics of villages in Jinzhong Prefecture are summarized. But the remains of many areas have not been systematically discovered and analyzed. This study found that the reasons for the formation of villages are natural, cultural, traffic and economic reasons. It mainly includes water, mountain, and developed business culture during the Ming and Qing Dynasties. By analyzing the evolution characteristics of each period, the characteristics and remains of the existing villages are explained in detail. These types of relics mainly include courtyards, fortresses, and Exchange shops. This study can provide systematic guidance on the protection of future village remains.

Keywords: Jinzhong Prefecture, village, features, remains.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 697
4827 Iterative Joint Power Control and Partial Crosstalk Cancellation in Upstream VDSL

Authors: H. Bagheri, H. Emami, M. R. Pakravan

Abstract:

Crosstalk is the major limiting issue in very high bit-rate digital subscriber line (VDSL) systems in terms of bit-rate or service coverage. At the central office side, joint signal processing accompanied by appropriate power allocation enables complex multiuser processors to provide near capacity rates. Unfortunately complexity grows with the square of the number of lines within a binder, so by taking into account that there are only a few dominant crosstalkers who contribute to main part of crosstalk power, the canceller structure can be simplified which resulted in a much lower run-time complexity. In this paper, a multiuser power control scheme, namely iterative waterfilling, is combined with previously proposed partial crosstalk cancellation approaches to demonstrate the best ever achieved performance which is verified by simulation results.

Keywords: iterative waterfilling, partial crosstalk cancellation, run-time complexity, VDSL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1380
4826 A Genetic Algorithm for Clustering on Image Data

Authors: Qin Ding, Jim Gasvoda

Abstract:

Clustering is the process of subdividing an input data set into a desired number of subgroups so that members of the same subgroup are similar and members of different subgroups have diverse properties. Many heuristic algorithms have been applied to the clustering problem, which is known to be NP Hard. Genetic algorithms have been used in a wide variety of fields to perform clustering, however, the technique normally has a long running time in terms of input set size. This paper proposes an efficient genetic algorithm for clustering on very large data sets, especially on image data sets. The genetic algorithm uses the most time efficient techniques along with preprocessing of the input data set. We test our algorithm on both artificial and real image data sets, both of which are of large size. The experimental results show that our algorithm outperforms the k-means algorithm in terms of running time as well as the quality of the clustering.

Keywords: Clustering, data mining, genetic algorithm, image data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2024
4825 Liquid Chromatography Microfluidics for Detection and Quantification of Urine Albumin Using Linear Regression Method

Authors: Patricia B. Cruz, Catrina Jean G. Valenzuela, Analyn N. Yumang

Abstract:

Nearly a hundred per million of the Filipino population is diagnosed with Chronic Kidney Disease (CKD). The early stage of CKD has no symptoms and can only be discovered once the patient undergoes urinalysis. Over the years, different methods were discovered and used for the quantification of the urinary albumin such as the immunochemical assays where most of these methods require large machinery that has a high cost in maintenance and resources, and a dipstick test which is yet to be proven and is still debated as a reliable method in detecting early stages of microalbuminuria. This research study involves the use of the liquid chromatography concept in microfluidic instruments with biosensor as a means of separation and detection respectively, and linear regression to quantify human urinary albumin. The researchers’ main objective was to create a miniature system that quantifies and detect patients’ urinary albumin while reducing the amount of volume used per five test samples. For this study, 30 urine samples of unknown albumin concentrations were tested using VITROS Analyzer and the microfluidic system for comparison. Based on the data shared by both methods, the actual vs. predicted regression were able to create a positive linear relationship with an R2 of 0.9995 and a linear equation of y = 1.09x + 0.07, indicating that the predicted values and actual values are approximately equal. Furthermore, the microfluidic instrument uses 75% less in total volume – sample and reagents combined, compared to the VITROS Analyzer per five test samples.

Keywords: Chronic kidney disease, microfluidics, linear regression, VITROS analyzer, urinary albumin.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 828
4824 Fast Calculation for Particle Interactions in SPH Simulations: Outlined Sub-domain Technique

Authors: Buntara Sthenly Gan, Naohiro Kawada

Abstract:

A simple and easy algorithm is presented for a fast calculation of kernel functions which required in fluid simulations using the Smoothed Particle Hydrodynamic (SPH) method. Present proposed algorithm improves the Linked-list algorithm and adopts the Pair-Wise Interaction technique, which are widely used for evaluating kernel functions in fluid simulations using the SPH method. The algorithm is easy to be implemented without any complexities in programming. Some benchmark examples are used to show the simulation time saved by using the proposed algorithm. Parametric studies on the number of divisions for sub-domains, smoothing length and total amount of particles are conducted to show the effectiveness of the present technique. A compact formulation is proposed for practical usage.

Keywords: Technique, fluid simulation, smoothing particle hydrodynamic (SPH), particle interaction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1611
4823 Effect of Magnetic Field on Mixed Convection Boundary Layer Flow over an Exponentially Shrinking Vertical Sheet with Suction

Authors: S. S. P. M. Isa, N. M. Arifin, R. Nazar, N. Bachok, F. M. Ali, I. Pop

Abstract:

A theoretical study has been presented to describe the boundary layer flow and heat transfer on an exponentially shrinking sheet with a variable wall temperature and suction, in the presence of magnetic field. The governing nonlinear partial differential equations are converted into ordinary differential equations by similarity transformation, which are then solved numerically using the shooting method. Results for the skin friction coefficient, local Nusselt number, velocity profiles as well as temperature profiles are presented through graphs and tables for several sets of values of the parameters. The effects of the governing parameters on the flow and heat transfer characteristics are thoroughly examined.

Keywords: Exponentially shrinking sheet, magnetic field, mixed convection, suction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2460
4822 Simulation Study of DFIG Wind Turbine under Grid Fault

Authors: N. Zerzouri, H. Labar, S. Kechida

Abstract:

During recent years wind turbine technology has undergone rapid developments. Growth in size and the optimization of wind turbines has enabled wind energy to become increasingly competitive with conventional energy sources. As a result today-s wind turbines participate actively in the power production of several countries around the world. These developments raise a number of challenges to be dealt with now and in the future. The penetration of wind energy in the grid raises questions about the compatibility of the wind turbine power production with the grid. In particular, the contribution to grid stability, power quality and behavior during fault situations plays therefore as important a role as the reliability. In the present work, we addressed two fault situations that have shown their influence on the generator and the behavior of the wind over the defects which are briefly discussed based on simulation results.

Keywords: Doubly fed induction generator (DFIG), Wind energy, grid fault

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2394