Search results for: Large%20eddy%20simulation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2198

Search results for: Large%20eddy%20simulation

1808 Easy Shopping by Electronic Credit

Authors: M. Kargar, A. Isazadeh, F. Fartash, T. Saderi

Abstract:

In this paper we suggest a method for setting electronic credits for the customers. In this method banks and market-sites help each other to make doing large shopping through internet so easy. By developing this system, the people who have less money to buy most of the things they want, become able to buy all of them just through a credit. This credit is given by market-sites through a banking control on it. The method suggested can stop being imprisoned because of banking debts.

Keywords: E-Business, E-Credit, Market-site, Buy-site, Bank, E-Commerce.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1458
1807 Megalopolisation: An Effect of Large Scale Urbanisation in Post-Reform China

Authors: Siqing Chen

Abstract:

Megalopolis is a group of densely populated metropolitan areas that combine to form an urban complex. Since China introduced the economic reforms in late 1970s, the Chinese urban system has experienced unprecedented growth. The process of urbanisation prevailed in the 1980s, and the process of predominantly large city growth appeared to continue through 1990s and 2000s. In this study, the magnitude and pattern of urbanisation in China during 1990s were examined using remotely sensed imagery acquired by TM/ETM+ sensor onboard the Landsat satellites. The development of megalopolis areas in China was also studied based on the GIS analysis of the increases of urban and built-up area from 1990 to 2000. The analysis suggests that in the traditional agricultural zones in China, e.g., Huang-Huai-Hai Plains, Changjiang River Delta, Pearl River Delta and Sichuan Basin, the urban and built-up areas increased by 1.76 million hectares, of which 0.82 million hectares are expansion of urban areas, an increase of 24.78% compared with 1990 at the national scale. The Yellow River Delta, Changjiang River Delta and Pearl River Delta also saw an increase of urban and built-up area by 63.9%, 66.2% and 83.0% respectively. As a result, three major megalopolises were developed in China: the Guangzhou-Shenzhen-Hong Kong- Macau (Pearl River Delta: PRD) megalopolis area, the Shanghai- Nanjing-Hangzhou (Changjiang River Delta: CRD) megalopolis area and the Beijing-Tianjing-Tangshan-Qinhuangdao (Yellow River Delta-Bohai Sea Ring: YRD) megalopolis area. The relationship between the processed of megalopolisation and the inter-provincial population flow was also explored in the context of social-economic and transport infrastructure development in Post-reform China.

Keywords: Megalopolisation, Land use change, Spatial analysis, Post-reform China

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1516
1806 Shear Capacity of Rectangular Duct Panel Experiencing Internal Pressure

Authors: K. S. Sivakumaran, T. Thanga, B. Halabieh

Abstract:

The end panels of a large rectangular industrial duct, which experience significant internal pressures, also experience considerable transverse shear due to transfer of gravity loads to the supports. The current design practice of such thin plate panels for shear load is based on methods used for the design of plate girder webs. The structural arrangements, the loadings and the resulting behavior associated with the industrial duct end panels are, however, significantly different from those of the web of a plate girder. The large aspect ratio of the end panels gives rise to multiple bands of tension fields, whereas the plate girder web design is based on one tension field. In addition to shear, the industrial end panels are subjected to internal pressure which in turn produces significant membrane action. This paper reports a study which was undertaken to review the current industrial analysis and design methods and to propose a comprehensive method of designing industrial duct end panels for shear resistance. In this investigation, a nonlinear finite element model was developed to simulate the behavior of industrial duct end panel, along with the associated edge stiffeners, subjected to transverse shear and internal pressures. The model considered the geometric imperfections and constitutive relations for steels. Six scale independent dimensionless parameters that govern the behavior of such end panel were identified and were then used in a parametric study. It was concluded that the plate slenderness dominates the shear strength of stockier end panels, and whereas, both the plate slenderness and the aspect ratio influence the shear strength of slender end panels. Based on these studies, this paper proposes design aids for estimating the shear strength of rectangular duct end panels.

Keywords: Thin plate, transverse shear, tension field, finite element analysis, parametric study, design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930
1805 Surface and Bulk Magnetization Behavior of Isolated Ferromagnetic NiFe Nanowires

Authors: Musaab Salman Sultan

Abstract:

The surface and bulk magnetization behavior of template released isolated ferromagnetic Ni60Fe40 nanowires of relatively thick diameters (~200 nm), deposited from a dilute suspension onto pre-patterned insulating chips have been investigated experimentally, using a highly sensitive Magneto-Optical Ker Effect (MOKE) magnetometry and Magneto-Resistance (MR) measurements, respectively. The MR data were consistent with the theoretical predictions of the anisotropic magneto-resistance (AMR) effect. The MR measurements, in all the angles of investigations, showed large features and a series of nonmonotonic "continuous small features" in the resistance profiles. The extracted switching fields from these features and from MOKE loops were compared with each other and with the switching fields reported in the literature that adopted the same analytical techniques on the similar compositions and dimensions of nanowires. A large difference between MOKE and MR measurments was noticed. The disparate between MOKE and MR results is attributed to the variance in the micro-magnetic structure of the surface and the bulk of such ferromagnetic nanowires. This result was ascertained using micro-magnetic simulations on an individual: cylindrical and rectangular cross sections NiFe nanowires, with the same diameter/thickness of the experimental wires, using the Object Oriented Micro-magnetic Framework (OOMMF) package where the simulated loops showed different switching events, indicating that such wires have different magnetic states in the reversal process and the micro-magnetic spin structures during switching behavior was complicated. These results further supported the difference between surface and bulk magnetization behavior in these nanowires. This work suggests that a combination of MOKE and MR measurements is required to fully understand the magnetization behavior of such relatively thick isolated cylindrical ferromagnetic nanowires.

Keywords: MOKE magnetometry, MR measurements, OOMMF package, micro-magnetic simulations, ferromagnetic nanowires, surface magnetic properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 719
1804 The Genesis of the Anomalous Sernio Fan, Valtellina, Northern Italy

Authors: E. De Finis, P. Gattinoni, L. Scesi

Abstract:

Massive rock avalanches formed some of the largest landslide deposits on Earth and they represent one of the major geohazards in high-relief mountains. This paper interprets a very large sedimentary fan (the Sernio fan, Valtellina, Northern Italy), located 20 Km SW from Val Pola Rock avalanche (1987), as the deposit of a partial collapse of a Deep Seated Gravitational Slope Deformation (DSGSD), afterwards eroded and buried by debris flows. The proposed emplacement sequence has been reconstructed based on geomorphological, structural and mechanical evidences. The Sernio fan is actually considered anomalous with reference to the very high ratio between the fan area (≈ 4.5km2) and the basin area (≈ 3km2). The morphology of the fan area is characterised by steep slopes (dip ≈ 20%) and the fan apex is extended for 1.8 km inside the small catchment basin. This sedimentary fan was originated by a landslide that interested a part of a large deep-seated gravitational slope deformation, involving a wide area of about 55 km². The main controlling factor is tectonic and it is related to the proximity to regional fault systems and the consequent occurrence of fault weak rocks (GSI locally lower than 10 with compressive stress lower than 20MPa). Moreover, the fan deposit shows sedimentary evidences of recent debris flow events. The best current explanation of the Sernio fan involves an initial failure of some hundreds of Mm3. The run-out was quite limited because of the morphology of Valtellina’s valley floor, and the deposit filled the main valley forming a landslide dam, as confirmed by the lacustrine deposits detected upstream the fan. Nowadays the debris flow events represent the main hazard in the study area.

Keywords: Anomalous sedimentary fans, debris flow, deep seated gravitational slope deformation, Italy, rock avalanche.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1707
1803 Effects of Functional Protein on Osteoblasts in Rat

Authors: Jie Sun, Guoyou Yin, Xianqing Zhang, Qiusheng She, Zhaohui Xie, Lanying Chen, Anfang Zhao

Abstract:

To assess the effects of functional protein on osteoblast, Large quantity of high-purity osteoblasts had been cultivated successfully by adopting sequential enzyme digestion. The growth curve of osteoblasts was protracted by cell counting. Proliferation of osteoblasts was assessed by MTT colorimetry. The experimental results show the functional protein can enhance proliferation, the properties of adhesion and discuss the effect of osteopontin on osteoblast.

Keywords: functional protein, osteoblast, MTT

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1209
1802 Genetic Algorithms with Oracle for the Traveling Salesman Problem

Authors: Robin Gremlich, Andreas Hamfelt, Héctor de Pereda, Vladislav Valkovsky

Abstract:

By introducing the concept of Oracle we propose an approach for improving the performance of genetic algorithms for large-scale asymmetric Traveling Salesman Problems. The results have shown that the proposed approach allows overcoming some traditional problems for creating efficient genetic algorithms.

Keywords: Genetic algorithms, Traveling Salesman Problem, optimal decision distribution, oracle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685
1801 Text Mining Technique for Data Mining Application

Authors: M. Govindarajan

Abstract:

Text Mining is around applying knowledge discovery techniques to unstructured text is termed knowledge discovery in text (KDT), or Text data mining or Text Mining. In decision tree approach is most useful in classification problem. With this technique, tree is constructed to model the classification process. There are two basic steps in the technique: building the tree and applying the tree to the database. This paper describes a proposed C5.0 classifier that performs rulesets, cross validation and boosting for original C5.0 in order to reduce the optimization of error ratio. The feasibility and the benefits of the proposed approach are demonstrated by means of medial data set like hypothyroid. It is shown that, the performance of a classifier on the training cases from which it was constructed gives a poor estimate by sampling or using a separate test file, either way, the classifier is evaluated on cases that were not used to build and evaluate the classifier are both are large. If the cases in hypothyroid.data and hypothyroid.test were to be shuffled and divided into a new 2772 case training set and a 1000 case test set, C5.0 might construct a different classifier with a lower or higher error rate on the test cases. An important feature of see5 is its ability to classifiers called rulesets. The ruleset has an error rate 0.5 % on the test cases. The standard errors of the means provide an estimate of the variability of results. One way to get a more reliable estimate of predictive is by f-fold –cross- validation. The error rate of a classifier produced from all the cases is estimated as the ratio of the total number of errors on the hold-out cases to the total number of cases. The Boost option with x trials instructs See5 to construct up to x classifiers in this manner. Trials over numerous datasets, large and small, show that on average 10-classifier boosting reduces the error rate for test cases by about 25%.

Keywords: C5.0, Error Ratio, text mining, training data, test data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2446
1800 SURF Based Image Matching from Different Angle of Viewpoints using Rectification and Simplified Orientation Correction

Authors: K.M.Goh, M.M.Mokji, S.A.R. Abu-Bakar

Abstract:

Speeded-Up Robust Feature (SURF) is commonly used for feature matching in stereovision because of their robustness towards scale changes and rotational changes. However, SURF feature cannot cope with large viewpoint changes or skew distortion. This paper introduces a method which can help to improve the wide baseline-s matching performance in term of accuracy by rectifying the image using two vanishing points. Simplified orientation correction was used to remove the false matching..

Keywords: Affine, orientation, projective, SURF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2281
1799 Implementing an Intuitive Reasoner with a Large Weather Database

Authors: Yung-Chien Sun, O. Grant Clark

Abstract:

In this paper, the implementation of a rule-based intuitive reasoner is presented. The implementation included two parts: the rule induction module and the intuitive reasoner. A large weather database was acquired as the data source. Twelve weather variables from those data were chosen as the “target variables" whose values were predicted by the intuitive reasoner. A “complex" situation was simulated by making only subsets of the data available to the rule induction module. As a result, the rules induced were based on incomplete information with variable levels of certainty. The certainty level was modeled by a metric called "Strength of Belief", which was assigned to each rule or datum as ancillary information about the confidence in its accuracy. Two techniques were employed to induce rules from the data subsets: decision tree and multi-polynomial regression, respectively for the discrete and the continuous type of target variables. The intuitive reasoner was tested for its ability to use the induced rules to predict the classes of the discrete target variables and the values of the continuous target variables. The intuitive reasoner implemented two types of reasoning: fast and broad where, by analogy to human thought, the former corresponds to fast decision making and the latter to deeper contemplation. . For reference, a weather data analysis approach which had been applied on similar tasks was adopted to analyze the complete database and create predictive models for the same 12 target variables. The values predicted by the intuitive reasoner and the reference approach were compared with actual data. The intuitive reasoner reached near-100% accuracy for two continuous target variables. For the discrete target variables, the intuitive reasoner predicted at least 70% as accurately as the reference reasoner. Since the intuitive reasoner operated on rules derived from only about 10% of the total data, it demonstrated the potential advantages in dealing with sparse data sets as compared with conventional methods.

Keywords: Artificial intelligence, intuition, knowledge acquisition, limited certainty.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1351
1798 Waste Burial to the Pressure Deficit Areas in the Eastern Siberia

Authors: L. Abukova, O. Abramova, A. Goreva, Y. Yakovlev

Abstract:

Important executive decisions on oil and gas production stimulation in Eastern Siberia have been recently taken. There are unique and large fields of oil, gas, and gas-condensate in Eastern Siberia. The Talakan, Koyumbinskoye, Yurubcheno-Tahomskoye, Kovykta, Chayadinskoye fields are supposed to be developed first. It will result in an abrupt increase in environmental load on the nature of Eastern Siberia. In Eastern Siberia, the introduction of ecological imperatives in hydrocarbon production is still realistic. Underground water movement is the one of the most important factors of the ecosystems condition management. Oil and gas production is associated with the forced displacement of huge water masses, mixing waters of different composition, and origin that determines the extent of anthropogenic impact on water drive systems and their protective reaction. An extensive hydrogeological system of the depression type is identified in the pre-salt deposits here. Pressure relieve here is steady up to the basement. The decrease of the hydrodynamic potential towards the basement with such a gradient resulted in reformation of the fields in process of historical (geological) development of the Nepsko-Botuobinskaya anteclise. The depression hydrodynamic systems are characterized by extremely high isolation and can only exist under such closed conditions. A steady nature of water movement due to a strictly negative gradient of reservoir pressure makes it quite possible to use environmentally-harmful liquid substances instead of water. Disposal of the most hazardous wastes is the most expedient in the deposits of the crystalline basement in certain structures distant from oil and gas fields. The time period for storage of environmentally-harmful liquid substances may be calculated by means of the geological time scales ensuring their complete prevention from releasing into environment or air even during strong earthquakes. Disposal of wastes of chemical and nuclear industries is a matter of special consideration. The existing methods of storage and disposal of wastes are very expensive. The methods applied at the moment for storage of nuclear wastes at the depth of several meters, even in the most durable containers, constitute a potential danger. The enormous size of the depression system of the Nepsko-Botuobinskaya anteclise makes it possible to easily identify such objects at the depth below 1500 m where nuclear wastes will be stored indefinitely without any environmental impact. Thus, the water drive system of the Nepsko-Botuobinskaya anteclise is the ideal object for large-volume injection of environmentally harmful liquid substances even if there are large oil and gas accumulations in the subsurface. Specific geological and hydrodynamic conditions of the system allow the production of hydrocarbons from the subsurface simultaneously with the disposal of industrial wastes of oil and gas, mining, chemical, and nuclear industries without any environmental impact.

Keywords: Eastern Siberia, formation pressure, underground water, waste burial.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 969
1797 Analytical and Numerical Approaches in Coagulation of Particles

Authors: Bilal Barakeh

Abstract:

In this paper we discuss the effect of unbounded particle interaction operator on particle growth and we study how this can address the choice of appropriate time steps of the numerical simulation. We provide also rigorous mathematical proofs showing that large particles become dominating with increasing time while small particles contribute negligibly. Second, we discuss the efficiency of the algorithm by performing numerical simulations tests and by comparing the simulated solutions with some known analytic solutions to the Smoluchowski equation.

Keywords: Stochastic processes, coagulation of particles, numerical scheme.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1468
1796 Considering Aerosol Processes in Nuclear Transport Package Containment Safety Cases

Authors: Andrew Cummings, Rhianne Boag, Sarah Bryson, Gordon Turner

Abstract:

Packages designed for transport of radioactive material must satisfy rigorous safety regulations specified by the International Atomic Energy Agency (IAEA). Higher Activity Waste (HAW) transport packages have to maintain containment of their contents during normal and accident conditions of transport (NCT and ACT). To ensure containment criteria is satisfied these packages are required to be leak-tight in all transport conditions to meet allowable activity release rates. Package design safety reports are the safety cases that provide the claims, evidence and arguments to demonstrate that packages meet the regulations and once approved by the competent authority (in the UK this is the Office for Nuclear Regulation) a licence to transport radioactive material is issued for the package(s). The standard approach to demonstrating containment in the RWM transport safety case is set out in BS EN ISO 12807. In this document a method for measuring a leak rate from the package is explained by way of a small interspace test volume situated between two O-ring seals on the underside of the package lid. The interspace volume is pressurised and a pressure drop measured. A small interspace test volume makes the method more sensitive enabling the measurement of smaller leak rates. By ascertaining the activity of the contents, identifying a releasable fraction of material and by treating that fraction of material as a gas, allowable leak rates for NCT and ACT are calculated. The adherence to basic safety principles in ISO12807 is very pessimistic and current practice in the demonstration of transport safety, which is accepted by the UK regulator. It is UK government policy that management of HAW will be through geological disposal. It is proposed that the intermediate level waste be transported to the geological disposal facility (GDF) in large cuboid packages. This poses a challenge for containment demonstration because such packages will have long seals and therefore large interspace test volumes. There is also uncertainty on the releasable fraction of material within the package ullage space. This is because the waste may be in many different forms which makes it difficult to define the fraction of material released by the waste package. Additionally because of the large interspace test volume, measuring the calculated leak rates may not be achievable. For this reason a justification for a lower releasable fraction of material is sought. This paper considers the use of aerosol processes to reduce the releasable fraction for both NCT and ACT. It reviews the basic coagulation and removal processes and applies the dynamic aerosol balance equation. The proposed solution includes only the most well understood physical processes namely; Brownian coagulation and gravitational settling. Other processes have been eliminated either on the basis that they would serve to reduce the release to the environment further (pessimistically in keeping with the essence of nuclear transport safety cases) or that they are not credible in the conditions of transport considered.

Keywords: Aerosol processes, Brownian coagulation, gravitational settling, transport regulations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 600
1795 An Improved Design of Area Efficient Two Bit Comparator

Authors: Shashank Gautam, Pramod Sharma

Abstract:

In present era, development of digital circuits, signal processors and other integrated circuits, magnitude comparators are challenged by large area and more power consumption. Comparator is most basic circuit that performs comparison. This paper presents a technique to design a two bit comparator which consumes less area and power. DSCH and MICROWIND version 3 are used to design the schematic and design the layout of the schematic, observe the performance parameters at different nanometer technologies respectively.

Keywords: Chip design, consumed power, layout area, two bit comparator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1183
1794 Hopf Bifurcation Analysis for a Delayed Predator–prey System with Stage Structure

Authors: Kejun Zhuang

Abstract:

In this paper, a delayed predator–prey system with stage structure is investigated. Sufficient conditions for the system to have multiple periodic solutions are obtained when the delay is sufficiently large by applying Bendixson-s criterion. Further, some numerical examples are given.

Keywords: Predator-prey system, Stage structure, Hopf bifurcation, Periodic solutions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1304
1793 Exact Solution of Some Helical Flows of Newtonian Fluids

Authors: Imran Siddique

Abstract:

This paper deals with the helical flow of a Newtonian fluid in an infinite circular cylinder, due to both longitudinal and rotational shear stress. The velocity field and the resulting shear stress are determined by means of the Laplace and finite Hankel transforms and satisfy all imposed initial and boundary conditions. For large times, these solutions reduce to the well-known steady-state solutions.

Keywords: Newtonian fluids, Velocity field, Exact solutions, Shear stress, Cylindrical domains.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1421
1792 Characterization of the Near-Wake of an Ahmed Body Profile

Authors: Stéphanie Pellerin, Bérengére Podvin, Luc Pastur

Abstract:

In aerovehicles context, the flow around an Ahmed body profile is simulated using the velocity-vorticity formulation of the Navier-Stokes equations, associated to a penalization method for solids and Large Eddy Simulation for turbulence. The study focuses both on the ground influence on the flow and on the dissymetry of the wake, observed for a ground clearance greater than 10% of the body height H. Unsteady and mean flows are presented and analyzed. POD study completes the analysis and gives information on the most energetic structures of the flow.

Keywords: Ahmed body, bi-stability, LES, near wake.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1832
1791 A Comparison of Inflow Generation Methods for Large-Eddy Simulation

Authors: Francois T. Pronk, Steven J. Hulshoff

Abstract:

A study of various turbulent inflow generation methods was performed to compare their relative effectiveness for LES computations of turbulent boundary layers. This study confirmed the quality of the turbulent information produced by the family of recycling and rescaling methods which take information from within the computational domain. Furthermore, more general inflow methods also proved applicable to such simulations, with a precursor-like inflow and a random inflow augmented with forcing planes showing promising results.

Keywords: Boundary layer, Flat plate, Inflow modeling, LES

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
1790 Fast Algorithm of Shot Cut Detection

Authors: Lenka Krulikovská, Jaroslav Polec, Tomáš Hirner

Abstract:

In this paper we present a novel method, which reduces the computational complexity of abrupt cut detection. We have proposed fast algorithm, where the similarity of frames within defined step is evaluated instead of comparing successive frames. Based on the results of simulation on large video collection, the proposed fast algorithm is able to achieve 80% reduction of needed frames comparisons compared to actually used methods without the shot cut detection accuracy degradation.

Keywords: Abrupt cut, fast algorithm, shot cut detection, Pearson correlation coefficient.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1717
1789 Agents Network on a Grid: An Approach with the Set of Circulant Operators

Authors: Babiga Birregah, Prosper K. Doh, Kondo H. Adjallah

Abstract:

In this work we present some matrix operators named circulant operators and their action on square matrices. This study on square matrices provides new insights into the structure of the space of square matrices. Moreover it can be useful in various fields as in agents networking on Grid or large-scale distributed self-organizing grid systems.

Keywords: Pascal matrices, Binomial Recursion, Circulant Operators, Square Matrix Bipartition, Local Network, Parallel networks of agents.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1068
1788 Green Bridges and Their Migration Potential

Authors: Jaroslav Žák, Aleš Florian

Abstract:

Green bridges enable wildlife to pass through linear structures, especially freeways. The term migration potential is used to quantify their functionality. The proposed methodology for determining migration potential eliminates the mathematical, systematic and ecological inaccuracies of previous methodologies and provides a reliable tool for designers and environmentalists. The methodology is suited especially to medium-sized and large mammals, is mathematically correct, and its correspondence with reality was tested by monitoring existing green bridges. 

Keywords: Green bridges, migration potential, partial probabilities, wildlife migration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1715
1787 Quantifying Uncertainties in an Archetype-Based Building Stock Energy Model by Use of Individual Building Models

Authors: Morten Brøgger, Kim Wittchen

Abstract:

Focus on reducing energy consumption in existing buildings at large scale, e.g. in cities or countries, has been increasing in recent years. In order to reduce energy consumption in existing buildings, political incentive schemes are put in place and large scale investments are made by utility companies. Prioritising these investments requires a comprehensive overview of the energy consumption in the existing building stock, as well as potential energy-savings. However, a building stock comprises thousands of buildings with different characteristics making it difficult to model energy consumption accurately. Moreover, the complexity of the building stock makes it difficult to convey model results to policymakers and other stakeholders. In order to manage the complexity of the building stock, building archetypes are often employed in building stock energy models (BSEMs). Building archetypes are formed by segmenting the building stock according to specific characteristics. Segmenting the building stock according to building type and building age is common, among other things because this information is often easily available. This segmentation makes it easy to convey results to non-experts. However, using a single archetypical building to represent all buildings in a segment of the building stock is associated with loss of detail. Thermal characteristics are aggregated while other characteristics, which could affect the energy efficiency of a building, are disregarded. Thus, using a simplified representation of the building stock could come at the expense of the accuracy of the model. The present study evaluates the accuracy of a conventional archetype-based BSEM that segments the building stock according to building type- and age. The accuracy is evaluated in terms of the archetypes’ ability to accurately emulate the average energy demands of the corresponding buildings they were meant to represent. This is done for the buildings’ energy demands as a whole as well as for relevant sub-demands. Both are evaluated in relation to the type- and the age of the building. This should provide researchers, who use archetypes in BSEMs, with an indication of the expected accuracy of the conventional archetype model, as well as the accuracy lost in specific parts of the calculation, due to use of the archetype method.

Keywords: Building stock energy modelling, energy-savings, archetype.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 694
1786 Parameter Selections of Fuzzy C-Means Based on Robust Analysis

Authors: Kuo-Lung Wu

Abstract:

The weighting exponent m is called the fuzzifier that can have influence on the clustering performance of fuzzy c-means (FCM) and mÎ[1.5,2.5] is suggested by Pal and Bezdek [13]. In this paper, we will discuss the robust properties of FCM and show that the parameter m will have influence on the robustness of FCM. According to our analysis, we find that a large m value will make FCM more robust to noise and outliers. However, if m is larger than the theoretical upper bound proposed by Yu et al. [14], the sample mean will become the unique optimizer. Here, we suggest to implement the FCM algorithm with mÎ[1.5,4] under the restriction when m is smaller than the theoretical upper bound.

Keywords: Fuzzy c-means, robust, fuzzifier.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614
1785 Multidimensional Performance Tracking

Authors: C. Ardil

Abstract:

In this study, a model, together with a software tool that implements it, has been developed to determine the performance ratings of employees in an organization operating in the information technology sector using the indicators obtained from employees' online study data. Weighted Sum (WS) Method and Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) method based on multidimensional decision making approach were used in the study. WS and TOPSIS methods provide multidimensional decision making (MDDM) methods that allow all dimensions to be evaluated together considering specific weights, allowing employees to objectively evaluate the problem of online performance tracking. The application of WS and TOPSIS mathematical methods, which can combine alternatives with a large number of dimensions and reach simultaneous solution, has been implemented through an online performance tracking software. In the application of WS and TOPSIS methods, objective dimension weights were calculated by using entropy information (EI) and standard deviation (SD) methods from the data obtained by employees' online performance tracking method, decision matrix was formed by using performance scores for each employee, and a single performance score was calculated for each employee. Based on the calculated performance score, employees were given a performance evaluation decision. The results of Pareto set evidence and comparative mathematical analysis validate that employees' performance preference rankings in WS and TOPSIS methods are closely related. This suggests the compatibility, applicability, and validity of the proposed method to the MDDM problems in which a large number of alternative and dimension types are taken into account. With this study, an objective, realistic, feasible and understandable mathematical method, together with a software tool that implements it has been demonstrated. This is considered to be preferable because of the subjectivity, limitations and high cost of the methods traditionally used in the measurement and performance appraisal in the information technology sector.

Keywords: Weighted sum, entropy ınformation, standard deviation, online performance tracking, performance evaluation, performance management, multidimensional decision making.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1065
1784 Relation of Optimal Pilot Offsets in the Shifted Constellation-Based Method for the Detection of Pilot Contamination Attacks

Authors: Dimitriya A. Mihaylova, Zlatka V. Valkova-Jarvis, Georgi L. Iliev

Abstract:

One possible approach for maintaining the security of communication systems relies on Physical Layer Security mechanisms. However, in wireless time division duplex systems, where uplink and downlink channels are reciprocal, the channel estimate procedure is exposed to attacks known as pilot contamination, with the aim of having an enhanced data signal sent to the malicious user. The Shifted 2-N-PSK method involves two random legitimate pilots in the training phase, each of which belongs to a constellation, shifted from the original N-PSK symbols by certain degrees. In this paper, legitimate pilots’ offset values and their influence on the detection capabilities of the Shifted 2-N-PSK method are investigated. As the implementation of the technique depends on the relation between the shift angles rather than their specific values, the optimal interconnection between the two legitimate constellations is investigated. The results show that no regularity exists in the relation between the pilot contamination attacks (PCA) detection probability and the choice of offset values. Therefore, an adversary who aims to obtain the exact offset values can only employ a brute-force attack but the large number of possible combinations for the shifted constellations makes such a type of attack difficult to successfully mount. For this reason, the number of optimal shift value pairs is also studied for both 100% and 98% probabilities of detecting pilot contamination attacks. Although the Shifted 2-N-PSK method has been broadly studied in different signal-to-noise ratio scenarios, in multi-cell systems the interference from the signals in other cells should be also taken into account. Therefore, the inter-cell interference impact on the performance of the method is investigated by means of a large number of simulations. The results show that the detection probability of the Shifted 2-N-PSK decreases inversely to the signal-to-interference-plus-noise ratio.

Keywords: Channel estimation, inter-cell interference, pilot contamination attacks, wireless communications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 640
1783 How to Build and Evaluate a Solution Method: An Illustration for the Vehicle Routing Problem

Authors: Nicolas Zufferey

Abstract:

The vehicle routing problem (VRP) is a famous combinatorial optimization problem. Because of its well-known difficulty, metaheuristics are the most appropriate methods to tackle large and realistic instances. The goal of this paper is to highlight the key ideas for designing VRP metaheuristics according to the following criteria: efficiency, speed, robustness, and ability to take advantage of the problem structure. Such elements can obviously be used to build solution methods for other combinatorial optimization problems, at least in the deterministic field.

Keywords: Vehicle routing problem, Metaheuristics, Combinatorial optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2045
1782 Clustered Signatures for Modeling and Recognizing 3D Rigid Objects

Authors: H. B. Darbandi, M. R. Ito, J. Little

Abstract:

This paper describes a probabilistic method for three-dimensional object recognition using a shared pool of surface signatures. This technique uses flatness, orientation, and convexity signatures that encode the surface of a free-form object into three discriminative vectors, and then creates a shared pool of data by clustering the signatures using a distance function. This method applies the Bayes-s rule for recognition process, and it is extensible to a large collection of three-dimensional objects.

Keywords: Object recognition, modeling, classification, computer vision.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1250
1781 Global Existence of Periodic Solutions in a Delayed Tri–neuron Network

Authors: Kejun Zhuang, Zhaohui Wen

Abstract:

In this paper, a tri–neuron network model with time delay is investigated. By using the Bendixson-s criterion for high– dimensional ordinary differential equations and global Hopf bifurcation theory for functional differential equations, sufficient conditions for existence of periodic solutions when the time delay is sufficiently large are established.

Keywords: Delay, global Hopf bifurcation, neural network, periodicsolutions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1448
1780 Performance Evaluation of Bluetooth Links in the Presence of Specific Types of Interference

Authors: Radosveta Sokullu, Engin Karatepe

Abstract:

In the last couple of years Bluetooth has gained a large share in the market of home and personal appliances. It is now a well established technology a short range supplement to the wireless world of 802.11. The two main trends of research that have sprung from these developments are directed towards the coexistence and performance issues of Bluetooth and 802.11 as well as the co-existence in the very short range of multiple Bluetooth devices. Our work aims at thoroughly investigating different aspects of co-channel interference and effects of transmission power, distance and 802.11 interference on Bluetooth connections.

Keywords: Bluetooth, co-channel interference, 802.11, performance analysis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1724
1779 Levenberg-Marquardt Algorithm for Karachi Stock Exchange Share Rates Forecasting

Authors: Syed Muhammad Aqil Burney, Tahseen Ahmed Jilani, C. Ardil

Abstract:

Financial forecasting is an example of signal processing problems. A number of ways to train/learn the network are available. We have used Levenberg-Marquardt algorithm for error back-propagation for weight adjustment. Pre-processing of data has reduced much of the variation at large scale to small scale, reducing the variation of training data.

Keywords: Gradient descent method, jacobian matrix.Levenberg-Marquardt algorithm, quadratic error surfaces,

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2435