Search results for: Expensive function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2403

Search results for: Expensive function

1683 Assertion-Driven Test Repair Based on Priority Criteria

Authors: Ruilian Zhao, Shukai Zhang, Yan Wang, Weiwei Wang

Abstract:

Repairing broken test cases is an expensive and challenging task in evolving software systems. Although an automated repair technique with intent-preservation has been proposed, it does not take into account the association between test repairs and assertions, leading a large number of irrelevant candidates and decreasing the repair capability. This paper proposes a assertion-driven test repair approach. Furthermore, a intent-oriented priority criterion is raised to guide the repair candidate generation, making the repairs closer to the intent of the test. In more detail, repair targets are determined through post-dominance relations between assertions and the methods that directly cause compilation errors. Then, test repairs are generated from the target in a bottom-up way, guided by the the intent-oriented priority criteria. Finally, the generated repair candidates are prioritized to match the original test intent. The approach is implemented and evaluated on the benchmark of 4 open-source programs and 91 broken test cases. The result shows that the approach can fix 89% (81/91) broken test cases, which are more effective than the existing intent-preserved test repair approach, and our intent-oriented priority criteria work well.

Keywords: Test repair, test intent, software test, test case evolution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 168
1682 Optimal Selling Prices for Small Sized Poultry Farmers

Authors: Hidefumi Kawakatsu, Dong Li, Kosuke Kato

Abstract:

In Japan, meat-type chickens are mainly classified into three categories: (1) Broilers, (2) Branded chickens, and (3) Jidori (Free-range local traditional pedigree chickens). The Jidori chickens are certified by the Japanese Ministry of Agriculture, whilst, for the Branded chickens, there is no regulation with respect to their breed (genotype) or methods for rearing them. It is, therefore, relatively easy for poultry farmers to introduce Branded than Jidori chickens. The Branded chickens are normally fed a low-calorie diet with ingredients such as herbs, which lengthens their breeding period (compared with that of the Broilers) and increases their market value. In the field of inventory management, fast-growing animals such as broilers are categorised as ameliorating items. To the best of our knowledge, there are no previous studies that have explicitly considered smaller sized poultry farmers with limited breeding areas. This study develops an inventory model for a small sized poultry farmer that produces both the Broilers (Product 1) and the Branded chickens (Product 2) with different amelioration rates. The poultry farmer’s total profit per unit of time is formulated as a function of selling prices by using a price-dependent demand function. The existence of a unique optimal selling price for each product, which maximises the total profit, established. It has also been confirmed through numerical examples that, when the breeding area is fixed, the total profit could increase if the poultry farmer reduced the product quantity of Product 1 to introduce Product 2.

Keywords: Amelioration, deterioration, small sized poultry farmers, optimal price.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 811
1681 Digital Manufacturing: Evolution and a Process Oriented Approach to Align with Business Strategy

Authors: Abhimanyu Pati, Prabir K. Bandyopadhyay

Abstract:

The paper intends to highlight the significance of Digital Manufacturing (DM) strategy in support and achievement of business strategy and goals of any manufacturing organization. Towards this end, DM initiatives have been given a process perspective, while not undermining its technological significance, with a view to link its benefits directly with fulfilment of customer needs and expectations in a responsive and cost-effective manner. A digital process model has been proposed to categorize digitally enabled organizational processes with a view to create synergistic groups, which adopt and use digital tools having similar characteristics and functionalities. This will throw future opportunities for researchers and developers to create a unified technology environment for integration and orchestration of processes. Secondly, an effort has been made to apply “what” and “how” features of Quality Function Deployment (QFD) framework to establish the relationship between customers’ needs – both for external and internal customers, and the features of various digital processes, which support for the achievement of these customer expectations. The paper finally concludes that in the present highly competitive environment, business organizations cannot thrive to sustain unless they understand the significance of digital strategy and integrate it with their business strategy with a clearly defined implementation roadmap. A process-oriented approach to DM strategy will help business executives and leaders to appreciate its value propositions and its direct link to organization’s competitiveness.

Keywords: Digital manufacturing, digital process model, quality function deployment, business strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1323
1680 Evaluating the Effect of Domestic Price on Rice Production in an African Setting: A Typical Evidence of the Sierra Leone Case

Authors: Alhaji M. H. Conteh, Xiangbin Yan, Alfred V Gborie

Abstract:

Rice, which is the staple food in Sierra Leone, is consumed on a daily basis. It is the most imperative food crop extensively grown by farmers across all ecologies in the country. Though much attention is now given to rice grain production through the small holder commercialization programme (SHCP), however, no attention has been given in investigating the limitations faced by rice producers. This paper will contribute to attempts to overcome the development challenges caused by food insecurity. The objective of this paper is thus, to analysis the relationship between rice production and the domestic retail price of rice. The study employed a log linear model in which, the quantity of rice produced is the dependent variable, quantity of rice imported, price of imported rice and price of domestic rice as explanatory variables. Findings showed that, locally produced rice is even more expensive than the imported rice per ton, and almost all the inhabitants in the capital city which hosts about 65% of the entire population of the country favor imported rice, as it is free from stones with other impurities. On the other hand, to control price and simultaneously increase rice production, the government should purchase the rice from the farmers and then sell to private retailers.

Keywords: Domestic price of rice, Econometric model, Rice production, Sierra Leone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2483
1679 Use of Vegetation and Geo-Jute in Erosion Control of Slopes in a Sub-Tropical Climate

Authors: Mohammad Shariful Islam, Shamima Nasrin, Md. Shahidul Islam, Farzana Rahman Moury

Abstract:

Protection of slope and embankment from erosion has become an important issue in Bangladesh. The constructions of strong structures require large capital, integrated designing, high maintenance cost. Strong structure methods have negative impact on the environment and sometimes not function for the design period. Plantation of vetiver system along the slopes is an alternative solution. Vetiver not only serves the purpose of slope protection but also adds green environment reducing pollution. Vetiver is available in almost all the districts of Bangladesh. This paper presents the application of vetiver system with geo-jute, for slope protection and erosion control of embankments and slopes. In-situ shear tests have been conducted on vetiver rooted soil system to find the shear strength. The shear strength and effective soil cohesion of vetiver rooted soil matrix are respectively 2.0 times and 2.1 times higher than that of the bared soil. Similar trends have been found in direct shear tests conducted on laboratory reconstituted samples. Field trials have been conducted in road embankment and slope protection with vetiver at different sites. During the time of vetiver root growth the soil protection has been accomplished by geo-jute. As the geo-jute degrades with time, vetiver roots grow and take over the function of geo-jutes. Slope stability analyses showed that vegetation increase the factor of safety significantly.

Keywords: Erosion, geo-jute, green technology, vegetation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4168
1678 FEM Simulation of HE Blast-Fragmentation Warhead and the Calculation of Lethal Range

Authors: G. Tanapornraweekit, W. Kulsirikasem

Abstract:

This paper presents the simulation of fragmentation warhead using a hydrocode, Autodyn. The goal of this research is to determine the lethal range of such a warhead. This study investigates the lethal range of warheads with and without steel balls as preformed fragments. The results from the FE simulation, i.e. initial velocities and ejected spray angles of fragments, are further processed using an analytical approach so as to determine a fragment hit density and probability of kill of a modelled warhead. In order to simulate a plenty of preformed fragments inside a warhead, the model requires expensive computation resources. Therefore, this study attempts to model the problem in an alternative approach by considering an equivalent mass of preformed fragments to the mass of warhead casing. This approach yields approximately 7% and 20% difference of fragment velocities from the analytical results for one and two layers of preformed fragments, respectively. The lethal ranges of the simulated warheads are 42.6 m and 56.5 m for warheads with one and two layers of preformed fragments, respectively, compared to 13.85 m for a warhead without preformed fragment. These lethal ranges are based on the requirement of fragment hit density. The lethal ranges which are based on the probability of kill are 27.5 m, 61 m and 70 m for warheads with no preformed fragment, one and two layers of preformed fragments, respectively.

Keywords: Lethal Range, Natural Fragment, Preformed Fragment, Warhead.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4328
1677 Viscosity Reduction and Upgrading of Athabasca Oilsands Bitumen by Natural Zeolite Cracking

Authors: Abu S.M. Junaid, Wei Wang, Christopher Street, Moshfiqur Rahman, Matt Gersbach, Sarah Zhou, William McCaffrey, Steven M. Kuznicki

Abstract:

Oilsands bitumen is an extremely important source of energy for North America. However, due to the presence of large molecules such as asphaltenes, the density and viscosity of the bitumen recovered from these sands are much higher than those of conventional crude oil. As a result the extracted bitumen has to be diluted with expensive solvents, or thermochemically upgraded in large, capital-intensive conventional upgrading facilities prior to pipeline transport. This study demonstrates that globally abundant natural zeolites such as clinoptilolite from Saint Clouds, New Mexico and Ca-chabazite from Bowie, Arizona can be used as very effective reagents for cracking and visbreaking of oilsands bitumen. Natural zeolite cracked oilsands bitumen products are highly recoverable (up to ~ 83%) using light hydrocarbons such as pentane, which indicates substantial conversion of heavier fractions to lighter components. The resultant liquid products are much less viscous, and have lighter product distribution compared to those produced from pure thermal treatment. These natural minerals impart similar effect on industrially extracted Athabasca bitumen.

Keywords: Natural Zeolites, Oilsands Bitumen, Cracking, Viscosity Reduction, Upgrading.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2861
1676 Tool for Analysing the Sensitivity and Tolerance of Mechatronic Systems in Matlab GUI

Authors: Bohuslava Juhasova, Martin Juhas, Renata Masarova, Zuzana Sutova

Abstract:

The article deals with the tool in Matlab GUI form that is designed to analyse a mechatronic system sensitivity and tolerance. In the analysed mechatronic system, a torque is transferred from the drive to the load through a coupling containing flexible elements. Different methods of control system design are used. The classic form of the feedback control is proposed using Naslin method, modulus optimum criterion and inverse dynamics method. The cascade form of the control is proposed based on combination of modulus optimum criterion and symmetric optimum criterion. The sensitivity is analysed on the basis of absolute and relative sensitivity of system function to the change of chosen parameter value of the mechatronic system, as well as the control subsystem. The tolerance is analysed in the form of determining the range of allowed relative changes of selected system parameters in the field of system stability. The tool allows to analyse an influence of torsion stiffness, torsion damping, inertia moments of the motor and the load and controller(s) parameters. The sensitivity and tolerance are monitored in terms of the impact of parameter change on the response in the form of system step response and system frequency-response logarithmic characteristics. The Symbolic Math Toolbox for expression of the final shape of analysed system functions was used. The sensitivity and tolerance are graphically represented as 2D graph of sensitivity or tolerance of the system function and 3D/2D static/interactive graph of step/frequency response.

Keywords: Mechatronic systems, Matlab GUI, sensitivity, tolerance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2053
1675 Improved Feature Processing for Iris Biometric Authentication System

Authors: Somnath Dey, Debasis Samanta

Abstract:

Iris-based biometric authentication is gaining importance in recent times. Iris biometric processing however, is a complex process and computationally very expensive. In the overall processing of iris biometric in an iris-based biometric authentication system, feature processing is an important task. In feature processing, we extract iris features, which are ultimately used in matching. Since there is a large number of iris features and computational time increases as the number of features increases, it is therefore a challenge to develop an iris processing system with as few as possible number of features and at the same time without compromising the correctness. In this paper, we address this issue and present an approach to feature extraction and feature matching process. We apply Daubechies D4 wavelet with 4 levels to extract features from iris images. These features are encoded with 2 bits by quantizing into 4 quantization levels. With our proposed approach it is possible to represent an iris template with only 304 bits, whereas existing approaches require as many as 1024 bits. In addition, we assign different weights to different iris region to compare two iris templates which significantly increases the accuracy. Further, we match the iris template based on a weighted similarity measure. Experimental results on several iris databases substantiate the efficacy of our approach.

Keywords: Iris recognition, biometric, feature processing, patternrecognition, pattern matching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2143
1674 Manual to Automated Testing: An Effort-Based Approach for Determining the Priority of Software Test Automation

Authors: Peter Sabev, Katalina Grigorova

Abstract:

Test automation allows performing difficult and time consuming manual software testing tasks efficiently, quickly and repeatedly. However, development and maintenance of automated tests is expensive, so it needs a proper prioritization what to automate first. This paper describes a simple yet efficient approach for such prioritization of test cases based on the effort needed for both manual execution and software test automation. The suggested approach is very flexible because it allows working with a variety of assessment methods, and adding or removing new candidates at any time. The theoretical ideas presented in this article have been successfully applied in real world situations in several software companies by the authors and their colleagues including testing of real estate websites, cryptographic and authentication solutions, OSGi-based middleware framework that has been applied in various systems for smart homes, connected cars, production plants, sensors, home appliances, car head units and engine control units (ECU), vending machines, medical devices, industry equipment and other devices that either contain or are connected to an embedded service gateway.

Keywords: Automated Testing, Manual Testing, Test Automation, Software testing, Test Prioritization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3392
1673 Optimal Duty-Cycle Modulation Scheme for Analog-To-Digital Conversion Systems

Authors: G. Sonfack, J. Mbihi, B. Lonla Moffo

Abstract:

This paper presents an optimal duty-cycle modulation (ODCM) scheme for analog-to-digital conversion (ADC) systems. The overall ODCM-Based ADC problem is decoupled into optimal DCM and digital filtering sub-problems, while taking into account constraints of mutual design parameters between the two. Using a set of three lemmas and four morphological theorems, the ODCM sub-problem is modelled as a nonlinear cost function with nonlinear constraints. Then, a weighted least pth norm of the error between ideal and predicted frequency responses is used as a cost function for the digital filtering sub-problem. In addition, MATLAB fmincon and MATLAB iirlnorm tools are used as optimal DCM and least pth norm solvers respectively. Furthermore, the virtual simulation scheme of an overall prototyping ODCM-based ADC system is implemented and well tested with the help of Simulink tool according to relevant set of design data, i.e., 3 KHz of modulating bandwidth, 172 KHz of maximum modulation frequency and 25 MHZ of sampling frequency. Finally, the results obtained and presented show that the ODCM-based ADC achieves under 3 KHz of modulating bandwidth: 57 dBc of SINAD (signal-to-noise and distorsion), 58 dB of SFDR (Surpious free dynamic range) -80 dBc of THD (total harmonic distorsion), and 10 bits of minimum resolution. These performance levels appear to be a great challenge within the class of oversampling ADC topologies, with 2nd order IIR (infinite impulse response) decimation filter.

Keywords: Digital IIR filter, morphological lemmas and theorems, optimal DCM-based DAC, virtual simulation, weighted least pth norm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 937
1672 A Study of the Planning and Designing of the Built Environment under the Green Transit-Oriented Development

Authors: Wann-Ming Wey

Abstract:

In recent years, the problems of global climate change and natural disasters have induced the concerns and attentions of environmental sustainability issues for the public. Aside from the environmental planning efforts done for human environment, Transit-Oriented Development (TOD) has been widely used as one of the future solutions for the sustainable city development. In order to be more consistent with the urban sustainable development, the development of the built environment planning based on the concept of Green TOD which combines both TOD and Green Urbanism is adapted here. The connotation of the urban development under the green TOD including the design toward environment protect, the maximum enhancement resources and the efficiency of energy use, use technology to construct green buildings and protected areas, natural ecosystems and communities linked, etc. Green TOD is not only to provide the solution to urban traffic problems, but to direct more sustainable and greener consideration for future urban development planning and design. In this study, we use both the TOD and Green Urbanism concepts to proceed to the study of the built environment planning and design. Fuzzy Delphi Technique (FDT) is utilized to screen suitable criteria of the green TOD. Furthermore, Fuzzy Analytic Network Process (FANP) and Quality Function Deployment (QFD) were then developed to evaluate the criteria and prioritize the alternatives. The study results can be regarded as the future guidelines of the built environment planning and designing under green TOD development in Taiwan.

Keywords: Green transit-oriented development, built environment, fuzzy Delphi technique, quality function deployment, fuzzy analytic network process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1524
1671 Near Shore Wave Manipulation for Electricity Generation

Authors: K. D. R. Jagath-Kumara, D. D. Dias

Abstract:

The sea waves carry thousands of GWs of power globally. Although there are a number of different approaches to harness offshore energy, they are likely to be expensive, practically challenging, and vulnerable to storms. Therefore, this paper considers using the near shore waves for generating mechanical and electrical power. It introduces two new approaches, the wave manipulation and using a variable duct turbine, for intercepting very wide wave fronts and coping with the fluctuations of the wave height and the sea level, respectively. The first approach effectively allows capturing much more energy yet with a much narrower turbine rotor. The second approach allows using a rotor with a smaller radius but captures energy of higher wave fronts at higher sea levels yet preventing it from totally submerging. To illustrate the effectiveness of the first approach, the paper contains a description and the simulation results of a scale model of a wave manipulator. Then, it includes the results of testing a physical model of the manipulator and a single duct, axial flow turbine in a wave flume in the laboratory. The paper also includes comparisons of theoretical predictions, simulation results, and wave flume tests with respect to the incident energy, loss in wave manipulation, minimal loss, brake torque, and the angular velocity.

Keywords: Near-shore sea waves, Renewable energy, Wave energy conversion, Wave manipulation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1995
1670 Systems Engineering Management Using Transdisciplinary Quality System Development Lifecycle Model

Authors: Mohamed Asaad Abdelrazek, Amir Taher El-Sheikh, M. Zayan, A.M. Elhady

Abstract:

The successful realization of complex systems is dependent not only on the technology issues and the process for implementing them, but on the management issues as well. Managing the systems development lifecycle requires technical management. Systems engineering management is the technical management. Systems engineering management is accomplished by incorporating many activities. The three major activities are development phasing, systems engineering process and lifecycle integration. Systems engineering management activities are performed across the system development lifecycle. Due to the ever-increasing complexity of systems as well the difficulty of managing and tracking the development activities, new ways to achieve systems engineering management activities are required. This paper presents a systematic approach used as a design management tool applied across systems engineering management roles. In this approach, Transdisciplinary System Development Lifecycle (TSDL) Model has been modified and integrated with Quality Function Deployment. Hereinafter, the name of the systematic approach is the Transdisciplinary Quality System Development Lifecycle (TQSDL) Model. The QFD translates the voice of customers (VOC) into measurable technical characteristics. The modified TSDL model is based on Axiomatic Design developed by Suh which is applicable to all designs: products, processes, systems and organizations. The TQSDL model aims to provide a robust structure and systematic thinking to support the implementation of systems engineering management roles. This approach ensures that the customer requirements are fulfilled as well as satisfies all the systems engineering manager roles and activities.

Keywords: Axiomatic design, quality function deployment, systems engineering management, system development lifecycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1764
1669 Simulation of Utility Accrual Scheduling and Recovery Algorithm in Multiprocessor Environment

Authors: A. Idawaty, O. Mohamed, A. Z. Zuriati

Abstract:

This paper presents the development of an event based Discrete Event Simulation (DES) for a recovery algorithm known Backward Recovery Global Preemptive Utility Accrual Scheduling (BR_GPUAS). This algorithm implements the Backward Recovery (BR) mechanism as a fault recovery solution under the existing Time/Utility Function/ Utility Accrual (TUF/UA) scheduling domain for multiprocessor environment. The BR mechanism attempts to take the faulty tasks back to its initial safe state and then proceeds to re-execute the affected section of the faulty tasks to enable recovery. Considering that faults may occur in the components of any system; a fault tolerance system that can nullify the erroneous effect is necessary to be developed. Current TUF/UA scheduling algorithm uses the abortion recovery mechanism and it simply aborts the erroneous task as their fault recovery solution. None of the existing algorithm in TUF/UA scheduling domain in multiprocessor scheduling environment have considered the transient fault and implement the BR mechanism as a fault recovery mechanism to nullify the erroneous effect and solve the recovery problem in this domain. The developed BR_GPUAS simulator has derived the set of parameter, events and performance metrics according to a detailed analysis of the base model. Simulation results revealed that BR_GPUAS algorithm can saved almost 20-30% of the accumulated utilities making it reliable and efficient for the real-time application in the multiprocessor scheduling environment.

Keywords: Time Utility Function/ Utility Accrual (TUF/UA) scheduling, Real-time system (RTS), Backward Recovery, Multiprocessor, Discrete Event Simulation (DES).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 970
1668 Economic Evaluation of Bowland Shale Gas Wells Development in the UK

Authors: Elijah Acquah-Andoh

Abstract:

The UK has had its fair share of the shale gas revolutionary waves blowing across the global oil and gas industry at present. Although, its exploitation is widely agreed to have been delayed, shale gas was looked upon favorably by the UK Parliament when they recognized it as genuine energy source and granted licenses to industry to search and extract the resource. This, although a significant progress by industry, there yet remains another test the UK fracking resource must pass in order to render shale gas extraction feasible – it must be economically extractible and sustainably so. Developing unconventional resources is much more expensive and risky, and for shale gas wells, producing in commercial volumes is conditional upon drilling horizontal wells and hydraulic fracturing, techniques which increase CAPEX. Meanwhile, investment in shale gas development projects is sensitive to gas price and technical and geological risks. Using a Two-Factor Model, the economics of the Bowland shale wells were analyzed and the operational conditions under which fracking is profitable in the UK was characterized. We find that there is a great degree of flexibility about Opex spending; hence Opex does not pose much threat to the fracking industry in the UK. However, we discover Bowland shale gas wells fail to add value at gas price of $8/ Mmbtu. A minimum gas price of $12/Mmbtu at Opex of no more than $2/ Mcf and no more than $14.95M Capex are required to create value within the present petroleum tax regime, in the UK fracking industry.

Keywords: Capex, economical, investment, profitability, shale gas development, sustainable.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2722
1667 Proxisch: An Optimization Approach of Large-Scale Unstable Proxy Servers Scheduling

Authors: Xiaoming Jiang, Jinqiao Shi, Qingfeng Tan, Wentao Zhang, Xuebin Wang, Muqian Chen

Abstract:

Nowadays, big companies such as Google, Microsoft, which have adequate proxy servers, have perfectly implemented their web crawlers for a certain website in parallel. But due to lack of expensive proxy servers, it is still a puzzle for researchers to crawl large amounts of information from a single website in parallel. In this case, it is a good choice for researchers to use free public proxy servers which are crawled from the Internet. In order to improve efficiency of web crawler, the following two issues should be considered primarily: (1) Tasks may fail owing to the instability of free proxy servers; (2) A proxy server will be blocked if it visits a single website frequently. In this paper, we propose Proxisch, an optimization approach of large-scale unstable proxy servers scheduling, which allow anyone with extremely low cost to run a web crawler efficiently. Proxisch is designed to work efficiently by making maximum use of reliable proxy servers. To solve second problem, it establishes a frequency control mechanism which can ensure the visiting frequency of any chosen proxy server below the website’s limit. The results show that our approach performs better than the other scheduling algorithms.

Keywords: Proxy server, priority queue, optimization approach, distributed web crawling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2894
1666 Application of RP Technology with Polycarbonate Material for Wind Tunnel Model Fabrication

Authors: A. Ahmadi Nadooshan, S. Daneshmand, C. Aghanajafi

Abstract:

Traditionally, wind tunnel models are made of metal and are very expensive. In these years, everyone is looking for ways to do more with less. Under the right test conditions, a rapid prototype part could be tested in a wind tunnel. Using rapid prototype manufacturing techniques and materials in this way significantly reduces time and cost of production of wind tunnel models. This study was done of fused deposition modeling (FDM) and their ability to make components for wind tunnel models in a timely and cost effective manner. This paper discusses the application of wind tunnel model configuration constructed using FDM for transonic wind tunnel testing. A study was undertaken comparing a rapid prototyping model constructed of FDM Technologies using polycarbonate to that of a standard machined steel model. Testing covered the Mach range of Mach 0.3 to Mach 0.75 at an angle-ofattack range of - 2° to +12°. Results from this study show relatively good agreement between the two models and rapid prototyping Method reduces time and cost of production of wind tunnel models. It can be concluded from this study that wind tunnel models constructed using rapid prototyping method and materials can be used in wind tunnel testing for initial baseline aerodynamic database development.

Keywords: Polycarbonate, Fabrication, FDM, Model, RapidPrototyping, Wind Tunnel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450
1665 Deterioration Assessment Models for Water Pipelines

Authors: L. Parvizsedghy, I. Gkountis, A. Senouci, T. Zayed, M. Alsharqawi, H. El Chanati, M. El-Abbasy, F. Mosleh

Abstract:

The aging and deterioration of water pipelines in cities worldwide result in more frequent water main breaks, water service disruptions, and flooding damage. Therefore, there is an urgent need for undertaking proper maintenance procedures to avoid breaks and disastrous failures. However, due to budget limitations, the maintenance of water pipeline networks needs to be prioritized through efficient deterioration assessment models. Previous studies focused on the development of structural or physical deterioration assessment models, which require expensive inspection data. But, this paper aims at developing deterioration assessment models for water pipelines using statistical techniques. Several deterioration models were developed based on pipeline size, material type, and soil type using linear regression analysis. The categorical nature of some variables affecting pipeline deterioration was considered through developing several categorical models. The developed models were validated with an average validity percentage greater than 95%. Moreover, sensitivity analysis was carried out against different classifications and it displayed higher importance of age of pipes compared to other factors. The developed models will be helpful for the water municipalities and asset managers to assess the condition of their pipes and prioritize them for maintenance and inspection purposes.

Keywords: Water pipelines, deterioration assessment models, regression analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1203
1664 Unsupervised Classification of DNA Barcodes Species Using Multi-Library Wavelet Networks

Authors: Abdesselem Dakhli, Wajdi Bellil, Chokri Ben Amar

Abstract:

DNA Barcode provides good sources of needed information to classify living species. The classification problem has to be supported with reliable methods and algorithms. To analyze species regions or entire genomes, it becomes necessary to use the similarity sequence methods. A large set of sequences can be simultaneously compared using Multiple Sequence Alignment which is known to be NP-complete. However, all the used methods are still computationally very expensive and require significant computational infrastructure. Our goal is to build predictive models that are highly accurate and interpretable. In fact, our method permits to avoid the complex problem of form and structure in different classes of organisms. The empirical data and their classification performances are compared with other methods. Evenly, in this study, we present our system which is consisted of three phases. The first one, is called transformation, is composed of three sub steps; Electron-Ion Interaction Pseudopotential (EIIP) for the codification of DNA Barcodes, Fourier Transform and Power Spectrum Signal Processing. Moreover, the second phase step is an approximation; it is empowered by the use of Multi Library Wavelet Neural Networks (MLWNN). Finally, the third one, is called the classification of DNA Barcodes, is realized by applying the algorithm of hierarchical classification.

Keywords: DNA Barcode, Electron-Ion Interaction Pseudopotential, Multi Library Wavelet Neural Networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1969
1663 An Analysis of New Service Interchange Designs

Authors: Joseph E. Hummer

Abstract:

An efficient freeway system will be essential to the development of Africa, and interchanges are a key to that efficiency. Around the world, many interchanges between freeways and surface streets, called service interchanges, are of the diamond configuration, and interchanges using roundabouts or loop ramps are also popular. However, many diamond interchanges have serious operational problems, interchanges with roundabouts fail at high demand levels, and loops use lots of expensive land. Newer service interchange designs provide other options. The most popular new interchange design in the US at the moment is the double crossover diamond (DCD), also known as the diverging diamond. The DCD has enormous potential, but also has several significant limitations. The objectives of this paper are to review new service interchange options and to highlight some of the main features of those alternatives. The paper tests four conventional and seven unconventional designs using seven measures related to efficiency, cost, and safety. The results show that there is no superior design in all measures investigated. The DCD is better than most designs tested on most measures examined. However, the DCD was only superior to all other designs for bridge width. The DCD performed relatively poorly for capacity and for serving pedestrians. Based on the results, African freeway designers are encouraged to investigate the full range of alternatives that could work at the spot of interest. Diamonds and DCDs have their niches, but some of the other designs investigated could be optimum at some spots.

Keywords: Alternative, design, diverging diamond, freeway, interchange.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2279
1662 Reverse Engineering of Agricultural Machinery: A Key to Food Sufficiency in Nigeria

Authors: Williams S. Ebhota, Virginia Chika Ebhota, Samuel A. Ilupeju

Abstract:

Agriculture employs about three-quarter of Nigeria's workforce and yet food sufficiency is a challenge in the country. This is largely due to poor and outdated pre-harvest and post-harvest farming practices. The land fallow system is still been practised as fertiliser production in the country is grossly inadequate and expensive. The few available post-harvest processing facilities are faced with ageing and are inefficient. Also, use of modern processing equipment is limited by farmers' lack of fund, adequate capacity to operate and maintain modern farming equipment. This paper, therefore, examines key barriers to agricultural products processing equipment in the country. These barriers include over-dependence on foreign technologies and expertise; poor and inadequate manufacturing infrastructure; and lack of political will by political leaders; lack of funds; and lack of adequate technical skills. This paper, however, sees the increase in the domestic manufacturing of pre-harvest and post-harvest machinery and equipment through reverse engineering approach as a key to food production sufficiency in Nigeria.

Keywords: Agricultural machinery, domestic manufacturing, forward engineering, production reverse engineering, technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1078
1661 A Mixing Matrix Estimation Algorithm for Speech Signals under the Under-Determined Blind Source Separation Model

Authors: Jing Wu, Wei Lv, Yibing Li, Yuanfan You

Abstract:

The separation of speech signals has become a research hotspot in the field of signal processing in recent years. It has many applications and influences in teleconferencing, hearing aids, speech recognition of machines and so on. The sounds received are usually noisy. The issue of identifying the sounds of interest and obtaining clear sounds in such an environment becomes a problem worth exploring, that is, the problem of blind source separation. This paper focuses on the under-determined blind source separation (UBSS). Sparse component analysis is generally used for the problem of under-determined blind source separation. The method is mainly divided into two parts. Firstly, the clustering algorithm is used to estimate the mixing matrix according to the observed signals. Then the signal is separated based on the known mixing matrix. In this paper, the problem of mixing matrix estimation is studied. This paper proposes an improved algorithm to estimate the mixing matrix for speech signals in the UBSS model. The traditional potential algorithm is not accurate for the mixing matrix estimation, especially for low signal-to noise ratio (SNR).In response to this problem, this paper considers the idea of an improved potential function method to estimate the mixing matrix. The algorithm not only avoids the inuence of insufficient prior information in traditional clustering algorithm, but also improves the estimation accuracy of mixing matrix. This paper takes the mixing of four speech signals into two channels as an example. The results of simulations show that the approach in this paper not only improves the accuracy of estimation, but also applies to any mixing matrix.

Keywords: Clustering algorithm, potential function, speech signal, the UBSS model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 684
1660 Decision Analysis Module for Excel

Authors: Radomir Perzina, Jaroslav Ramik

Abstract:

The Analytic Hierarchy Process is frequently used approach for solving decision making problems. There exists wide range of software programs utilizing that approach. Their main disadvantage is that they are relatively expensive and missing intermediate calculations. This work introduces a Microsoft Excel add-in called DAME – Decision Analysis Module for Excel. Comparing to other computer programs DAME is free, can work with scenarios or multiple decision makers and displays intermediate calculations. Users can structure their decision models into three levels – scenarios/users, criteria and variants. Items on all levels can be evaluated either by weights or pair-wise comparisons. There are provided three different methods for the evaluation of the weights of criteria, the variants as well as the scenarios – Saaty’s Method, Geometric Mean Method and Fuller’s Triangle Method. Multiplicative and additive syntheses are supported. The proposed software package is demonstrated on couple of illustrating examples of real life decision problems.

Keywords: Analytic hierarchy process, multi-criteria decision making, pair-wise comparisons, Microsoft Excel, Scenarios.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3407
1659 Generating State-Based Testing Models for Object-Oriented Framework Interface Classes

Authors: Jehad Al Dallal, Paul Sorenson

Abstract:

An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define the Framework Interface Classes (FICs) and the specifications of their methods. As part of the development life cycle, it is required to test the implementations of the FICs. Building a testing model to express the behavior of a class is an essential step for the generation of the class-based test cases. The testing model has to be consistent with the specifications provided for the hooks. State-based models consisting of states and transitions are testing models well suited to objectoriented software. Typically, hand-construction of a state-based model of a class behavior is expensive, error-prone, and may result in constructing an inconsistent model with the specifications of the class methods, which misleads verification results. In this paper, a technique is introduced to automatically synthesize a state-based testing model for FICs using the specifications provided for the hooks. A tool that supports the proposed technique is introduced.

Keywords: Framework interface classes, hooks, state-basedtesting, testing model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1229
1658 Applying p-Balanced Energy Technique to Solve Liouville-Type Problems in Calculus

Authors: Lina Wu, Ye Li, Jia Liu

Abstract:

We are interested in solving Liouville-type problems to explore constancy properties for maps or differential forms on Riemannian manifolds. Geometric structures on manifolds, the existence of constancy properties for maps or differential forms, and energy growth for maps or differential forms are intertwined. In this article, we concentrate on discovery of solutions to Liouville-type problems where manifolds are Euclidean spaces (i.e. flat Riemannian manifolds) and maps become real-valued functions. Liouville-type results of vanishing properties for functions are obtained. The original work in our research findings is to extend the q-energy for a function from finite in Lq space to infinite in non-Lq space by applying p-balanced technique where q = p = 2. Calculation skills such as Hölder's Inequality and Tests for Series have been used to evaluate limits and integrations for function energy. Calculation ideas and computational techniques for solving Liouville-type problems shown in this article, which are utilized in Euclidean spaces, can be universalized as a successful algorithm, which works for both maps and differential forms on Riemannian manifolds. This innovative algorithm has a far-reaching impact on research work of solving Liouville-type problems in the general settings involved with infinite energy. The p-balanced technique in this algorithm provides a clue to success on the road of q-energy extension from finite to infinite.

Keywords: Differential Forms, Hölder Inequality, Liouville-type problems, p-balanced growth, p-harmonic maps, q-energy growth, tests for series.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 833
1657 Design and Fabrication of a Parabolic Trough Collector and Experimental Investigation of Wind Impact on Direct Steam Production in Tehran

Authors: H. Akhbari, M. Bidi, A. Bakhtiari, S. Eslami

Abstract:

The present paper aims to the techno-economic feasibility of enhancing low-cost parabolic trough collectors in the light of developing the use of solar energy in under-developed regions where expensive high-tech solar devices cannot be afforded. Moreover, the collector is aimed to produce steam so that its performance is based on heat which can be discovered. In this regard, the manufacturing process and the detailed design models in Solidworks software are elaborated. Furthermore, the colletor’s material is chosen in a way to minimize the costs. Finally, to assess the performance of the built collector, it is installed in the site of Shahid Beheshti University, Tehran, and the values of the effective peripheral parameters, such as temperature, wind speed, and most importantly, solar irradiance, are recorded simultaneously in June. According to the results obtained, the manufactured collector with the aperture area of 2 m2 (1×2 m) is capable of producing 350 ml.h-1 steam. Also, the wind influence is comprehensively investigated in this paper. As a case in point, it was measured that as the wind speed maximized to 9.77 km/h, the amount of steam outlet is minimized to 580 ml.

Keywords: Direct steam production, design and fabrication parabolic trough collector, solar water heater, wind impact, experimental investigation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 996
1656 Phenomenological Ductile Fracture Criteria Applied to the Cutting Process

Authors: František Šebek, Petr Kubík, Jindřich Petruška, Jiří Hůlka

Abstract:

Present study is aimed on the cutting process of circular cross-section rods where the fracture is used to separate one rod into two pieces. Incorporating the phenomenological ductile fracture model into the explicit formulation of finite element method, the process can be analyzed without the necessity of realizing too many real experiments which could be expensive in case of repetitive testing in different conditions. In the present paper, the steel AISI 1045 was examined and the tensile tests of smooth and notched cylindrical bars were conducted together with biaxial testing of the notched tube specimens to calibrate material constants of selected phenomenological ductile fracture models. These were implemented into the Abaqus/Explicit through user subroutine VUMAT and used for cutting process simulation. As the calibration process is based on variables which cannot be obtained directly from experiments, numerical simulations of fracture tests are inevitable part of the calibration. Finally, experiments regarding the cutting process were carried out and predictive capability of selected fracture models is discussed. Concluding remarks then make the summary of gained experience both with the calibration and application of particular ductile fracture criteria.

Keywords: Ductile fracture, phenomenological criteria, cutting process, explicit formulation, AISI 1045 steel.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2596
1655 The Impact of Alumina Cement on Properties of Portland Cement Slurries and Mortars

Authors: Krzysztof Zieliński, Dariusz Kierzek

Abstract:

The addition of a small amount of alumina cement to Portland cement results in immediate setting, a rapid increase in the compressive strength and a clear increase of the adhesion to concrete substrate. This phenomenon is used, among others, for the production of liquid floor self-levelling compounds. Alumina cement is several times more expensive than Portland cement and is a component having a significant impact on prices of products manufactured with its use. For the production of liquid floor self-levelling compounds, low-alumina cement containing approximately 40% Al2O3 is normally used. The aim of the study was to determine the impact of Portland cement with the addition of alumina cement on the basic physical and mechanical properties of cement slurries and mortars. CEM I 42.5R and three types of alumina cement containing 40%, 50% and 70% of Al2O3 were used for the tests. Mixes containing 4%, 6%, 8%, 10% and 12% of different varieties of alumina cement were prepared; for which, the time of initial and final setting, compressive and flexural strength and adhesion to concrete substrate were determined. The analysis of the obtained test results showed that a similar immediate setting effect and clearly better adhesion strength can be obtained using the addition of 6% of high-alumina cement than 12% of low-alumina cement. As the prices of these cements are similar, this can give significant financial savings in the production of liquid floor self-levelling compounds.

Keywords: Alumina cement, immediate setting, compression strength, adhesion to substrate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 667
1654 Outdoor Anomaly Detection with a Spectroscopic Line Detector

Authors: O. J. G. Somsen

Abstract:

One of the tasks of optical surveillance is to detect anomalies in large amounts of image data. However, if the size of the anomaly is very small, limited information is available to distinguish it from the surrounding environment. Spectral detection provides a useful source of additional information and may help to detect anomalies with a size of a few pixels or less. Unfortunately, spectral cameras are expensive because of the difficulty of separating two spatial in addition to one spectral dimension. We investigate the possibility of modifying a simple spectral line detector for outdoor detection. This may be especially useful if the area of interest forms a line, such as the horizon. We use a monochrome CCD that also enables detection into the near infrared. A simple camera is attached to the setup to determine which part of the environment is spectrally imaged. Our preliminary results indicate that sensitive detection of very small targets is indeed possible. Spectra could be taken from the various targets by averaging columns in the line image. By imaging a set of lines of various widths we found narrow lines that could not be seen in the color image but remained visible in the spectral line image. A simultaneous analysis of the entire spectra can produce better results than visual inspection of the line spectral image. We are presently developing calibration targets for spatial and spectral focusing and alignment with the spatial camera. This will present improved results and more use in outdoor application.

Keywords: Anomaly detection, spectroscopic line imaging, image analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1650