Search results for: Construction Project
347 User-Friendly Task Creation Using a CAD Integrated Robotic System on a Real Workcell
Authors: Alireza Changizi, Arash Rezaei, Jamal Muhammad, Jyrki Latokartano, Minna Lanz
Abstract:
Offline programming (OLP) is a new method in robot programming which is used widely in the industry nowadays which is a simulation base method that can produce the robot codes for motion according to virtual world in the simulation software. In this project Delmia v5 is used as simulation software. First the work cell component was modelled by Catia v5 and all of them was imported to a process file in Delmia and placed roughly to form the virtual work cell. Then robot was added to the work cell from the Delmia library. Work cell was calibrated corresponding to real world work cell to have accurate code. Tool calibration is the first step of calibration scheme and then work cell equipment can be calibrated using 6 point calibration method. Finally generated code needs to be reformed to match related controller code instruction. At the last stage IO were set to accomplish robots cooperation and make their motion synchronized. The pros and cons also will be discussed to clarify the presented results show the feasibility of the method and its effect on production line efficiency. Finally the positive and negative points of the implementation will be discussed.
Keywords: Component, robotic, automated, production, offline programming, CAD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1112346 Comparisons of Co-Seismic Gravity Changes between GRACE Observations and the Predictions from the Finite-Fault Models for the 2012 Mw = 8.6 Indian Ocean Earthquake Off-Sumatra
Authors: Armin Rahimi
Abstract:
The Gravity Recovery and Climate Experiment (GRACE) has been a very successful project in determining math redistribution within the Earth system. Large deformations caused by earthquakes are in the high frequency band. Unfortunately, GRACE is only capable to provide reliable estimate at the low-to-medium frequency band for the gravitational changes. In this study, we computed the gravity changes after the 2012 Mw8.6 Indian Ocean earthquake off-Sumatra using the GRACE Level-2 monthly spherical harmonic (SH) solutions released by the University of Texas Center for Space Research (UTCSR). Moreover, we calculated gravity changes using different fault models derived from teleseismic data. The model predictions showed non-negligible discrepancies in gravity changes. However, after removing high-frequency signals, using Gaussian filtering 350 km commensurable GRACE spatial resolution, the discrepancies vanished, and the spatial patterns of total gravity changes predicted from all slip models became similar at the spatial resolution attainable by GRACE observations, and predicted-gravity changes were consistent with the GRACE-detected gravity changes. Nevertheless, the fault models, in which give different slip amplitudes, proportionally lead to different amplitude in the predicted gravity changes.Keywords: Undersea earthquake, GRACE observation, gravity change, dislocation model, slip distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 991345 Sensitivity and Reliability Analysis of Masonry Infilled Frames
Authors: Avadhoot Bhosale, Robin Davis P., Pradip Sarkar
Abstract:
The seismic performance of buildings with irregular distribution of mass, stiffness and strength along the height may be significantly different from that of regular buildings with masonry infill. Masonry infilled reinforced concrete (RC) frames are very common structural forms used for multi-storey building construction. These structures are found to perform better in past earthquakes owing to additional strength, stiffness and energy dissipation in the infill walls. The seismic performance of a building depends on the variation of material, structural and geometrical properties. The sensitivity of these properties affects the seismic response of the building. The main objective of the sensitivity analysis is to found out the most sensitive parameter that affects the response of the building. This paper presents a sensitivity analysis by considering 5% and 95% probability value of random variable in the infills characteristics, trying to obtain a reasonable range of results representing a wide number of possible situations that can be met in practice by using pushover analysis. The results show that the strength-related variation values of concrete and masonry, with the exception of tensile strength of the concrete, have shown a significant effect on the structural performance and that this effect increases with the progress of damage condition for the concrete. The seismic risk assessments of the selected frames are expressed in terms of reliability index.Keywords: Fragility curve, sensitivity analysis, reliability index, RC frames.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1205344 A Machine Learning Approach for Anomaly Detection in Environmental IoT-Driven Wastewater Purification Systems
Authors: Giovanni Cicceri, Roberta Maisano, Nathalie Morey, Salvatore Distefano
Abstract:
The main goal of this paper is to present a solution for a water purification system based on an Environmental Internet of Things (EIoT) platform to monitor and control water quality and machine learning (ML) models to support decision making and speed up the processes of purification of water. A real case study has been implemented by deploying an EIoT platform and a network of devices, called Gramb meters and belonging to the Gramb project, on wastewater purification systems located in Calabria, south of Italy. The data thus collected are used to control the wastewater quality, detect anomalies and predict the behaviour of the purification system. To this extent, three different statistical and machine learning models have been adopted and thus compared: Autoregressive Integrated Moving Average (ARIMA), Long Short Term Memory (LSTM) autoencoder, and Facebook Prophet (FP). The results demonstrated that the ML solution (LSTM) out-perform classical statistical approaches (ARIMA, FP), in terms of both accuracy, efficiency and effectiveness in monitoring and controlling the wastewater purification processes.Keywords: EIoT, machine learning, anomaly detection, environment monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1027343 Performance Comparisons between PID and Adaptive PID Controllers for Travel Angle Control of a Bench-Top Helicopter
Authors: H. Mansor, S. B. Mohd-Noor, T. S. Gunawan, S. Khan, N. I. Othman, N. Tazali, R. B. Islam
Abstract:
This paper provides a comparative study on the performances of standard PID and adaptive PID controllers tested on travel angle of a 3-Degree-of-Freedom (3-DOF) Quanser bench-top helicopter. Quanser, a well-known manufacturer of educational bench-top helicopter has developed Proportional Integration Derivative (PID) controller with Linear Quadratic Regulator (LQR) for all travel, pitch and yaw angle of the bench-top helicopter. The performance of the PID controller is relatively good; however, its performance could also be improved if the controller is combined with adaptive element. The objective of this research is to design adaptive PID controller and then compare the performances of the adaptive PID with the standard PID. The controller design and test is focused on travel angle control only. Adaptive method used in this project is self-tuning controller, which controller’s parameters are updated online. Two adaptive algorithms those are pole-placement and deadbeat have been chosen as the method to achieve optimal controller’s parameters. Performance comparisons have shown that the adaptive (deadbeat) PID controller has produced more desirable performance compared to standard PID and adaptive (poleplacement). The adaptive (deadbeat) PID controller attained very fast settling time (5 seconds) and very small percentage of overshoot (5% to 7.5%) for 10° to 30° step change of travel angle.
Keywords: Adaptive control, bench-top helicopter, deadbeat, pole-placement, self-tuning control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3312342 Evaluation of Antioxidant Activities of Rice Paddy Herb (Limnophila aromatica (Lam.) Merr.)
Authors: Rutanachai Thaipratum
Abstract:
Free radicals are atoms or molecules with unpaired electrons. Many diseases are caused by free radicals. Normally, free radical formation is controlled naturally by various beneficial compounds known as antioxidants. Several analytical methods have been used for qualitative and quantitative determination of antioxidants, and each has its own specificity. This project aimed to evaluate antioxidant activity of ethanolic and aqueous extracts from the rice paddy herb (Limnophila aromatica (Lam.) Merr.) measured by DPPH and Hydroxyl radical scavenging method. The results showed that averaged antioxidant activity measured in ethanolic extract (µmol Ascorbic acid equivalent/g fresh mass) were 67.09± 4.99 and 15.55±4.82 as determined by DPPH and Hydroxyl radical scavenging activity assays, respectively. Averaged antioxidant activity measured in aqueous extract (µmol Ascorbic acid equivalent/g fresh mass) were 21.08±1.25 and 10.14±3.94 as determined by DPPH and Hydroxyl radical scavenging activity assays respectively.
Keywords: Free radical, antioxidant, rice paddy herb, Limnophila aromatica (Lam.) Merr.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2180341 A Combined Approach of a Sequential Life Testing and an Accelerated Life Testing Applied to a Low-Alloy High Strength Steel Component
Authors: D. I. De Souza, D. R. Fonseca, G. P. Azevedo
Abstract:
Sometimes the amount of time available for testing could be considerably less than the expected lifetime of the component. To overcome such a problem, there is the accelerated life-testing alternative aimed at forcing components to fail by testing them at much higher-than-intended application conditions. These models are known as acceleration models. One possible way to translate test results obtained under accelerated conditions to normal using conditions could be through the application of the “Maxwell Distribution Law.” In this paper we will apply a combined approach of a sequential life testing and an accelerated life testing to a low alloy high-strength steel component used in the construction of overpasses in Brazil. The underlying sampling distribution will be three-parameter Inverse Weibull model. To estimate the three parameters of the Inverse Weibull model we will use a maximum likelihood approach for censored failure data. We will be assuming a linear acceleration condition. To evaluate the accuracy (significance) of the parameter values obtained under normal conditions for the underlying Inverse Weibull model we will apply to the expected normal failure times a sequential life testing using a truncation mechanism. An example will illustrate the application of this procedure.
Keywords: Sequential Life Testing, Accelerated Life Testing, Underlying Three-Parameter Weibull Model, Maximum Likelihood Approach, Hypothesis Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1639340 Feasibility Studies through Quantitative Methods: The Revamping of a Tourist Railway Line in Italy
Authors: Armando Cartenì, Ilaria Henke
Abstract:
Recently, the Italian government has approved a new law for public contracts and has been laying the groundwork for restarting a planning phase. The government has adopted the indications given by the European Commission regarding the estimation of the external costs within the Cost-Benefit Analysis, and has been approved the ‘Guidelines for assessment of Investment Projects’. In compliance with the new Italian law, the aim of this research was to perform a feasibility study applying quantitative methods regarding the revamping of an Italian tourist railway line. A Cost-Benefit Analysis was performed starting from the quantification of the passengers’ demand potentially interested in using the revamped rail services. The benefits due to the external costs reduction were also estimated (quantified) in terms of variations (with respect to the not project scenario): climate change, air pollution, noises, congestion, and accidents. Estimations results have been proposed in terms of the Measure of Effectiveness underlying a positive Net Present Value equal to about 27 million of Euros, an Internal Rate of Return much greater the discount rate, a benefit/cost ratio equal to 2 and a PayBack Period of 15 years.
Keywords: Cost-benefit analysis, evaluation analysis, demand management, external cost, transport planning, quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 882339 Construction of Attitude Reference Benchmark for Test of Star Sensor Based on Precise Timing
Authors: Tingting Lu, Yonghai Wang, Haiyong Wang, Jiaqi Liu
Abstract:
To satisfy the need of outfield tests of star sensors, a method is put forward to construct the reference attitude benchmark. Firstly, its basic principle is introduced; Then, all the separate conversion matrixes are deduced, which include: the conversion matrix responsible for the transformation from the Earth Centered Inertial frame i to the Earth-centered Earth-fixed frame w according to the time of an atomic clock, the conversion matrix from frame w to the geographic frame t, and the matrix from frame t to the platform frame p, so the attitude matrix of the benchmark platform relative to the frame i can be obtained using all the three matrixes as the multiplicative factors; Next, the attitude matrix of the star sensor relative to frame i is got when the mounting matrix from frame p to the star sensor frame s is calibrated, and the reference attitude angles for star sensor outfield tests can be calculated from the transformation from frame i to frame s; Finally, the computer program is finished to solve the reference attitudes, and the error curves are drawn about the three axis attitude angles whose absolute maximum error is just 0.25ÔÇ│. The analysis on each loop and the final simulating results manifest that the method by precise timing to acquire the absolute reference attitude is feasible for star sensor outfield tests.Keywords: Atomic time, attitude determination, coordinate conversion, inertial coordinate system, star sensor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1200338 Comparison between Deterministic and Probabilistic Stability Analysis, Featuring Consequent Risk Assessment
Authors: Isabela Moreira Queiroz
Abstract:
Slope stability analyses are largely carried out by deterministic methods and evaluated through a single security factor. Although it is known that the geotechnical parameters can present great dispersal, such analyses are considered fixed and known. The probabilistic methods, in turn, incorporate the variability of input key parameters (random variables), resulting in a range of values of safety factors, thus enabling the determination of the probability of failure, which is an essential parameter in the calculation of the risk (probability multiplied by the consequence of the event). Among the probabilistic methods, there are three frequently used methods in geotechnical society: FOSM (First-Order, Second-Moment), Rosenblueth (Point Estimates) and Monte Carlo. This paper presents a comparison between the results from deterministic and probabilistic analyses (FOSM method, Monte Carlo and Rosenblueth) applied to a hypothetical slope. The end was held to evaluate the behavior of the slope and consequent risk analysis, which is used to calculate the risk and analyze their mitigation and control solutions. It can be observed that the results obtained by the three probabilistic methods were quite close. It should be noticed that the calculation of the risk makes it possible to list the priority to the implementation of mitigation measures. Therefore, it is recommended to do a good assessment of the geological-geotechnical model incorporating the uncertainty in viability, design, construction, operation and closure by means of risk management.Keywords: Probabilistic methods, risk assessment, risk management, slope stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1739337 An ACO Based Algorithm for Distribution Networks Including Dispersed Generations
Authors: B. Bahmani Firouzi, T. Niknam, M. Nayeripour
Abstract:
With Power system movement toward restructuring along with factors such as life environment pollution, problems of transmission expansion and with advancement in construction technology of small generation units, it is expected that small units like wind turbines, fuel cells, photovoltaic, ... that most of the time connect to the distribution networks play a very essential role in electric power industry. With increase in developing usage of small generation units, management of distribution networks should be reviewed. The target of this paper is to present a new method for optimal management of active and reactive power in distribution networks with regard to costs pertaining to various types of dispersed generations, capacitors and cost of electric energy achieved from network. In other words, in this method it-s endeavored to select optimal sources of active and reactive power generation and controlling equipments such as dispersed generations, capacitors, under load tapchanger transformers and substations in a way that firstly costs in relation to them are minimized and secondly technical and physical constraints are regarded. Because the optimal management of distribution networks is an optimization problem with continuous and discrete variables, the new evolutionary method based on Ant Colony Algorithm has been applied. The simulation results of the method tested on two cases containing 23 and 34 buses exist and will be shown at later sections.
Keywords: Distributed Generation, Optimal Operation Management of distribution networks, Ant Colony Optimization(ACO).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1710336 Optimum Design of Steel Space Frames by Hybrid Teaching-Learning Based Optimization and Harmony Search Algorithms
Authors: Alper Akın, İbrahim Aydoğdu
Abstract:
This study presents a hybrid metaheuristic algorithm to obtain optimum designs for steel space buildings. The optimum design problem of three-dimensional steel frames is mathematically formulated according to provisions of LRFD-AISC (Load and Resistance factor design of American Institute of Steel Construction). Design constraints such as the strength requirements of structural members, the displacement limitations, the inter-story drift and the other structural constraints are derived from LRFD-AISC specification. In this study, a hybrid algorithm by using teachinglearning based optimization (TLBO) and harmony search (HS) algorithms is employed to solve the stated optimum design problem. These algorithms are two of the recent additions to metaheuristic techniques of numerical optimization and have been an efficient tool for solving discrete programming problems. Using these two algorithms in collaboration creates a more powerful tool and mitigates each other’s weaknesses. To demonstrate the powerful performance of presented hybrid algorithm, the optimum design of a large scale steel building is presented and the results are compared to the previously obtained results available in the literature.Keywords: Optimum structural design, hybrid techniques, teaching-learning based optimization, harmony search algorithm, minimum weight, steel space frame.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2450335 Comparative Analysis of Ranunculus muricatus and Typha latifolia as Wetland Plants Applied for Domestic Wastewater Treatment in a Mesocosm Scale Study
Authors: S. Aziz, M. Ali, S. Asghar, S. Ahmed
Abstract:
Comparing other methods of waste water treatment, constructed wetlands are one of the most fascinating practices because being a natural process they are eco-friendly have low construction and maintenance cost and have considerable capability of wastewater treatment. The current research was focused mainly on comparison of Ranunculus muricatus and Typha latifolia as wetland plants for domestic wastewater treatment by designing and constructing efficient pilot scale horizontal subsurface flow mesocosms. Parameters like chemical oxygen demand, biological oxygen demand, phosphates, sulphates, nitrites, nitrates, and pathogenic indicator microbes were studied continuously with successive treatments. Treatment efficiency of the system increases with passage of time and with increase in temperature. Efficiency of T. latifolia planted setups in open environment was fairly good for parameters like COD and BOD5 which was showing reduction up to 82.5% for COD and 82.6% for BOD5 while DO was increased up to 125%. Efficiency of R. muricatus vegetated setup was also good but lowers than that of T. latifolia planted showing 80.95% removal of COD and BOD5. Ranunculus muricatus was found effective in reducing bacterial count in wastewater. Both macrophytes were found promising in wastewater treatment.
Keywords: Biological oxygen demand, chemical oxygen demand, horizontal subsurface flow, Total suspended solids, Wetland.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2635334 Discrete Element Modeling of the Effect of Particle Shape on Creep Behavior of Rockfills
Authors: Yunjia Wang, Zhihong Zhao, Erxiang Song
Abstract:
Rockfills are widely used in civil engineering, such as dams, railways, and airport foundations in mountain areas. A significant long-term post-construction settlement may affect the serviceability or even the safety of rockfill infrastructures. The creep behavior of rockfills is influenced by a number of factors, such as particle size, strength and shape, water condition and stress level. However, the effect of particle shape on rockfill creep still remains poorly understood, which deserves a careful investigation. Particle-based discrete element method (DEM) was used to simulate the creep behavior of rockfills under different boundary conditions. Both angular and rounded particles were considered in this numerical study, in order to investigate the influence of particle shape. The preliminary results showed that angular particles experience more breakages and larger creep strains under one-dimensional compression than rounded particles. On the contrary, larger creep strains were observed in he rounded specimens in the direct shear test. The mechanism responsible for this difference is that the possibility of the existence of key particle in rounded particles is higher than that in angular particles. The above simulations demonstrate that the influence of particle shape on the creep behavior of rockfills can be simulated by DEM properly. The method of DEM simulation may facilitate our understanding of deformation properties of rockfill materials.
Keywords: Rockfills, creep behavior, particle crushing, discrete element method, boundary conditions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1080333 Field-Programmable Gate Array Based Tester for Protective Relay
Authors: H. Bentarzi, A. Zitouni
Abstract:
The reliability of the power grid depends on the successful operation of thousands of protective relays. The failure of one relay to operate as intended may lead the entire power grid to blackout. In fact, major power system failures during transient disturbances may be caused by unnecessary protective relay tripping rather than by the failure of a relay to operate. Adequate relay testing provides a first defense against false trips of the relay and hence improves power grid stability and prevents catastrophic bulk power system failures. The goal of this research project is to design and enhance the relay tester using a technology such as Field Programmable Gate Array (FPGA) card NI 7851. A PC based tester framework has been developed using Simulink power system model for generating signals under different conditions (faults or transient disturbances) and LabVIEW for developing the graphical user interface and configuring the FPGA. Besides, the interface system has been developed for outputting and amplifying the signals without distortion. These signals should be like the generated ones by the real power system and large enough for testing the relay’s functionality. The signals generated that have been displayed on the scope are satisfactory. Furthermore, the proposed testing system can be used for improving the performance of protective relay.
Keywords: Amplifier class D, FPGA, protective relay, tester.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 804332 Quality Determinants of Client Satisfaction: A Case Study of Australian Consulting Engineers, Sydney, Australia
Authors: Elham S. Hasham, Anthony S. Hasham
Abstract:
The construction industry is one of Australia’s fastest growing industries and its success is a result of a firm’s client satisfaction with focus on product determinants such as price and quality. Ensuring quality at every phase is a must and building rapport with the client will go a long way. To capitalize on the growing demand for Engineering Consulting Firms (ECFs), it is imperative to stress the implication of customer satisfaction and excellence in standards and performance. Consequently, the emphasis should be on improving employee skills through various training provisions. Clients seek consistency and thus expect that all services should be similar in respect to quality and the ability of the service to meet their needs. This calls for empowerment and comfortable work conditions to motivate employees and give them incentive to deliver quality and excellent output. The methodology utilized is triangulation - a combination of both quantitative and qualitative research. The case study - Australian Consulting Engineers (ACE) - was established in Australia in 1995 and has operations throughout Australia, the Philippines, Europe, UAE, KSA, and with a branch in Lebanon. ACE is affiliated with key agencies and support organizations in the engineering industry with International Organization for Standardization (ISO) certifications in Safety and Quality Management. The objective of this study, conducted in Australia, is significant as it sheds light on employee motivation and client satisfaction as imperative determinants of the success of an organization.
Keywords: Organizational behaviour, leadership, satisfaction, motivation, quality service.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46331 Probabilistic Wavelet Neural Network Based Vibration Analysis of Induction Motor Drive
Authors: K. Jayakumar, S. Thangavel
Abstract:
In this paper proposed the effective fault detection of industrial drives by using Biorthogonal Posterior Vibration Signal-Data Probabilistic Wavelet Neural Network (BPPVS-WNN) system. This system was focused to reducing the current flow and to identify faults with lesser execution time with harmonic values obtained through fifth derivative. Initially, the construction of Biorthogonal vibration signal-data based wavelet transform in BPPVS-WNN system localizes the time and frequency domain. The Biorthogonal wavelet approximates the broken bearing using double scaling and factor, identifies the transient disturbance due to fault on induction motor through approximate coefficients and detailed coefficient. Posterior Probabilistic Neural Network detects the final level of faults using the detailed coefficient till fifth derivative and the results obtained through it at a faster rate at constant frequency signal on the industrial drive. Experiment through the Simulink tool detects the healthy and unhealthy motor on measuring parametric factors such as fault detection rate based on time, current flow rate, and execution time.Keywords: Biorthogonal Wavelet Transform, Posterior Probabilistic Neural Network, Induction Motor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1019330 Forecasting Optimal Production Program Using Profitability Optimization by Genetic Algorithm and Neural Network
Authors: Galal H. Senussi, Muamar Benisa, Sanja Vasin
Abstract:
In our business field today, one of the most important issues for any enterprises is cost minimization and profit maximization. Second issue is how to develop a strong and capable model that is able to give us desired forecasting of these two issues. Many researches deal with these issues using different methods. In this study, we developed a model for multi-criteria production program optimization, integrated with Artificial Neural Network.
The prediction of the production cost and profit per unit of a product, dealing with two obverse functions at same time can be extremely difficult, especially if there is a great amount of conflict information about production parameters.
Feed-Forward Neural Networks are suitable for generalization, which means that the network will generate a proper output as a result to input it has never seen. Therefore, with small set of examples the network will adjust its weight coefficients so the input will generate a proper output.
This essential characteristic is of the most important abilities enabling this network to be used in variety of problems spreading from engineering to finance etc.
From our results as we will see later, Feed-Forward Neural Networks has a strong ability and capability to map inputs into desired outputs.
Keywords: Project profitability, multi-objective optimization, genetic algorithm, Pareto set, Neural Networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2057329 AJcFgraph - AspectJ Control Flow Graph Builder for Aspect-Oriented Software
Authors: Reza Meimandi Parizi, Abdul Azim Abdul Ghani
Abstract:
The ever-growing usage of aspect-oriented development methodology in the field of software engineering requires tool support for both research environments and industry. So far, tool support for many activities in aspect-oriented software development has been proposed, to automate and facilitate their development. For instance, the AJaTS provides a transformation system to support aspect-oriented development and refactoring. In particular, it is well established that the abstract interpretation of programs, in any paradigm, pursued in static analysis is best served by a high-level programs representation, such as Control Flow Graph (CFG). This is why such analysis can more easily locate common programmatic idioms for which helpful transformation are already known as well as, association between the input program and intermediate representation can be more closely maintained. However, although the current researches define the good concepts and foundations, to some extent, for control flow analysis of aspectoriented programs but they do not provide a concrete tool that can solely construct the CFG of these programs. Furthermore, most of these works focus on addressing the other issues regarding Aspect- Oriented Software Development (AOSD) such as testing or data flow analysis rather than CFG itself. Therefore, this study is dedicated to build an aspect-oriented control flow graph construction tool called AJcFgraph Builder. The given tool can be applied in many software engineering tasks in the context of AOSD such as, software testing, software metrics, and so forth.Keywords: Aspect-Oriented Software Development, AspectJ, Control Flow Graph, Data Flow Analysis
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2100328 X-Ray Intensity Measurement Using Frequency Output Sensor for Computed Tomography
Authors: R. M. Siddiqui, D. Z. Moghaddam, T. R. Turlapati, S. H. Khan, I. Ul Ahad
Abstract:
Quality of 2D and 3D cross-sectional images produce by Computed Tomography primarily depend upon the degree of precision of primary and secondary X-Ray intensity detection. Traditional method of primary intensity detection is apt to errors. Recently the X-Ray intensity measurement system along with smart X-Ray sensors is developed by our group which is able to detect primary X-Ray intensity unerringly. In this study a new smart X-Ray sensor is developed using Light-to-Frequency converter TSL230 from Texas Instruments which has numerous advantages in terms of noiseless data acquisition and transmission. TSL230 construction is based on a silicon photodiode which converts incoming X-Ray radiation into the proportional current signal. A current to frequency converter is attached to this photodiode on a single monolithic CMOS integrated circuit which provides proportional frequency count to incoming current signal in the form of the pulse train. The frequency count is delivered to the center of PICDEM FS USB board with PIC18F4550 microcontroller mounted on it. With highly compact electronic hardware, this Demo Board efficiently read the smart sensor output data. The frequency output approaches overcome nonlinear behavior of sensors with analog output thus un-attenuated X-Ray intensities could be measured precisely and better normalization could be acquired in order to attain high resolution.Keywords: Computed tomography, detector technology, X-Ray intensity measurement
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2609327 Genetic Polymorphism of Main Lactoproteins of Romanian Grey Steppe Breed in Preservation
Authors: Şt. Creangâ, V. Maciuc, A.V. Bâlteanu, S.S. Chelmu
Abstract:
The paper presents a part of the results obtained in a complex research project on Romanian Grey Steppe breed, owner of some remarkable qualities such as hardiness, longevity, adaptability, special resistance to ban weather and diseases and included in the genetic fund (G.D. no. 822/2008.) from Romania. Following the researches effectuated, we identified alleles of six loci, codifying the six types of major milk proteins: alpha-casein S1 (α S1-cz); beta-casein (β-cz); kappa-casein (K-cz); beta-lactoglobulin (β-lg); alpha-lactalbumin (α-la) and alpha-casein S2 (α S2-cz). In system αS1-cz allele αs1-Cn B has the highest frequency (0.700), in system β-cz allele β-Cn A2 ( 0.550 ), in system K-cz allele k-CnA2 ( 0.583 ) and heterozygote genotype AB ( 0.416 ) and BB (0.375), in system β-lg allele β-lgA1 has the highest frequency (0.542 ) and heterozygote genotype AB ( 0.500 ), in system α-la there is monomorphism for allele α-la B and similarly in system αS2-cz for allele αs2-Cn A. The milk analysis by the isoelectric focalization technique (I.E.F.) allowed the identification of a new allele for locus αS1-casein, for two of the individuals under analysis, namely allele called αS1-casein IRV. When experiments were repeated, we noticed that this is not a proteolysis band and it really was a new allele that has not been registered in the specialized literature so far. We identified two heterozygote individuals, carriers of this allele, namely: BIRV and CIRV. This discovery is extremely important if focus is laid on the national genetic patrimony.Keywords: allele, breed, genetic preservation, lactoproteins, Romanian Grey Steppe
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1692326 Nuclear Medical Image Treatment System Based On FPGA in Real Time
Authors: B. Mahmoud, M.H. Bedoui, R. Raychev, H. Essabbah
Abstract:
We present in this paper an acquisition and treatment system designed for semi-analog Gamma-camera. It consists of a nuclear medical Image Acquisition, Treatment and Display chain(IATD) ensuring the acquisition, the treatment of the signals(resulting from the Gamma-camera detection head) and the scintigraphic image construction in real time. This chain is composed by an analog treatment board and a digital treatment board. We describe the designed systems and the digital treatment algorithms in which we have improved the performance and the flexibility. The digital treatment algorithms are implemented in a specific reprogrammable circuit FPGA (Field Programmable Gate Array).interface for semi-analog cameras of Sopha Medical Vision(SMVi) by taking as example SOPHY DS7. The developed system consists of an Image Acquisition, Treatment and Display (IATD) ensuring the acquisition and the treatment of the signals resulting from the DH. The developed chain is formed by a treatment analog board and a digital treatment board designed around a DSP [2]. In this paper we have presented the architecture of a new version of our chain IATD in which the integration of the treatment algorithms is executed on an FPGA (Field Programmable Gate Array)
Keywords: Nuclear medical image, scintigraphic image, digitaltreatment, linearity, spectrometry, FPGA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1676325 Indoor and Outdoor Concentration of Particulate Matter at Domestic Homes
Authors: B. Karakas, S. Lakestani, C. Guler, B. Guciz Dogan, S. Acar Vaizoglu, A. Taner, B. Sekerel, R. Tıpırdamaz, G. Gullu
Abstract:
Particulate matter (PM) in ambient air is responsible for adverse health effects in adults and children. Relatively little is known about the concentrations, sources and health effects of PM in indoor air. A monitoring study was conducted in Ankara by three campaigns in order to measure PM levels in indoor and outdoor environments to identify and quantify associations between sources and concentrations. Approximately 82 homes (1st campaign for 42, 2nd campaign for 12, and 3rd campaign for 28), three rooms (living room, baby-s room and living room used as a baby-s room) and outdoor ambient at each home were sampled with Grimm Environmental Dust Monitoring (EDM) 107, during different seasonal periods of 2011 and 2012. In this study, the relationship between indoor and outdoor PM levels for particulate matter less than 10 micrometer (.m) (PM10), particulate matter less than 2.5.m (PM2.5) and particulate matter less than 1.0.m (PM1) were investigated. The mean concentration of PM10, PM2.5, and PM1.0 at living room used as baby-s room is higher than living and baby-s room (or bedroom) for three sampling campaigns. It is concluded that the household activities and environmental conditions are very important for PM concentrations in the indoor environments during the sampling periods. The amount of smokers, being near a main street and/or construction activities increased the PM concentration. This study is based on the assessment the relationship between indoor and outdoor PM levels and the household activities and environmental conditionsKeywords: Indoor air quality, particulate matter (PM), PM10, PM2.5, PM1.0.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3236324 A Comparison of Air Quality in Arid and Temperate Climatic Conditions – A Case Study of Leeds and Makkah
Authors: Turki M. Habeebullah, Said Munir, Karl Ropkins, Essam A. Morsy, Atef M. F. Mohammed, Abdulaziz R. Seroji
Abstract:
In this paper air quality conditions in Makkah and Leeds are compared. These two cities have totally different climatic conditions. Makkah climate is characterised as hot and dry (arid) whereas that of Leeds is characterised as cold and wet (temperate). This study uses air quality data from 2012 collected in Makkah, Saudi Arabia and Leeds, UK. The concentrations of all pollutants, except NO are higher in Makkah. Most notable, the concentrations of PM10 are much higher in Makkah than in Leeds. This is probably due to the arid nature of climatic conditions in Makkah and not solely due to anthropogenic emission sources, otherwise like PM10 some of the other pollutants, such as CO, NO, and SO2 would have shown much greater difference between Leeds and Makkah. Correlation analysis is performed between different pollutants at the same site and the same pollutants at different sites. In Leeds the correlation between PM10 and other pollutants is significantly stronger than in Makkah. Weaker correlation in Makkah is probably due to the fact that in Makkah most of the gaseous pollutants are emitted by combustion processes, whereas most of the PM10 is generated by other sources, such as windblown dust, re-suspension, and construction activities. This is in contrast to Leeds where all pollutants including PM10 are predominantly emitted by combustions, such as road traffic. Furthermore, in Leeds frequent rains wash out most of the atmospheric particulate matter and suppress re-suspension of dust. Temporal trends of various pollutants are compared and discussed. This study emphasises the role of climatic conditions in managing air quality, and hence the need for region-specific controlling strategies according to the local climatic and meteorological conditions.Keywords: Air pollution, climatic conditions, particulate matter, Makkah, Leeds.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2564323 A New Intelligent, Dynamic and Real Time Management System of Sewerage
Authors: R. Tlili Yaakoubi, H. Nakouri, O. Blanpain, S. Lallahem
Abstract:
The current tools for real time management of sewer systems are based on two software tools: the software of weather forecast and the software of hydraulic simulation. The use of the first ones is an important cause of imprecision and uncertainty, the use of the second requires temporal important steps of decision because of their need in times of calculation. This way of proceeding fact that the obtained results are generally different from those waited. The major idea of this project is to change the basic paradigm by approaching the problem by the "automatic" face rather than by that "hydrology". The objective is to make possible the realization of a large number of simulations at very short times (a few seconds) allowing to take place weather forecasts by using directly the real time meditative pluviometric data. The aim is to reach a system where the decision-making is realized from reliable data and where the correction of the error is permanent. A first model of control laws was realized and tested with different return-period rainfalls. The gains obtained in rejecting volume vary from 19 to 100 %. The development of a new algorithm was then used to optimize calculation time and thus to overcome the subsequent combinatorial problem in our first approach. Finally, this new algorithm was tested with 16- year-rainfall series. The obtained gains are 40 % of total volume rejected to the natural environment and of 65 % in the number of discharges.Keywords: Automation, optimization, paradigm, RTC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1488322 Facility Location Selection using Preference Programming
Authors: C. Ardil
Abstract:
This paper presents preference programming technique based multiple criteria decision making analysis for selecting a facility location for a new organization or expansion of an existing facility which is of vital importance for a decision support system and strategic planning process. The implementation of decision support systems is considered crucial to sustain competitive advantage and profitability persistence in turbulent environment. As an effective strategic management and decision making is necessary, multiple criteria decision making analysis supports the decision makers to formulate and implement the right strategy. The investment cost associated with acquiring the property and facility construction makes the facility location selection problem a long-term strategic investment decision, which rationalize the best location selection which results in higher economic benefits through increased productivity and optimal distribution network. Selecting the proper facility location from a given set of alternatives is a difficult task, as many potential qualitative and quantitative multiple conflicting criteria are to be considered. This paper solves a facility location selection problem using preference programming, which is an effective multiple criteria decision making analysis tool applied to deal with complex decision problems in the operational research environment. The ranking results of preference programming are compared with WSM, TOPSIS and VIKOR methods.
Keywords: Facility Location Selection, Multiple Criteria Decision Making, Multiple Criteria Decision Making Analysis, Preference Programming, Location Selection, WSM, TOPSIS, VIKOR
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 541321 Experimental Evaluation of Drilling Damage on the Strength of Cores Extracted from RC Buildings
Authors: A. Masi, A. Digrisolo, G. Santarsiero
Abstract:
Concrete strength evaluated from compression tests on cores is affected by several factors causing differences from the in-situ strength at the location from which the core specimen was extracted. Among the factors, there is the damage possibly occurring during the drilling phase that generally leads to underestimate the actual in-situ strength. In order to quantify this effect, in this study two wide datasets have been examined, including: (i) about 500 core specimens extracted from Reinforced Concrete existing structures, and (ii) about 600 cube specimens taken during the construction of new structures in the framework of routine acceptance control. The two experimental datasets have been compared in terms of compression strength and specific weight values, accounting for the main factors affecting a concrete property, that is type and amount of cement, aggregates' grading, type and maximum size of aggregates, water/cement ratio, placing and curing modality, concrete age. The results show that the magnitude of the strength reduction due to drilling damage is strongly affected by the actual properties of concrete, being inversely proportional to its strength. Therefore, the application of a single value of the correction coefficient, as generally suggested in the technical literature and in structural codes, appears inappropriate. A set of values of the drilling damage coefficient is suggested as a function of the strength obtained from compressive tests on cores.
Keywords: RC Buildings, Assessment, In-situ concrete strength, Core testing, Drilling damage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2060320 Heat Treatment of Aluminum Alloy 7449
Authors: Suleiman E. Al-lubani, Mohammad E. Matarneh, Hussien M. Al-Wedyan, Ala M. Rayes
Abstract:
Aluminum alloy has an extensive range of industrial application due to its consistent mechanical properties and structural integrity. The heat treatment by precipitation technique affected the Magnesium, Silicon Manganese and copper crystals dissolved in the Aluminum alloy. The crystals dislocated to precipitate on the crystal’s boundaries of the Aluminum alloy when given a thermal energy increased its hardness. In this project various times and temperature were varied to find out the best combination of these variables to increase the precipitation of the metals on the Aluminum crystal’s boundaries which will lead to get the highest hardness. These specimens are then tested for their hardness and tensile strength. It is noticed that when the temperature increases, the precipitation increases and consequently the hardness increases. A threshold temperature value (264C0) of Aluminum alloy should not be reached due to the occurrence of recrystalization which causes the crystal to grow. This recrystalization process affected the ductility of the alloy and decrease hardness. In addition, and while increasing the temperature the alloy’s mechanical properties will decrease. The mechanical properties, namely tensile and hardness properties are investigated according to standard procedures. In this research, different temperature and time have been applied to increase hardening.The highest hardness at 100°c in 6 hours equals to 207.31 HBR, while at the same temperature and time the lowest elongation equals to 146.5.Keywords: Aluminum alloy, recrystalization process, heat treatment, hardness properties, precipitation, intergranular breakage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4076319 Concept for Determining the Focus of Technology Monitoring Activities
Authors: Guenther Schuh, Christina Koenig, Nico Schoen, Markus Wellensiek
Abstract:
Identification and selection of appropriate product and manufacturing technologies are key factors for competitiveness and market success of technology-based companies. Therefore, many companies perform technology intelligence (TI) activities to ensure the identification of evolving technologies at the right time. Technology monitoring is one of the three base activities of TI, besides scanning and scouting. As the technological progress is accelerating, more and more technologies are being developed. Against the background of limited resources it is therefore necessary to focus TI activities. In this paper we propose a concept for defining appropriate search fields for technology monitoring. This limitation of search space leads to more concentrated monitoring activities. The concept will be introduced and demonstrated through an anonymized case study conducted within an industry project at the Fraunhofer Institute for Production Technology IPT. The described concept provides a customized monitoring approach, which is suitable for use in technology-oriented companies. It is shown in this paper that the definition of search fields and search tasks are suitable methods to define topics of interest and thus to align monitoring activities. Current as well as planned product, production and material technologies and existing skills, capabilities and resources form the basis for derivation of relevant search areas. To further improve the concept of technology monitoring the proposed concept should be extended during future research e.g. by the definition of relevant monitoring parameters.
Keywords: Monitoring radar, search field, technology intelligence, technology monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3254318 Optimizing Forecasting for Indonesia's Coal and Palm Oil Exports: A Comparative Analysis of ARIMA, ANN, and LSTM Methods
Authors: Mochammad Dewo, Sumarsono Sudarto
Abstract:
The Exponential Triple Smoothing Algorithm approach nowadays, which is used to anticipate the export value of Indonesia's two major commodities, coal and palm oil, has a Mean Percentage Absolute Error (MAPE) value of 30-50%, which may be considered as a "reasonable" forecasting mistake. Forecasting errors of more than 30% shall have a domino effect on industrial output, as extra production adds to raw material, manufacturing and storage expenses. Whereas, reaching an "excellent" classification with an error value of less than 10% will provide new investors and exporters with confidence in the commercial development of related sectors. Industrial growth will bring out a positive impact on economic development. It can be applied for other commodities if the forecast error is less than 10%. The purpose of this project is to create a forecasting technique that can produce precise forecasting results with an error of less than 10%. This research analyzes forecasting methods such as ARIMA (Autoregressive Integrated Moving Average), ANN (Artificial Neural Network) and LSTM (Long-Short Term Memory). By providing a MAPE of 1%, this study reveals that ANN is the most successful strategy for forecasting coal and palm oil commodities in Indonesia.
Keywords: ANN, Artificial Neural Network, ARIMA, Autoregressive Integrated Moving Average, export value, forecast, LSTM, Long Short Term Memory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 224