Search results for: mark-up approach.
365 Evaluation of Stormwater Quantity and Quality Control through Constructed Mini Wet Pond: A Case Study
Authors: Y. S. Liew, K. A. Puteh Ariffin, M. A. Mohd Nor
Abstract:
One of the Best Management Practices (BMPs) promoted in Urban Stormwater Management Manual for Malaysia (MSMA) published by the Department of Irrigation and Drainage (DID) in 2001 is through the construction of wet ponds in new development projects for water quantity and quality control. Therefore, this paper aims to demonstrate a case study on evaluation of a constructed mini wet pond located at Sekolah Rendah Kebangsaan Seksyen 2, Puchong, Selangor, Malaysia in both stormwater quantity and quality aspect particularly to reduce the peak discharge by temporary storing and gradual release of stormwater runoff from an outlet structure or other release mechanism. The evaluation technique will be using InfoWorks Collection System (CS) as the numerical modeling approach for water quantity aspect. Statistical test by comparing the correlation coefficient (R2), mean error (ME), mean absolute error (MAE) and root mean square error (RMSE) were used to evaluate the model in simulating the peak discharge changes. Results demonstrated that there will be a reduction in peak flow at 11 % to 15% and time to peak flow is slower by 5 minutes through a wet pond. For water quality aspect, a survey on biological indicator of water quality carried out depicts that the pond is within the range of rather clean to clean water with the score of 5.3. This study indicates that a constructed wet pond with wetland facilities is able to help in managing water quantity and stormwater generated pollution at source, towards achieving ecologically sustainable development in urban areas.
Keywords: Wet pond, Retention Facilities, Best Management Practices (BMP), Urban Stormwater Management Manual for Malaysia (MSMA).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2527364 Efficient Design Optimization of Multi-State Flow Network for Multiple Commodities
Authors: Yu-Cheng Chou, Po Ting Lin
Abstract:
The network of delivering commodities has been an important design problem in our daily lives and many transportation applications. The delivery performance is evaluated based on the system reliability of delivering commodities from a source node to a sink node in the network. The system reliability is thus maximized to find the optimal routing. However, the design problem is not simple because (1) each path segment has randomly distributed attributes; (2) there are multiple commodities that consume various path capacities; (3) the optimal routing must successfully complete the delivery process within the allowable time constraints. In this paper, we want to focus on the design optimization of the Multi-State Flow Network (MSFN) for multiple commodities. We propose an efficient approach to evaluate the system reliability in the MSFN with respect to randomly distributed path attributes and find the optimal routing subject to the allowable time constraints. The delivery rates, also known as delivery currents, of the path segments are evaluated and the minimal-current arcs are eliminated to reduce the complexity of the MSFN. Accordingly, the correct optimal routing is found and the worst-case reliability is evaluated. It has been shown that the reliability of the optimal routing is at least higher than worst-case measure. Two benchmark examples are utilized to demonstrate the proposed method. The comparisons between the original and the reduced networks show that the proposed method is very efficient.
Keywords: Multiple Commodities, Multi-State Flow Network (MSFN), Time Constraints, Worst-Case Reliability (WCR)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1450363 A Technical Perspective on Roadway Safety in Eastern Province: Data Evaluation and Spatial Analysis
Authors: Muhammad Farhan, Sayed Faruque, Amr Mohammed, Sami Osman, Omar Al-Jabari, Abdul Almojil
Abstract:
Saudi Arabia in recent years has seen drastic increase in traffic related crashes. With population of over 29 million, Saudi Arabia is considered as a fast growing and emerging economy. The rapid population increase and economic growth has resulted in rapid expansion of transportation infrastructure, which has led to increase in road crashes. Saudi Ministry of Interior reported more than 7,000 people killed and 68,000 injured in 2011 ranking Saudi Arabia to be one of the worst worldwide in traffic safety. The traffic safety issues in the country also result in distress to road users and cause and economic loss exceeding 3.7 billion Euros annually. Keeping this in view, the researchers in Saudi Arabia are investigating ways to improve traffic safety conditions in the country. This paper presents a multilevel approach to collect traffic safety related data required to do traffic safety studies in the region. Two highway corridors including King Fahd Highway 39 kilometre and Gulf Cooperation Council Highway 42 kilometre long connecting the cities of Dammam and Khobar were selected as a study area. Traffic data collected included traffic counts, crash data, travel time data, and speed data. The collected data was analysed using geographic information system to evaluate any correlation. Further research is needed to investigate the effectiveness of traffic safety related data when collected in a concerted effort.
Keywords: Crash Data, Data Collection, Traffic Safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2352362 LCA and Multi-Criteria Analysis of Fly Ash Concrete Pavements
Authors: M. Ondova, A. Estokova
Abstract:
Rapid industrialization results in increased use of natural resources bring along serious ecological and environmental imbalance due to the dumping of industrial wastes. Principles of sustainable construction have to be accepted with regard to the consumption of natural resources and the production of harmful emissions. Cement is a great importance raw material in the building industry and today is its large amount used in the construction of concrete pavements. Concerning raw materials cost and producing CO2 emission the replacing of cement in concrete mixtures with more sustainable materials is necessary. To reduce this environmental impact people all over the world are looking for a solution. Over a period of last ten years, the image of fly ash has completely been changed from a polluting waste to resource material and it can solve the major problems of cement use. Fly ash concretes are proposed as a potential approach for achieving substantial reductions in cement. It is known that it improves the workability of concrete, extends the life cycle of concrete roads, and reduces energy use and greenhouse gas as well as amount of coal combustion products that must be disposed in landfills.
Life cycle assessment also proved that a concrete pavement with fly ash cement replacement is considerably more environmentally friendly compared to standard concrete roads. In addition, fly ash is cheap raw material, and the costs saving are guaranteed. The strength properties, resistance to a frost or de-icing salts, which are important characteristics in the construction of concrete pavements, have reached the required standards as well. In terms of human health it can´t be stated that a concrete cover with fly ash could be dangerous compared with a cover without fly ash. Final Multi-criteria analysis also pointed that a concrete with fly ash is a clearly proper solution.
Keywords: Life cycle assessment, fly ash, waste, concrete pavements
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3503361 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation
Authors: H. Khanfari, M. Johari Fard
Abstract:
Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.
Keywords: Carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1791360 FPGA Hardware Implementation and Evaluation of a Micro-Network Architecture for Multi-Core Systems
Authors: Yahia Salah, Med Lassaad Kaddachi, Rached Tourki
Abstract:
This paper presents the design, implementation and evaluation of a micro-network, or Network-on-Chip (NoC), based on a generic pipeline router architecture. The router is designed to efficiently support traffic generated by multimedia applications on embedded multi-core systems. It employs a simplest routing mechanism and implements the round-robin scheduling strategy to resolve output port contentions and minimize latency. A virtual channel flow control is applied to avoid the head-of-line blocking problem and enhance performance in the NoC. The hardware design of the router architecture has been implemented at the register transfer level; its functionality is evaluated in the case of the two dimensional Mesh/Torus topology, and performance results are derived from ModelSim simulator and Xilinx ISE 9.2i synthesis tool. An example of a multi-core image processing system utilizing the NoC structure has been implemented and validated to demonstrate the capability of the proposed micro-network architecture. To reduce complexity of the image compression and decompression architecture, the system use image processing algorithm based on classical discrete cosine transform with an efficient zonal processing approach. The experimental results have confirmed that both the proposed image compression scheme and NoC architecture can achieve a reasonable image quality with lower processing time.
Keywords: Generic Pipeline Network-on-Chip Router Architecture, JPEG Image Compression, FPGA Hardware Implementation, Performance Evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3097359 Constraint Based Frequent Pattern Mining Technique for Solving GCS Problem
Authors: First G.M. Karthik, Second Ramachandra.V.Pujeri, Dr.
Abstract:
Generalized Center String (GCS) problem are generalized from Common Approximate Substring problem and Common substring problems. GCS are known to be NP-hard allowing the problems lies in the explosion of potential candidates. Finding longest center string without concerning the sequence that may not contain any motifs is not known in advance in any particular biological gene process. GCS solved by frequent pattern-mining techniques and known to be fixed parameter tractable based on the fixed input sequence length and symbol set size. Efficient method known as Bpriori algorithms can solve GCS with reasonable time/space complexities. Bpriori 2 and Bpriori 3-2 algorithm are been proposed of any length and any positions of all their instances in input sequences. In this paper, we reduced the time/space complexity of Bpriori algorithm by Constrained Based Frequent Pattern mining (CBFP) technique which integrates the idea of Constraint Based Mining and FP-tree mining. CBFP mining technique solves the GCS problem works for all center string of any length, but also for the positions of all their mutated copies of input sequence. CBFP mining technique construct TRIE like with FP tree to represent the mutated copies of center string of any length, along with constraints to restraint growth of the consensus tree. The complexity analysis for Constrained Based FP mining technique and Bpriori algorithm is done based on the worst case and average case approach. Algorithm's correctness compared with the Bpriori algorithm using artificial data is shown.Keywords: Constraint Based Mining, FP tree, Data mining, GCS problem, CBFP mining technique.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1702358 Estimation of Hysteretic Damping in Steel Dual Systems with Buckling Restrained Brace and Moment Resisting Frame
Authors: Seyed Saeid Tabaee, Omid Bahar
Abstract:
Nowadays, energy dissipation devices are commonly used in structures. High rate of energy absorption during earthquakes is the benefit of using such devices, which results in damage reduction of structural elements, specifically columns. The hysteretic damping capacity of energy dissipation devices is the key point that it may adversely make analysis and design process complicated. This effect may be generally represented by Equivalent Viscous Damping (EVD). The equivalent viscous damping might be obtained from the expected hysteretic behavior regarding to the design or maximum considered displacement of a structure. In this paper, the hysteretic damping coefficient of a steel Moment Resisting Frame (MRF), which its performance is enhanced by a Buckling Restrained Brace (BRB) system has been evaluated. Having foresight of damping fraction between BRB and MRF is inevitable for seismic design procedures like Direct Displacement-Based Design (DDBD) method. This paper presents an approach to calculate the damping fraction for such systems by carrying out the dynamic nonlinear time history analysis (NTHA) under harmonic loading, which is tuned to the natural system frequency. Two MRF structures, one equipped with BRB and the other without BRB are simultaneously studied. Extensive analysis shows that proportion of each system damping fraction may be calculated by its shear story portion. In this way, contribution of each BRB in the floors and their general contribution in the structural performance may be clearly recognized, in advance.Keywords: Buckling restrained brace, Direct displacement based design, Dual systems, Hysteretic damping, Moment resisting frames.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2475357 PeliGRIFF: A Parallel DEM-DLM/FD Method for DNS of Particulate Flows with Collisions
Authors: Anthony Wachs, Guillaume Vinay, Gilles Ferrer, Jacques Kouakou, Calin Dan, Laurence Girolami
Abstract:
An original Direct Numerical Simulation (DNS) method to tackle the problem of particulate flows at moderate to high concentration and finite Reynolds number is presented. Our method is built on the framework established by Glowinski and his coworkers [1] in the sense that we use their Distributed Lagrange Multiplier/Fictitious Domain (DLM/FD) formulation and their operator-splitting idea but differs in the treatment of particle collisions. The novelty of our contribution relies on replacing the simple artificial repulsive force based collision model usually employed in the literature by an efficient Discrete Element Method (DEM) granular solver. The use of our DEM solver enables us to consider particles of arbitrary shape (at least convex) and to account for actual contacts, in the sense that particles actually touch each other, in contrast with the simple repulsive force based collision model. We recently upgraded our serial code, GRIFF 1 [2], to full MPI capabilities. Our new code, PeliGRIFF 2, is developed under the framework of the full MPI open source platform PELICANS [3]. The new MPI capabilities of PeliGRIFF open new perspectives in the study of particulate flows and significantly increase the number of particles that can be considered in a full DNS approach: O(100000) in 2D and O(10000) in 3D. Results on the 2D/3D sedimentation/fluidization of isometric polygonal/polyedral particles with collisions are presented.
Keywords: Particulate flow, distributed lagrange multiplier/fictitious domain method, discrete element method, polygonal shape, sedimentation, distributed computing, MPI
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2124356 Reducing CO2 Emission Using EDA and Weighted Sum Model in Smart Parking System
Authors: Rahman Ali, Muhammad Sajjad, Farkhund Iqbal, Muhammad Sadiq Hassan Zada, Mohammed Hussain
Abstract:
Emission of Carbon Dioxide (CO2) has adversely affected the environment. One of the major sources of CO2 emission is transportation. In the last few decades, the increase in mobility of people using vehicles has enormously increased the emission of CO2 in the environment. To reduce CO2 emission, sustainable transportation system is required in which smart parking is one of the important measures that need to be established. To contribute to the issue of reducing the amount of CO2 emission, this research proposes a smart parking system. A cloud-based solution is provided to the drivers which automatically searches and recommends the most preferred parking slots. To determine preferences of the parking areas, this methodology exploits a number of unique parking features which ultimately results in the selection of a parking that leads to minimum level of CO2 emission from the current position of the vehicle. To realize the methodology, a scenario-based implementation is considered. During the implementation, a mobile application with GPS signals, vehicles with a number of vehicle features and a list of parking areas with parking features are used by sorting, multi-level filtering, exploratory data analysis (EDA, Analytical Hierarchy Process (AHP)) and weighted sum model (WSM) to rank the parking areas and recommend the drivers with top-k most preferred parking areas. In the EDA process, “2020testcar-2020-03-03”, a freely available dataset is used to estimate CO2 emission of a particular vehicle. To evaluate the system, results of the proposed system are compared with the conventional approach, which reveal that the proposed methodology supersedes the conventional one in reducing the emission of CO2 into the atmosphere.
Keywords: CO2 emission, IoT, EDA, Weighted Sum Model, WSM, regression, smart parking system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 743355 Elliptical Features Extraction Using Eigen Values of Covariance Matrices, Hough Transform and Raster Scan Algorithms
Authors: J. Prakash, K. Rajesh
Abstract:
In this paper, we introduce a new method for elliptical object identification. The proposed method adopts a hybrid scheme which consists of Eigen values of covariance matrices, Circular Hough transform and Bresenham-s raster scan algorithms. In this approach we use the fact that the large Eigen values and small Eigen values of covariance matrices are associated with the major and minor axial lengths of the ellipse. The centre location of the ellipse can be identified using circular Hough transform (CHT). Sparse matrix technique is used to perform CHT. Since sparse matrices squeeze zero elements and contain a small number of nonzero elements they provide an advantage of matrix storage space and computational time. Neighborhood suppression scheme is used to find the valid Hough peaks. The accurate position of circumference pixels is identified using raster scan algorithm which uses the geometrical symmetry property. This method does not require the evaluation of tangents or curvature of edge contours, which are generally very sensitive to noise working conditions. The proposed method has the advantages of small storage, high speed and accuracy in identifying the feature. The new method has been tested on both synthetic and real images. Several experiments have been conducted on various images with considerable background noise to reveal the efficacy and robustness. Experimental results about the accuracy of the proposed method, comparisons with Hough transform and its variants and other tangential based methods are reported.Keywords: Circular Hough transform, covariance matrix, Eigen values, ellipse detection, raster scan algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2641354 Exploring Utility and Intrinsic Value among UAE Arabic Teachers in Integrating M-Learning
Authors: Dina Tareq Ismail, Alexandria A. Proff
Abstract:
The United Arab Emirates (UAE) is a nation seeking to advance in all fields, particularly education. One area of focus for UAE 2021 agenda is to restructure UAE schools and universities by equipping them with highly developed technology. The agenda also advises educational institutions to prepare students with applicable and transferrable Information and Communication Technology (ICT) skills. Despite the emphasis on ICT and computer literacy skills, there exists limited empirical data on the use of M-Learning in the literature. This qualitative study explores the motivation of higher primary Arabic teachers in private schools toward implementing and integrating M-Learning apps in their classrooms. This research employs a phenomenological approach through the use of semistructured interviews with nine purposefully selected Arabic teachers. The data were analyzed using a content analysis via multiple stages of coding: open, axial, and thematic. Findings reveal three primary themes: (1) Arabic teachers with high levels of procedural knowledge in ICT are more motivated to implement M-Learning; (2) Arabic teachers' perceptions of self-efficacy influence their motivation toward implementation of M-Learning; (3) Arabic teachers implement M-Learning when they possess high utility and/or intrinsic value in these applications. These findings indicate a strong need for further training, equipping, and creating buy-in among Arabic teachers to enhance their ICT skills in implementing M-Learning. Further, given the limited availability of M-Learning apps designed for use in the Arabic language on the market, it is imperative that developers consider designing M-Learning tools that Arabic teachers, and Arabic-speaking students, can use and access more readily. This study contributes to closing the knowledge gap on teacher-motivation for implementing M-Learning in their classrooms in the UAE.Keywords: ICT Skills, M-Learning, self-efficacy, teachermotivation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 481353 The Performance of Natural Light by Roof Systems in Cultural Buildings
Authors: Ana Paula Esteves, Diego S. Caetano, Louise L. B. Lomardo
Abstract:
This paper presents an approach to the performance of the natural lighting, when the use of appropriated solar lighting systems on the roof is applied in cultural buildings such as museums and foundations. The roofs, as a part of contact between the building and the external environment, require special attention in projects that aim at energy efficiency, being an important element for the capture of natural light in greater quantity, but also for being the most important point of generation of photovoltaic solar energy, even semitransparent, allowing the partial passage of light. Transparent elements in roofs, as well as superior protection of the building, can also play other roles, such as: meeting the needs of natural light for the accomplishment of the internal tasks, attending to the visual comfort; to bring benefits to the human perception and about the interior experience in a building. When these resources are well dimensioned, they also contribute to the energy efficiency and consequent character of sustainability of the building. Therefore, when properly designed and executed, a roof light system can bring higher quality natural light to the interior of the building, which is related to the human health and well-being dimension. Furthermore, it can meet the technologic, economic and environmental yearnings, making possible the more efficient use of that primordial resource, which is the light of the Sun. The article presents the analysis of buildings that used zenith light systems in search of better lighting performance in museums and foundations: the Solomon R. Guggenheim Museum in the United States, the Iberê Camargo Foundation in Brazil, the Museum of Fine Arts in Castellón in Spain and the Pinacoteca of São Paulo.
Keywords: Natural lighting, roof lighting systems, natural lighting in museums, comfort lighting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1071352 Parallel Pipelined Conjugate Gradient Algorithm on Heterogeneous Platforms
Authors: Sergey Kopysov, Nikita Nedozhogin, Leonid Tonkov
Abstract:
The article presents a parallel iterative solver for large sparse linear systems which can be used on a heterogeneous platform. Traditionally, the problem of solving linear systems do not scale well on cluster containing multiple Central Processing Units (multi-CPUs cluster) or cluster containing multiple Graphics Processing Units (multi-GPUs cluster). For example, most of the attempts to implement the classical conjugate gradient method were at best counted in the same amount of time as the problem was enlarged. The paper proposes the pipelined variant of the conjugate gradient method (PCG), a formulation that is potentially better suited for hybrid CPU/GPU computing since it requires only one synchronization point per one iteration, instead of two for standard CG (Conjugate Gradient). The standard and pipelined CG methods need the vector entries generated by current GPU and other GPUs for matrix-vector product. So the communication between GPUs becomes a major performance bottleneck on miltiGPU cluster. The article presents an approach to minimize the communications between parallel parts of algorithms. Additionally, computation and communication can be overlapped to reduce the impact of data exchange. Using pipelined version of the CG method with one synchronization point, the possibility of asynchronous calculations and communications, load balancing between the CPU and GPU for solving the large linear systems allows for scalability. The algorithm is implemented with the combined use of technologies: MPI, OpenMP and CUDA. We show that almost optimum speed up on 8-CPU/2GPU may be reached (relatively to a one GPU execution). The parallelized solver achieves a speedup of up to 5.49 times on 16 NVIDIA Tesla GPUs, as compared to one GPU.
Keywords: Conjugate Gradient, GPU, parallel programming, pipelined algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 371351 A Content Analysis of Sustainability Reporting to Frame the Heterogeneity in Corporate Environment Sustainability Practices
Authors: Venkataraman Sankaranarayanan, Sougata Ray
Abstract:
While extant research has examined many aspects of differential corporate environmental outcomes and behavior, a holistic and integrated view of heterogeneity in corporate environment sustainability (CES) practices remains a puzzle to be fully unraveled – its extent and nature, its relationship to macro or micro level influences, or strategic orientations. Such a perspective would be meaningful for the field given notable strides in CES practices and the corporate social responsibility agenda over the last two decades, in the backdrop of altered global socio-political sensitivities and technological advances. To partly address this gap, this exploratory research adopted a content analysis approach to code patterns in the sustainability disclosures of the 160 largest global firms spread over 8 years. The sample of firms spanned seven industries, nine countries and three continents thereby presenting data rich and diverse enough in several dimensions to be representative of global heterogeneity in CES practices. Through a factor analysis of the coded data, four strategic CES orientations were extracted through the analysis, that effectively straddles most of the variation observed in current CES practices – one that seeks to reduce environmental damage on account of the firm’s operations, another that prioritizes minimalism, a third that focuses on broader ecological status quo, and a final one that champions the ‘business of green’, extending the CES agenda beyond the firm’s boundaries. These environment sustainability strategy orientations are further examined to elicit prominent patterns and explore plausible antecedents.Keywords: Corporate sustainability, corporate social responsibility, corporate environmental management, environmental strategy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1544350 Changing Roles and Skills of Urban Planners in the Turkish Planning System
Authors: Fatih Eren
Abstract:
This research aims to find an answer to the question of which knowledge and skills do the Turkish urban planners need in their business practice. Understanding change in cities, making a prediction, making an urban decision and putting it into practice, working together with actors from different organizations from various academic disciplines, persuading people to accept something and developing good personal and professional relationships have become very complex and difficult in today’s world. The truth is that urban planners work in many institutions under various positions which are not similar to each other by field of activity and all planners are forced to develop some knowledge and skills for success in their business in Turkey. This study targets to explore what urban planners do in the global information age. The study is the product of a comprehensive nation-wide research. In-depth interviews were conducted with 174 experienced urban planners, who work in different public institutions and private companies under varied positions in the Turkish Planning System, to find out knowledge and skills needed by next-generation urban planners. The main characteristics of next-generation urban planners are defined; skills that planners needed today are explored in this paper. Findings show that the positivist (traditional) planning approach has given place to anti-positivist planning approaches in the Turkish Planning System so next-generation urban planners who seek success and want to carve out a niche for themselves in business life have to equip themselves with innovative skills. The result section also includes useful and instructive findings for planners about what is the meaning of being an urban planner and what is the ideal content and context of planning education at universities in the global age.Keywords: The global information age, urban planners, innovative job skills, planning education.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1259349 A Zero-Cost Collar Option Applied to Materials Procurement Contracts to Reduce Price Fluctuation Risks in Construction
Authors: H. L. Yim, S. H. Lee, S. K. Yoo, J. J. Kim
Abstract:
This study proposes a materials procurement contracts model to which the zero-cost collar option is applied for heading price fluctuation risks in construction.The material contract model based on the collar option that consists of the call option striking zone of the construction company(the buyer) following the materials price increase andthe put option striking zone of the material vendor(the supplier) following a materials price decrease. This study first determined the call option strike price Xc of the construction company by a simple approach: it uses the predicted profit at the project starting point and then determines the strike price of put option Xp that has an identical option value, which completes the zero-cost material contract.The analysis results indicate that the cost saving of the construction company increased as Xc decreased. This was because the critical level of the steel materials price increasewas set at a low level. However, as Xc decreased, Xpof a put option that had an identical option value gradually increased. Cost saving increased as Xc decreased. However, as Xp gradually increased, the risk of loss from a construction company increased as the steel materials price decreased. Meanwhile, cost saving did not occur for the construction company, because of volatility. This result originated in the zero-cost features of the two-way contract of the collar option. In the case of the regular one-way option, the transaction cost had to be subtracted from the cost saving. The transaction cost originated from an option value that fluctuated with the volatility. That is, the cost saving of the one-way option was affected by the volatility. Meanwhile, even though the collar option with zero transaction cost cut the connection between volatility and cost saving, there was a risk of exercising the put option.Keywords: Construction materials, Supply chain management, Procurement, Payment, Collar option
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2522348 Building Information Modelling for Construction Delay Management
Authors: Essa Alenazi, Zulfikar Adamu
Abstract:
The Kingdom of Saudi Arabia (KSA) is not an exception in relying on the growth of its construction industry to support rapid population growth. However, its need for infrastructure development is constrained by low productivity levels and cost overruns caused by factors such as delays to project completion. Delays in delivering a construction project are a global issue and while theories such as Optimism Bias have been used to explain such delays, in KSA, client-related causes of delays are also significant. The objective of this paper is to develop a framework-based approach to explore how the country’s construction industry can manage and reduce delays in construction projects through building information modelling (BIM) in order to mitigate the cost consequences of such delays. It comprehensively and systematically reviewed the global literature on the subject and identified gaps, critical delay factors and the specific benefits that BIM can deliver for the delay management. A case study comprising of nine hospital projects that have experienced delay and cost overruns was also carried out. Five critical delay factors related to the clients were identified as candidates that can be mitigated through BIM’s benefits. These factors are: Ineffective planning and scheduling of the project; changes during construction by the client; delay in progress payment; slowness in decision making by the client; and poor communication between clients and other stakeholders. In addition, data from the case study projects strongly suggest that optimism bias is present in many of the hospital projects. Further validation via key stakeholder interviews and documentations are planned.
Keywords: BIM, client perspective, delay management, optimism bias, public sector projects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2467347 Evaluation of Easy-to-Use Energy Building Design Tools for Solar Access Analysis in Urban Contexts: Comparison of Friendly Simulation Design Tools for Architectural Practice in the Early Design Stage
Abstract:
Current building sector is focused on reduction of energy requirements, on renewable energy generation and on regeneration of existing urban areas. These targets need to be solved with a systemic approach, considering several aspects simultaneously such as climate conditions, lighting conditions, solar radiation, PV potential, etc. The solar access analysis is an already known method to analyze the solar potentials, but in current years, simulation tools have provided more effective opportunities to perform this type of analysis, in particular in the early design stage. Nowadays, the study of the solar access is related to the easiness of the use of simulation tools, in rapid and easy way, during the design process. This study presents a comparison of three simulation tools, from the point of view of the user, with the aim to highlight differences in the easy-to-use of these tools. Using a real urban context as case study, three tools; Ecotect, Townscope and Heliodon, are tested, performing models and simulations and examining the capabilities and output results of solar access analysis. The evaluation of the ease-to-use of these tools is based on some detected parameters and features, such as the types of simulation, requirements of input data, types of results, etc. As a result, a framework is provided in which features and capabilities of each tool are shown. This framework shows the differences among these tools about functions, features and capabilities. The aim of this study is to support users and to improve the integration of simulation tools for solar access with the design process.
Keywords: Solar access analysis, energy building design tools, urban planning, solar potential.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2068346 Large Eddy Simulation of Hydrogen Deflagration in Open Space and Vented Enclosure
Authors: T. Nozu, K. Hibi, T. Nishiie
Abstract:
This paper discusses the applicability of the numerical model for a damage prediction method of the accidental hydrogen explosion occurring in a hydrogen facility. The numerical model was based on an unstructured finite volume method (FVM) code “NuFD/FrontFlowRed”. For simulating unsteady turbulent combustion of leaked hydrogen gas, a combination of Large Eddy Simulation (LES) and a combustion model were used. The combustion model was based on a two scalar flamelet approach, where a G-equation model and a conserved scalar model expressed a propagation of premixed flame surface and a diffusion combustion process, respectively. For validation of this numerical model, we have simulated the previous two types of hydrogen explosion tests. One is open-space explosion test, and the source was a prismatic 5.27 m3 volume with 30% of hydrogen-air mixture. A reinforced concrete wall was set 4 m away from the front surface of the source. The source was ignited at the bottom center by a spark. The other is vented enclosure explosion test, and the chamber was 4.6 m × 4.6 m × 3.0 m with a vent opening on one side. Vent area of 5.4 m2 was used. Test was performed with ignition at the center of the wall opposite the vent. Hydrogen-air mixtures with hydrogen concentrations close to 18% vol. were used in the tests. The results from the numerical simulations are compared with the previous experimental data for the accuracy of the numerical model, and we have verified that the simulated overpressures and flame time-of-arrival data were in good agreement with the results of the previous two explosion tests.
Keywords: Deflagration, Large Eddy Simulation, Turbulent combustion, Vented enclosure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1477345 Thermo-mechanical Deformation Behavior of Functionally Graded Rectangular Plates Subjected to Various Boundary Conditions and Loadings
Authors: Mohammad Talha, B. N. Singh
Abstract:
This paper deals with the thermo-mechanical deformation behavior of shear deformable functionally graded ceramicmetal (FGM) plates. Theoretical formulations are based on higher order shear deformation theory with a considerable amendment in the transverse displacement using finite element method (FEM). The mechanical properties of the plate are assumed to be temperaturedependent and graded in the thickness direction according to a powerlaw distribution in terms of the volume fractions of the constituents. The temperature field is supposed to be a uniform distribution over the plate surface (XY plane) and varied in the thickness direction only. The fundamental equations for the FGM plates are obtained using variational approach by considering traction free boundary conditions on the top and bottom faces of the plate. A C0 continuous isoparametric Lagrangian finite element with thirteen degrees of freedom per node have been employed to accomplish the results. Convergence and comparison studies have been performed to demonstrate the efficiency of the present model. The numerical results are obtained for different thickness ratios, aspect ratios, volume fraction index and temperature rise with different loading and boundary conditions. Numerical results for the FGM plates are provided in dimensionless tabular and graphical forms. The results proclaim that the temperature field and the gradient in the material properties have significant role on the thermo-mechanical deformation behavior of the FGM plates.
Keywords: Functionally graded material, higher order shear deformation theory, finite element method, independent field variables.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2334344 A Two-Step, Temperature-Staged Direct Coal Liquefaction Process
Authors: Reyna Singh, David Lokhat, Milan Carsky
Abstract:
The world crude oil demand is projected to rise to 108.5 million bbl/d by the year 2035. With reserves estimated at 869 billion tonnes worldwide, coal remains an abundant resource. The aim of this work was to produce a high value hydrocarbon liquid product using a Direct Coal Liquefaction (DCL) process at, relatively mild operating conditions. Via hydrogenation, the temperature-staged approach was investigated in a dual reactor lab-scale pilot plant facility. The objectives included maximising thermal dissolution of the coal in the presence of tetralin as the hydrogen donor solvent in the first stage with 2:1 and 3:1 solvent: coal ratios. Subsequently, in the second stage, hydrogen saturation, in particular, hydrodesulphurization (HDS) performance was assessed. Two commercial hydrotreating catalysts were investigated viz. NickelMolybdenum (Ni-Mo) and Cobalt-Molybdenum (Co-Mo). GC-MS results identified 77 compounds and various functional groups present in the first and second stage liquid product. In the first stage 3:1 ratios and liquid product yields catalysed by magnetite were favoured. The second stage product distribution showed an increase in the BTX (Benzene, Toluene, Xylene) quality of the liquid product, branched chain alkanes and a reduction in the sulphur concentration. As an HDS performer and selectivity to the production of long and branched chain alkanes, Ni-Mo had an improved performance over Co-Mo. Co-Mo is selective to a higher concentration of cyclohexane. For 16 days on stream each, Ni-Mo had a higher activity than Co-Mo. The potential to cover the demand for low–sulphur, crude diesel and solvents from the production of high value hydrocarbon liquid in the said process, is thus demonstrated.
Keywords: Catalyst, coal, liquefaction, temperature-staged.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1641343 Reducing Defects through Organizational Learning within a Housing Association Environment
Authors: T. Hopkin, S. Lu, P. Rogers, M. Sexton
Abstract:
Housing Associations (HAs) contribute circa 20% of the UK’s housing supply. HAs are however under increasing pressure as a result of funding cuts and rent reductions. Due to the increased pressure, a number of processes are currently being reviewed by HAs, especially how they manage and learn from defects. Learning from defects is considered a useful approach to achieving defect reduction within the UK housebuilding industry. This paper contributes to our understanding of how HAs learn from defects by undertaking an initial round table discussion with key HA stakeholders as part of an ongoing collaborative research project with the National House Building Council (NHBC) to better understand how house builders and HAs learn from defects to reduce their prevalence. The initial discussion shows that defect information runs through a number of groups, both internal and external of a HA during both the defects management process and organizational learning (OL) process. Furthermore, HAs are reliant on capturing and recording defect data as the foundation for the OL process. During the OL process defect data analysis is the primary enabler to recognizing a need for a change to organizational routines. When a need for change has been recognized, new options are typically pursued to design out defects via updates to a HAs Employer’s Requirements. Proposed solutions are selected by a review board and committed to organizational routine. After implementing a change, both structured and unstructured feedback is sought to establish the change’s success. The findings from the HA discussion demonstrates that OL can achieve defect reduction within the house building sector in the UK. The paper concludes by outlining a potential ‘learning from defects model’ for the housebuilding industry as well as describing future work.
Keywords: Defects, new homes, housing associations, organizational learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897342 Land Suitability Prediction Modelling for Agricultural Crops Using Machine Learning Approach: A Case Study of Khuzestan Province, Iran
Authors: Saba Gachpaz, Hamid Reza Heidari
Abstract:
The sharp increase in population growth leads to more pressure on agricultural areas to satisfy the food supply. This necessitates increased resource consumption and underscores the importance of addressing sustainable agriculture development along with other environmental considerations. Land-use management is a crucial factor in obtaining optimum productivity. Machine learning is a widely used technique in the agricultural sector, from yield prediction to customer behavior. This method focuses on learning and provides patterns and correlations from our data set. In this study, nine physical control factors, namely, soil classification, electrical conductivity, normalized difference water index (NDWI), groundwater level, elevation, annual precipitation, pH of water, annual mean temperature, and slope in the alluvial plain in Khuzestan (an agricultural hotspot in Iran) are used to decide the best agricultural land use for both rainfed and irrigated agriculture for 10 different crops. For this purpose, each variable was imported into Arc GIS, and a raster layer was obtained. In the next level, by using training samples, all layers were imported into the python environment. A random forest model was applied, and the weight of each variable was specified. In the final step, results were visualized using a digital elevation model, and the importance of all factors for each one of the crops was obtained. Our results show that despite 62% of the study area being allocated to agricultural purposes, only 42.9% of these areas can be defined as a suitable class for cultivation purposes.
Keywords: Land suitability, machine learning, random forest, sustainable agriculture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 283341 Evaluation of Ensemble Classifiers for Intrusion Detection
Authors: M. Govindarajan
Abstract:
One of the major developments in machine learning in the past decade is the ensemble method, which finds highly accurate classifier by combining many moderately accurate component classifiers. In this research work, new ensemble classification methods are proposed with homogeneous ensemble classifier using bagging and heterogeneous ensemble classifier using arcing and their performances are analyzed in terms of accuracy. A Classifier ensemble is designed using Radial Basis Function (RBF) and Support Vector Machine (SVM) as base classifiers. The feasibility and the benefits of the proposed approaches are demonstrated by the means of standard datasets of intrusion detection. The main originality of the proposed approach is based on three main parts: preprocessing phase, classification phase, and combining phase. A wide range of comparative experiments is conducted for standard datasets of intrusion detection. The performance of the proposed homogeneous and heterogeneous ensemble classifiers are compared to the performance of other standard homogeneous and heterogeneous ensemble methods. The standard homogeneous ensemble methods include Error correcting output codes, Dagging and heterogeneous ensemble methods include majority voting, stacking. The proposed ensemble methods provide significant improvement of accuracy compared to individual classifiers and the proposed bagged RBF and SVM performs significantly better than ECOC and Dagging and the proposed hybrid RBF-SVM performs significantly better than voting and stacking. Also heterogeneous models exhibit better results than homogeneous models for standard datasets of intrusion detection.Keywords: Data mining, ensemble, radial basis function, support vector machine, accuracy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1701340 Comparative Review of Modulation Techniques for Harmonic Minimization in Multilevel Inverter
Authors: M. Suresh Kumar, K. Ramani
Abstract:
This paper proposed the comparison made between Multi-Carrier Pulse Width Modulation, Sinusoidal Pulse Width Modulation and Selective Harmonic Elimination Pulse Width Modulation technique for minimization of Total Harmonic Distortion in Cascaded H-Bridge Multi-Level Inverter. In Multicarrier Pulse Width Modulation method by using Alternate Position of Disposition scheme for switching pulse generation to Multi-Level Inverter. Another carrier based approach; Sinusoidal Pulse Width Modulation method is also implemented to define the switching pulse generation system in the multi-level inverter. In Selective Harmonic Elimination method using Genetic Algorithm and Particle Swarm Optimization algorithm for define the required switching angles to eliminate low order harmonics from the inverter output voltage waveform and reduce the total harmonic distortion value. So, the results validate that the Selective Harmonic Elimination Pulse Width Modulation method does capably eliminate a great number of precise harmonics and minimize the Total Harmonic Distortion value in output voltage waveform in compared with Multi-Carrier Pulse Width Modulation method, Sinusoidal Pulse Width Modulation method. In this paper, comparison of simulation results shows that the Selective Harmonic Elimination method can attain optimal harmonic minimization solution better than Multi-Carrier Pulse Width Modulation method, Sinusoidal Pulse Width Modulation method.Keywords: Multi-level inverter, Selective Harmonic Elimination Pulse Width Modulation, Multi-Carrier Pulse Width Modulation, Total Harmonic Distortion, Genetic Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2975339 Disparities versus Similarities: WHO GPPQCL and ISO/IEC 17025:2017 International Standards for Quality Management Systems in Pharmaceutical Laboratories
Authors: M. A. Okezue, K. L. Clase, S. R. Byrn, P. Shivanand
Abstract:
Medicines regulatory authorities expect pharmaceutical companies and contract research organizations to seek ways to certify that their laboratory control measurements are reliable. Establishing and maintaining laboratory quality standards are essential in ensuring the accuracy of test results. ‘ISO/IEC 17025:2017’ and ‘WHO Good Practices for Pharmaceutical Quality Control Laboratories (GPPQCL)’ are two quality standards commonly employed in developing laboratory quality systems. A review was conducted on the two standards to elaborate on areas on convergence and divergence. The goal was to understand how differences in each standard's requirements may influence laboratories' choices as to which document is easier to adopt for quality systems. A qualitative review method compared similar items in the two standards while mapping out areas where there were specific differences in the requirements of the two documents. The review also provided a detailed description of the clauses and parts covering management and technical requirements in these laboratory standards. The review showed that both documents share requirements for over ten critical areas covering objectives, infrastructure, management systems, and laboratory processes. There were, however, differences in standard expectations where GPPQCL emphasizes system procedures for planning and future budgets that will ensure continuity. Conversely, ISO 17025 was more focused on the risk management approach to establish laboratory quality systems. Elements in the two documents form common standard requirements to assure the validity of laboratory test results that promote mutual recognition. The ISO standard currently has more global patronage than GPPQCL.
Keywords: ISO/IEC 17025:2017, laboratory standards, quality control, WHO GPPQCL
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1124338 Improvement of Overall Equipment Effectiveness through Total Productive Maintenance
Abstract:
Frequent machine breakdowns, low plant availability and increased overtime are a great threat to a manufacturing plant as they increase operating costs of an industry. The main aim of this study was to improve Overall Equipment Effectiveness (OEE) at a manufacturing company through the implementation of innovative maintenance strategies. A case study approach was used. The paper focuses on improving the maintenance in a manufacturing set up using an innovative maintenance regime mix to improve overall equipment effectiveness. Interviews, reviewing documentation and historical records, direct and participatory observation were used as data collection methods during the research. Usually production is based on the total kilowatt of motors produced per day. The target kilowatt at 91% availability is 75 Kilowatts a day. Reduced demand and lack of raw materials particularly imported items are adversely affecting the manufacturing operations. The company had to reset its targets from the usual figure of 250 Kilowatt per day to mere 75 per day due to lower availability of machines as result of breakdowns as well as lack of raw materials. The price reductions and uncertainties as well as general machine breakdowns further lowered production. Some recommendations were given. For instance, employee empowerment in the company will enhance responsibility and authority to improve and totally eliminate the six big losses. If the maintenance department is to realise its proper function in a progressive, innovative industrial society, then its personnel must be continuously trained to meet current needs as well as future requirements. To make the maintenance planning system effective, it is essential to keep track of all the corrective maintenance jobs and preventive maintenance inspections. For large processing plants these cannot be handled manually. It was therefore recommended that the company implement (Computerised Maintenance Management System) CMMS.
Keywords: Maintenance, Manufacturing, Overall Equipment Effectiveness
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3988337 Performances Assessment of Direct Torque Controlled IM Drives Using Fuzzy Logic Control and Space Vector Modulation Strategy
Authors: L. Moussaoui, L. Rahmani
Abstract:
This paper deals with the direct torque control (DTC) of the induction motor. This type of control allows decoupling control between the flux and the torque without the need for a transformation of coordinates. However, as with other hysteresis-based systems, the classical DTC scheme represents a high ripple, in both the electromagnetic torque and the stator flux and a distortion in the stator current. As well, it suffers from variable switching frequency. To solve these problems various modifications, in conventional DTC scheme, have been made during the last decade. Indeed the DTC based on space vector modulation (SVM) has proved to generate very low ripples in torque and flux with constant switching frequency. It also shows almost the same dynamic performances as the classical DTC system. On the other hand, fuzzy logic is considered as an interesting alternative approach for its advantages: Analysis close to the exigencies of user, ability of nonlinear systems control, best dynamic performances and inherent quality of robustness.
Therefore, two fuzzy direct torque control approaches, for the induction motor fed by SVM-voltage source inverter, are proposed in this paper. By using these two approaches of DTC, the advantages of fuzzy logic control, space vector modulation, and direct torque control method are combined. The performances of these DTC schemes are evaluated through digital simulation using Matlab/Simulink platform and fuzzy logic tools. Simulation results illustrate the effectiveness and the superiority of the proposed Fuzzy DTC-SVM schemes in comparison to the classical DTC.
Keywords: Direct torque control, Fuzzy logic control, Induction motor, Switching frequency, Space vector modulation, Torque and flux ripples.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2398336 Novel Adaptive Channel Equalization Algorithms by Statistical Sampling
Authors: János Levendovszky, András Oláh
Abstract:
In this paper, novel statistical sampling based equalization techniques and CNN based detection are proposed to increase the spectral efficiency of multiuser communication systems over fading channels. Multiuser communication combined with selective fading can result in interferences which severely deteriorate the quality of service in wireless data transmission (e.g. CDMA in mobile communication). The paper introduces new equalization methods to combat interferences by minimizing the Bit Error Rate (BER) as a function of the equalizer coefficients. This provides higher performance than the traditional Minimum Mean Square Error equalization. Since the calculation of BER as a function of the equalizer coefficients is of exponential complexity, statistical sampling methods are proposed to approximate the gradient which yields fast equalization and superior performance to the traditional algorithms. Efficient estimation of the gradient is achieved by using stratified sampling and the Li-Silvester bounds. A simple mechanism is derived to identify the dominant samples in real-time, for the sake of efficient estimation. The equalizer weights are adapted recursively by minimizing the estimated BER. The near-optimal performance of the new algorithms is also demonstrated by extensive simulations. The paper has also developed a (Cellular Neural Network) CNN based approach to detection. In this case fast quadratic optimization has been carried out by t, whereas the task of equalizer is to ensure the required template structure (sparseness) for the CNN. The performance of the method has also been analyzed by simulations.
Keywords: Cellular Neural Network, channel equalization, communication over fading channels, multiuser communication, spectral efficiency, statistical sampling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1520