Search results for: Alara Stream
72 The Effect of Blockage Factor on Savonius Hydrokinetic Turbine Performance
Authors: Thochi Seb Rengma, Mahendra Kumar Gupta, P. M. V. Subbarao
Abstract:
Hydrokinetic turbines can be used to produce power in inaccessible villages located near rivers. The hydrokinetic turbine uses the kinetic energy of the water and maybe put it directly into the natural flow of water without dams. For off-grid power production, the Savonius-type vertical axis turbine is the easiest to design and manufacture. This proposal uses three-dimensional Computational Fluid Dynamics (CFD) simulations to measure the considerable interaction and complexity of turbine blades. Savonius hydrokinetic turbine (SHKT) performance is affected by a blockage in the river, canals, and waterways. Putting a large object in a water channel causes water obstruction and raises local free stream velocity. The blockage correction factor or velocity increment measures the impact of velocity on the performance. SHKT performance is evaluated by comparing power coefficient (Cp) with tip-speed ratio (TSR) at various blockage ratios. The maximum Cp was obtained at a TSR of 1.1 with a blockage ratio of 45%, whereas TSR of 0.8 yielded the highest Cp without blockage. The greatest Cp of 0.29 was obtained with a 45% blockage ratio compared to a Cp max of 0.18 without a blockage.
Keywords: Savonius hydrokinetic turbine, blockage ratio, vertical axis turbine, power coefficient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17271 Detached-Eddy Simulation of Vortex Generator Jet Using Chimera Grids
Authors: Saqib Mahmood, Rolf Radespiel
Abstract:
This paper aims at numerically analysing the effect of an active flow control (AFC) by a vortex generator jet (VGJ) submerged in a boundary layer via Chimera Grids and Detached- Eddy Simulation (DES). The performance of DES results are judged against Reynolds-Averaged Navier-Stokes (RANS) and compared with the experiments that showed an unsteady vortex motion downstream of VGJ. Experimental results showed that the mechanism of embedding logitudinal vortex structure in the main stream flow is quite effective in increasing the near wall momentum of separated aircraft wing. In order to simulate such a flow configuration together with the VGJ, an efficient numerical approach is required. This requirement is fulfilled by performing the DES simulation over the flat plate using the DLR TAU Code. The DES predictions identify the vortex region via smooth hybrid length scale and predict the unsteady vortex motion observed in the experiments. The DES results also showed that the sufficient grid refinement in the vortex region resolves the turbulent scales downstream of the VGJ, the spatial vortex core postion and nondimensional momentum coefficient RVx .Keywords: VGJ, Chimera Grid, DES, RANS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 248170 The Dependence of the Liquid Application on the Coverage of the Sprayed Objects in Terms of the Characteristics of the Sprayed Object during Spraying
Authors: Beata Cieniawska, Deta Łuczycka, Katarzyna Dereń
Abstract:
When assessing the quality of the spraying procedure, three indicators are used: uneven distribution of precipitation of liquid sprayed, degree of coverage of sprayed surfaces, and deposition of liquid spraying However, there is a lack of information on the relationship between the quality parameters of the procedure. Therefore, the research was carried out at the Institute of Agricultural Engineering of Wrocław University of Environmental and Life Sciences. The aim of the study was to determine the relationship between the degree of coverage of sprayed surfaces and the deposition of liquid in the aspect of the parametric characteristics of the protected plant using selected single and double stream nozzles. Experiments were conducted under laboratory conditions. The carrier of nozzles acted as an independent self-propelled sprayer used for spraying, whereas the parametric characteristics of plants were determined using artificial plants as the ratio of the vertical projection surface and the horizontal projection surface. The results and their analysis showed a strong and very strong correlation between the analyzed parameters in terms of the characteristics of the sprayed object.
Keywords: Degree of coverage, deposition of liquid, nozzle, spraying.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 76569 Estimating Marine Tidal Power Potential in Kenya
Authors: Lucy Patricia Onundo, Wilfred Njoroge Mwema
Abstract:
The rapidly diminishing fossil fuel reserves, their exorbitant cost and the increasingly apparent negative effect of fossil fuels to climate changes is a wake-up call to explore renewable energy. Wind, bio-fuel and solar power have already become staples of Kenyan electricity mix. The potential of electric power generation from marine tidal currents is enormous, with oceans covering more than 70% of the earth. However, attempts to harness marine tidal energy in Kenya, has yet to be studied thoroughly due to its promising, cyclic, reliable and predictable nature and the vast energy contained within it. The high load factors resulting from the fluid properties and the predictable resource characteristics make marine currents particularly attractive for power generation and advantageous when compared to others. Global-level resource assessments and oceanographic literature and data have been compiled in an analysis of the technology-specific requirements for tidal energy technologies and the physical resources. Temporal variations in resource intensity as well as the differences between small-scale applications are considered.
Keywords: Energy data assessment, environmental legislation, renewable energy, tidal-in-stream turbines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 136668 Rainfall–Runoff Simulation Using WetSpa Model in Golestan Dam Basin, Iran
Authors: M. R. Dahmardeh Ghaleno, M. Nohtani, S. Khaledi
Abstract:
Flood simulation and prediction is one of the most active research areas in surface water management. WetSpa is a distributed, continuous, and physical model with daily or hourly time step that explains precipitation, runoff, and evapotranspiration processes for both simple and complex contexts. This model uses a modified rational method for runoff calculation. In this model, runoff is routed along the flow path using Diffusion-Wave equation which depends on the slope, velocity, and flow route characteristics. Golestan Dam Basin is located in Golestan province in Iran and it is passing over coordinates 55° 16´ 50" to 56° 4´ 25" E and 37° 19´ 39" to 37° 49´ 28"N. The area of the catchment is about 224 km2, and elevations in the catchment range from 414 to 2856 m at the outlet, with average slope of 29.78%. Results of the simulations show a good agreement between calculated and measured hydrographs at the outlet of the basin. Drawing upon Nash-Sutcliffe model efficiency coefficient for calibration periodic model estimated daily hydrographs and maximum flow rate with an accuracy up to 59% and 80.18%, respectively.
Keywords: Watershed simulation, WetSpa, stream flow, flood prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 103267 High Level Synthesis of Canny Edge Detection Algorithm on Zynq Platform
Authors: Hanaa M. Abdelgawad, Mona Safar, Ayman M. Wahba
Abstract:
Real time image and video processing is a demand in many computer vision applications, e.g. video surveillance, traffic management and medical imaging. The processing of those video applications requires high computational power. Thus, the optimal solution is the collaboration of CPU and hardware accelerators. In this paper, a Canny edge detection hardware accelerator is proposed. Edge detection is one of the basic building blocks of video and image processing applications. It is a common block in the pre-processing phase of image and video processing pipeline. Our presented approach targets offloading the Canny edge detection algorithm from processing system (PS) to programmable logic (PL) taking the advantage of High Level Synthesis (HLS) tool flow to accelerate the implementation on Zynq platform. The resulting implementation enables up to a 100x performance improvement through hardware acceleration. The CPU utilization drops down and the frame rate jumps to 60 fps of 1080p full HD input video stream.
Keywords: High Level Synthesis, Canny edge detection, Hardware accelerators, and Computer Vision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 543166 Examination of Flood Runoff Reproductivity for Different Rainfall Sources in Central Vietnam
Authors: Do Hoai Nam, Keiko Udo, Akira Mano
Abstract:
This paper presents the combination of different precipitation data sets and the distributed hydrological model, in order to examine the flood runoff reproductivity of scattered observation catchments. The precipitation data sets were obtained from observation using rain-gages, satellite based estimate (TRMM), and numerical weather prediction model (NWP), then were coupled with the super tank model. The case study was conducted in three basins (small, medium, and large size) located in Central Vietnam. Calculated hydrographs based on ground observation rainfall showed best fit to measured stream flow, while those obtained from TRMM and NWP showed high uncertainty of peak discharges. However, calculated hydrographs using the adjusted rainfield depicted a promising alternative for the application of TRMM and NWP in flood modeling for scattered observation catchments, especially for the extension of forecast lead time.
Keywords: Flood forecast, rainfall-runoff model, satellite rainfall estimate, numerical weather prediction, quantitative precipitation forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 160965 Modeling and Parametric Study for CO2/CH4 Separation Using Membrane Processes
Authors: Faizan Ahmad, Lau Kok Keong, Azmi Mohd. Shariff
Abstract:
The upgrading of low quality crude natural gas (NG) is attracting interest due to high demand of pipeline-grade gas in recent years. Membrane processes are commercially proven technology for the removal of impurities like carbon dioxide from NG. In this work, cross flow mathematical model has been suggested to be incorporated with ASPEN HYSYS as a user defined unit operation in order to design the membrane system for CO2/CH4 separation. The effect of operating conditions (such as feed composition and pressure) and membrane selectivity on the design parameters (methane recovery and total membrane area required for the separation) has been studied for different design configurations. These configurations include single stage (with and without recycle) and double stage membrane systems (with and without permeate or retentate recycle). It is shown that methane recovery can be improved by recycling permeate or retentate stream as well as by using double stage membrane systems. The ASPEN HYSYS user defined unit operation proposed in the study has potential to be applied for complex membrane system design and optimization.
Keywords: CO2/CH4 Separation, Membrane Process, Membrane modeling, Natural Gas Processing
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 385864 Water Boundary Layer Flow Over Rotating Sphere with Mass Transfer
Authors: G. Revathi, P. Saikrishnan
Abstract:
An analysis is performed to study the influence of nonuniform double slot suction on a steady laminar boundary layer flow over a rotating sphere when fluid properties such as viscosity and Prandtl number are inverse linear functions of temperature. Nonsimilar solutions have been obtained from the starting point of the streamwise co-ordinate to the exact point of separation. The difficulties arising at the starting point of the streamwise co-ordinate, at the edges of the slot and at the point of separation have been overcome by applying an implicit finite difference scheme in combination with the quasi-linearization technique and an appropriate selection of the finer step sizes along the stream-wise direction. The present investigation shows that the point of ordinary separation can be delayed by nonuniform double slot suction if the mass transfer rate is increased and also if the slots are positioned further downstream. In addition, the investigation reveals that double slot suction is found to be more effective compared to a single slot suction in delaying ordinary separation. As rotation parameter increase the point of separation moves upstream direction.
Keywords: Boundary layer, suction, mass transfer, rotating sphere.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 637463 Treatment of Wastewater from Wet Scrubbers in Secondary Lead Smelters for Recycling and Lead Recovery
Authors: Mahmoud A. Rabah
Abstract:
The present study shows a method to recover lead metal from wastewater of wet scrubber in secondary lead smelter. The wastewater is loaded with 42,000 ppm of insoluble lead compounds (TSP) submicron in diameter. The technical background benefits the use of cationic polyfloc solution to flocculate these colloidal solids before press filtration. The polymer solution is injected in the wastewater stream in a countercurrent flow design. The study demonstrates the effect of polymer dose, temperature, pH, flow velocity of the wastewater and different filtration media on the filtration extent. Results indicated that filtration rate (¦r), quality of purified water, purifying efficiency (¦e) and floc diameter decrease regularly with increase in mass flow rate and velocity up to turbulence of 0.5 m.sec-1. Laminar flow is in favor of flocculation. Polyfloc concentration of 0.75 – 1.25 g/m3 wastewater is convenient. Increasing temperature of the wastewater and pneumatic pressure of filtration enhances ¦r. High pH value deforms floc formation and assists degradation of the filtration fabric. The overall efficiency of the method amounts to 93.2 %. Lead metal was recovered from the filtrate cake using carbon as a reducing agent at 900°C.
Keywords: Wastewater, wet scrubbers, filtration, secondary lead.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 344362 Innovation in Lean Thinking to Achieve Rapid Construction
Authors: Muhamad Azani Yahya, Vikneswaran Munikanan, Mohammed Alias Yusof
Abstract:
Lean thinking holds the potential for improving the construction sector, and therefore, it is a concept that should be adopted by construction sector players and academicians in the real industry. Bridging from that, a learning process for construction sector players regarding this matter should be the agenda in gaining the knowledge in preparation for their career. Lean principles offer opportunities for reducing lead times, eliminating non-value adding activities, reducing variability, and are facilitated by methods such as pull scheduling, simplified operations and buffer reduction. Thus, the drive for rapid construction, which is a systematic approach in enhancing efficiency to deliver a project using time reduction, while lean is the continuous process of eliminating waste, meeting or exceeding all customer requirements, focusing on the entire value stream and pursuing perfection in the execution of a constructed project. The methodology presented is shown to be valid through literature, interviews and questionnaire. The results show that the majority of construction sector players unfamiliar with lean thinking and they agreed that it can improve the construction process flow. With this background knowledge established and identified, best practices and recommended action are drawn.
Keywords: Construction improvement, rapid construction, time reduction, lean construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 130661 Mooring Analysis of Duct-Type Tidal Current Power System in Shallow Water
Authors: Chul H. Jo, Do Y. Kim, Bong K. Cho, Myeong J. Kim
Abstract:
The exhaustion of oil and the environmental pollution from the use of fossil fuel are increasing. Tidal current power (TCP) has been proposed as an alternative energy source because of its predictability and reliability. By applying a duct and single point mooring (SPM) system, a TCP device can amplify the generating power and keep its position properly. Because the generating power is proportional to cube of the current stream velocity, amplifying the current speed by applying a duct to a TCP system is an effective way to improve the efficiency of the power device. An SPM system can be applied at any water depth and is highly cost effective. Simple installation and maintenance procedures are also merits of an SPM system. In this study, we designed an SPM system for a duct-type TCP device for use in shallow water. Motions of the duct are investigated to obtain the response amplitude operator (RAO) as the magnitude of the transfer function. Parameters affecting the stability of the SPM system such as the fairlead departure angle, current velocity, and the number of clamp weights are analyzed and/or optimized. Wadam and OrcaFlex commercial software is used to design the mooring line.
Keywords: Mooring design, parametric analysis, response amplitude operator, single point mooring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 217460 Comparison of Different k-NN Models for Speed Prediction in an Urban Traffic Network
Authors: Seyoung Kim, Jeongmin Kim, Kwang Ryel Ryu
Abstract:
A database that records average traffic speeds measured at five-minute intervals for all the links in the traffic network of a metropolitan city. While learning from this data the models that can predict future traffic speed would be beneficial for the applications such as the car navigation system, building predictive models for every link becomes a nontrivial job if the number of links in a given network is huge. An advantage of adopting k-nearest neighbor (k-NN) as predictive models is that it does not require any explicit model building. Instead, k-NN takes a long time to make a prediction because it needs to search for the k-nearest neighbors in the database at prediction time. In this paper, we investigate how much we can speed up k-NN in making traffic speed predictions by reducing the amount of data to be searched for without a significant sacrifice of prediction accuracy. The rationale behind this is that we had a better look at only the recent data because the traffic patterns not only repeat daily or weekly but also change over time. In our experiments, we build several different k-NN models employing different sets of features which are the current and past traffic speeds of the target link and the neighbor links in its up/down-stream. The performances of these models are compared by measuring the average prediction accuracy and the average time taken to make a prediction using various amounts of data.Keywords: Big data, k-NN, machine learning, traffic speed prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 137659 Simulation of Fluid Flow and Heat Transfer in Inclined Cavity using Lattice Boltzmann Method
Authors: Arash Karimipour, A. Hossein Nezhad, E. Shirani, A. Safaei
Abstract:
In this paper, Lattice Boltzmann Method (LBM) is used to study laminar flow with mixed convection heat transfer inside a two-dimensional inclined lid-driven rectangular cavity with aspect ratio AR = 3. Bottom wall of the cavity is maintained at lower temperature than the top lid, and its vertical walls are assumed insulated. Top lid motion results in fluid motion inside the cavity. Inclination of the cavity causes horizontal and vertical components of velocity to be affected by buoyancy force. To include this effect, calculation procedure of macroscopic properties by LBM is changed and collision term of Boltzmann equation is modified. A computer program is developed to simulate this problem using BGK model of lattice Boltzmann method. The effects of the variations of Richardson number and inclination angle on the thermal and flow behavior of the fluid inside the cavity are investigated. The results are presented as velocity and temperature profiles, stream function contours and isotherms. It is concluded that LBM has good potential to simulate mixed convection heat transfer problems.
Keywords: gravity, inclined lid driven cavity, lattice Boltzmannmethod, mixed convection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 195458 Evaluating Portfolio Performance by Highlighting Network Property and the Sharpe Ratio in the Stock Market
Authors: Zahra Hatami, Hesham Ali, David Volkman
Abstract:
Selecting a portfolio for investing is a crucial decision for individuals and legal entities. In the last two decades, with economic globalization, a stream of financial innovations has rushed to the aid of financial institutions. The importance of selecting stocks for the portfolio is always a challenging task for investors. This study aims to create a financial network to identify optimal portfolios using network centralities metrics. This research presents a community detection technique of superior stocks that can be described as an optimal stock portfolio to be used by investors. By using the advantages of a network and its property in extracted communities, a group of stocks was selected for each of the various time periods. The performance of the optimal portfolios was compared to the famous index. Their Sharpe ratio was calculated in a timely manner to evaluate their profit for making decisions. The analysis shows that the selected potential portfolio from stocks with low centrality measurement can outperform the market; however, they have a lower Sharpe ratio than stocks with high centrality scores. In other words, stocks with low centralities could outperform the S&P500 yet have a lower Sharpe ratio than high central stocks.
Keywords: Portfolio management performance, network analysis, centrality measurements, Sharpe ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40857 A Fast Neural Algorithm for Serial Code Detection in a Stream of Sequential Data
Authors: Hazem M. El-Bakry, Qiangfu Zhao
Abstract:
In recent years, fast neural networks for object/face detection have been introduced based on cross correlation in the frequency domain between the input matrix and the hidden weights of neural networks. In our previous papers [3,4], fast neural networks for certain code detection was introduced. It was proved in [10] that for fast neural networks to give the same correct results as conventional neural networks, both the weights of neural networks and the input matrix must be symmetric. This condition made those fast neural networks slower than conventional neural networks. Another symmetric form for the input matrix was introduced in [1-9] to speed up the operation of these fast neural networks. Here, corrections for the cross correlation equations (given in [13,15,16]) to compensate for the symmetry condition are presented. After these corrections, it is proved mathematically that the number of computation steps required for fast neural networks is less than that needed by classical neural networks. Furthermore, there is no need for converting the input data into symmetric form. Moreover, such new idea is applied to increase the speed of neural networks in case of processing complex values. Simulation results after these corrections using MATLAB confirm the theoretical computations.
Keywords: Fast Code/Data Detection, Neural Networks, Cross Correlation, real/complex values.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162656 Lean Environmental Management Integration System (LEMIS) Framework Development
Authors: Puvanasvaran, A. P., Suresh V., N. Norazlin
Abstract:
The Lean Environmental Management Integration System (LEMIS) framework development is integration between lean core element and ISO 14001. The curiosity on the relationship between continuous improvement and sustainability of lean implementation has influenced this study toward LEMIS. Characteristic of ISO 14001 standard clauses and core elements of lean principles are explored from past studies and literature reviews. Survey was carried out on ISO 14001 certified companies to examine continual improvement by implementing the ISO 14001 standard. The study found that there is a significant and positive relationship between Lean Principles: value, value stream, flow, pull and perfection with the ISO 14001 requirements. LEMIS is significant to support the continuous improvement and sustainability. The integration system can be implemented to any manufacturing company. It gives awareness on the importance on why organizations need to sustain its environmental management system. In the meantime, the lean principle can be adapted in order to streamline daily activities of the company. Throughout the study, it had proven that there is no sacrifice or trade-off between lean principles with ISO 14001 requirements. The framework developed in the study can be further simplified in the future, especially the method of crossing each sub requirements of ISO 14001 standard with the core elements of Lean principles in this study.
Keywords: LEMIS, ISO 14001, integration, framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 238355 Separation of Polyphenolics and Sugar by Ultrafiltration: Effects of Operating Conditions on Fouling and Diafiltration
Authors: Diqiao S. Wei, M. Hossain, Zaid S. Saleh
Abstract:
Polyphenolics and sugar are the components of many fruit juices. In this work, the performance of ultra-filtration (UF) for separating phenolic compounds from apple juice was studied by performing batch experiments in a membrane module with an area of 0.1 m2 and fitted with a regenerated cellulose membrane of 1 kDa MWCO. The effects of various operating conditions: transmembrane pressure (3, 4, 5 bar), temperature (30, 35, 40 ºC), pH (2, 3, 4, 5), feed concentration (3, 5, 7, 10, 15 ºBrix for apple juice) and feed flow rate (1, 1.5, 1.8 L/min) on the performance were determined. The optimum operating conditions were: transmembrane pressure 4 bar, temperature 30 ºC, feed flow rate 1 – 1.8 L/min, pH 3 and 10 Brix (apple juice). After performing ultrafiltration under these conditions, the concentration of polyphenolics in retentate was increased by a factor of up to 2.7 with up to 70% recovered in the permeate and with approx. 20% of the sugar in that stream.. Application of diafiltration (addition of water to the concentrate) can regain the flux by a factor of 1.5, which has been decreased due to fouling. The material balance performed on the process has shown the amount of deposits on the membrane and the extent of fouling in the system. In conclusion, ultrafiltration has been demonstrated as a potential technology to separate the polyphenolics and sugars from their mixtures and can be applied to remove sugars from fruit juice.Keywords: Fouling, membrane, polyphenols, ultrafiltration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 338754 An Analysis of Digital Forensic Laboratory Development among Malaysia’s Law Enforcement Agencies
Authors: Sarah K. Taylor, Miratun M. Saharuddin, Zabri A. Talib
Abstract:
Cybercrime is on the rise, and yet many Law Enforcement Agencies (LEAs) in Malaysia have no Digital Forensics Laboratory (DFL) to assist them in the attrition and analysis of digital evidence. From the estimated number of 30 LEAs in Malaysia, sadly, only eight of them owned a DFL. All of the DFLs are concentrated in the capital of Malaysia and none at the state level. LEAs are still depending on the national DFL (CyberSecurity Malaysia) even for simple and straightforward cases. A survey was conducted among LEAs in Malaysia owning a DFL to understand their history of establishing the DFL, the challenges that they faced and the significance of the DFL to their case investigation. The results showed that the while some LEAs faced no challenge in establishing a DFL, some of them took seven to 10 years to do so. The reason was due to the difficulty in convincing their management because of the high costs involved. The results also revealed that with the establishment of a DFL, LEAs were better able to get faster forensic result and to meet agency’s timeline expectation. It is also found that LEAs were also able to get more meaningful forensic results on cases that require niche expertise, compared to sending off cases to the national DFL. Other than that, cases are getting more complex, and hence, a continuous stream of budget for equipment and training is inevitable. The result derived from the study is hoped to be used by other LEAs in justifying to their management the benefits of establishing an in-house DFL.
Keywords: Digital forensics, digital forensics laboratory, digital evidence, law enforcement agency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 149753 RFU Based Computational Unit Design For Reconfigurable Processors
Authors: M. Aqeel Iqbal
Abstract:
Fully customized hardware based technology provides high performance and low power consumption by specializing the tasks in hardware but lacks design flexibility since any kind of changes require re-design and re-fabrication. Software based solutions operate with software instructions due to which a great flexibility is achieved from the easy development and maintenance of the software code. But this execution of instructions introduces a high overhead in performance and area consumption. In past few decades the reconfigurable computing domain has been introduced which overcomes the traditional trades-off between flexibility and performance and is able to achieve high performance while maintaining a good flexibility. The dramatic gains in terms of chip performance and design flexibility achieved through the reconfigurable computing systems are greatly dependent on the design of their computational units being integrated with reconfigurable logic resources. The computational unit of any reconfigurable system plays vital role in defining its strength. In this research paper an RFU based computational unit design has been presented using the tightly coupled, multi-threaded reconfigurable cores. The proposed design has been simulated for VLIW based architectures and a high gain in performance has been observed as compared to the conventional computing systems.
Keywords: Configuration Stream, Configuration overhead, Configuration Controller, Reconfigurable devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162152 Ultimately Bounded Takagi-Sugeno Fuzzy Management in Urban Traffic Stream Mechanism: Multi-Agent Modeling Approach
Authors: Reza Ghasemi, Negin Amiri Hazaveh
Abstract:
In this paper, control methodology based on the selection of the type of traffic light and the period of the green phase to accomplish an optimum balance at intersections is proposed. This balance should be flexible to the static behavior of time, and randomness in a traffic situation; the goal of the proposed method is to reduce traffic volume in transportation, the average delay for each vehicle, and control over the crash of cars. The proposed method was specifically investigated at the intersection through an appropriate timing of traffic lights by sampling a multi-agent system. It consists of a large number of intersections, each of which is considered as an independent agent that exchanges information with each other, and the stability of each agent is provided separately. The robustness against uncertainties, scalability, and stability of the closed-loop overall system are the main merits of the proposed methodology. The simulation results show that the fuzzy intelligent controller in this multi-factor system which is a Takagi-Sugeno (TS) fuzzy is more useful than scheduling in the fixed-time method and it reduces the lengths of vehicles queuing.
Keywords: Fuzzy intelligent controller, traffic-light control, multi-agent systems, state space equations, stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55451 Lean Manufacturing: Systematic Layout Planning Application to an Assembly Line Layout of a Welding Industry
Authors: Fernando Augusto Ullmann Tobe, Moacyr Amaral Domingues, Figueiredo, Stephany Rie Yamamoto Gushiken
Abstract:
The purpose of this paper is to present the process of elaborating the layout of an assembly line of a welding industry using the principles of lean manufacturing as the main driver. The objective of this paper is relevant since the current layout of the assembly line causes non-productive times for operators, being related to the lean waste of unnecessary movements. The methodology used for the project development was Project-based Learning (PBL), which is an active way of learning focused on real problems. The process of selecting the methodology for layout planning was developed considering three criteria to evaluate the most relevant one for this paper's goal. As a result of this evaluation, Systematic Layout Planning was selected, and three steps were added to it – Value Stream Mapping for the current situation and after layout changed and the definition of lean tools and layout type. This inclusion was to consider lean manufacturing in the layout redesign of the industry. The layout change resulted in an increase in the value-adding time of operations carried out in the sector, reduction in movement times between previous and final assemblies, and in cost savings regarding the man-hour value of the employees, which can be invested in productive hours instead of movement times.
Keywords: Assembly line, layout, lean manufacturing, systematic layout planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82650 Production of Energetic Nanomaterials by Spray Flash Evaporation
Authors: Martin Klaumünzer, Jakob Hübner, Denis Spitzer
Abstract:
Within this paper, latest results on processing of energetic nanomaterials by means of the Spray Flash Evaporation technique are presented. This technology constitutes a highly effective and continuous way to prepare fascinating materials on the nano- and micro-scale. Within the process, a solution is set under high pressure and sprayed into an evacuated atomization chamber. Subsequent ultrafast evaporation of the solvent leads to an aerosol stream, which is separated by cyclones or filters. No drying gas is required, so the present technique should not be confused with spray dying. Resulting nanothermites, insensitive explosives or propellants and compositions are foreseen to replace toxic (according to REACH) and very sensitive matter in military and civil applications. Diverse examples are given in detail: nano-RDX (n-Cyclotrimethylentrinitramin) and nano-aluminum based systems, mixtures (n-RDX/n-TNT - trinitrotoluene) or even cocrystalline matter like n-CL-20/HMX (Hexanitrohexaazaisowurtzitane/ Cyclotetra-methylentetranitramin). These nanomaterials show reduced sensitivity by trend without losing effectiveness and performance. An analytical study for material characterization was performed by using Atomic Force Microscopy, X-Ray Diffraction, and combined techniques as well as spectroscopic methods. As a matter of course, sensitivity tests regarding electrostatic discharge, impact, and friction are provided.
Keywords: Continuous synthesis, energetic material, nanoscale, nanothermite, nanoexplosive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 143649 Masquerade and “What Comes Behind Six Is More Than Seven”: Thoughts on Art History and Visual Culture Research Methods
Authors: Osa D Egonwa
Abstract:
In the 21st century, the disciplinary boundaries of past centuries that we often create through mainstream art historical classification, techniques and sources may have been eroded by visual culture, which seems to provide a more inclusive umbrella for the new ways artists go about the creative process and its resultant commodities. Over the past four decades, artists in Africa have resorted to new materials, techniques and themes which have affected our ways of research on these artists and their art. Frontline artists such as El Anatsui, Yinka Shonibare, Erasmus Onyishi are demonstrating that any material is just suitable for artistic expression. Most of times, these materials come with their own techniques/effects and visual syntax: a combination of materials compounds techniques, formal aesthetic indexes, halo effects, and iconography. This tends to challenge the categories and we lean on to view, think and talk about them. This renders our main stream art historical research methods inadequate, thus suggesting new discursive concepts, terms and theories. This paper proposed the Africanist eclectic methods derived from the dual framework of Masquerade Theory and What Comes Behind Six is More Than Seven. This paper shares thoughts/research on art historical methods, terminological re-alignments on classification/source data, presentational format and interpretation arising from the emergent trends in our subject. The outcome provides useful tools to mediate new thoughts and experiences in recent African art and visual culture.
Keywords: Art Historical Methods, Classifications, Concepts , Re-alignment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 63848 Use of Caffeine and Human Pharmaceutical Compounds to Identify Sewage Contamination
Authors: Jingming Wu, Junqi Yue, Ruikang Hu, Zhaoguang Yang, Lifeng Zhang
Abstract:
Fecal coliform bacteria are widely used as indicators of sewage contamination in surface water. However, there are some disadvantages in these microbial techniques including time consuming (18-48h) and inability in discriminating between human and animal fecal material sources. Therefore, it is necessary to seek a more specific indicator of human sanitary waste. In this study, the feasibility was investigated to apply caffeine and human pharmaceutical compounds to identify the human-source contamination. The correlation between caffeine and fecal coliform was also explored. Surface water samples were collected from upstream, middle-stream and downstream points respectively, along Rochor Canal, as well as 8 locations of Marina Bay. Results indicate that caffeine is a suitable chemical tracer in Singapore because of its easy detection (in the range of 0.30-2.0 ng/mL), compared with other chemicals monitored. Relative low concentrations of human pharmaceutical compounds (< 0.07 ng/mL) in Rochor Canal and Marina Bay water samples make them hard to be detected and difficult to be chemical tracer. However, their existence can help to validate sewage contamination. In addition, it was discovered the high correlation exists between caffeine concentration and fecal coliform density in the Rochor Canal water samples, demonstrating that caffeine is highly related to the human-source contamination.Keywords: Caffeine, Human Pharmaceutical Compounds, Chemical Tracer, Sewage Contamination.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 251247 Event Information Extraction System (EIEE): FSM vs HMM
Authors: Shaukat Wasi, Zubair A. Shaikh, Sajid Qasmi, Hussain Sachwani, Rehman Lalani, Aamir Chagani
Abstract:
Automatic Extraction of Event information from social text stream (emails, social network sites, blogs etc) is a vital requirement for many applications like Event Planning and Management systems and security applications. The key information components needed from Event related text are Event title, location, participants, date and time. Emails have very unique distinctions over other social text streams from the perspective of layout and format and conversation style and are the most commonly used communication channel for broadcasting and planning events. Therefore we have chosen emails as our dataset. In our work, we have employed two statistical NLP methods, named as Finite State Machines (FSM) and Hidden Markov Model (HMM) for the extraction of event related contextual information. An application has been developed providing a comparison among the two methods over the event extraction task. It comprises of two modules, one for each method, and works for both bulk as well as direct user input. The results are evaluated using Precision, Recall and F-Score. Experiments show that both methods produce high performance and accuracy, however HMM was good enough over Title extraction and FSM proved to be better for Venue, Date, and time.Keywords: Emails, Event Extraction, Event Detection, Finite state machines, Hidden Markov Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 231746 A Four-Step Ortho-Rectification Procedure for Geo-Referencing Video Streams from a Low-Cost UAV
Authors: B. O. Olawale, C. R. Chatwin, R. C. D. Young, P. M. Birch, F. O. Faithpraise, A. O. Olukiran
Abstract:
In this paper, we present a four-step ortho-rectification procedure for real-time geo-referencing of video data from a low-cost UAV equipped with a multi-sensor system. The basic procedures for the real-time ortho-rectification are: (1) decompilation of the video stream into individual frames; (2) establishing the interior camera orientation parameters; (3) determining the relative orientation parameters for each video frame with respect to each other; (4) finding the absolute orientation parameters, using a self-calibration bundle and adjustment with the aid of a mathematical model. Each ortho-rectified video frame is then mosaicked together to produce a mosaic image of the test area, which is then merged with a well referenced existing digital map for the purpose of geo-referencing and aerial surveillance. A test field located in Abuja, Nigeria was used to evaluate our method. Video and telemetry data were collected for about fifteen minutes, and they were processed using the four-step ortho-rectification procedure. The results demonstrated that the geometric measurement of the control field from ortho-images is more accurate when compared with those from original perspective images when used to pin point the exact location of targets on the video imagery acquired by the UAV. The 2-D planimetric accuracy when compared with the 6 control points measured by a GPS receiver is between 3 to 5 metres.Keywords: Geo-referencing, ortho-rectification, video frame, self-calibration, UAV, target tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 161445 Catchment Yield Prediction in an Ungauged Basin Using PyTOPKAPI
Authors: B. S. Fatoyinbo, D. Stretch, O. T. Amoo, D. Allopi
Abstract:
This study extends the use of the Drainage Area Regionalization (DAR) method in generating synthetic data and calibrating PyTOPKAPI stream yield for an ungauged basin at a daily time scale. The generation of runoff in determining a river yield has been subjected to various topographic and spatial meteorological variables, which integers form the Catchment Characteristics Model (CCM). Many of the conventional CCM models adapted in Africa have been challenged with a paucity of adequate, relevance and accurate data to parameterize and validate the potential. The purpose of generating synthetic flow is to test a hydrological model, which will not suffer from the impact of very low flows or very high flows, thus allowing to check whether the model is structurally sound enough or not. The employed physically-based, watershed-scale hydrologic model (PyTOPKAPI) was parameterized with GIS-pre-processing parameters and remote sensing hydro-meteorological variables. The validation with mean annual runoff ratio proposes a decent graphical understanding between observed and the simulated discharge. The Nash-Sutcliffe efficiency and coefficient of determination (R²) values of 0.704 and 0.739 proves strong model efficiency. Given the current climate variability impact, water planner can now assert a tool for flow quantification and sustainable planning purposes.
Keywords: Ungauged Basin, Catchment Characteristics Model, Synthetic data, GIS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 131144 Wet Flue Gas Desulfurization Using a New O-Element Design Which Replaces the Venturi Scrubber
Authors: P. Lestinsky, D. Jecha, V. Brummer, P. Stehlik
Abstract:
Scrubbing by a liquid spraying is one of the most effective processes used for removal of fine particles and soluble gas pollutants (such as SO2, HCl, HF) from the flue gas. There are many configurations of scrubbers designed to provide contact between the liquid and gas stream for effectively capturing particles or soluble gas pollutants, such as spray plates, packed bed towers, jet scrubbers, cyclones, vortex and venturi scrubbers. The primary function of venturi scrubber is the capture of fine particles as well as HCl, HF or SO2 removal with effect of the flue gas temperature decrease before input to the absorption column. In this paper, sulfur dioxide (SO2) from flue gas was captured using new design replacing venturi scrubber (1st degree of wet scrubbing). The flue gas was prepared by the combustion of the carbon disulfide solution in toluene (1:1 vol.) in the flame in the reactor. Such prepared flue gas with temperature around 150°C was processed in designed laboratory O-element scrubber. Water was used as absorbent liquid. The efficiency of SO2 removal, pressure drop and temperature drop were measured on our experimental device. The dependence of these variables on liquid-gas ratio was observed. The average temperature drop was in the range from 150°C to 40°C. The pressure drop was increased with increasing of a liquid-gas ratio, but no too much as for the common venturi scrubber designs. The efficiency of SO2 removal was up to 70 %. The pressure drop of our new designed wet scrubber is similar to commonly used venturi scrubbers; nevertheless the influence of amount of the liquid on pressure drop is not so significant.
Keywords: Desulphurization, absorption, flue gas, modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 289043 Lagrangian Flow Skeletons Captured in the Wake of a Swimming Nematode C. elegans Using an Immersed Boundary Fluid-Structure Interaction Approach
Authors: Arash Taheri
Abstract:
In this paper, Lagrangian coherent structure (LCS) concept is applied to wake flows generated in the up/down-stream of a swimming nematode C. elegans in an intermediate Re number range, i.e., 250-1200. It materializes Lagrangian hidden structures depicting flow transport barriers. To pursue the goals, nematode swimming in a quiescent fluid flow environment is numerically simulated by a two-way fluid-structure interaction (FSI) approach with the aid of immersed boundary method (IBM). In this regard, incompressible Navier-Stokes equations, fully-coupled with Lagrangian deformation equations for the immersed body, are solved using IB2d code. For all simulations, nematode’s body is modeled with a parametrized spring-fiber built-in case available in the computational code. Reverse von-Kármán vortex street formation and vortex shedding characteristics are studied and discussed in details via LCS approach, including grid resolution, integration time and Reynolds number effects. Results unveil presence of different flow regions with distinct fluid particle fates in the swimming animal’s wake and formation of so-called ‘mushroom-shaped’ structures in attracting LCS identities.
Keywords: Lagrangian coherent structure, nematode swimming, fluid-structure interaction, immersed boundary method, bionics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 994