Search results for: spatial audio processing
4744 Mitigating Urban Flooding through Spatial Planning Interventions: A Case of Bhopal City
Authors: Rama Umesh Pandey, Jyoti Yadav
Abstract:
Flooding is one of the waterborne disasters that causes extensive destruction in urban areas. Developing countries are at a higher risk of such damage and more than half of the global flooding events take place in Asian countries including India. Urban flooding is more of a human-induced disaster rather than natural. This is highly influenced by the anthropogenic factors, besides metrological and hydrological causes. Unplanned urbanization and poor management of cities enhance the impact manifold and cause huge loss of life and property in urban areas. It is an irony that urban areas have been facing water scarcity in summers and flooding during monsoon. This paper is an attempt to highlight the factors responsible for flooding in a city especially from an urban planning perspective and to suggest mitigating measures through spatial planning interventions. Analysis has been done in two stages; first is to assess the impacts of previous flooding events and second to analyze the factors responsible for flooding at macro and micro level in cities. Bhopal, a city in Central India having nearly two million population, has been selected for the study. The city has been experiencing flooding during heavy rains in monsoon. The factors responsible for urban flooding were identified through literature review as well as various case studies from different cities across the world and India. The factors thus identified were analyzed for both macro and micro level influences. For macro level, the previous flooding events that have caused huge destructions were analyzed and the most affected areas in Bhopal city were identified. Since the identified area was falling within the catchment of a drain so the catchment area was delineated for the study. The factors analyzed were: rainfall pattern to calculate the return period using Weibull’s formula; imperviousness through mapping in ArcGIS; runoff discharge by using Rational method. The catchment was divided into micro watersheds and the micro watershed having maximum impervious surfaces was selected to analyze the coverage and effect of physical infrastructure such as: storm water management; sewerage system; solid waste management practices. The area was further analyzed to assess the extent of violation of ‘building byelaws’ and ‘development control regulations’ and encroachment over the natural water streams. Through analysis, the study has revealed that the main issues have been: lack of sewerage system; inadequate storm water drains; inefficient solid waste management in the study area; violation of building byelaws through extending building structures ether on to the drain or on the road; encroachments by slum dwellers along or on to the drain reducing the width and capacity of the drain. Other factors include faulty culvert’s design resulting in back water effect. Roads are at higher level than the plinth of houses which creates submersion of their ground floors. The study recommends spatial planning interventions for mitigating urban flooding and strategies for management of excess rain water during monsoon season. Recommendations have also been made for efficient land use management to mitigate water logging in areas vulnerable to flooding.Keywords: mitigating strategies, spatial planning interventions, urban flooding, violation of development control regulations
Procedia PDF Downloads 3294743 Spatial Variability of Heavy Metals in Sediments of Two Streams of the Olifants River System, South Africa
Authors: Abraham Addo-Bediako, Sophy Nukeri, Tebatso Mmako
Abstract:
Many freshwater ecosystems have been subjected to prolonged and cumulative pollution as a result of human activities such as mining, agricultural, industrial and human settlements in their catchments. The objective of this study was to investigate spatial variability of heavy metal pollution of sediments and possible sources of pollutants in two streams of the Olifants River System, South Africa. Stream sediments were collected and analysed for Arsenic (As), Cadmium (Cd), Chromium (Cr), Copper (Cu), Lead (Pb), Nickel (Ni) and Zinc (Zn) concentrations using inductively coupled plasma-mass mass spectrometry (ICP-MS). In both rivers, As, Cd, Cu, Pb and Zn fell within the concentration ranges recommended by CCME and ANZECC, while the concentrations of Cr and Ni exceeded the standards; the results indicated that Cr and Ni in the sediments originated from human activities and not from natural geological background. The index of geo-accumulation (Igeo) was used to assess the degree of pollution. The results of the geo-accumulation index evaluation showed that Cr and Ni were present in the sediments of the rivers at moderately to extremely polluted levels, while As, Cd, Cu, Pb and Zn existed at unpolluted to moderately polluted levels. Generally, heavy metal concentrations increased along the gradient in the rivers. The high concentrations of Cr and Ni in both rivers are of great concern, as previously these two rivers were classified to be supplying the Olifants River with water of good quality. There is a critical need, therefore to monitor heavy metal concentrations and distributions, as well as a comprehensive plan to prevent health risks, especially those communities still reliant on untreated water from the rivers, as sediment pollution may pose a risk of secondary water pollution under sediment disturbance and/or changes in the geo-chemistry of sediments.Keywords: geo-accumulation index, heavy metals, sediment pollution, water quality
Procedia PDF Downloads 1644742 Application of Regularized Spatio-Temporal Models to the Analysis of Remote Sensing Data
Authors: Salihah Alghamdi, Surajit Ray
Abstract:
Space-time data can be observed over irregularly shaped manifolds, which might have complex boundaries or interior gaps. Most of the existing methods do not consider the shape of the data, and as a result, it is difficult to model irregularly shaped data accommodating the complex domain. We used a method that can deal with space-time data that are distributed over non-planner shaped regions. The method is based on partial differential equations and finite element analysis. The model can be estimated using a penalized least squares approach with a regularization term that controls the over-fitting. The model is regularized using two roughness penalties, which consider the spatial and temporal regularities separately. The integrated square of the second derivative of the basis function is used as temporal penalty. While the spatial penalty consists of the integrated square of Laplace operator, which is integrated exclusively over the domain of interest that is determined using finite element technique. In this paper, we applied a spatio-temporal regression model with partial differential equations regularization (ST-PDE) approach to analyze a remote sensing data measuring the greenness of vegetation, measure by an index called enhanced vegetation index (EVI). The EVI data consist of measurements that take values between -1 and 1 reflecting the level of greenness of some region over a period of time. We applied (ST-PDE) approach to irregular shaped region of the EVI data. The approach efficiently accommodates the irregular shaped regions taking into account the complex boundaries rather than smoothing across the boundaries. Furthermore, the approach succeeds in capturing the temporal variation in the data.Keywords: irregularly shaped domain, partial differential equations, finite element analysis, complex boundray
Procedia PDF Downloads 1414741 Spatio-Temporal Analysis of Drought in Cholistan Region, Pakistan: An Application of Standardized Precipitation Index
Authors: Qurratulain Safdar
Abstract:
Drought is a temporary aberration in contrast to aridity, as it is a permanent feature of climate. Virtually, it takes place in all types of climatic regions that range from high to low rainfall areas. Due to the wide latitudinal extent of Pakistan, there is seasonal and annual variability in rainfall. The south-central part of the country is arid and hyper-arid. This study focuses on the spatio-temporal analysis of droughts in arid and hyperarid region of Cholistan using the standardized precipitation index (SPI) approach. This study has assessed the extent of recurrences of drought and its temporal vulnerability to drought in Cholistan region. Initially, the paper described the geographic setup of the study area along with a brief description of the drought conditions that prevail in Pakistan. The study also provides a scientific foundation for preparing literature and theoretical framework in-line with the selected parameters and indicators. Data were collected both from primary and secondary data sources. Rainfall and temperature data were obtained from Pakistan Meteorology Department. By applying geostatistical approach, a standardized precipitation index (SPI) was calculated for the study region, and the value of spatio-temporal variability of drought and its severity was explored. As a result, in-depth spatial analysis of drought conditions in Cholistan area was found. Parallel to this, drought-prone areas with seasonal variation were also identified using Kriging spatial interpolation techniques in a GIS environment. The study revealed that there is temporal variation in droughts' occurrences both in time series and SPI values. The paper is finally concluded, and strategic plan was suggested to minimize the impacts of drought.Keywords: Cholistan desert, climate anomalies, metrological droughts, standardized precipitation index
Procedia PDF Downloads 2134740 A Survey and Analysis on Inflammatory Pain Detection and Standard Protocol Selection Using Medical Infrared Thermography from Image Processing View Point
Authors: Mrinal Kanti Bhowmik, Shawli Bardhan Jr., Debotosh Bhattacharjee
Abstract:
Human skin containing temperature value more than absolute zero, discharges infrared radiation related to the frequency of the body temperature. The difference in infrared radiation from the skin surface reflects the abnormality present in human body. Considering the difference, detection and forecasting the temperature variation of the skin surface is the main objective of using Medical Infrared Thermography(MIT) as a diagnostic tool for pain detection. Medical Infrared Thermography(MIT) is a non-invasive imaging technique that records and monitors the temperature flow in the body by receiving the infrared radiated from the skin and represent it through thermogram. The intensity of the thermogram measures the inflammation from the skin surface related to pain in human body. Analysis of thermograms provides automated anomaly detection associated with suspicious pain regions by following several image processing steps. The paper represents a rigorous study based survey related to the processing and analysis of thermograms based on the previous works published in the area of infrared thermal imaging for detecting inflammatory pain diseases like arthritis, spondylosis, shoulder impingement, etc. The study also explores the performance analysis of thermogram processing accompanied by thermogram acquisition protocols, thermography camera specification and the types of pain detected by thermography in summarized tabular format. The tabular format provides a clear structural vision of the past works. The major contribution of the paper introduces a new thermogram acquisition standard associated with inflammatory pain detection in human body to enhance the performance rate. The FLIR T650sc infrared camera with high sensitivity and resolution is adopted to increase the accuracy of thermogram acquisition and analysis. The survey of previous research work highlights that intensity distribution based comparison of comparable and symmetric region of interest and their statistical analysis assigns adequate result in case of identifying and detecting physiological disorder related to inflammatory diseases.Keywords: acquisition protocol, inflammatory pain detection, medical infrared thermography (MIT), statistical analysis
Procedia PDF Downloads 3434739 Suggestion of Methodology to Detect Building Damage Level Collectively with Flood Depth Utilizing Geographic Information System at Flood Disaster in Japan
Authors: Munenari Inoguchi, Keiko Tamura
Abstract:
In Japan, we were suffered by earthquake, typhoon, and flood disaster in 2019. Especially, 38 of 47 prefectures were affected by typhoon #1919 occurred in October 2019. By this disaster, 99 people were dead, three people were missing, and 484 people were injured as human damage. Furthermore, 3,081 buildings were totally collapsed, 24,998 buildings were half-collapsed. Once disaster occurs, local responders have to inspect damage level of each building by themselves in order to certificate building damage for survivors for starting their life reconstruction process. At that disaster, the total number to be inspected was so high. Based on this situation, Cabinet Office of Japan approved the way to detect building damage level efficiently, that is collectively detection. However, they proposed a just guideline, and local responders had to establish the concrete and infallible method by themselves. Against this issue, we decided to establish the effective and efficient methodology to detect building damage level collectively with flood depth. Besides, we thought that the flood depth was relied on the land height, and we decided to utilize GIS (Geographic Information System) for analyzing the elevation spatially. We focused on the analyzing tool of spatial interpolation, which is utilized to survey the ground water level usually. In establishing the methodology, we considered 4 key-points: 1) how to satisfy the condition defined in the guideline approved by Cabinet Office for detecting building damage level, 2) how to satisfy survivors for the result of building damage level, 3) how to keep equitability and fairness because the detection of building damage level was executed by public institution, 4) how to reduce cost of time and human-resource because they do not have enough time and human-resource for disaster response. Then, we proposed a methodology for detecting building damage level collectively with flood depth utilizing GIS with five steps. First is to obtain the boundary of flooded area. Second is to collect the actual flood depth as sampling over flooded area. Third is to execute spatial analysis of interpolation with sampled flood depth to detect two-dimensional flood depth extent. Fourth is to divide to blocks by four categories of flood depth (non-flooded, over the floor to 100 cm, 100 cm to 180 cm and over 180 cm) following lines of roads for getting satisfaction from survivors. Fifth is to put flood depth level to each building. In Koriyama city of Fukushima prefecture, we proposed the methodology of collectively detection for building damage level as described above, and local responders decided to adopt our methodology at typhoon #1919 in 2019. Then, we and local responders detect building damage level collectively to over 1,000 buildings. We have received good feedback that the methodology was so simple, and it reduced cost of time and human-resources.Keywords: building damage inspection, flood, geographic information system, spatial interpolation
Procedia PDF Downloads 1244738 Detection of Powdery Mildew Disease in Strawberry Using Image Texture and Supervised Classifiers
Authors: Sultan Mahmud, Qamar Zaman, Travis Esau, Young Chang
Abstract:
Strawberry powdery mildew (PM) is a serious disease that has a significant impact on strawberry production. Field scouting is still a major way to find PM disease, which is not only labor intensive but also almost impossible to monitor disease severity. To reduce the loss caused by PM disease and achieve faster automatic detection of the disease, this paper proposes an approach for detection of the disease, based on image texture and classified with support vector machines (SVMs) and k-nearest neighbors (kNNs). The methodology of the proposed study is based on image processing which is composed of five main steps including image acquisition, pre-processing, segmentation, features extraction and classification. Two strawberry fields were used in this study. Images of healthy leaves and leaves infected with PM (Sphaerotheca macularis) disease under artificial cloud lighting condition. Colour thresholding was utilized to segment all images before textural analysis. Colour co-occurrence matrix (CCM) was introduced for extraction of textural features. Forty textural features, related to a physiological parameter of leaves were extracted from CCM of National television system committee (NTSC) luminance, hue, saturation and intensity (HSI) images. The normalized feature data were utilized for training and validation, respectively, using developed classifiers. The classifiers have experimented with internal, external and cross-validations. The best classifier was selected based on their performance and accuracy. Experimental results suggested that SVMs classifier showed 98.33%, 85.33%, 87.33%, 93.33% and 95.0% of accuracy on internal, external-I, external-II, 4-fold cross and 5-fold cross-validation, respectively. Whereas, kNNs results represented 90.0%, 72.00%, 74.66%, 89.33% and 90.3% of classification accuracy, respectively. The outcome of this study demonstrated that SVMs classified PM disease with a highest overall accuracy of 91.86% and 1.1211 seconds of processing time. Therefore, overall results concluded that the proposed study can significantly support an accurate and automatic identification and recognition of strawberry PM disease with SVMs classifier.Keywords: powdery mildew, image processing, textural analysis, color co-occurrence matrix, support vector machines, k-nearest neighbors
Procedia PDF Downloads 1204737 A Comparative Study of Cognitive Functions in Relapsing-Remitting Multiple Sclerosis Patients, Secondary-Progressive Multiple Sclerosis Patients and Normal People
Authors: Alireza Pirkhaefi
Abstract:
Background: Multiple sclerosis (MS) is one of the most common diseases of the central nervous system (brain and spinal cord). Given the importance of cognitive disorders in patients with multiple sclerosis, the present study was in order to compare cognitive functions (Working memory, Attention and Centralization, and Visual-spatial perception) in patients with relapsing- remitting multiple sclerosis (RRMS) and secondary progressive multiple sclerosis (SPMS). Method: Present study was performed as a retrospective study. This research was conducted with Ex-Post Facto method. The samples of research consisted of 60 patients with multiple sclerosis (30 patients relapsing-retrograde and 30 patients secondary progressive), who were selected from Tehran Community of MS Patients Supported as convenience sampling. 30 normal persons were also selected as a comparison group. Montreal Cognitive Assessment (MOCA) was used to assess cognitive functions. Data were analyzed using multivariate analysis of variance. Results: The results showed that there were significant differences among cognitive functioning in patients with RRMS, SPMS, and normal individuals. There were not significant differences in working memory between two groups of patients with RRMS and SPMS; while significant differences in these variables were seen between the two groups and normal individuals. Also, results showed significant differences in attention and centralization and visual-spatial perception among three groups. Conclusions: Results showed that there are differences between cognitive functions of RRMS and SPMS patients so that the functions of RRMS patients are better than SPMS patients. These results have a critical role in improvement of cognitive functions; reduce the factors causing disability due to cognitive impairment, and especially overall health of society.Keywords: multiple sclerosis, cognitive function, secondary-progressive, normal subjects
Procedia PDF Downloads 2394736 Analysis of Underground Logistics Transportation Technology and Planning Research: Based on Xiong'an New Area, China
Authors: Xia Luo, Cheng Zeng
Abstract:
Under the promotion of the Central Committee of the Communist Party of China and the State Council in 2017, Xiong'an New Area is the third crucial new area in China established after Shenzhen and Shanghai. Its constructions' significance lies in mitigating Beijing's non-capital functions and exploring a new mode of optimizing development in densely populated and economically intensive areas. For this purpose, developing underground logistics can assume the role of goods distribution in the capital, relieve the road transport pressure in Beijing-Tianjin-Hebei Urban Agglomeration, adjust and optimize the urban layout and spatial structure of it. Firstly, the construction planning of Xiong'an New Area and underground logistics development are summarized, especially the development status abroad, the development trend, and bottlenecks of underground logistics in China. This paper explores the technicality, feasibility, and necessity of four modes of transportation. There are pneumatic capsule pipeline (PCP) technology, the CargoCap technology, cable hauled mule, and automatic guided vehicle (AGV). The above technical parameters and characteristics are introduced to relevant experts or scholars. Through establishing an indicator system, carrying out a questionnaire survey with the Delphi method, the final suggestion is obtained: China should develop logistics vehicles similar to CargoCap, adopting rail mode and driverless mode. Based on China's temporal and spatial logistics demand and the geographical pattern of Xiong'an New Area, the construction scale, technical parameters, node location, and other vital parameters of underground logistics are planned. In this way, we hope to speed up the new area's construction and the logistics industry's innovation.Keywords: the Xiong'an new area, underground logistics, contrastive analysis, CargoCap, logistics planning
Procedia PDF Downloads 1294735 Detection and Classification of Rubber Tree Leaf Diseases Using Machine Learning
Authors: Kavyadevi N., Kaviya G., Gowsalya P., Janani M., Mohanraj S.
Abstract:
Hevea brasiliensis, also known as the rubber tree, is one of the foremost assets of crops in the world. One of the most significant advantages of the Rubber Plant in terms of air oxygenation is its capacity to reduce the likelihood of an individual developing respiratory allergies like asthma. To construct such a system that can properly identify crop diseases and pests and then create a database of insecticides for each pest and disease, we must first give treatment for the illness that has been detected. We shall primarily examine three major leaf diseases since they are economically deficient in this article, which is Bird's eye spot, algal spot and powdery mildew. And the recommended work focuses on disease identification on rubber tree leaves. It will be accomplished by employing one of the superior algorithms. Input, Preprocessing, Image Segmentation, Extraction Feature, and Classification will be followed by the processing technique. We will use time-consuming procedures that they use to detect the sickness. As a consequence, the main ailments, underlying causes, and signs and symptoms of diseases that harm the rubber tree are covered in this study.Keywords: image processing, python, convolution neural network (CNN), machine learning
Procedia PDF Downloads 764734 The New Media and Their Economic and Socio-Political Imperatives for Africa: A Study of Nigeria
Authors: Chukwukelue Uzodinma Umenyilorah
Abstract:
The advent of the New Media as enabled by information and communication technology from the 19th through the 21st century has no doubt taken its toll on all fronts of human existence; especially in Africa. Apart from shortening the distance between all parts of the world, technology and the new media has also succeeded in making the world a global village. Hence, it is now easy to relay live audio and visual signals across the length and breadth of the world in real time. People now contract and execute businesses across countries, conferences are held and ideas are shared with a simple push of a button. Likewise, political leaders and diplomats are now just a click away from reaching those important decisions that take their country’s fortunes to the next level. On the flip side, ICT and the New Media have also contributed in no small measure in aiding global terrorism and general insecurity around the world. More interesting is the fact that as developing economies, African countries have massively embraced the information technology and this has helped them in keeping up with the trends in the polity of other model democracies around the world. This paper is therefore designed to determine the how much effect ICT and the New Media has exerted on the economic, social and political lives of African. Nigeria shall be used as a case in point for the purpose of this paper.Keywords: Africa, ICT, new media, Nigeria
Procedia PDF Downloads 2554733 Structure and Properties of Meltblown Polyetherimide as High Temperature Filter Media
Authors: Gajanan Bhat, Vincent Kandagor, Daniel Prather, Ramesh Bhave
Abstract:
Polyetherimide (PEI), an engineering plastic with very high glass transition temperature and excellent chemical and thermal stability, has been processed into a controlled porosity filter media of varying pore size, performance, and surface characteristics. A special grade of the PEI was processed by melt blowing to produce microfiber nonwovens suitable as filter media. The resulting microfiber webs were characterized to evaluate their structure and properties. The fiber webs were further modified by hot pressing, a post processing technique, which reduces the pore size in order to improve the barrier properties of the resulting membranes. This ongoing research has shown that PEI can be a good candidate for filter media requiring high temperature and chemical resistance with good mechanical properties. Also, by selecting the appropriate processing conditions, it is possible to achieve desired filtration performance from this engineering plastic.Keywords: nonwovens, melt blowing, polyehterimide, filter media, microfibers
Procedia PDF Downloads 3154732 Drying of Agro-Industrial Wastes Using a Cabinet Type Solar Dryer
Authors: N. Metidji, O. Badaoui, A. Djebli, H. Bendjebbas, R. Sellami
Abstract:
The agro-industry is considered as one of the most waste producing industrial fields as a result of food processing. Upgrading and reuse of these wastes as animal or poultry food seems to be a promising alternative. Combined with the use of clean energy resources, the recovery process would contribute more to the environment protection. It is in this framework that a new solar dryer has been designed in the Unit of Solar Equipment Development. Direct solar drying has, also, many advantages compared to natural sun drying. In fact, the first does not cause product degradation as it is protected by the drying chamber from direct sun, insects and exterior environment. The aim of this work is to study the drying kinetics of waste, generated during the processing of pepper, by using a direct natural convection solar dryer at 35◦C and 55◦C. The rate of moisture removal from the product to be dried has been found to be directly related to temperature, humidity and flow rate. The characterization of these parameters has allowed the determination of the appropriate drying time for this product namely peppers waste.Keywords: solar energy, solar dryer, energy conversion, pepper drying, forced convection solar dryer
Procedia PDF Downloads 4114731 Working Memory and Phonological Short-Term Memory in the Acquisition of Academic Formulaic Language
Authors: Zhicheng Han
Abstract:
This study examines the correlation between knowledge of formulaic language, working memory (WM), and phonological short-term memory (PSTM) in Chinese L2 learners of English. This study investigates if WM and PSTM correlate differently to the acquisition of formulaic language, which may be relevant for the discourse around the conceptualization of formulas. Connectionist approaches have lead scholars to argue that formulas are form-meaning connections stored whole, making PSTM significant in the acquisitional process as it pertains to the storage and retrieval of chunk information. Generativist scholars, on the other hand, argued for active participation of interlanguage grammar in the acquisition and use of formulaic language, where formulas are represented in the mind but retain the internal structure built around a lexical core. This would make WM, especially the processing component of WM an important cognitive factor since it plays a role in processing and holding information for further analysis and manipulation. The current study asked L1 Chinese learners of English enrolled in graduate programs in China to complete a preference raking task where they rank their preference for formulas, grammatical non-formulaic expressions, and ungrammatical phrases with and without the lexical core in academic contexts. Participants were asked to rank the options in order of the likeliness of them encountering these phrases in the test sentences within academic contexts. Participants’ syntactic proficiency is controlled with a cloze test and grammar test. Regression analysis found a significant relationship between the processing component of WM and preference of formulaic expressions in the preference ranking task while no significant correlation is found for PSTM or syntactic proficiency. The correlational analysis found that WM, PSTM, and the two proficiency test scores have significant covariates. However, WM and PSTM have different predictor values for participants’ preference for formulaic language. Both storage and processing components of WM are significantly correlated with the preference for formulaic expressions while PSTM is not. These findings are in favor of the role of interlanguage grammar and syntactic knowledge in the acquisition of formulaic expressions. The differing effects of WM and PSTM suggest that selective attention to and processing of the input beyond simple retention play a key role in successfully acquiring formulaic language. Similar correlational patterns were found for preferring the ungrammatical phrase with the lexical core of the formula over the ones without the lexical core, attesting to learners’ awareness of the lexical core around which formulas are constructed. These findings support the view that formulaic phrases retain internal syntactic structures that are recognized and processed by the learners.Keywords: formulaic language, working memory, phonological short-term memory, academic language
Procedia PDF Downloads 634730 Assessment of Air Quality Around Western Refinery in Libya: Mobile Monitoring
Authors: A. Elmethnani, A. Jroud
Abstract:
This coastal crude oil refinery is situated north of a big city west of Tripoli; the city then could be highly prone to downwind refinery emissions where the NNE wind direction is prevailing through most seasons of the year. Furthermore, due to the absence of an air quality monitoring network and scarce emission data available for the neighboring community, nearby residents have serious worries about the impacts of the oil refining operations on local air quality. In responding to these concerns, a short term survey has performed for three consecutive days where a semi-continues mobile monitoring approach has developed effectively in this study; the monitoring station (Compact AQM 65 AeroQual) was mounted on a vehicle to move quickly between locations, measurements of 10 minutes averaging of 60 seconds then been taken at each fixed sampling point. The downwind ambient concentration of CO, H₂S, NOₓ, NO₂, SO₂, PM₁, PM₂.₅ PM₁₀, and TSP were measured at carefully chosen sampling locations, ranging from 200m nearby the fence-line passing through the city center up to 4.7 km east to attain best spatial coverage. Results showed worrying levels of PM₂.₅ PM₁₀, and TSP at one sampling location in the city center, southeast of the refinery site, with an average mean of 16.395μg/m³, 33.021μg/m³, and 42.426μg/m³ respectively, which could be attributed to road traffic. No significant concentrations have been detected for other pollutants of interest over the study area, as levels observed for CO, SO₂, H₂S, NOₓ, and NO₂ haven’t respectively exceeded 1.707 ppm, 0.021ppm, 0.134 ppm, 0.4582 ppm, and 0.0018 ppm, which was at the same sampling locations as well. Although it wasn’t possible to compare the results with the Libyan air quality standards due to the difference in the averaging time period, the technique was adequate for the baseline air quality screening procedure. Overall, findings primarily suggest modeling of dispersion of the refinery emissions to assess the likely impact and spatial-temporal distribution of air pollutants.Keywords: air quality, mobil monitoring, oil refinery
Procedia PDF Downloads 964729 Analysing the Mesoscale Variations of 7Be and 210Pb Concentrations in a Complex Orography, Guadalquivir Valley, Southern Spain
Authors: M. A. Hernández-Ceballos, E. G. San Miguel, C. Galán, J. P. Bolívar
Abstract:
The evolution of 7Be and 210Pb activity concentrations in surface air along the Guadalquivir valley (southern Iberian Peninsula) is presented in this study. Samples collected for 48 h, every fifteen days, from September 2012 to November 2013 at two sampling sites (Huelva city in the mouth and Cordoba city in the middle (located 250 km far away)), are used to 1) analysing the spatial variability and 2) understanding the influence of wind conditions on 7Be and 210Pb. Similar average concentrations were registered along the valley. The mean 7Be activity concentration was 4.46 ± 0.21 mBq/m3 at Huelva and 4.33 ± 0.20 mBq/m3 at Cordoba, although registering higher maximum and minimum values at Cordoba (9.44 mBq/m3 and 1.80 mBq/m3) than at Huelva (7.95 mBq/m3 and 1.04 mBq/m3). No significant differences were observed in the 210Pb mean activity concentrations between Cordoba (0.40 ± 0.04 mBq/m3) and Huelva (0.35 ± 0.04 mBq/m3), although the maximum (1.10 mBq/m3 and 0.87 mBq/m3) and minimum (0.02 mBq/m3 and 0.04 mBq/m3) values were recorded in Cordoba. Although similar average concentrations were obtained in both sites, the temporal evolution of both natural radionuclides presents differences between them. The meteorological analysis of two sampling periods, in which large differences on 7Be and 210Pb concentrations are observed, indicates the different impact of surface and upper wind dynamics. The analysis reveals the different impact of the two sea-land breeze patterns usually observed along the valley (pure and non-pure) and the corresponding air masses at higher layers associated with each one. The pure, with short development (around 30 km inland) and increasing accumulation process, favours high concentrations of both radionuclides in Huelva (coastal site), while the non-pure, with winds sweeping the valley until arrive to Cordoba (250 km far away), causes high activity values at this site. These results reveal the impact of mesoscale conditions on these two natural radionuclides, and the importance of these circulations on its spatial and temporal variability.Keywords: 7Be, 210Pb, air masses, mesoscale process
Procedia PDF Downloads 4094728 A Study on the Different Components of a Typical Back-Scattered Chipless RFID Tag Reflection
Authors: Fatemeh Babaeian, Nemai Chandra Karmakar
Abstract:
Chipless RFID system is a wireless system for tracking and identification which use passive tags for encoding data. The advantage of using chipless RFID tag is having a planar tag which is printable on different low-cost materials like paper and plastic. The printed tag can be attached to different items in the labelling level. Since the price of chipless RFID tag can be as low as a fraction of a cent, this technology has the potential to compete with the conventional optical barcode labels. However, due to the passive structure of the tag, data processing of the reflection signal is a crucial challenge. The captured reflected signal from a tag attached to an item consists of different components which are the reflection from the reader antenna, the reflection from the item, the tag structural mode RCS component and the antenna mode RCS of the tag. All these components are summed up in both time and frequency domains. The effect of reflection from the item and the structural mode RCS component can distort/saturate the frequency domain signal and cause difficulties in extracting the desired component which is the antenna mode RCS. Therefore, it is required to study the reflection of the tag in both time and frequency domains to have a better understanding of the nature of the captured chipless RFID signal. The other benefits of this study can be to find an optimised encoding technique in tag design level and to find the best processing algorithm the chipless RFID signal in decoding level. In this paper, the reflection from a typical backscattered chipless RFID tag with six resonances is analysed, and different components of the signal are separated in both time and frequency domains. Moreover, the time domain signal corresponding to each resonator of the tag is studied. The data for this processing was captured from simulation in CST Microwave Studio 2017. The outcome of this study is understanding different components of a measured signal in a chipless RFID system and a discovering a research gap which is a need to find an optimum detection algorithm for tag ID extraction.Keywords: antenna mode RCS, chipless RFID tag, resonance, structural mode RCS
Procedia PDF Downloads 2004727 Wobbled Laser Beam Welding for Macro-to Micro-Fabrication Process
Authors: Farzad Vakili-Farahani, Joern Lungershausen, Kilian Wasmer
Abstract:
Wobbled laser beam welding, fast oscillations of a tiny laser beam within a designed path (weld geometry) during the laser pulse illumination, opens new possibilities to improve the marco-to micro-manufacturing process. The present work introduces the wobbled laser beam welding as a robust welding strategy for improving macro-to micro-fabrication process, e.g., the laser processing for gap-bridging and packaging industry. The typical requisites and relevant equipment for the development of a wobbled laser processing unit are addressed, including a suitable laser source, light delivery system, optics, proper beam deflection system and the design geometry. In addition, experiments have been carried out on titanium plate to compare the results of wobbled laser welding with conventional pulsed laser welding. As compared to the pulsed laser welding, the wobbled laser welding offers a much greater fusion area (i.e. additional molten material) while minimizing the HAZ and provides a better confinement of the material microstructural changes.Keywords: wobbled laser beam welding, wobbling function, beam oscillation, micro welding
Procedia PDF Downloads 3284726 Soil Salinity Mapping using Electromagnetic Induction Measurements
Authors: Fethi Bouksila, Nessrine Zemni, Fairouz Slama, Magnus Persson, Ronny Berndasson, Akissa Bahri
Abstract:
Electromagnetic sensor EM 38 was used to predict and map soil salinity (ECe) in arid oasis. Despite the high spatial variation of soil moisture and shallow watertable, significant ECe-EM relationships were developed. The low drainage network efficiency is the main factor of soil salinizationKeywords: soil salinity map, electromagnetic induction, EM38, oasis, shallow watertable
Procedia PDF Downloads 1874725 A Multivariate Statistical Approach for Water Quality Assessment of River Hindon, India
Authors: Nida Rizvi, Deeksha Katyal, Varun Joshi
Abstract:
River Hindon is an important river catering the demand of highly populated rural and industrial cluster of western Uttar Pradesh, India. Water quality of river Hindon is deteriorating at an alarming rate due to various industrial, municipal and agricultural activities. The present study aimed at identifying the pollution sources and quantifying the degree to which these sources are responsible for the deteriorating water quality of the river. Various water quality parameters, like pH, temperature, electrical conductivity, total dissolved solids, total hardness, calcium, chloride, nitrate, sulphate, biological oxygen demand, chemical oxygen demand and total alkalinity were assessed. Water quality data obtained from eight study sites for one year has been subjected to the two multivariate techniques, namely, principal component analysis and cluster analysis. Principal component analysis was applied with the aim to find out spatial variability and to identify the sources responsible for the water quality of the river. Three Varifactors were obtained after varimax rotation of initial principal components using principal component analysis. Cluster analysis was carried out to classify sampling stations of certain similarity, which grouped eight different sites into two clusters. The study reveals that the anthropogenic influence (municipal, industrial, waste water and agricultural runoff) was the major source of river water pollution. Thus, this study illustrates the utility of multivariate statistical techniques for analysis and elucidation of multifaceted data sets, recognition of pollution sources/factors and understanding temporal/spatial variations in water quality for effective river water quality management.Keywords: cluster analysis, multivariate statistical techniques, river Hindon, water quality
Procedia PDF Downloads 4674724 Soil Degradati̇on Mapping Using Geographic Information System, Remote Sensing and Laboratory Analysis in the Oum Er Rbia High Basin, Middle Atlas, Morocco
Authors: Aafaf El Jazouli, Ahmed Barakat, Rida Khellouk
Abstract:
Mapping of soil degradation is derived from field observations, laboratory measurements, and remote sensing data, integrated quantitative methods to map the spatial characteristics of soil properties at different spatial and temporal scales to provide up-to-date information on the field. Since soil salinity, texture and organic matter play a vital role in assessing topsoil characteristics and soil quality, remote sensing can be considered an effective method for studying these properties. The main objective of this research is to asses soil degradation by combining remote sensing data and laboratory analysis. In order to achieve this goal, the required study of soil samples was taken at 50 locations in the upper basin of Oum Er Rbia in the Middle Atlas in Morocco. These samples were dried, sieved to 2 mm and analyzed in the laboratory. Landsat 8 OLI imagery was analyzed using physical or empirical methods to derive soil properties. In addition, remote sensing can serve as a supporting data source. Deterministic potential (Spline and Inverse Distance weighting) and probabilistic interpolation methods (ordinary kriging and universal kriging) were used to produce maps of each grain size class and soil properties using GIS software. As a result, a correlation was found between soil texture and soil organic matter content. This approach developed in ongoing research will improve the prospects for the use of remote sensing data for mapping soil degradation in arid and semi-arid environments.Keywords: Soil degradation, GIS, interpolation methods (spline, IDW, kriging), Landsat 8 OLI, Oum Er Rbia high basin
Procedia PDF Downloads 1654723 Strategic Innovation of Nanotechnology: Novel Applications of Biomimetics and Microfluidics in Food Safety
Authors: Boce Zhang
Abstract:
Strategic innovation of nanotechnology to promote food safety has drawn tremendous attentions among research groups, which includes the need for research support during the implementation of the Food Safety Modernization Act (FSMA) in the United States. There are urgent demands and knowledge gaps to the understanding of a) food-water-bacteria interface as for how pathogens persist and transmit during food processing and storage; b) minimum processing requirement needed to prevent pathogen cross-contamination in the food system. These knowledge gaps are of critical importance to the food industry. However, the lack of knowledge is largely hindered by the limitations of research tools. Our groups recently endeavored two novel engineering systems with biomimetics and microfluidics as a holistic approach to hazard analysis and risk mitigation, which provided unprecedented research opportunities to study pathogen behavior, in particular, contamination, and cross-contamination, at the critical food-water-pathogen interface. First, biomimetically-patterned surfaces (BPS) were developed to replicate the identical surface topography and chemistry of a natural food surface. We demonstrated that BPS is a superior research tool that empowers the study of a) how pathogens persist through sanitizer treatment, b) how to apply fluidic shear-force and surface tension to increase the vulnerability of the bacterial cells, by detaching them from a protected area, etc. Secondly, microfluidic devices were designed and fabricated to study the bactericidal kinetics in the sub-second time frame (0.1~1 second). The sub-second kinetics is critical because the cross-contamination process, which includes detachment, migration, and reattachment, can occur in a very short timeframe. With this microfluidic device, we were able to simulate and study these sub-second cross-contamination scenarios, and to further investigate the minimum sanitizer concentration needed to sufficiently prevent pathogen cross-contamination during the food processing. We anticipate that the findings from these studies will provide critical insight on bacterial behavior at the food-water-cell interface, and the kinetics of bacterial inactivation from a broad range of sanitizers and processing conditions, thus facilitating the development and implementation of science-based food safety regulations and practices to mitigate the food safety risks.Keywords: biomimetic materials, microbial food safety, microfluidic device, nanotechnology
Procedia PDF Downloads 3594722 Detection and Tracking for the Protection of the Elderly and Socially Vulnerable People in the Video Surveillance System
Authors: Mobarok Hossain Bhuyain
Abstract:
Video surveillance processing has attracted various security fields transforming it into one of the leading research fields. Today's demand for detection and tracking of human mobility for security is very useful for human security, such as in crowded areas. Accordingly, video surveillance technology has seen a rapid advancement in recent years, with algorithms analyzing the behavior of people under surveillance automatically. The main motivation of this research focuses on the detection and tracking of the elderly and socially vulnerable people in crowded areas. Degenerate people are a major health concern, especially for elderly people and socially vulnerable people. One major disadvantage of video surveillance is the need for continuous monitoring, especially in crowded areas. To assist the security monitoring live surveillance video, image processing, and artificial intelligence methods can be used to automatically send warning signals to the monitoring officers about elderly people and socially vulnerable people.Keywords: human detection, target tracking, neural network, particle filter
Procedia PDF Downloads 1664721 “Octopub”: Geographical Sentiment Analysis Using Named Entity Recognition from Social Networks for Geo-Targeted Billboard Advertising
Authors: Oussama Hafferssas, Hiba Benyahia, Amina Madani, Nassima Zeriri
Abstract:
Although data nowadays has multiple forms; from text to images, and from audio to videos, yet text is still the most used one at a public level. At an academical and research level, and unlike other forms, text can be considered as the easiest form to process. Therefore, a brunch of Data Mining researches has been always under its shadow, called "Text Mining". Its concept is just like data mining’s, finding valuable patterns in data, from large collections and tremendous volumes of data, in this case: Text. Named entity recognition (NER) is one of Text Mining’s disciplines, it aims to extract and classify references such as proper names, locations, expressions of time and dates, organizations and more in a given text. Our approach "Octopub" does not aim to find new ways to improve named entity recognition process, rather than that it’s about finding a new, and yet smart way, to use NER in a way that we can extract sentiments of millions of people using Social Networks as a limitless information source, and Marketing for product promotion as the main domain of application.Keywords: textmining, named entity recognition(NER), sentiment analysis, social media networks (SN, SMN), business intelligence(BI), marketing
Procedia PDF Downloads 5894720 The Usage of Nitrogen Gas and Alum for Sludge Dewatering
Authors: Mamdouh Yousef Saleh, Medhat Hosny El-Zahar, Shymaa El-Dosoky
Abstract:
In most cases, the associated processing cost of dewatering sludge increase with the solid particles concentration. All experiments in this study were conducted on biological sludge type. All experiments help to reduce the greenhouse gases in addition, the technology used was faster in time and less in cost compared to other methods. First, the bubbling pressure was used to dissolve N₂ gas into the sludge, second alum was added to accelerate the process of coagulation of the sludge particles and facilitate their flotation, and third nitrogen gas was used to help floating the sludge particles and reduce the processing time because of the nitrogen gas from the inert gases. The conclusions of this experiment were as follows: first, the best conditions were obtained when the bubbling pressure was 0.6 bar. Second, the best alum dose was determined to help the sludge agglomerate and float. During the experiment, the best alum dose was 80 mg/L. It increased concentration of the sludge by 7-8 times. Third, the economic dose of nitrogen gas was 60 mg/L with separation efficiency of 85%. The sludge concentration was about 8-9 times. That happened due to the gas released tiny bubbles which adhere to the suspended matter causing them to float to the surface of the water where it could be then removed.Keywords: nitrogen gas, biological treatment, alum, dewatering sludge, greenhouse gases
Procedia PDF Downloads 2174719 Rethinking Urban Floodplain Management: The Case of Colombo, Sri Lanka
Authors: Malani Herath, Sohan Wijesekera, Jagath Munasingha
Abstract:
The impact of recent floods become significant, and the extraordinary flood events cause considerable damage to lives, properties, environment and negatively affect the whole development of Colombo urban region. Even though the Colombo urban region experiences recurrent flood impacts, several spatial planning interventions have been taken from time to time since early 20th century. All past plans have adopted a traditional approach to flood management, using infrastructural measures to reduce the chance of flooding together with rigid planning regulations. The existing flood risk management practices do not operate to be acceptable by the local community particular the urban poor. Researchers have constantly reported the differences in estimations of flood risk, priorities, concerns of experts and the local community. Risk-based decision making in flood management is not only a matter of technical facts; it has a significant bearing on how flood risk is viewed by local community and individuals. Moreover, sustainable flood management is an integrated approach, which highlights joint actions of experts and community. This indicates the necessity of further societal discussion on the acceptable level of flood risk indicators to prioritize and identify the appropriate flood management measures in Colombo. The understanding and evaluation of flood risk by local people are important to integrate in the decision-making process. This research questioned about the gap between the acceptable level of flood risk to spatial planners and to the local communities in Colombo. A comprehensive literature review was conducted to prepare a framework to analyze the public perception in Colombo. This research work identifies the factors that affect the variation of flood risk and acceptable levels to both local community and planning authorities.Keywords: Colombo basin, public perception, urban flood risk, multi-criteria analysis
Procedia PDF Downloads 3144718 The Impact of Recurring Events in Fake News Detection
Authors: Ali Raza, Shafiq Ur Rehman Khan, Raja Sher Afgun Usmani, Asif Raza, Basit Umair
Abstract:
Detection of Fake news and missing information is gaining popularity, especially after the advancement in social media and online news platforms. Social media platforms are the main and speediest source of fake news propagation, whereas online news websites contribute to fake news dissipation. In this study, we propose a framework to detect fake news using the temporal features of text and consider user feedback to identify whether the news is fake or not. In recent studies, the temporal features in text documents gain valuable consideration from Natural Language Processing and user feedback and only try to classify the textual data as fake or true. This research article indicates the impact of recurring and non-recurring events on fake and true news. We use two models BERT and Bi-LSTM to investigate, and it is concluded from BERT we get better results and 70% of true news are recurring and rest of 30% are non-recurring.Keywords: natural language processing, fake news detection, machine learning, Bi-LSTM
Procedia PDF Downloads 234717 Efficient Reconstruction of DNA Distance Matrices Using an Inverse Problem Approach
Authors: Boris Melnikov, Ye Zhang, Dmitrii Chaikovskii
Abstract:
We continue to consider one of the cybernetic methods in computational biology related to the study of DNA chains. Namely, we are considering the problem of reconstructing the not fully filled distance matrix of DNA chains. When applied in a programming context, it is revealed that with a modern computer of average capabilities, creating even a small-sized distance matrix for mitochondrial DNA sequences is quite time-consuming with standard algorithms. As the size of the matrix grows larger, the computational effort required increases significantly, potentially spanning several weeks to months of non-stop computer processing. Hence, calculating the distance matrix on conventional computers is hardly feasible, and supercomputers are usually not available. Therefore, we started publishing our variants of the algorithms for calculating the distance between two DNA chains; then, we published algorithms for restoring partially filled matrices, i.e., the inverse problem of matrix processing. In this paper, we propose an algorithm for restoring the distance matrix for DNA chains, and the primary focus is on enhancing the algorithms that shape the greedy function within the branches and boundaries method framework.Keywords: DNA chains, distance matrix, optimization problem, restoring algorithm, greedy algorithm, heuristics
Procedia PDF Downloads 1194716 Effect of Thermal Treatment on Mechanical Properties of Reduced Activation Ferritic/Martensitic Eurofer Steel Grade
Authors: Athina Puype, Lorenzo Malerba, Nico De Wispelaere, Roumen Petrov, Jilt Sietsma
Abstract:
Reduced activation ferritic/martensitic (RAFM) steels like EUROFER97 are primary candidate structural materials for first wall application in the future demonstration (DEMO) fusion reactor. Existing steels of this type obtain their functional properties by a two-stage heat treatment, which consists of an annealing stage at 980°C for thirty minutes followed by quenching and an additional tempering stage at 750°C for two hours. This thermal quench and temper (Q&T) treatment creates a microstructure of tempered martensite with, as main precipitates, M23C6 carbides, with M = Fe, Cr and carbonitrides of MX type, e.g. TaC and VN. The resulting microstructure determines the mechanical properties of the steel. The ductility is largely determined by the tempered martensite matrix, while the resistance to mechanical degradation, determined by the spatial and size distribution of precipitates and the martensite crystals, plays a key role in the high temperature properties of the steel. Unfortunately, the high temperature response of EUROFER97 is currently insufficient for long term use in fusion reactors, due to instability of the matrix phase and coarsening of the precipitates at prolonged high temperature exposure. The objective of this study is to induce grain refinement by appropriate modifications of the processing route in order to increase the high temperature strength of a lab-cast EUROFER RAFM steel grade. The goal of the work is to obtain improved mechanical behavior at elevated temperatures with respect to conventionally heat treated EUROFER97. A dilatometric study was conducted to study the effect of the annealing temperature on the mechanical properties after a Q&T treatment. The microstructural features were investigated with scanning electron microscopy (SEM), electron back-scattered diffraction (EBSD) and transmission electron microscopy (TEM). Additionally, hardness measurements, tensile tests at elevated temperatures and Charpy V-notch impact testing of KLST-type MCVN specimens were performed to study the mechanical properties of the furnace-heated lab-cast EUROFER RAFM steel grade. A significant prior austenite grain (PAG) refinement was obtained by lowering the annealing temperature of the conventionally used Q&T treatment for EUROFER97. The reduction of the PAG results in finer martensitic constituents upon quenching, which offers more nucleation sites for carbide and carbonitride formation upon tempering. The ductile-to-brittle transition temperature (DBTT) was found to decrease with decreasing martensitic block size. Additionally, an increased resistance against high temperature degradation was accomplished in the fine grained martensitic materials with smallest precipitates obtained by tailoring the annealing temperature of the Q&T treatment. It is concluded that the microstructural refinement has a pronounced effect on the DBTT without significant loss of strength and ductility. Further investigation into the optimization of the processing route is recommended to improve the mechanical behavior of RAFM steels at elevated temperatures.Keywords: ductile-to-brittle transition temperature (DBTT), EUROFER, reduced activation ferritic/martensitic (RAFM) steels, thermal treatments
Procedia PDF Downloads 3004715 IIROC's Enforcement Performance: Funnel in, Funnel out, and Funnel away
Authors: Mark Lokanan
Abstract:
The paper analyzes the processing of complaints against investment brokers and dealer members through the Investment Industry Regulatory Organization of Canada (IIROC) from 2008 to 2017. IIROC is the self-regulatory organization (SRO) that is responsible for policing investment dealers and brokerage firms that trade in Canada’s securities market. Data from the study came from IIROC's enforcement annual reports for the years examined. The case processing is evaluated base on the misconduct funnel that was originally designed for street crime and applies to the enforcement of investment fraud. The misconduct funnel is used as a framework to examine IIROC’s claim that it brought in more complaints (funnel in) than government regulators and shows how these complaints are funneled out and funneled away as they are processed through IIROC’s enforcement system. The results indicate that IIROC is ineffective in disciplining its members and is unable to handle the more serious quasi-criminal and improper sales practices offenses. It is hard not to see the results of the paper being used by the legislator in Ottawa to show the importance of a federal securities regulatory agency such as the Securities and Exchange Commission (SEC) in the United States.Keywords: investment fraud, securities regulation, compliance, enforcement
Procedia PDF Downloads 160