Search results for: continuous spatial range queries
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10839

Search results for: continuous spatial range queries

10029 Development of Web-Based Iceberg Detection Using Deep Learning

Authors: A. Kavya Sri, K. Sai Vineela, R. Vanitha, S. Rohith

Abstract:

Large pieces of ice that break from the glaciers are known as icebergs. The threat that icebergs pose to navigation, production of offshore oil and gas services, and underwater pipelines makes their detection crucial. In this project, an automated iceberg tracking method using deep learning techniques and satellite images of icebergs is to be developed. With a temporal resolution of 12 days and a spatial resolution of 20 m, Sentinel-1 (SAR) images can be used to track iceberg drift over the Southern Ocean. In contrast to multispectral images, SAR images are used for analysis in meteorological conditions. This project develops a web-based graphical user interface to detect and track icebergs using sentinel-1 images. To track the movement of the icebergs by using temporal images based on their latitude and longitude values and by comparing the center and area of all detected icebergs. Testing the accuracy is done by precision and recall measures.

Keywords: synthetic aperture radar (SAR), icebergs, deep learning, spatial resolution, temporal resolution

Procedia PDF Downloads 90
10028 Application of Advanced Remote Sensing Data in Mineral Exploration in the Vicinity of Heavy Dense Forest Cover Area of Jharkhand and Odisha State Mining Area

Authors: Hemant Kumar, R. N. K. Sharma, A. P. Krishna

Abstract:

The study has been carried out on the Saranda in Jharkhand and a part of Odisha state. Geospatial data of Hyperion, a remote sensing satellite, have been used. This study has used a wide variety of patterns related to image processing to enhance and extract the mining class of Fe and Mn ores.Landsat-8, OLI sensor data have also been used to correctly explore related minerals. In this way, various processes have been applied to increase the mineralogy class and comparative evaluation with related frequency done. The Hyperion dataset for hyperspectral remote sensing has been specifically verified as an effective tool for mineral or rock information extraction within the band range of shortwave infrared used. The abundant spatial and spectral information contained in hyperspectral images enables the differentiation of different objects of any object into targeted applications for exploration such as exploration detection, mining.

Keywords: Hyperion, hyperspectral, sensor, Landsat-8

Procedia PDF Downloads 122
10027 Perceptual Image Coding by Exploiting Internal Generative Mechanism

Authors: Kuo-Cheng Liu

Abstract:

In the perceptual image coding, the objective is to shape the coding distortion such that the amplitude of distortion does not exceed the error visibility threshold, or to remove perceptually redundant signals from the image. While most researches focus on color image coding, the perceptual-based quantizer developed for luminance signals are always directly applied to chrominance signals such that the color image compression methods are inefficient. In this paper, the internal generative mechanism is integrated into the design of a color image compression method. The internal generative mechanism working model based on the structure-based spatial masking is used to assess the subjective distortion visibility thresholds that are visually consistent to human eyes better. The estimation method of structure-based distortion visibility thresholds for color components is further presented in a locally adaptive way to design quantization process in the wavelet color image compression scheme. Since the lowest subband coefficient matrix of images in the wavelet domain preserves the local property of images in the spatial domain, the error visibility threshold inherent in each coefficient of the lowest subband for each color component is estimated by using the proposed spatial error visibility threshold assessment. The threshold inherent in each coefficient of other subbands for each color component is then estimated in a local adaptive fashion based on the distortion energy allocation. By considering that the error visibility thresholds are estimated using predicting and reconstructed signals of the color image, the coding scheme incorporated with locally adaptive perceptual color quantizer does not require side information. Experimental results show that the entropies of three color components obtained by using proposed IGM-based color image compression scheme are lower than that obtained by using the existing color image compression method at perceptually lossless visual quality.

Keywords: internal generative mechanism, structure-based spatial masking, visibility threshold, wavelet domain

Procedia PDF Downloads 247
10026 Assessment on the Collective Memory after Alteration of Urban Heritage: Case Study of Hengshan Mansions in Shanghai

Authors: Yueying Chen

Abstract:

A city can be developed through memory, and memory is one of the most important elements for urban contexts. Collective memory is a collection of personal memories that can be preserved with objects, places, and events of heritage, expressing culture through spatial changes. These preserved forms can evoke a sense of community and certain emotions. Collective memory in cities reflects urban spatial alterations and historical developments. It can be preserved and reflected by revitalisation projects. A major current focus in collective memory research is how to identify and preserve memory in an intangible way. The influential elements within the preservation of collective memory mainly include institutions and objects. However, current research lacks the assessment of the collective memory after alterations of urban heritage. The assessment of urban heritage lacks visualization and qualitative methods. The emergence of the application of space syntax can fill in this gap. Hengshan Mansions was a new project in 2015. The original residential area has been replaced with a comprehensive commercial area integrating boutique shopping, upscale restaurants, and creative offices. Hengshan Mansions is located in the largest historic area in Shanghai, and its development is the epitome of the traditional culture in Shanghai. Its alteration is the newest project in this area and presents the new concept of revitalisation of urban heritage. For its physical parts, modern vitality is created, and historical information is preserved at the same time. However, most of the local people are moved away, and its functions are altered a lot. The preservation of its collective memory needs to discuss furtherly. Thus, the article builds a framework to assess the collective memory of urban heritage, including spatial configuration, spatial interaction, and cultural cognition. Then, it selects Hengshan Mansions in Shanghai as a case to analyse the assessed framework. Space syntax can be applied to visualize the assessment. Based on the analysis, the article will explore the influential reasons for the collective memory after alterations and proposes relevant advice for the preservation of the collective memory of urban heritage.

Keywords: collective memory, alternation of urban heritage, space syntax, Hengshan Mansions

Procedia PDF Downloads 137
10025 HD-WSComp: Hypergraph Decomposition for Web Services Composition Based on QoS

Authors: Samah Benmerbi, Kamal Amroun, Abdelkamel Tari

Abstract:

The increasing number of Web service (WS)providers throughout the globe, have produced numerous Web services providing the same or similar functionality. Therefore, there is a need of tools developing the best answer of queries by selecting and composing services with total transparency. This paper reviews various QoS based Web service selection mechanisms and architectures which facilitate qualitatively optimal selection, in other fact Web service composition is required when a request cannot be fulfilled by a single web service. In such cases, it is preferable to integrate existing web services to satisfy user’s request. We introduce an automatic Web service composition method based on hypergraph decomposition using hypertree decomposition method. The problem of selection and the composition of the web services is transformed into a resolution in a hypertree by exploring the relations of dependency between web services to get composite web service via employing an execution order of WS satisfying global request.

Keywords: web service, web service selection, web service composition, QoS, hypergraph decomposition, BE hypergraph decomposition, hypertree resolution

Procedia PDF Downloads 507
10024 Exploring Artistic Creation and Autoethnography in the Spatial Context of Geography

Authors: Sinem Tas

Abstract:

This research paper attempts to study the perspective of personal experience in relation to spatial dynamics and artistic outcomes within the realm of cultural identity. This article serves as a partial analysis within a broader PhD investigation that focuses on the cultural dynamics and political structures behind cultural identity through an autoethnography of narrative while presenting its correlation with artistic creation in the context of space and people. Focusing on the artistic/creative practice project AUTRUI, the primary goal is to analyse and understand the influence of personal experiences and culturally constructed identity as an artist in resulting in the compositional modality of the last image considering self-reflective experience. Referencing the works of Joyce Davidson and Christine Milligan - the scholars who emphasise the importance of emotion and spatial experience in geographical studies contribute to this work as they highlight the significance of emotion across various spatial scales in their work Embodying Emotion Sensing Space: Introducing Emotional Geographies (2004). Their perspective suggests that understanding emotions within different spatial contexts is crucial for comprehending human experiences and interactions with space. Incorporating the insights of scholars like Yi-Fu Tuan, particularly his seminal work Space and Place: The Perspective of Experience (1979), is important for creating an in-depth frame of geographical experience. Tuan's humanistic perspective on space and place provides a valuable theoretical framework for understanding the interplay between personal experiences and spatial contexts. A substantial contextualisation of the geopolitics of Turkey - the implications for national identity and cohesion - will be addressed by drawing an outline of the political and geographical frame as a methodological strategy to understand the dynamics behind this research. Besides the bibliographical reading, the methods used to study this relation are participatory observation, memory work along with memoir analysis, personal interviews, and discussion of photographs and news. The utilisation of the self as data requires the analysis of the written sources with personal engagement. By delving into written sources such as written communications or diaries as well as memoirs, the research gains a firsthand perspective, enriching the analytical depth of the study. Furthermore, the examination of photography and news articles serves as a valuable means of contextualising experiences from a journalist's background within specific geographical settings. The inclusion of interviews with close family members access provides firsthand perspectives and intimate insights rooted in shared experiences within similar geographical contexts, offering complementary insights and diversified viewpoints, enhancing the comprehensiveness of the investigation.

Keywords: art, autoethnography, place and space, Turkey

Procedia PDF Downloads 48
10023 Research Methodology of Living Environment of Modern Residential Development in St. Petersburg

Authors: Kalina Alina Aidarovna, Khayrullina Yulia Sergeevna

Abstract:

The question of forming quality housing and living environment remains a vexed problem in the current situation of high-rise apartment building in big cities of Russia. At this start up stage of the modern so-called "mass housing" market it needs to identify key quality characteristics on a different scale from apartments to the district. This paper describes the methodology of qualitative assessment of modern mass housing construction, made on the basis of the ITMO university in cooperation with the institute of spatial planning "Urbanika," based on the case study of St. Petersburg’s residential mass housing built in 2011-2014. The methodology of the study of housing and living environment goes back to the native and foreign urbanists of 60s - 80s, such Jane Jacobs, Jan Gehl, Oscar Newman, Krasheninnikov, as well as Sommer, Stools, Kohnen and Sherrod, Krasilnikova, Sychev, Zhdanov, Tinyaeva considering spatial features of living environment in a wide range of its characteristics (environmental control, territorial and personalization, privacy, etc.). Assessment is carrying out on the proposed system of criteria developed for each residential environment scale-district, quarter, courtyard, building surrounding grounds, houses, and flats. Thus the objects of study are planning unit of residential development areas (residential area, neighborhood, quarter) residential units areas (living artist, a house), and households (apartments) consisting of residential units. As a product of identified methodology, after the results of case studies of more than 700 residential complexes in St. Petersburg, we intend the creation of affordable online resource that would allow conducting a detailed qualitative evaluation or comparative characteristics of residential complexes for all participants of the construction market-developers, designers, realtors and buyers. Thereby the main objective of the rating may be achieved to improve knowledge, requirements, and demand for quality housing and living environment among the major stakeholders of the construction market.

Keywords: methodology of living environment, qualitative assessment of mass housing, scale-district, vexed problem

Procedia PDF Downloads 458
10022 Spatial Interpolation of Aerosol Optical Depth Pollution: Comparison of Methods for the Development of Aerosol Distribution

Authors: Sahabeh Safarpour, Khiruddin Abdullah, Hwee San Lim, Mohsen Dadras

Abstract:

Air pollution is a growing problem arising from domestic heating, high density of vehicle traffic, electricity production, and expanding commercial and industrial activities, all increasing in parallel with urban population. Monitoring and forecasting of air quality parameters are important due to health impact. One widely available metric of aerosol abundance is the aerosol optical depth (AOD). The AOD is the integrated light extinction coefficient over a vertical atmospheric column of unit cross section, which represents the extent to which the aerosols in that vertical profile prevent the transmission of light by absorption or scattering. Seasonal aerosol optical depth (AOD) values at 550 nm derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor onboard NASA’s Terra satellites, for the 10 years period of 2000-2010 were used to test 7 different spatial interpolation methods in the present study. The accuracy of estimations was assessed through visual analysis as well as independent validation based on basic statistics, such as root mean square error (RMSE) and correlation coefficient. Based on the RMSE and R values of predictions made using measured values from 2000 to 2010, Radial Basis Functions (RBFs) yielded the best results for spring, summer, and winter and ordinary kriging yielded the best results for fall.

Keywords: aerosol optical depth, MODIS, spatial interpolation techniques, Radial Basis Functions

Procedia PDF Downloads 406
10021 Date Pits Oil Used as Potential Source for Synthesizing Jet Fuel and Green Diesel Fractions

Authors: Farrukh Jamil, Ala'a H. Al-Muhtaseb, Lamya Al-Haj, Mohab A. Al-Hinai

Abstract:

Date pits are major agricultural waste produced in Oman. Current work was conducted to produce jet fuel and green diesel from hydrodeoxygenation of Date pits oil in the presence of Pd/C catalyst. The hydrodeoxygenation of Date pits oil occurred to be highly efficient at following mild operating conditions such as conditions temperature 300°C pressure 10bar with continuous stirring at 500rpm. Detailed product characterization revealed that large fraction of paraffinic hydrocarbons was found which accounts up to 91.1 % which attributed due to efficient hydrodeoxygenation. Based on the type of components in product oil, it was calculated that the maximum fraction of hydrocarbons formed lies within the range of green diesel 72.0 % then jet fuel 30.4% by using Pd/C catalysts. The densities of product oil were 0.88 kg/m³, the viscosity of products calculated was 3.49 mm²/s. Calorific values for products obtained were 44.11 MJ/kg when Pd/C catalyst was used for hydrodeoxygenation. Based on products analysis it can conclude that Date pits oil could successfully utilize for synthesizing green diesel and jet fuel fraction.

Keywords: biomass, jet fuel, green diesel, catalyst

Procedia PDF Downloads 293
10020 Experimental Study on Flooding Phenomena in a Three-Phase Direct Contact Heat Exchanger for the Utilisation in Solar Pond Applications

Authors: Hameed B. Mahood, Ali Sh. Baqir, Alasdair N. Campbell

Abstract:

Experiments to study the limitation of flooding inception of three-phase direct contact condenser have been carried out in a counter-current small diameter vertical condenser. The total column height was 70 cm and 4 cm diameter. Only 48 cm has been used as an active three-phase direct contact condenser height. Vapour pentane with three different initial temperatures (40, 43.5 and 47.5 °C) and water with a constant temperature (19 °C) have been used as a dispersed phase and a continuous phase respectively. Five different continuous phase mass flow rate and four different dispersed phase mass flow rate have been tested throughout the experiments. Dimensionless correlation based on the previous common flooding correlation is proposed to calculate the up flow flooding inception of the three-phase direct contact condenser.

Keywords: Three-phase heat exchanger, condenser, solar energy, flooding phenomena

Procedia PDF Downloads 337
10019 Nest-Building Using Place Cells for Spatial Navigation in an Artificial Neural Network

Authors: Thomas E. Portegys

Abstract:

An animal behavior problem is presented in the form of a nest-building task that involves two cooperating virtual birds, a male and female. The female builds a nest into which she lays an egg. The male's job is to forage in a forest for food for both himself and the female. In addition, the male must fetch stones from a nearby desert for the female to use as nesting material. The task is completed when the nest is built, and an egg is laid in it. A goal-seeking neural network and a recurrent neural network were trained and tested with little success. The goal-seeking network was then enhanced with “place cells”, allowing the birds to spatially navigate the world, building the nest while keeping themselves fed. Place cells are neurons in the hippocampus that map space.

Keywords: artificial animal intelligence, artificial life, goal-seeking neural network, nest-building, place cells, spatial navigation

Procedia PDF Downloads 57
10018 Evaluating the Performance of Existing Full-Reference Quality Metrics on High Dynamic Range (HDR) Video Content

Authors: Maryam Azimi, Amin Banitalebi-Dehkordi, Yuanyuan Dong, Mahsa T. Pourazad, Panos Nasiopoulos

Abstract:

While there exists a wide variety of Low Dynamic Range (LDR) quality metrics, only a limited number of metrics are designed specifically for the High Dynamic Range (HDR) content. With the introduction of HDR video compression standardization effort by international standardization bodies, the need for an efficient video quality metric for HDR applications has become more pronounced. The objective of this study is to compare the performance of the existing full-reference LDR and HDR video quality metrics on HDR content and identify the most effective one for HDR applications. To this end, a new HDR video data set is created, which consists of representative indoor and outdoor video sequences with different brightness, motion levels and different representing types of distortions. The quality of each distorted video in this data set is evaluated both subjectively and objectively. The correlation between the subjective and objective results confirm that VIF quality metric outperforms all to their tested metrics in the presence of the tested types of distortions.

Keywords: HDR, dynamic range, LDR, subjective evaluation, video compression, HEVC, video quality metrics

Procedia PDF Downloads 522
10017 A Feasibility Study of Producing Biofuels from Textile Sludge by Torrefaction Technology

Authors: Hua-Shan Tai, Yu-Ting Zeng

Abstract:

In modern and industrial society, enormous amounts of sludge from various of industries are constantly produced; currently, most of the sludge are treated by landfill and incineration. However, both treatments are not ideal because of the limited land for landfill and the secondary pollution caused by incineration. Consequently, treating industrial sludge appropriately has become an urgent issue of environmental protection. In order to solve the problem of the massive sludge, this study uses textile sludge which is the major source of waste sludge in Taiwan as raw material for torrefaction treatments. To investigate the feasibility of producing biofuels from textile sludge by torrefaction, the experiments were conducted with temperatures at 150, 200, 250, 300, and 350°C, with heating rates of 15, 20, 25 and 30°C/min, and with residence time of 30 and 60 minutes. The results revealed that the mass yields after torrefaction were approximately in the range of 54.9 to 93.4%. The energy densification ratios were approximately in the range of 0.84 to 1.10, and the energy yields were approximately in the range of 45.9 to 98.3%. The volumetric densities were approximately in the range of 0.78 to 1.14, and the volumetric energy densities were approximately in the range of 0.65 to 1.18. To sum up, the optimum energy yield (98.3%) can be reached with terminal temperature at 150 °C, heating rate of 20°C/min, and residence time of 30 minutes, and the mass yield, energy densification ratio as well as volumetric energy density were 92.2%, 1.07, and 1.15, respectively. These results indicated that the solid products after torrefaction are easy to preserve, which not only enhance the quality of the product, but also achieve the purpose of developing the material into fuel.

Keywords: biofuel, biomass energy, textile sludge, torrefaction

Procedia PDF Downloads 319
10016 Bayesian Parameter Inference for Continuous Time Markov Chains with Intractable Likelihood

Authors: Randa Alharbi, Vladislav Vyshemirsky

Abstract:

Systems biology is an important field in science which focuses on studying behaviour of biological systems. Modelling is required to produce detailed description of the elements of a biological system, their function, and their interactions. A well-designed model requires selecting a suitable mechanism which can capture the main features of the system, define the essential components of the system and represent an appropriate law that can define the interactions between its components. Complex biological systems exhibit stochastic behaviour. Thus, using probabilistic models are suitable to describe and analyse biological systems. Continuous-Time Markov Chain (CTMC) is one of the probabilistic models that describe the system as a set of discrete states with continuous time transitions between them. The system is then characterised by a set of probability distributions that describe the transition from one state to another at a given time. The evolution of these probabilities through time can be obtained by chemical master equation which is analytically intractable but it can be simulated. Uncertain parameters of such a model can be inferred using methods of Bayesian inference. Yet, inference in such a complex system is challenging as it requires the evaluation of the likelihood which is intractable in most cases. There are different statistical methods that allow simulating from the model despite intractability of the likelihood. Approximate Bayesian computation is a common approach for tackling inference which relies on simulation of the model to approximate the intractable likelihood. Particle Markov chain Monte Carlo (PMCMC) is another approach which is based on using sequential Monte Carlo to estimate intractable likelihood. However, both methods are computationally expensive. In this paper we discuss the efficiency and possible practical issues for each method, taking into account the computational time for these methods. We demonstrate likelihood-free inference by performing analysing a model of the Repressilator using both methods. Detailed investigation is performed to quantify the difference between these methods in terms of efficiency and computational cost.

Keywords: Approximate Bayesian computation(ABC), Continuous-Time Markov Chains, Sequential Monte Carlo, Particle Markov chain Monte Carlo (PMCMC)

Procedia PDF Downloads 201
10015 Towards a Better Understanding of Planning for Urban Intensification: Case Study of Auckland, New Zealand

Authors: Wen Liu, Errol Haarhoff, Lee Beattie

Abstract:

In 2010, New Zealand’s central government re-organise the local governments arrangements in Auckland, New Zealand by amalgamating its previous regional council and seven supporting local government units into a single unitary council, the Auckland Council. The Auckland Council is charged with providing local government services to approximately 1.5 million people (a third of New Zealand’s total population). This includes addressing Auckland’s strategic urban growth management and setting its urban planning policy directions for the next 40 years. This is expressed in the first ever spatial plan in the region – the Auckland Plan (2012). The Auckland plan supports implementing a compact city model by concentrating the larger part of future urban growth and development in, and around, existing and proposed transit centres, with the intention of Auckland to become globally competitive city and achieving ‘the most liveable city in the world’. Turning that vision into reality is operatized through the statutory land use plan, the Auckland Unitary Plan. The Unitary plan replaced the previous regional and local statutory plans when it became operative in 2016, becoming the ‘rule book’ on how to manage and develop the natural and built environment, using land use zones and zone standards. Common to the broad range of literature on urban growth management, one significant issue stands out about intensification. The ‘gap’ between strategic planning and what has been achieved is evident in the argument for the ‘compact’ urban form. Although the compact city model may have a wide range of merits, the extent to which these are actualized largely rely on how intensification actually is delivered. The transformation of the rhetoric of the residential intensification model into reality is of profound influence, yet has enjoyed limited empirical analysis. In Auckland, the establishment of the Auckland Plan set up the strategies to deliver intensification into diversified arenas. Nonetheless, planning policy itself does not necessarily achieve the envisaged objectives, delivering the planning system and high capacity to enhance and sustain plan implementation is another demanding agenda. Though the Auckland Plan provides a wide ranging strategic context, its actual delivery is beholden on the Unitary Plan. However, questions have been asked if the Unitary Plan has the necessary statutory tools to deliver the Auckland Plan’s policy outcomes. In Auckland, there is likely to be continuing tension between the strategies for intensification and their envisaged objectives, and made it doubtful whether the main principles of the intensification strategies could be realized. This raises questions over whether the Auckland Plan’s policy goals can be achieved in practice, including delivering ‘quality compact city’ and residential intensification. Taking Auckland as an example of traditionally sprawl cities, this article intends to investigate the efficacy plan making and implementation directed towards higher density development. This article explores the process of plan development, plan making and implementation frameworks of the first ever spatial plan in Auckland, so as to explicate the objectives and processes involved, and consider whether this will facilitate decision making processes to realize the anticipated intensive urban development.

Keywords: urban intensification, sustainable development, plan making, governance and implementation

Procedia PDF Downloads 555
10014 Identification of High-Rise Buildings Using Object Based Classification and Shadow Extraction Techniques

Authors: Subham Kharel, Sudha Ravindranath, A. Vidya, B. Chandrasekaran, K. Ganesha Raj, T. Shesadri

Abstract:

Digitization of urban features is a tedious and time-consuming process when done manually. In addition to this problem, Indian cities have complex habitat patterns and convoluted clustering patterns, which make it even more difficult to map features. This paper makes an attempt to classify urban objects in the satellite image using object-oriented classification techniques in which various classes such as vegetation, water bodies, buildings, and shadows adjacent to the buildings were mapped semi-automatically. Building layer obtained as a result of object-oriented classification along with already available building layers was used. The main focus, however, lay in the extraction of high-rise buildings using spatial technology, digital image processing, and modeling, which would otherwise be a very difficult task to carry out manually. Results indicated a considerable rise in the total number of buildings in the city. High-rise buildings were successfully mapped using satellite imagery, spatial technology along with logical reasoning and mathematical considerations. The results clearly depict the ability of Remote Sensing and GIS to solve complex problems in urban scenarios like studying urban sprawl and identification of more complex features in an urban area like high-rise buildings and multi-dwelling units. Object-Oriented Technique has been proven to be effective and has yielded an overall efficiency of 80 percent in the classification of high-rise buildings.

Keywords: object oriented classification, shadow extraction, high-rise buildings, satellite imagery, spatial technology

Procedia PDF Downloads 154
10013 Impact of Light Intensity, Illumation Strategy and Self-Shading on Sustainable Algal Growth in Photo Bioreactors

Authors: Amritanshu Shriwastav, Purnendu Bose

Abstract:

Algal photo bioreactors were operated at incident light intensities of 0.24, 2.52 and 5.96 W L-1 to determine the impact of light on algal growth. Low specific Chlorophyll-a content of algae was a strong indicator of light induced stress on algal cells. It was concluded that long term operation of photo bioreactors in the continuous illumination mode was infeasible under the range of incident light intensities examined and provision of a dark period after each light period was necessary for algal cells to recover from light-induced stress. Long term operation of photo bioreactors in the intermittent illumination mode was however possible at light intensities of 0.24 and 2.52 W L-1. Further, the incident light intensity in the photo bioreactors was found to decline exponentially with increase in algal concentration in the reactor due to algal ‘self-shading’. This may be an important determinant for photo bioreactor performance at higher algal concentrations.

Keywords: Algae, algal growth, photo bioreactor, photo-inhibition, ‘self-shading’

Procedia PDF Downloads 317
10012 Mixed Model Sequencing in Painting Production Line

Authors: Unchalee Inkampa, Tuanjai Somboonwiwat

Abstract:

Painting process of automobiles and automobile parts, which is a continuous process based on EDP (Electrode position paint, EDP). Through EDP, all work pieces will be continuously sent to the painting process. Work process can be divided into 2 groups based on the running time: Painting Room 1 and Painting Room 2. This leads to continuous operation. The problem that arises is waiting for workloads onto Painting Room. The grading process EDP to Painting Room is a major problem. Therefore, this paper aim to develop production sequencing method by applying EDP to painting process. It also applied fixed rate launching for painting room and earliest due date (EDD) for EDP process and swap pairwise interchange for waiting time to a minimum of machine. The result found that the developed method could improve painting reduced waiting time, on time delivery, meeting customers wants and improved productivity of painting unit.

Keywords: sequencing, mixed model lines, painting process, electrode position paint

Procedia PDF Downloads 418
10011 An Algorithm of Set-Based Particle Swarm Optimization with Status Memory for Traveling Salesman Problem

Authors: Takahiro Hino, Michiharu Maeda

Abstract:

Particle swarm optimization (PSO) is an optimization approach that achieves the social model of bird flocking and fish schooling. PSO works in continuous space and can solve continuous optimization problem with high quality. Set-based particle swarm optimization (SPSO) functions in discrete space by using a set. SPSO can solve combinatorial optimization problem with high quality and is successful to apply to the large-scale problem. In this paper, we present an algorithm of SPSO with status memory to decide the position based on the previous position for solving traveling salesman problem (TSP). In order to show the effectiveness of our approach. We examine SPSOSM for TSP compared to the existing algorithms.

Keywords: combinatorial optimization problems, particle swarm optimization, set-based particle swarm optimization, traveling salesman problem

Procedia PDF Downloads 550
10010 Evaluation of Low Power Wi-Fi Modules in Simulated Ocean Environments

Authors: Gabriel Chenevert, Abhilash Arora, Zeljko Pantic

Abstract:

The major problem underwater acoustic communication faces is the low data rate due to low signal frequency. By contrast, the Wi-Fi communication protocol offers high throughput but limited operating range due to the attenuation effect of the sea and ocean medium. However, short-range near-field underwater wireless power transfer systems offer an environment where Wi-Fi communication can be effectively integrated to collect data and deliver instructions to sensors in underwater sensor networks. In this paper, low-power, low-cost off-the-shelf Wi-Fi modules are explored experimentally for four selected parameters for different distances between units and water salinities. The results reveal a shorter operating range and stronger dependence on water salinity than reported so far for high-end Wi-Fi modules.

Keywords: Wi-Fi, wireless power transfer, underwater communications, ESP

Procedia PDF Downloads 115
10009 Spatial Rank-Based High-Dimensional Monitoring through Random Projection

Authors: Chen Zhang, Nan Chen

Abstract:

High-dimensional process monitoring becomes increasingly important in many application domains, where usually the process distribution is unknown and much more complicated than the normal distribution, and the between-stream correlation can not be neglected. However, since the process dimension is generally much bigger than the reference sample size, most traditional nonparametric multivariate control charts fail in high-dimensional cases due to the curse of dimensionality. Furthermore, when the process goes out of control, the influenced variables are quite sparse compared with the whole dimension, which increases the detection difficulty. Targeting at these issues, this paper proposes a new nonparametric monitoring scheme for high-dimensional processes. This scheme first projects the high-dimensional process into several subprocesses using random projections for dimension reduction. Then, for every subprocess with the dimension much smaller than the reference sample size, a local nonparametric control chart is constructed based on the spatial rank test to detect changes in this subprocess. Finally, the results of all the local charts are fused together for decision. Furthermore, after an out-of-control (OC) alarm is triggered, a diagnostic framework is proposed. using the square-root LASSO. Numerical studies demonstrate that the chart has satisfactory detection power for sparse OC changes and robust performance for non-normally distributed data, The diagnostic framework is also effective to identify truly changed variables. Finally, a real-data example is presented to demonstrate the application of the proposed method.

Keywords: random projection, high-dimensional process control, spatial rank, sequential change detection

Procedia PDF Downloads 297
10008 Depletion Behavior of Potassium by Continuous Cropping Using Rice as a Test Crop

Authors: Rafeza Begum, Mohammad Mokhlesur Rahman, Safikul Moula, Rafiqul Islam

Abstract:

Potassium (K) is crucial for healthy soil and plant growth. However, K fertilization is either disregarded or poorly underutilized in Bangladesh agriculture, despite the great demand for crops. This could eventually result in a significant depletion of the soil's potassium reserves, irreversible alteration of the minerals that contain potassium, and detrimental effects on crop productivity. Soil K mining in Bangladesh is a worrying problem, and we need to evaluate it thoroughly and find remedies. A pot culture experiment was conducted in the greenhouse of Bangladesh Institute of Nuclear Agriculture (BINA) using eleven soil series of Bangladesh in order to see the depletion behaviour of potassium (K) by continuous cropping using rice (var. Iratom-24) as the test crop. The soil series were Ranishankhail, Kaonia. Sonatala, Silmondi, Gopalpur, Ishurdi, Sara, Kongsha, Nunni, Lauta and Amnura on which four successive rice plants (45 days duration) were raised with (100 ppm K) or without addition of potassium. Nitrogen, phosphorus, sulfur and zinc were applied as basal to all pots. Potassium application resulted in higher dry matter yield, increased K concentration and uptake in all the soils compared with no K treatment; which gradually decreased in the subsequent harvests. Furthermore, plant takes up K not only from exchangeable pool but also from non-exchangeable sites and a minimum replenishment of K from the soil reserve was observed. Continuous cropping has resulted in the depletion of available K of the soil. The result indicated that in order to sustain higher crop yield under intensive cultivation, the addition of potash fertilizer is necessary.

Keywords: potassium, exchangeable pool, depletion behavior., Soil series

Procedia PDF Downloads 125
10007 A Two-Stage Bayesian Variable Selection Method with the Extension of Lasso for Geo-Referenced Data

Authors: Georgiana Onicescu, Yuqian Shen

Abstract:

Due to the complex nature of geo-referenced data, multicollinearity of the risk factors in public health spatial studies is a commonly encountered issue, which leads to low parameter estimation accuracy because it inflates the variance in the regression analysis. To address this issue, we proposed a two-stage variable selection method by extending the least absolute shrinkage and selection operator (Lasso) to the Bayesian spatial setting, investigating the impact of risk factors to health outcomes. Specifically, in stage I, we performed the variable selection using Bayesian Lasso and several other variable selection approaches. Then, in stage II, we performed the model selection with only the selected variables from stage I and compared again the methods. To evaluate the performance of the two-stage variable selection methods, we conducted a simulation study with different distributions for the risk factors, using geo-referenced count data as the outcome and Michigan as the research region. We considered the cases when all candidate risk factors are independently normally distributed, or follow a multivariate normal distribution with different correlation levels. Two other Bayesian variable selection methods, Binary indicator, and the combination of Binary indicator and Lasso were considered and compared as alternative methods. The simulation results indicated that the proposed two-stage Bayesian Lasso variable selection method has the best performance for both independent and dependent cases considered. When compared with the one-stage approach, and the other two alternative methods, the two-stage Bayesian Lasso approach provides the highest estimation accuracy in all scenarios considered.

Keywords: Lasso, Bayesian analysis, spatial analysis, variable selection

Procedia PDF Downloads 142
10006 Validation of Mapping Historical Linked Data to International Committee for Documentation (CIDOC) Conceptual Reference Model Using Shapes Constraint Language

Authors: Ghazal Faraj, András Micsik

Abstract:

Shapes Constraint Language (SHACL), a World Wide Web Consortium (W3C) language, provides well-defined shapes and RDF graphs, named "shape graphs". These shape graphs validate other resource description framework (RDF) graphs which are called "data graphs". The structural features of SHACL permit generating a variety of conditions to evaluate string matching patterns, value type, and other constraints. Moreover, the framework of SHACL supports high-level validation by expressing more complex conditions in languages such as SPARQL protocol and RDF Query Language (SPARQL). SHACL includes two parts: SHACL Core and SHACL-SPARQL. SHACL Core includes all shapes that cover the most frequent constraint components. While SHACL-SPARQL is an extension that allows SHACL to express more complex customized constraints. Validating the efficacy of dataset mapping is an essential component of reconciled data mechanisms, as the enhancement of different datasets linking is a sustainable process. The conventional validation methods are the semantic reasoner and SPARQL queries. The former checks formalization errors and data type inconsistency, while the latter validates the data contradiction. After executing SPARQL queries, the retrieved information needs to be checked manually by an expert. However, this methodology is time-consuming and inaccurate as it does not test the mapping model comprehensively. Therefore, there is a serious need to expose a new methodology that covers the entire validation aspects for linking and mapping diverse datasets. Our goal is to conduct a new approach to achieve optimal validation outcomes. The first step towards this goal is implementing SHACL to validate the mapping between the International Committee for Documentation (CIDOC) conceptual reference model (CRM) and one of its ontologies. To initiate this project successfully, a thorough understanding of both source and target ontologies was required. Subsequently, the proper environment to run SHACL and its shape graphs were determined. As a case study, we performed SHACL over a CIDOC-CRM dataset after running a Pellet reasoner via the Protégé program. The applied validation falls under multiple categories: a) data type validation which constrains whether the source data is mapped to the correct data type. For instance, checking whether a birthdate is assigned to xsd:datetime and linked to Person entity via crm:P82a_begin_of_the_begin property. b) Data integrity validation which detects inconsistent data. For instance, inspecting whether a person's birthdate occurred before any of the linked event creation dates. The expected results of our work are: 1) highlighting validation techniques and categories, 2) selecting the most suitable techniques for those various categories of validation tasks. The next plan is to establish a comprehensive validation model and generate SHACL shapes automatically.

Keywords: SHACL, CIDOC-CRM, SPARQL, validation of ontology mapping

Procedia PDF Downloads 251
10005 Volunteered Geographic Information Coupled with Wildfire Fire Progression Maps: A Spatial and Temporal Tool for Incident Storytelling

Authors: Cassandra Hansen, Paul Doherty, Chris Ferner, German Whitley, Holly Torpey

Abstract:

Wildfire is a natural and inevitable occurrence, yet changing climatic conditions have increased the severity, frequency, and risk to human populations in the wildland/urban interface (WUI) of the Western United States. Rapid dissemination of accurate wildfire information is critical to both the Incident Management Team (IMT) and the affected community. With the advent of increasingly sophisticated information systems, GIS can now be used as a web platform for sharing geographic information in new and innovative ways, such as virtual story map applications. Crowdsourced information can be extraordinarily useful when coupled with authoritative information. Information abounds in the form of social media, emergency alerts, radio, and news outlets, yet many of these resources lack a spatial component when first distributed. In this study, we describe how twenty-eight volunteer GIS professionals across nine Geographic Area Coordination Centers (GACC) sourced, curated, and distributed Volunteered Geographic Information (VGI) from authoritative social media accounts focused on disseminating information about wildfires and public safety. The combination of fire progression maps with VGI incident information helps answer three critical questions about an incident, such as: where the first started. How and why the fire behaved in an extreme manner and how we can learn from the fire incident's story to respond and prepare for future fires in this area. By adding a spatial component to that shared information, this team has been able to visualize shared information about wildfire starts in an interactive map that answers three critical questions in a more intuitive way. Additionally, long-term social and technical impacts on communities are examined in relation to situational awareness of the disaster through map layers and agency links, the number of views in a particular region of a disaster, community involvement and sharing of this critical resource. Combined with a GIS platform and disaster VGI applications, this workflow and information become invaluable to communities within the WUI and bring spatial awareness for disaster preparedness, response, mitigation, and recovery. This study highlights progression maps as the ultimate storytelling mechanism through incident case studies and demonstrates the impact of VGI and sophisticated applied cartographic methodology make this an indispensable resource for authoritative information sharing.

Keywords: storytelling, wildfire progression maps, volunteered geographic information, spatial and temporal

Procedia PDF Downloads 175
10004 Combining ASTER Thermal Data and Spatial-Based Insolation Model for Identification of Geothermal Active Areas

Authors: Khalid Hussein, Waleed Abdalati, Pakorn Petchprayoon, Khaula Alkaabi

Abstract:

In this study, we integrated ASTER thermal data with an area-based spatial insolation model to identify and delineate geothermally active areas in Yellowstone National Park (YNP). Two pairs of L1B ASTER day- and nighttime scenes were used to calculate land surface temperature. We employed the Emissivity Normalization Algorithm which separates temperature from emissivity to calculate surface temperature. We calculated the incoming solar radiation for the area covered by each of the four ASTER scenes using an insolation model and used this information to compute temperature due to solar radiation. We then identified the statistical thermal anomalies using land surface temperature and the residuals calculated from modeled temperatures and ASTER-derived surface temperatures. Areas that had temperatures or temperature residuals greater than 2σ and between 1σ and 2σ were considered ASTER-modeled thermal anomalies. The areas identified as thermal anomalies were in strong agreement with the thermal areas obtained from the YNP GIS database. Also the YNP hot springs and geysers were located within areas identified as anomalous thermal areas. The consistency between our results and known geothermally active areas indicate that thermal remote sensing data, integrated with a spatial-based insolation model, provides an effective means for identifying and locating areas of geothermal activities over large areas and rough terrain.

Keywords: thermal remote sensing, insolation model, land surface temperature, geothermal anomalies

Procedia PDF Downloads 370
10003 Structured-Ness and Contextual Retrieval Underlie Language Comprehension

Authors: Yao-Ying Lai, Maria Pinango, Ashwini Deo

Abstract:

While grammatical devices are essential to language processing, how comprehension utilizes cognitive mechanisms is less emphasized. This study addresses this issue by probing the complement coercion phenomenon: an entity-denoting complement following verbs like begin and finish receives an eventive interpretation. For example, (1) “The queen began the book” receives an agentive reading like (2) “The queen began [reading/writing/etc.…] the book.” Such sentences engender additional processing cost in real-time comprehension. The traditional account attributes this cost to an operation that coerces the entity-denoting complement to an event, assuming that these verbs require eventive complements. However, in closer examination, examples like “Chapter 1 began the book” undermine this assumption. An alternative, Structured Individual (SI) hypothesis, proposes that the complement following aspectual verbs (AspV; e.g. begin, finish) is conceptualized as a structured individual, construed as an axis along various dimensions (e.g. spatial, eventive, temporal, informational). The composition of an animate subject and an AspV such as (1) engenders an ambiguity between an agentive reading along the eventive dimension like (2), and a constitutive reading along the informational/spatial dimension like (3) “[The story of the queen] began the book,” in which the subject is interpreted as a subpart of the complement denotation. Comprehenders need to resolve the ambiguity by searching contextual information, resulting in additional cost. To evaluate the SI hypothesis, a questionnaire was employed. Method: Target AspV sentences such as “Shakespeare began the volume.” were preceded by one of the following types of context sentence: (A) Agentive-biasing, in which an event was mentioned (…writers often read…), (C) Constitutive-biasing, in which a constitutive meaning was hinted (Larry owns collections of Renaissance literature.), (N) Neutral context, which allowed both interpretations. Thirty-nine native speakers of English were asked to (i) rate each context-target sentence pair from a 1~5 scale (5=fully understandable), and (ii) choose possible interpretations for the target sentence given the context. The SI hypothesis predicts that comprehension is harder for the Neutral condition, as compared to the biasing conditions because no contextual information is provided to resolve an ambiguity. Also, comprehenders should obtain the specific interpretation corresponding to the context type. Results: (A) Agentive-biasing and (C) Constitutive-biasing were rated higher than (N) Neutral conditions (p< .001), while all conditions were within the acceptable range (> 3.5 on the 1~5 scale). This suggests that when lacking relevant contextual information, semantic ambiguity decreases comprehensibility. The interpretation task shows that the participants selected the biased agentive/constitutive reading for condition (A) and (C) respectively. For the Neutral condition, the agentive and constitutive readings were chosen equally often. Conclusion: These findings support the SI hypothesis: the meaning of AspV sentences is conceptualized as a parthood relation involving structured individuals. We argue that semantic representation makes reference to spatial structured-ness (abstracted axis). To obtain an appropriate interpretation, comprehenders utilize contextual information to enrich the conceptual representation of the sentence in question. This study connects semantic structure to human’s conceptual structure, and provides a processing model that incorporates contextual retrieval.

Keywords: ambiguity resolution, contextual retrieval, spatial structured-ness, structured individual

Procedia PDF Downloads 331
10002 Spatial Analysis of Survival Pattern and Treatment Outcomes of Multi-Drug Resistant Tuberculosis (MDR-TB) Patients in Lagos, Nigeria

Authors: Akinsola Oluwatosin, Udofia Samuel, Odofin Mayowa

Abstract:

The study is aimed at assessing the Geographic Information System (GIS)-based spatial analysis of Survival Pattern and Treatment Outcomes of Multi-Drug Resistant Tuberculosis (MDR-TB) cases for Lagos, Nigeria, with an objective to inform priority areas for public health planning and resource allocation. Multi-drug resistant tuberculosis (MDR-TB) develops due to problems such as irregular drug supply, poor drug quality, inappropriate prescription, and poor adherence to treatment. The shapefile(s) for this study were already georeferenced to Minna datum. The patient’s information was acquired on MS Excel and later converted to . CSV file for easy processing to ArcMap from various hospitals. To superimpose the patient’s information the spatial data, the addresses was geocoded to generate the longitude and latitude of the patients. The database was used for the SQL query to the various pattern of the treatment. To show the pattern of disease spread, spatial autocorrelation analysis was used. The result was displayed in a graphical format showing the areas of dispersing, random and clustered of patients in the study area. Hot and cold spot analysis was analyzed to show high-density areas. The distance between these patients and the closest health facility was examined using the buffer analysis. The result shows that 22% of the points were successfully matched, while 15% were tied. However, the result table shows that a greater percentage of it was unmatched; this is evident in the fact that most of the streets within the State are unnamed, and then again, most of the patients are likely to supply the wrong addresses. MDR-TB patients of all age groups are concentrated within Lagos-Mainland, Shomolu, Mushin, Surulere, Oshodi-Isolo, and Ifelodun LGAs. MDR-TB patients between the age group of 30-47 years had the highest number and were identified to be about 184 in number. The outcome of patients on ART treatment revealed that a high number of patients (300) were not ART treatment while a paltry 45 patients were on ART treatment. The result shows the Z-score of the distribution is greater than 1 (>2.58), which means that the distribution is highly clustered at a significance level of 0.01.

Keywords: tuberculosis, patients, treatment, GIS, MDR-TB

Procedia PDF Downloads 152
10001 Time-Frequency Modelling and Analysis of Faulty Rotor

Authors: B. X. Tchomeni, A. A. Alugongo, T. B. Tengen

Abstract:

In this paper, de Laval rotor system has been characterized by a hinge model and its transient response numerically treated for a dynamic solution. The effect of the ensuing non-linear disturbances namely rub and breathing crack is numerically simulated. Subsequently, three analysis methods: Orbit Analysis, Fast Fourier Transform (FFT) and Wavelet Transform (WT) are employed to extract features of the vibration signal of the faulty system. An analysis of the system response orbits clearly indicates the perturbations due to the rotor-to-stator contact. The sensitivities of WT to the variation in system speed have been investigated by Continuous Wavelet Transform (CWT). The analysis reveals that features of crack, rubs and unbalance in vibration response can be useful for condition monitoring. WT reveals its ability to detect non-linear signal, and obtained results provide a useful tool method for detecting machinery faults.

Keywords: Continuous wavelet, crack, discrete wavelet, high acceleration, low acceleration, nonlinear, rotor-stator, rub

Procedia PDF Downloads 344
10000 Anaerobic Co-digestion of the Halophyte Salicornia Ramosissima and Pig Manure in Lab-Scale Batch and Semi-continuous Stirred Tank Reactors: Biomethane Production and Reactor Performance

Authors: Aadila Cayenne, Hinrich Uellendahl

Abstract:

Optimization of the anaerobic digestion (AD) process of halophytic plants is essential as the biomass contains a high salt content that can inhibit the AD process. Anaerobic co-digestion, together with manure, can resolve the inhibitory effects of saline biomass in order to dilute the salt concentration and establish favorable conditions for the microbial consortia of the AD process. The present laboratory study investigated the co-digestion of S. ramosissima (Sram), and pig manure (PM) in batch and semi-continuous stirred tank reactors (CSTR) under mesophilic (38oC) conditions. The 0.5L batch reactor experiments were in mono- and co-digestion of Sram: PM using different percent volatile solid (VS) based ratios (0:100, 15:85, 25:75, 35:65, 50:50, 100:0) with an inoculum to substate (I/R) ratio of 2. Two 5L CSTR systems (R1 and R2) were operated for 133 days with a feed of PM in a control reactor (R1) and with a co-digestion feed in an increasing Sram VS ratio of Sram: PM of 15:85, 25:75, 35:65 in reactor R2 at an organic loading rate (OLR) of 2 gVS/L/d and hydraulic retention time (HRT) of 20 days. After a start-up phase of 8 weeks for both reactors R1 and R2 with PM feed alone, the halophyte biomass Sram was added to the feed of R2 in an increasing ratio of 15 – 35 %VS Sram over an 11-week period. The process performance was monitored by pH, total solid (TS), VS, total nitrogen (TN), ammonium-nitrogen (NH4 – N), volatile fatty acids (VFA), and biomethane production. In the batch experiments, biomethane yields of 423, 418, 392, 365, 315, and 214 mL-CH4/gVS were achieved for mixtures of 0:100, 15:85, 25:75, 35:65, 50:50, 100:0 %VS Sram: PM, respectively. In the semi-continuous reactor processes, the average biomethane yields were 235, 387, and 365 mL-CH4/gVS for the phase of a co-digestion feed ratio in R2 of 15:85, 25:75, and 35:65 %VS Sram: PM, respectively. The methane yield of PM alone in R1 was in the corresponding phases on average 260, 388, and 446 mL-CH4/gVS. Accordingly, in the continuous AD process, the methane yield of the halophyte Sram was highest at 386 mL-CH4/gVS in the co-digestion ratio of 25:75%VS Sram: PM and significantly lower at 15:85 %VS Sram: PM (100 mL-CH4/gVS) and at 35:65 %VS Sram (214 mL-CH4/gVS). The co-digestion process showed no signs of inhibition at 2 – 4 g/L NH4 – N, 3.5 – 4.5 g/L TN, and total VFA of 0.45 – 2.6 g/L (based on Acetic, Propionic, Butyric and Valeric acid). This study demonstrates that a stable co-digestion process of S. ramosissima and pig manure can be achieved with a feed of 25%VS Sram at HRT of 20 d and OLR of 2 gVS/L/d.

Keywords: anaerobic co-digestion, biomethane production, halophytes, pig manure, salicornia ramosissima

Procedia PDF Downloads 150