Search results for: compressible boundary layers
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2698

Search results for: compressible boundary layers

28 Long-Term Tillage, Lime Matter and Cover Crop Effects under Heavy Soil Conditions in Northern Lithuania

Authors: Aleksandras Velykis, Antanas Satkus

Abstract:

Clay loam and clay soils are typical for northern Lithuania. These soils are susceptible to physical degradation in the case of intensive use of heavy machinery for field operations. However, clayey soils having poor physical properties by origin require more intensive tillage to maintain proper physical condition for grown crops. Therefore not only choice of suitable tillage system is very important for these soils in the region, but also additional search of other measures is essential for good soil physical state maintenance. Research objective: To evaluate the long-term effects of different intensity tillage as well as its combinations with supplementary agronomic practices on improvement of soil physical conditions and environmental sustainability. The experiment examined the influence of deep and shallow ploughing, ploughless tillage, combinations of ploughless tillage with incorporation of lime sludge and cover crop for green manure and application of the same cover crop for mulch without autumn tillage under spring and winter crop growing conditions on clay loam (27% clay, 50% silt, 23% sand) Endocalcaric Endogleyic Cambisol. Methods: The indicators characterizing the impact of investigated measures were determined using the following methods and devices: Soil dry bulk density – by Eijkelkamp cylinder (100 cm3), soil water content – by weighing, soil structure – by Retsch sieve shaker, aggregate stability – by Eijkelkamp wet sieving apparatus, soil mineral nitrogen – in 1 N KCL extract using colorimetric method. Results: Clay loam soil physical state (dry bulk density, structure, aggregate stability, water content) depends on tillage system and its combination with additional practices used. Application of cover crop winter mulch without tillage in autumn, ploughless tillage and shallow ploughing causes the compaction of bottom (15-25 cm) topsoil layer. However, due to ploughless tillage the soil dry bulk density in subsoil (25-35 cm) layer is less compared to deep ploughing. Soil structure in the upper (0-15 cm) topsoil layer and in the seedbed (0-5 cm), prepared for spring crops is usually worse when applying the ploughless tillage or cover crop mulch without autumn tillage. Application of lime sludge under ploughless tillage conditions helped to avoid the compaction and structure worsening in upper topsoil layer, as well as increase aggregate stability. Application of reduced tillage increased soil water content at upper topsoil layer directly after spring crop sowing. However, due to reduced tillage the water content in all topsoil markedly decreased when droughty periods lasted for a long time. Combination of reduced tillage with cover crop for green manure and winter mulch is significant for preserving the environment. Such application of cover crops reduces the leaching of mineral nitrogen into the deeper soil layers and environmental pollution. This work was supported by the National Science Program ‘The effect of long-term, different-intensity management of resources on the soils of different genesis and on other components of the agro-ecosystems’ [grant number SIT-9/2015] funded by the Research Council of Lithuania.

Keywords: clay loam, endocalcaric endogleyic cambisol, mineral nitrogen, physical state

Procedia PDF Downloads 226
27 Geovisualization of Human Mobility Patterns in Los Angeles Using Twitter Data

Authors: Linna Li

Abstract:

The capability to move around places is doubtless very important for individuals to maintain good health and social functions. People’s activities in space and time have long been a research topic in behavioral and socio-economic studies, particularly focusing on the highly dynamic urban environment. By analyzing groups of people who share similar activity patterns, many socio-economic and socio-demographic problems and their relationships with individual behavior preferences can be revealed. Los Angeles, known for its large population, ethnic diversity, cultural mixing, and entertainment industry, faces great transportation challenges such as traffic congestion, parking difficulties, and long commuting. Understanding people’s travel behavior and movement patterns in this metropolis sheds light on potential solutions to complex problems regarding urban mobility. This project visualizes people’s trajectories in Greater Los Angeles (L.A.) Area over a period of two months using Twitter data. A Python script was used to collect georeferenced tweets within the Greater L.A. Area including Ventura, San Bernardino, Riverside, Los Angeles, and Orange counties. Information associated with tweets includes text, time, location, and user ID. Information associated with users includes name, the number of followers, etc. Both aggregated and individual activity patterns are demonstrated using various geovisualization techniques. Locations of individual Twitter users were aggregated to create a surface of activity hot spots at different time instants using kernel density estimation, which shows the dynamic flow of people’s movement throughout the metropolis in a twenty-four-hour cycle. In the 3D geovisualization interface, the z-axis indicates time that covers 24 hours, and the x-y plane shows the geographic space of the city. Any two points on the z axis can be selected for displaying activity density surface within a particular time period. In addition, daily trajectories of Twitter users were created using space-time paths that show the continuous movement of individuals throughout the day. When a personal trajectory is overlaid on top of ancillary layers including land use and road networks in 3D visualization, the vivid representation of a realistic view of the urban environment boosts situational awareness of the map reader. A comparison of the same individual’s paths on different days shows some regular patterns on weekdays for some Twitter users, but for some other users, their daily trajectories are more irregular and sporadic. This research makes contributions in two major areas: geovisualization of spatial footprints to understand travel behavior using the big data approach and dynamic representation of activity space in the Greater Los Angeles Area. Unlike traditional travel surveys, social media (e.g., Twitter) provides an inexpensive way of data collection on spatio-temporal footprints. The visualization techniques used in this project are also valuable for analyzing other spatio-temporal data in the exploratory stage, thus leading to informed decisions about generating and testing hypotheses for further investigation. The next step of this research is to separate users into different groups based on gender/ethnic origin and compare their daily trajectory patterns.

Keywords: geovisualization, human mobility pattern, Los Angeles, social media

Procedia PDF Downloads 118
26 Development and Experimental Validation of Coupled Flow-Aerosol Microphysics Model for Hot Wire Generator

Authors: K. Ghosh, S. N. Tripathi, Manish Joshi, Y. S. Mayya, Arshad Khan, B. K. Sapra

Abstract:

We have developed a CFD coupled aerosol microphysics model in the context of aerosol generation from a glowing wire. The governing equations can be solved implicitly for mass, momentum, energy transfer along with aerosol dynamics. The computationally efficient framework can simulate temporal behavior of total number concentration and number size distribution. This formulation uniquely couples standard K-Epsilon scheme with boundary layer model with detailed aerosol dynamics through residence time. This model uses measured temperatures (wire surface and axial/radial surroundings) and wire compositional data apart from other usual inputs for simulations. The model predictions show that bulk fluid motion and local heat distribution can significantly affect the aerosol behavior when the buoyancy effect in momentum transfer is considered. Buoyancy generated turbulence was found to be affecting parameters related to aerosol dynamics and transport as well. The model was validated by comparing simulated predictions with results obtained from six controlled experiments performed with a laboratory-made hot wire nanoparticle generator. Condensation particle counter (CPC) and scanning mobility particle sizer (SMPS) were used for measurement of total number concentration and number size distribution at the outlet of reactor cell during these experiments. Our model-predicted results were found to be in reasonable agreement with observed values. The developed model is fast (fully implicit) and numerically stable. It can be used specifically for applications in the context of the behavior of aerosol particles generated from glowing wire technique and in general for other similar large scale domains. Incorporation of CFD in aerosol microphysics framework provides a realistic platform to study natural convection driven systems/ applications. Aerosol dynamics sub-modules (nucleation, coagulation, wall deposition) have been coupled with Navier Stokes equations modified to include buoyancy coupled K-Epsilon turbulence model. Coupled flow-aerosol dynamics equation was solved numerically and in the implicit scheme. Wire composition and temperature (wire surface and cell domain) were obtained/measured, to be used as input for the model simulations. Model simulations showed a significant effect of fluid properties on the dynamics of aerosol particles. The role of buoyancy was highlighted by observation and interpretation of nucleation zones in the planes above the wire axis. The model was validated against measured temporal evolution, total number concentration and size distribution at the outlet of hot wire generator cell. Experimentally averaged and simulated total number concentrations were found to match closely, barring values at initial times. Steady-state number size distribution matched very well for sub 10 nm particle diameters while reasonable differences were noticed for higher size ranges. Although tuned specifically for the present context (i.e., aerosol generation from hotwire generator), the model can also be used for diverse applications, e.g., emission of particles from hot zones (chimneys, exhaust), fires and atmospheric cloud dynamics.

Keywords: nanoparticles, k-epsilon model, buoyancy, CFD, hot wire generator, aerosol dynamics

Procedia PDF Downloads 143
25 Lessons Learned through a Bicultural Approach to Tsunami Education in Aotearoa New Zealand

Authors: Lucy H. Kaiser, Kate Boersen

Abstract:

Kura Kaupapa Māori (kura) and bilingual schools are primary schools in Aotearoa/New Zealand which operate fully or partially under Māori custom and have curricula developed to include Te Reo Māori and Tikanga Māori (Māori language and cultural practices). These schools were established to support Māori children and their families through reinforcing cultural identity by enabling Māori language and culture to flourish in the field of education. Māori kaupapa (values), Mātauranga Māori (Māori knowledge) and Te Reo are crucial considerations for the development of educational resources developed for kura, bilingual and mainstream schools. The inclusion of hazard risk in education has become an important issue in New Zealand due to the vulnerability of communities to a plethora of different hazards. Māori have an extensive knowledge of their local area and the history of hazards which is often not appropriately recognised within mainstream hazard education resources. Researchers from the Joint Centre for Disaster Research, Massey University and East Coast LAB (Life at the Boundary) in Napier were funded to collaboratively develop a toolkit of tsunami risk reduction activities with schools located in Hawke’s Bay’s tsunami evacuation zones. A Māori-led bicultural approach to developing and running the education activities was taken, focusing on creating culturally and locally relevant materials for students and schools as well as giving students a proactive role in making their communities better prepared for a tsunami event. The community-based participatory research is Māori-centred, framed by qualitative and Kaupapa Maori research methodologies and utilizes a range of data collection methods including interviews, focus groups and surveys. Māori participants, stakeholders and the researchers collaborated through the duration of the project to ensure the programme would align with the wider school curricula and kaupapa values. The education programme applied a tuakana/teina, Māori teaching and learning approach in which high school aged students (tuakana) developed tsunami preparedness activities to run with primary school students (teina). At the end of the education programme, high school students were asked to reflect on their participation, what they had learned and what they had enjoyed during the activities. This paper draws on lessons learned throughout this research project. As an exemplar, retaining a bicultural and bilingual perspective resulted in a more inclusive project as there was variability across the students’ levels of confidence using Te Reo and Māori knowledge and cultural frameworks. Providing a range of different learning and experiential activities including waiata (Māori songs), pūrākau (traditional stories) and games was important to ensure students had the opportunity to participate and contribute using a range of different approaches that were appropriate to their individual learning needs. Inclusion of teachers in facilitation also proved beneficial in assisting classroom behavioral management. Lessons were framed by the tikanga and kawa (protocols) of the school to maintain cultural safety for the researchers and the students. Finally, the tuakana/teina component of the education activities became the crux of the programme, demonstrating a path for Rangatahi to support their whānau and communities through facilitating disaster preparedness, risk reduction and resilience.

Keywords: school safety, indigenous, disaster preparedness, children, education, tsunami

Procedia PDF Downloads 122
24 Mapping Iron Content in the Brain with Magnetic Resonance Imaging and Machine Learning

Authors: Gabrielle Robertson, Matthew Downs, Joseph Dagher

Abstract:

Iron deposition in the brain has been linked with a host of neurological disorders such as Alzheimer’s, Parkinson’s, and Multiple Sclerosis. While some treatment options exist, there are no objective measurement tools that allow for the monitoring of iron levels in the brain in vivo. An emerging Magnetic Resonance Imaging (MRI) method has been recently proposed to deduce iron concentration through quantitative measurement of magnetic susceptibility. This is a multi-step process that involves repeated modeling of physical processes via approximate numerical solutions. For example, the last two steps of this Quantitative Susceptibility Mapping (QSM) method involve I) mapping magnetic field into magnetic susceptibility and II) mapping magnetic susceptibility into iron concentration. Process I involves solving an ill-posed inverse problem by using regularization via injection of prior belief. The end result from Process II highly depends on the model used to describe the molecular content of each voxel (type of iron, water fraction, etc.) Due to these factors, the accuracy and repeatability of QSM have been an active area of research in the MRI and medical imaging community. This work aims to estimate iron concentration in the brain via a single step. A synthetic numerical model of the human head was created by automatically and manually segmenting the human head on a high-resolution grid (640x640x640, 0.4mm³) yielding detailed structures such as microvasculature and subcortical regions as well as bone, soft tissue, Cerebral Spinal Fluid, sinuses, arteries, and eyes. Each segmented region was then assigned tissue properties such as relaxation rates, proton density, electromagnetic tissue properties and iron concentration. These tissue property values were randomly selected from a Probability Distribution Function derived from a thorough literature review. In addition to having unique tissue property values, different synthetic head realizations also possess unique structural geometry created by morphing the boundary regions of different areas within normal physical constraints. This model of the human brain is then used to create synthetic MRI measurements. This is repeated thousands of times, for different head shapes, volume, tissue properties and noise realizations. Collectively, this constitutes a training-set that is similar to in vivo data, but larger than datasets available from clinical measurements. This 3D convolutional U-Net neural network architecture was used to train data-driven Deep Learning models to solve for iron concentrations from raw MRI measurements. The performance was then tested on both synthetic data not used in training as well as real in vivo data. Results showed that the model trained on synthetic MRI measurements is able to directly learn iron concentrations in areas of interest more effectively than other existing QSM reconstruction methods. For comparison, models trained on random geometric shapes (as proposed in the Deep QSM method) are less effective than models trained on realistic synthetic head models. Such an accurate method for the quantitative measurement of iron deposits in the brain would be of important value in clinical studies aiming to understand the role of iron in neurological disease.

Keywords: magnetic resonance imaging, MRI, iron deposition, machine learning, quantitative susceptibility mapping

Procedia PDF Downloads 136
23 Thermally Conductive Polymer Nanocomposites Based on Graphene-Related Materials

Authors: Alberto Fina, Samuele Colonna, Maria del Mar Bernal, Orietta Monticelli, Mauro Tortello, Renato Gonnelli, Julio Gomez, Chiara Novara, Guido Saracco

Abstract:

Thermally conductive polymer nanocomposites are of high interest for several applications including low-temperature heat recovery, heat exchangers in a corrosive environment and heat management in electronics and flexible electronics. In this paper, the preparation of thermally conductive nanocomposites exploiting graphene-related materials is addressed, along with their thermal characterization. In particular, correlations between 1- chemical and physical features of the nanoflakes and 2- processing conditions with the heat conduction properties of nanocomposites is studied. Polymers are heat insulators; therefore, the inclusion of conductive particles is the typical solution to obtain a sufficient thermal conductivity. In addition to traditional microparticles such as graphite and ceramics, several nanoparticles have been proposed, including carbon nanotubes and graphene, for the use in polymer nanocomposites. Indeed, thermal conductivities for both carbon nanotubes and graphenes were reported in the wide range of about 1500 to 6000 W/mK, despite such property may decrease dramatically as a function of the size, number of layers, the density of topological defects, re-hybridization defects as well as on the presence of impurities. Different synthetic techniques have been developed, including mechanical cleavage of graphite, epitaxial growth on SiC, chemical vapor deposition, and liquid phase exfoliation. However, the industrial scale-up of graphene, defined as an individual, single-atom-thick sheet of hexagonally arranged sp2-bonded carbons still remains very challenging. For large scale bulk applications in polymer nanocomposites, some graphene-related materials such as multilayer graphenes (MLG), reduced graphene oxide (rGO) or graphite nanoplatelets (GNP) are currently the most interesting graphene-based materials. In this paper, different types of graphene-related materials were characterized for their chemical/physical as well as for thermal properties of individual flakes. Two selected rGOs were annealed at 1700°C in vacuum for 1 h to reduce defectiveness of the carbon structure. Thermal conductivity increase of individual GNP with annealing was assessed via scanning thermal microscopy. Graphene nano papers were prepared from both conventional RGO and annealed RGO flakes. Characterization of the nanopapers evidenced a five-fold increase in the thermal diffusivity on the nano paper plane for annealed nanoflakes, compared to pristine ones, demonstrating the importance of structural defectiveness reduction to maximize the heat dissipation performance. Both pristine and annealed RGO were used to prepare polymer nanocomposites, by melt reactive extrusion. Thermal conductivity showed two- to three-fold increase in the thermal conductivity of the nanocomposite was observed for high temperature treated RGO compared to untreated RGO, evidencing the importance of using low defectivity nanoflakes. Furthermore, the study of different processing paremeters (time, temperature, shear rate) during the preparation of poly (butylene terephthalate) nanocomposites evidenced a clear correlation with the dispersion and fragmentation of the GNP nanoflakes; which in turn affected the thermal conductivity performance. Thermal conductivity of about 1.7 W/mK, i.e. one order of magnitude higher than for pristine polymer, was obtained with 10%wt of annealed GNPs, which is in line with state of the art nanocomposites prepared by more complex and less upscalable in situ polymerization processes.

Keywords: graphene, graphene-related materials, scanning thermal microscopy, thermally conductive polymer nanocomposites

Procedia PDF Downloads 267
22 Quantitative Texture Analysis of Shoulder Sonography for Rotator Cuff Lesion Classification

Authors: Chung-Ming Lo, Chung-Chien Lee

Abstract:

In many countries, the lifetime prevalence of shoulder pain is up to 70%. In America, the health care system spends 7 billion per year about the healthy issues of shoulder pain. With respect to the origin, up to 70% of shoulder pain is attributed to rotator cuff lesions This study proposed a computer-aided diagnosis (CAD) system to assist radiologists classifying rotator cuff lesions with less operator dependence. Quantitative features were extracted from the shoulder ultrasound images acquired using an ALOKA alpha-6 US scanner (Hitachi-Aloka Medical, Tokyo, Japan) with linear array probe (scan width: 36mm) ranging from 5 to 13 MHz. During examination, the postures of the examined patients are standard sitting position and are followed by the regular routine. After acquisition, the shoulder US images were drawn out from the scanner and stored as 8-bit images with pixel value ranging from 0 to 255. Upon the sonographic appearance, the boundary of each lesion was delineated by a physician to indicate the specific pattern for analysis. The three lesion categories for classification were composed of 20 cases of tendon inflammation, 18 cases of calcific tendonitis, and 18 cases of supraspinatus tear. For each lesion, second-order statistics were quantified in the feature extraction. The second-order statistics were the texture features describing the correlations between adjacent pixels in a lesion. Because echogenicity patterns were expressed via grey-scale. The grey-scale co-occurrence matrixes with four angles of adjacent pixels were used. The texture metrics included the mean and standard deviation of energy, entropy, correlation, inverse different moment, inertia, cluster shade, cluster prominence, and Haralick correlation. Then, the quantitative features were combined in a multinomial logistic regression classifier to generate a prediction model of rotator cuff lesions. Multinomial logistic regression classifier is widely used in the classification of more than two categories such as the three lesion types used in this study. In the classifier, backward elimination was used to select a feature subset which is the most relevant. They were selected from the trained classifier with the lowest error rate. Leave-one-out cross-validation was used to evaluate the performance of the classifier. Each case was left out of the total cases and used to test the trained result by the remaining cases. According to the physician’s assessment, the performance of the proposed CAD system was shown by the accuracy. As a result, the proposed system achieved an accuracy of 86%. A CAD system based on the statistical texture features to interpret echogenicity values in shoulder musculoskeletal ultrasound was established to generate a prediction model for rotator cuff lesions. Clinically, it is difficult to distinguish some kinds of rotator cuff lesions, especially partial-thickness tear of rotator cuff. The shoulder orthopaedic surgeon and musculoskeletal radiologist reported greater diagnostic test accuracy than general radiologist or ultrasonographers based on the available literature. Consequently, the proposed CAD system which was developed according to the experiment of the shoulder orthopaedic surgeon can provide reliable suggestions to general radiologists or ultrasonographers. More quantitative features related to the specific patterns of different lesion types would be investigated in the further study to improve the prediction.

Keywords: shoulder ultrasound, rotator cuff lesions, texture, computer-aided diagnosis

Procedia PDF Downloads 284
21 Synthesis of Carbonyl Iron Particles Modified with Poly (Trimethylsilyloxyethyl Methacrylate) Nano-Grafts

Authors: Martin Cvek, Miroslav Mrlik, Michal Sedlacik, Tomas Plachy

Abstract:

Magnetorheological elastomers (MREs) are multi-phase composite materials containing micron-sized ferromagnetic particles dispersed in an elastomeric matrix. Their properties such as modulus, damping, magneto-striction, and electrical conductivity can be controlled by an external magnetic field and/or pressure. These features of the MREs are used in the development of damping devices, shock attenuators, artificial muscles, sensors or active elements of electric circuits. However, imperfections on the particle/matrix interfaces result in the lower performance of the MREs when compared with theoretical values. Moreover, magnetic particles are susceptible to corrosion agents such as acid rains or sea humidity. Therefore, the modification of particles is an effective tool for the improvement of MRE performance due to enhanced compatibility between particles and matrix as well as improvements of their thermo-oxidation and chemical stability. In this study, the carbonyl iron (CI) particles were controllably modified with poly(trimethylsilyloxyethyl methacrylate) (PHEMATMS) nano-grafts to develop magnetic core–shell structures exhibiting proper wetting with various elastomeric matrices resulting in improved performance within a frame of rheological, magneto-piezoresistance, pressure-piezoresistance, or radio-absorbing properties. The desired molecular weight of PHEMATMS nano-grafts was precisely tailored using surface-initiated atom transfer radical polymerization (ATRP). The CI particles were firstly functionalized using a 3-aminopropyltriethoxysilane agent, followed by esterification reaction with α-bromoisobutyryl bromide. The ATRP was performed in the anisole medium using ethyl α-bromoisobutyrate as a macroinitiator, N, N´, N´´, N´´-pentamethyldiethylenetriamine as a ligand, and copper bromide as an initiator. To explore the effect PHEMATMS molecular weights on final properties, two variants of core-shell structures with different nano-graft lengths were synthesized, while the reaction kinetics were designed through proper reactant feed ratios and polymerization times. The PHEMATMS nano-grafts were characterized by nuclear magnetic resonance and gel permeation chromatography proving information to their monomer conversions, molecular chain lengths, and low polydispersity indexes (1.28 and 1.35) as the results of the executed ATRP. The successful modifications were confirmed via Fourier transform infrared- and energy-dispersive spectroscopies while expected wavenumber outputs and element presences, respectively, of constituted PHEMATMS nano-grafts, were occurring in the spectra. The surface morphology of bare CI and their PHEMATMS-grafted analogues was further studied by scanning electron microscopy, and the thicknesses of grafted polymeric layers were directly observed by transmission electron microscopy. The contact angles as a measure of particle/matrix compatibility were investigated employing the static sessile drop method. The PHEMATMS nano-grafts enhanced compatibility of hydrophilic CI with low-surface-energy hydrophobic polymer matrix in terms of their wettability and dispersibility in an elastomeric matrix. Thus, the presence of possible defects at the particle/matrix interface is reduced, and higher performance of modified MREs is expected.

Keywords: atom transfer radical polymerization, core-shell, particle modification, wettability

Procedia PDF Downloads 200
20 A Case Study Report on Acoustic Impact Assessment and Mitigation of the Hyprob Research Plant

Authors: D. Bianco, A. Sollazzo, M. Barbarino, G. Elia, A. Smoraldi, N. Favaloro

Abstract:

The activities, described in the present paper, have been conducted in the framework of the HYPROB-New Program, carried out by the Italian Aerospace Research Centre (CIRA) promoted and funded by the Italian Ministry of University and Research (MIUR) in order to improve the National background on rocket engine systems for space applications. The Program has the strategic objective to improve National system and technology capabilities in the field of liquid rocket engines (LRE) for future Space Propulsion Systems applications, with specific regard to LOX/LCH4 technology. The main purpose of the HYPROB program is to design and build a Propulsion Test Facility (HIMP) allowing test activities on Liquid Thrusters. The development of skills in liquid rocket propulsion can only pass through extensive test campaign. Following its mission, CIRA has planned the development of new testing facilities and infrastructures for space propulsion characterized by adequate sizes and instrumentation. The IMP test cell is devoted to testing articles representative of small combustion chambers, fed with oxygen and methane, both in liquid and gaseous phase. This article describes the activities that have been carried out for the evaluation of the acoustic impact, and its consequent mitigation. The impact of the simulated acoustic disturbance has been evaluated, first, using an approximated method based on experimental data by Baumann and Coney, included in “Noise and Vibration Control Engineering” edited by Vér and Beranek. This methodology, used to evaluate the free-field radiation of jet in ideal acoustical medium, analyzes in details the jet noise and assumes sources acting at the same time. It considers as principal radiation sources the jet mixing noise, caused by the turbulent mixing of jet gas and the ambient medium. Empirical models, allowing a direct calculation of the Sound Pressure Level, are commonly used for rocket noise simulation. The model named after K. Eldred is probably one of the most exploited in this area. In this paper, an improvement of the Eldred Standard model has been used for a detailed investigation of the acoustical impact of the Hyprob facility. This new formulation contains an explicit expression for the acoustic pressure of each equivalent noise source, in terms of amplitude and phase, allowing the investigation of the sources correlation effects and their propagation through wave equations. In order to enhance the evaluation of the facility acoustic impact, including an assessment of the mitigation strategies to be set in place, a more advanced simulation campaign has been conducted using both an in-house code for noise propagation and scattering, and a commercial code for industrial noise environmental impact, CadnaA. The noise prediction obtained with the revised Eldred-based model has then been used for formulating an empirical/BEM (Boundary Element Method) hybrid approach allowing the evaluation of the barrier mitigation effect, at the design. This approach has been compared with the analogous empirical/ray-acoustics approach, implemented within CadnaA using a customized definition of sources and directivity factor. The resulting impact evaluation study is reported here, along with the design-level barrier optimization for noise mitigation.

Keywords: acoustic impact, industrial noise, mitigation, rocket noise

Procedia PDF Downloads 146
19 VIAN-DH: Computational Multimodal Conversation Analysis Software and Infrastructure

Authors: Teodora Vukovic, Christoph Hottiger, Noah Bubenhofer

Abstract:

The development of VIAN-DH aims at bridging two linguistic approaches: conversation analysis/interactional linguistics (IL), so far a dominantly qualitative field, and computational/corpus linguistics and its quantitative and automated methods. Contemporary IL investigates the systematic organization of conversations and interactions composed of speech, gaze, gestures, and body positioning, among others. These highly integrated multimodal behaviour is analysed based on video data aimed at uncovering so called “multimodal gestalts”, patterns of linguistic and embodied conduct that reoccur in specific sequential positions employed for specific purposes. Multimodal analyses (and other disciplines using videos) are so far dependent on time and resource intensive processes of manual transcription of each component from video materials. Automating these tasks requires advanced programming skills, which is often not in the scope of IL. Moreover, the use of different tools makes the integration and analysis of different formats challenging. Consequently, IL research often deals with relatively small samples of annotated data which are suitable for qualitative analysis but not enough for making generalized empirical claims derived quantitatively. VIAN-DH aims to create a workspace where many annotation layers required for the multimodal analysis of videos can be created, processed, and correlated in one platform. VIAN-DH will provide a graphical interface that operates state-of-the-art tools for automating parts of the data processing. The integration of tools that already exist in computational linguistics and computer vision, facilitates data processing for researchers lacking programming skills, speeds up the overall research process, and enables the processing of large amounts of data. The main features to be introduced are automatic speech recognition for the transcription of language, automatic image recognition for extraction of gestures and other visual cues, as well as grammatical annotation for adding morphological and syntactic information to the verbal content. In the ongoing instance of VIAN-DH, we focus on gesture extraction (pointing gestures, in particular), making use of existing models created for sign language and adapting them for this specific purpose. In order to view and search the data, VIAN-DH will provide a unified format and enable the import of the main existing formats of annotated video data and the export to other formats used in the field, while integrating different data source formats in a way that they can be combined in research. VIAN-DH will adapt querying methods from corpus linguistics to enable parallel search of many annotation levels, combining token-level and chronological search for various types of data. VIAN-DH strives to bring crucial and potentially revolutionary innovation to the field of IL, (that can also extend to other fields using video materials). It will allow the processing of large amounts of data automatically and, the implementation of quantitative analyses, combining it with the qualitative approach. It will facilitate the investigation of correlations between linguistic patterns (lexical or grammatical) with conversational aspects (turn-taking or gestures). Users will be able to automatically transcribe and annotate visual, spoken and grammatical information from videos, and to correlate those different levels and perform queries and analyses.

Keywords: multimodal analysis, corpus linguistics, computational linguistics, image recognition, speech recognition

Procedia PDF Downloads 108
18 Unknown Groundwater Pollution Source Characterization in Contaminated Mine Sites Using Optimal Monitoring Network Design

Authors: H. K. Esfahani, B. Datta

Abstract:

Groundwater is one of the most important natural resources in many parts of the world; however it is widely polluted due to human activities. Currently, effective and reliable groundwater management and remediation strategies are obtained using characterization of groundwater pollution sources, where the measured data in monitoring locations are utilized to estimate the unknown pollutant source location and magnitude. However, accurately identifying characteristics of contaminant sources is a challenging task due to uncertainties in terms of predicting source flux injection, hydro-geological and geo-chemical parameters, and the concentration field measurement. Reactive transport of chemical species in contaminated groundwater systems, especially with multiple species, is a complex and highly non-linear geochemical process. Although sufficient concentration measurement data is essential to accurately identify sources characteristics, available data are often sparse and limited in quantity. Therefore, this inverse problem-solving method for characterizing unknown groundwater pollution sources is often considered ill-posed, complex and non- unique. Different methods have been utilized to identify pollution sources; however, the linked simulation-optimization approach is one effective method to obtain acceptable results under uncertainties in complex real life scenarios. With this approach, the numerical flow and contaminant transport simulation models are externally linked to an optimization algorithm, with the objective of minimizing the difference between measured concentration and estimated pollutant concentration at observation locations. Concentration measurement data are very important to accurately estimate pollution source properties; therefore, optimal design of the monitoring network is essential to gather adequate measured data at desired times and locations. Due to budget and physical restrictions, an efficient and effective approach for groundwater pollutant source characterization is to design an optimal monitoring network, especially when only inadequate and arbitrary concentration measurement data are initially available. In this approach, preliminary concentration observation data are utilized for preliminary source location, magnitude and duration of source activity identification, and these results are utilized for monitoring network design. Further, feedback information from the monitoring network is used as inputs for sequential monitoring network design, to improve the identification of unknown source characteristics. To design an effective monitoring network of observation wells, optimization and interpolation techniques are used. A simulation model should be utilized to accurately describe the aquifer properties in terms of hydro-geochemical parameters and boundary conditions. However, the simulation of the transport processes becomes complex when the pollutants are chemically reactive. Three dimensional transient flow and reactive contaminant transport process is considered. The proposed methodology uses HYDROGEOCHEM 5.0 (HGCH) as the simulation model for flow and transport processes with chemically multiple reactive species. Adaptive Simulated Annealing (ASA) is used as optimization algorithm in linked simulation-optimization methodology to identify the unknown source characteristics. Therefore, the aim of the present study is to develop a methodology to optimally design an effective monitoring network for pollution source characterization with reactive species in polluted aquifers. The performance of the developed methodology will be evaluated for an illustrative polluted aquifer sites, for example an abandoned mine site in Queensland, Australia.

Keywords: monitoring network design, source characterization, chemical reactive transport process, contaminated mine site

Procedia PDF Downloads 231
17 Sustainable Urban Regenaration the New Vocabulary and the Timless Grammar of the Urban Tissue

Authors: Ruth Shapira

Abstract:

Introduction: The rapid urbanization of the last century confronts planners, regulatory bodies, developers and most of all the public with seemingly unsolved conflicts regarding values, capital, and wellbeing of the built and un-built urban space. There is an out of control change of scale of the urban form and of the rhythm of the urban life which has known no significant progress in the last 2-3 decades despite the on-growing urban population. It is the objective of this paper to analyze some of these fundamental issues through the case study of a relatively small town in the center of Israel (Kiryat-Ono, 36,000 inhabitants), unfold the deep structure of qualities versus disruptors, present some cure that we have developed to bridge over and humbly suggest a practice that may bring about a sustainable new urban environment based on timeless values of the past, an approach that can be generic for similar cases. Basic Methodologies:The object, the town of Kiryat Ono, shall be experimented upon in a series of four action processes: De-composition, Re-composition, the Centering process and, finally, Controlled Structural Disintegration. Each stage will be based on facts, analysis of previous multidisciplinary interventions on various layers – and the inevitable reaction of the OBJECT, leading to the conclusion based on innovative theoretical and practical methods that we have developed and that we believe are proper for the open ended network, setting the rules for the contemporary urban society to cluster by – thus – a new urban vocabulary based on the old structure of times passed. The Study: Kiryat Ono, was founded 70 years ago as an agricultural settlement and rapidly turned into an urban entity. In spite the massive intensification, the original DNA of the old small town was still deeply embedded, mostly in the quality of the public space and in the sense of clustered communities. In the past 20 years, the recent demand for housing has been addressed to on the national level with recent master plans and urban regeneration policies mostly encouraging individual economic initiatives. Unfortunately, due to the obsolete existing planning platform the present urban renewal is characterized by pressure of developers, a dramatic change in building scale and widespread disintegration of the existing urban and social tissue.Our office was commissioned to conceptualize two master plans for the two contradictory processes of Kiryat Ono’s future: intensification and conservation. Following a comprehensive investigation into the deep structures and qualities of the existing town, we developed a new vocabulary of conservation terms thus redefying the sense of PLACE. The main challenge was to create master plans that should offer a regulatory basis to the accelerated and sporadic development providing for the public good and preserving the characteristics of the place consisting of a tool box of design guidelines that will have the ability to reorganize space along the time axis in a sustainable way. In conclusion: The system of rules that we have developed can generate endless possible patterns making sure that at each implementation fragment an event is created, and a better place is revealed. It takes time and perseverance but it seems to be the way to provide a healthy and sustainable framework for the accelerated urbanization of our chaotic present.

Keywords: sustainable urban design, intensification, emergent urban patterns, sustainable housing, compact urban neighborhoods, sustainable regeneration, restoration, complexity, uncertainty, need for change, implications of legislation on local planning

Procedia PDF Downloads 388
16 Deep Learning Based on Image Decomposition for Restoration of Intrinsic Representation

Authors: Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Kensuke Nakamura, Dongeun Choi, Byung-Woo Hong

Abstract:

Artefacts are commonly encountered in the imaging process of clinical computed tomography (CT) where the artefact refers to any systematic discrepancy between the reconstructed observation and the true attenuation coefficient of the object. It is known that CT images are inherently more prone to artefacts due to its image formation process where a large number of independent detectors are involved, and they are assumed to yield consistent measurements. There are a number of different artefact types including noise, beam hardening, scatter, pseudo-enhancement, motion, helical, ring, and metal artefacts, which cause serious difficulties in reading images. Thus, it is desired to remove nuisance factors from the degraded image leaving the fundamental intrinsic information that can provide better interpretation of the anatomical and pathological characteristics. However, it is considered as a difficult task due to the high dimensionality and variability of data to be recovered, which naturally motivates the use of machine learning techniques. We propose an image restoration algorithm based on the deep neural network framework where the denoising auto-encoders are stacked building multiple layers. The denoising auto-encoder is a variant of a classical auto-encoder that takes an input data and maps it to a hidden representation through a deterministic mapping using a non-linear activation function. The latent representation is then mapped back into a reconstruction the size of which is the same as the size of the input data. The reconstruction error can be measured by the traditional squared error assuming the residual follows a normal distribution. In addition to the designed loss function, an effective regularization scheme using residual-driven dropout determined based on the gradient at each layer. The optimal weights are computed by the classical stochastic gradient descent algorithm combined with the back-propagation algorithm. In our algorithm, we initially decompose an input image into its intrinsic representation and the nuisance factors including artefacts based on the classical Total Variation problem that can be efficiently optimized by the convex optimization algorithm such as primal-dual method. The intrinsic forms of the input images are provided to the deep denosing auto-encoders with their original forms in the training phase. In the testing phase, a given image is first decomposed into the intrinsic form and then provided to the trained network to obtain its reconstruction. We apply our algorithm to the restoration of the corrupted CT images by the artefacts. It is shown that our algorithm improves the readability and enhances the anatomical and pathological properties of the object. The quantitative evaluation is performed in terms of the PSNR, and the qualitative evaluation provides significant improvement in reading images despite degrading artefacts. The experimental results indicate the potential of our algorithm as a prior solution to the image interpretation tasks in a variety of medical imaging applications. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: auto-encoder neural network, CT image artefact, deep learning, intrinsic image representation, noise reduction, total variation

Procedia PDF Downloads 190
15 Fabrication of Zeolite Modified Cu Doped ZnO Films and Their Response towards Nitrogen Monoxide

Authors: Irmak Karaduman, Tugba Corlu, Sezin Galioglu, Burcu Akata, M. Ali Yildirim, Aytunç Ateş, Selim Acar

Abstract:

Breath analysis represents a promising non-invasive, fast and cost-effective alternative to well-established diagnostic and monitoring techniques such as blood analysis, endoscopy, ultrasonic and tomographic monitoring. Portable, non-invasive, and low-cost breath analysis devices are becoming increasingly desirable for monitoring different diseases, especially asthma. Beacuse of this, NO gas sensing at low concentrations has attracted progressive attention for clinical analysis in asthma. Recently, nanomaterials based sensors are considered to be a promising clinical and laboratory diagnostic tool, because its large surface–to–volume ratio, controllable structure, easily tailored chemical and physical properties, which bring high sensitivity, fast dynamic processand even the increasing specificity. Among various nanomaterials, semiconducting metal oxides are extensively studied gas-sensing materials and are potential sensing elements for breathanalyzer due to their high sensitivity, simple design, low cost and good stability.The sensitivities of metal oxide semiconductor gas sensors can be enhanced by adding noble metals. Doping contents, distribution, and size of metallic or metal oxide catalysts are key parameters for enhancing gas selectivity as well as sensitivity. By manufacturing doping MOS structures, it is possible to develop more efficient sensor sensing layers. Zeolites are perhaps the most widely employed group of silicon-based nanoporous solids. Their well-defined pores of sub nanometric size have earned them the name of molecular sieves, meaning that operation in the size exclusion regime is possible by selecting, among over 170 structures available, the zeolite whose pores allow the pass of the desired molecule, while keeping larger molecules outside.In fact it is selective adsorption, rather than molecular sieving, the mechanism that explains most of the successful gas separations achieved with zeolite membranes. In view of their molecular sieving and selective adsorption properties, it is not surprising that zeolites have found use in a number of works dealing with gas sensing devices. In this study, the Cu doped ZnO nanostructure film was produced by SILAR method and investigated the NO gas sensing properties. To obtain the selectivity of the sample, the gases including CO,NH3,H2 and CH4 were detected to compare with NO. The maximum response is obtained at 85 C for 20 ppb NO gas. The sensor shows high response to NO gas. However, acceptable responses are calculated for CO and NH3 gases. Therefore, there are no responses obtain for H2 and CH4 gases. Enhanced to selectivity, Cu doped ZnO nanostructure film was coated with zeolite A thin film. It is found that the sample possess an acceptable response towards NO hardly respond to CO, NH3, H2 and CH4 at room temperature. This difference in the response can be expressed in terms of differences in the molecular structure, the dipole moment, strength of the electrostatic interaction and the dielectric constant. The as-synthesized thin film is considered to be one of the extremely promising candidate materials in electronic nose applications. This work is supported by The Scientific and Technological Research Council of Turkey (TUBİTAK) under Project No, 115M658 and Gazi University Scientific Research Fund under project no 05/2016-21.

Keywords: Cu doped ZnO, electrical characterization, gas sensing, zeolite

Procedia PDF Downloads 285
14 Erotic Subversions: Male Masochism, Power, and the Politics of Desire in Hong Kong’s BDSM Landscape

Authors: Maari Sugawara

Abstract:

This research critically engages with the erotic and political entanglements of male clientele of Dominatrices who identify as submissives (hereafter referred to as submissives) within Hong Kong's BDSM scene. Employing masochism as an analytical framework, it interrogates the intersections of capitalism, heteropatriarchy, postcolonialism, and commodified desire. BDSM (Bondage and Discipline, Dominance and Submission, Sadism and Masochism) encompasses practices that explore power, control, and subordination through both physical and psychological role-play, predicated on consent, negotiation, and boundary delineation. This makes BDSM a fertile site for examining how dominance and submission are mobilized, challenged, and reiterated. This study focuses on the dynamics between thirty male submissives and three professional Dominatrices active in Hong Kong since 2019. The predominance of male interviewees reflects the demographic reality that most clients engaging with professional Dominatrices are male. These submissives—men who willfully relinquish control—offer a critical lens for exploring how BDSM, as both practice and market, mirrors and destabilizes dominant power structures. BDSM relationships occasionally replicate the hierarchical logics of heterosexual marriage, particularly in the expectation that submissives engage exclusively with a single Dominatrix, reflecting a dynamic of devotion and fidelity akin to traditional marital structures. However, these relationships also function as counter-normative spaces where care and control are reconfigured, enabling the negotiation of alternative power configurations. By centering BDSM work rather than broader kink practices, this study foregrounds the commodification of intimacy as a key site where suppressed desires, economic forces, and political tensions converge. The submissives in this study are predominantly affluent, cisgender men, underscoring the socio-economic asymmetries in the BDSM market. Furthermore, the research examines how Hong Kong’s political turbulence—particularly the 2019 Yellow Umbrella Movement and the COVID-19 pandemic—has reverberated through the BDSM scene, reshaping the contours of desire, trust, and power in these intimate transactions. The increasing tensions with mainland China, alongside the erosion of public trust in state institutions, form a critical backdrop to this evolving landscape. Grounded in gender and sexuality theories, this research interrogates how the desires of male submissives are constructed within and resist heteronormative frameworks. BDSM practices, far from existing outside capitalist and colonial logics, often act as both a mirror and critique of these systems, revealing complex ways in which power is commodified, enacted, and contested. In their pursuit of emotional care and alternative forms of control, male submissives navigate a paradoxical terrain where their masochistic desires both subvert and perpetuate the socio-political status quo. By examining the intersections of desire, commodification, and the shifting socio-political landscape, this research provides a nuanced understanding of how BDSM functions as a site of negotiation for those navigating the turbulent crosscurrents of capitalist and colonial legacies. Ultimately, it uncovers the complex interplay between erotic practices and the structures of power and identity in a city undergoing profound transformation.

Keywords: masochism, Hong Kong, identity, BDSM, dominatrix, masculinity, gender studies

Procedia PDF Downloads 20
13 Simulation, Design, and 3D Print of Novel Highly Integrated TEG Device with Improved Thermal Energy Harvest Efficiency

Authors: Jaden Lu, Olivia Lu

Abstract:

Despite the remarkable advancement of solar cell technology, the challenge of optimizing total solar energy harvest efficiency persists, primarily due to significant heat loss. This excess heat not only diminishes solar panel output efficiency but also curtails its operational lifespan. A promising approach to address this issue is the conversion of surplus heat into electricity. In recent years, there is growing interest in the use of thermoelectric generators (TEG) as a potential solution. The integration of efficient TEG devices holds the promise of augmenting overall energy harvest efficiency while prolonging the longevity of solar panels. While certain research groups have proposed the integration of solar cells and TEG devices, a substantial gap between conceptualization and practical implementation remains, largely attributed to low thermal energy conversion efficiency of TEG devices. To bridge this gap and meet the requisites of practical application, a feasible strategy involves the incorporation of a substantial number of p-n junctions within a confined unit volume. However, the manufacturing of high-density TEG p-n junctions presents a formidable challenge. The prevalent solution often leads to large device sizes to accommodate enough p-n junctions, consequently complicating integration with solar cells. Recently, the adoption of 3D printing technology has emerged as a promising solution to address this challenge by fabricating high-density p-n arrays. Despite this, further developmental efforts are necessary. Presently, the primary focus is on the 3D printing of vertically layered TEG devices, wherein p-n junction density remains constrained by spatial limitations and the constraints of 3D printing techniques. This study proposes a novel device configuration featuring horizontally arrayed p-n junctions of Bi2Te3. The structural design of the device is subjected to simulation through the Finite Element Method (FEM) within COMSOL Multiphysics software. Various device configurations are simulated to identify optimal device structure. Based on the simulation results, a new TEG device is fabricated utilizing 3D Selective laser melting (SLM) printing technology. Fusion 360 facilitates the translation of the COMSOL device structure into a 3D print file. The horizontal design offers a unique advantage, enabling the fabrication of densely packed, three-dimensional p-n junction arrays. The fabrication process entails printing a singular row of horizontal p-n junctions using the 3D SLM printing technique in a single layer. Subsequently, successive rows of p-n junction arrays are printed within the same layer, interconnected by thermally conductive copper. This sequence is replicated across multiple layers, separated by thermal insulating glass. This integration created in a highly compact three-dimensional TEG device with high density p-n junctions. The fabricated TEG device is then attached to the bottom of the solar cell using thermal glue. The whole device is characterized, with output data closely matching with COMSOL simulation results. Future research endeavors will encompass the refinement of thermoelectric materials. This includes the advancement of high-resolution 3D printing techniques tailored to diverse thermoelectric materials, along with the optimization of material microstructures such as porosity and doping. The objective is to achieve an optimal and highly integrated PV-TEG device that can substantially increase the solar energy harvest efficiency.

Keywords: thermoelectric, finite element method, 3d print, energy conversion

Procedia PDF Downloads 62
12 The Semiotics of Soft Power; An Examination of the South Korean Entertainment Industry

Authors: Enya Trenholm-Jensen

Abstract:

This paper employs various semiotic methodologies to examine the mechanism of soft power. Soft power refers to a country’s global reputation and their ability to leverage that reputation to achieve certain aims. South Korea has invested heavily in their soft power strategy for a multitude of predominantly historical and geopolitical reasons. On account of this investment and the global prominence of their strategy, South Korea was considered to be the optimal candidate for the aims of this investigation. Having isolated the entertainment industry as one of the most heavily funded segments of the South Korean soft power strategy, the analysis restricted itself to this sector. Within this industry, two entertainment products were selected as case studies. The case studies were chosen based on commercial success according to metrics such as streams, purchases, and subsequent revenue. This criterion was deemed to be the most objective and verifiable indicator of the products general appeal. The entertainment products which met the chosen criterion were Netflix’ “Squid Game” and BTS’ hit single “Butter”. The methodologies employed were chosen according to the medium of the entertainment products. For “Squid Game,” an aesthetic analysis was carried out to investigate how multi- layered meanings were mobilized in a show popularized by its visual grammar. To examine “Butter”, both music semiology and linguistic analysis were employed. The music section featured an analysis underpinned by denotative and connotative music semiotic theories borrowing from scholars Theo van Leeuwen and Martin Irvine. The linguistic analysis focused on stance and semantic fields according to scholarship by George Yule and John W. DuBois. The aesthetic analysis of the first case study revealed intertextual references to famous artworks, which served to augment the emotional provocation of the Squid Game narrative. For the second case study, the findings exposed a set of musical meaning units arranged in a patchwork of familiar and futuristic elements to achieve a song that existed on the boundary between old and new. The linguistic analysis of the song’s lyrics found a deceptively innocuous surface level meaning that bore implications for authority, intimacy, and commercial success. Whether through means of visual metaphor, embedded auditory associations, or linguistic subtext, the collective findings of the three analyses exhibited a desire to conjure a form of positive arousal in the spectator. In the synthesis section, this process is likened to that of branding. Through an exploration of branding, the entertainment products can be understood as cogs in a larger operation aiming to create positive associations to Korea as a country and a concept. Limitations in the form of a timeframe biased perspective are addressed, and directions for future research are suggested. This paper employs semiotic methodologies to examine two entertainment products as mechanisms of soft power. Through means of visual metaphor, embedded auditory associations, or linguistic subtext, the findings reveal a desire to conjure positive arousal in the spectator. The synthesis finds similarities to branding, thus positioning the entertainment products as cogs in a larger operation aiming to create positive associations to Korea as a country and a concept.

Keywords: BTS, cognitive semiotics, entertainment, soft power, south korea, squid game

Procedia PDF Downloads 153
11 IEEE802.15.4e Based Scheduling Mechanisms and Systems for Industrial Internet of Things

Authors: Ho-Ting Wu, Kai-Wei Ke, Bo-Yu Huang, Liang-Lin Yan, Chun-Ting Lin

Abstract:

With the advances in advanced technology, wireless sensor network (WSN) has become one of the most promising candidates to implement the wireless industrial internet of things (IIOT) architecture. However, the legacy IEEE 802.15.4 based WSN technology such as Zigbee system cannot meet the stringent QoS requirement of low powered, real-time, and highly reliable transmission imposed by the IIOT environment. Recently, the IEEE society developed IEEE 802.15.4e Time Slotted Channel Hopping (TSCH) access mode to serve this purpose. Furthermore, the IETF 6TiSCH working group has proposed standards to integrate IEEE 802.15.4e with IPv6 protocol smoothly to form a complete protocol stack for IIOT. In this work, we develop key network technologies for IEEE 802.15.4e based wireless IIoT architecture, focusing on practical design and system implementation. We realize the OpenWSN-based wireless IIOT system. The system architecture is divided into three main parts: web server, network manager, and sensor nodes. The web server provides user interface, allowing the user to view the status of sensor nodes and instruct sensor nodes to follow commands via user-friendly browser. The network manager is responsible for the establishment, maintenance, and management of scheduling and topology information. It executes centralized scheduling algorithm, sends the scheduling table to each node, as well as manages the sensing tasks of each device. Sensor nodes complete the assigned tasks and sends the sensed data. Furthermore, to prevent scheduling error due to packet loss, a schedule inspection mechanism is implemented to verify the correctness of the schedule table. In addition, when network topology changes, the system will act to generate a new schedule table based on the changed topology for ensuring the proper operation of the system. To enhance the system performance of such system, we further propose dynamic bandwidth allocation and distributed scheduling mechanisms. The developed distributed scheduling mechanism enables each individual sensor node to build, maintain and manage the dedicated link bandwidth with its parent and children nodes based on locally observed information by exchanging the Add/Delete commands via two processes. The first process, termed as the schedule initialization process, allows each sensor node pair to identify the available idle slots to allocate the basic dedicated transmission bandwidth. The second process, termed as the schedule adjustment process, enables each sensor node pair to adjust their allocated bandwidth dynamically according to the measured traffic loading. Such technology can sufficiently satisfy the dynamic bandwidth requirement in the frequently changing environments. Last but not least, we propose a packet retransmission scheme to enhance the system performance of the centralized scheduling algorithm when the packet delivery rate (PDR) is low. We propose a multi-frame retransmission mechanism to allow every single network node to resend each packet for at least the predefined number of times. The multi frame architecture is built according to the number of layers of the network topology. Performance results via simulation reveal that such retransmission scheme is able to provide sufficient high transmission reliability while maintaining low packet transmission latency. Therefore, the QoS requirement of IIoT can be achieved.

Keywords: IEEE 802.15.4e, industrial internet of things (IIOT), scheduling mechanisms, wireless sensor networks (WSN)

Procedia PDF Downloads 160
10 Finite Element Method (FEM) Simulation, design and 3D Print of Novel Highly Integrated PV-TEG Device with Improved Solar Energy Harvest Efficiency

Authors: Jaden Lu, Olivia Lu

Abstract:

Despite the remarkable advancement of solar cell technology, the challenge of optimizing total solar energy harvest efficiency persists, primarily due to significant heat loss. This excess heat not only diminishes solar panel output efficiency but also curtails its operational lifespan. A promising approach to address this issue is the conversion of surplus heat into electricity. In recent years, there is growing interest in the use of thermoelectric generators (TEG) as a potential solution. The integration of efficient TEG devices holds the promise of augmenting overall energy harvest efficiency while prolonging the longevity of solar panels. While certain research groups have proposed the integration of solar cells and TEG devices, a substantial gap between conceptualization and practical implementation remains, largely attributed to low thermal energy conversion efficiency of TEG devices. To bridge this gap and meet the requisites of practical application, a feasible strategy involves the incorporation of a substantial number of p-n junctions within a confined unit volume. However, the manufacturing of high-density TEG p-n junctions presents a formidable challenge. The prevalent solution often leads to large device sizes to accommodate enough p-n junctions, consequently complicating integration with solar cells. Recently, the adoption of 3D printing technology has emerged as a promising solution to address this challenge by fabricating high-density p-n arrays. Despite this, further developmental efforts are necessary. Presently, the primary focus is on the 3D printing of vertically layered TEG devices, wherein p-n junction density remains constrained by spatial limitations and the constraints of 3D printing techniques. This study proposes a novel device configuration featuring horizontally arrayed p-n junctions of Bi2Te3. The structural design of the device is subjected to simulation through the Finite Element Method (FEM) within COMSOL Multiphysics software. Various device configurations are simulated to identify optimal device structure. Based on the simulation results, a new TEG device is fabricated utilizing 3D Selective laser melting (SLM) printing technology. Fusion 360 facilitates the translation of the COMSOL device structure into a 3D print file. The horizontal design offers a unique advantage, enabling the fabrication of densely packed, three-dimensional p-n junction arrays. The fabrication process entails printing a singular row of horizontal p-n junctions using the 3D SLM printing technique in a single layer. Subsequently, successive rows of p-n junction arrays are printed within the same layer, interconnected by thermally conductive copper. This sequence is replicated across multiple layers, separated by thermal insulating glass. This integration created in a highly compact three-dimensional TEG device with high density p-n junctions. The fabricated TEG device is then attached to the bottom of the solar cell using thermal glue. The whole device is characterized, with output data closely matching with COMSOL simulation results. Future research endeavors will encompass the refinement of thermoelectric materials. This includes the advancement of high-resolution 3D printing techniques tailored to diverse thermoelectric materials, along with the optimization of material microstructures such as porosity and doping. The objective is to achieve an optimal and highly integrated PV-TEG device that can substantially increase the solar energy harvest efficiency.

Keywords: thermoelectric, finite element method, 3d print, energy conversion

Procedia PDF Downloads 67
9 The Proposal for a Framework to Face Opacity and Discrimination ‘Sins’ Caused by Consumer Creditworthiness Machines in the EU

Authors: Diogo José Morgado Rebelo, Francisco António Carneiro Pacheco de Andrade, Paulo Jorge Freitas de Oliveira Novais

Abstract:

Not everything in AI-power consumer credit scoring turns out to be a wonder. When using AI in Creditworthiness Assessment (CWA), opacity and unfairness ‘sins’ must be considered to the task be deemed Responsible. AI software is not always 100% accurate, which can lead to misclassification. Discrimination of some groups can be exponentiated. A hetero personalized identity can be imposed on the individual(s) affected. Also, autonomous CWA sometimes lacks transparency when using black box models. However, for this intended purpose, human analysts ‘on-the-loop’ might not be the best remedy consumers are looking for in credit. This study seeks to explore the legality of implementing a Multi-Agent System (MAS) framework in consumer CWA to ensure compliance with the regulation outlined in Article 14(4) of the Proposal for an Artificial Intelligence Act (AIA), dated 21 April 2021 (as per the last corrigendum by the European Parliament on 19 April 2024), Especially with the adoption of Art. 18(8)(9) of the EU Directive 2023/2225, of 18 October, which will go into effect on 20 November 2026, there should be more emphasis on the need for hybrid oversight in AI-driven scoring to ensure fairness and transparency. In fact, the range of EU regulations on AI-based consumer credit will soon impact the AI lending industry locally and globally, as shown by the broad territorial scope of AIA’s Art. 2. Consequently, engineering the law of consumer’s CWA is imperative. Generally, the proposed MAS framework consists of several layers arranged in a specific sequence, as follows: firstly, the Data Layer gathers legitimate predictor sets from traditional sources; then, the Decision Support System Layer, whose Neural Network model is trained using k-fold Cross Validation, provides recommendations based on the feeder data; the eXplainability (XAI) multi-structure comprises Three-Step-Agents; and, lastly, the Oversight Layer has a 'Bottom Stop' for analysts to intervene in a timely manner. From the analysis, one can assure a vital component of this software is the XAY layer. It appears as a transparent curtain covering the AI’s decision-making process, enabling comprehension, reflection, and further feasible oversight. Local Interpretable Model-agnostic Explanations (LIME) might act as a pillar by offering counterfactual insights. SHapley Additive exPlanation (SHAP), another agent in the XAI layer, could address potential discrimination issues, identifying the contribution of each feature to the prediction. Alternatively, for thin or no file consumers, the Suggestion Agent can promote financial inclusion. It uses lawful alternative sources such as the share of wallet, among others, to search for more advantageous solutions to incomplete evaluation appraisals based on genetic programming. Overall, this research aspires to bring the concept of Machine-Centered Anthropocentrism to the table of EU policymaking. It acknowledges that, when put into service, credit analysts no longer exert full control over the data-driven entities programmers have given ‘birth’ to. With similar explanatory agents under supervision, AI itself can become self-accountable, prioritizing human concerns and values. AI decisions should not be vilified inherently. The issue lies in how they are integrated into decision-making and whether they align with non-discrimination principles and transparency rules.

Keywords: creditworthiness assessment, hybrid oversight, machine-centered anthropocentrism, EU policymaking

Procedia PDF Downloads 34
8 Settings of Conditions Leading to Reproducible and Robust Biofilm Formation in vitro in Evaluation of Drug Activity against Staphylococcal Biofilms

Authors: Adela Diepoltova, Klara Konecna, Ondrej Jandourek, Petr Nachtigal

Abstract:

A loss of control over antibiotic-resistant pathogens has become a global issue due to severe and often untreatable infections. This state is reflected in complicated treatment, health costs, and higher mortality. All these factors emphasize the urgent need for the discovery and development of new anti-infectives. One of the most common pathogens mentioned in the phenomenon of antibiotic resistance are bacteria of the genus Staphylococcus. These bacterial agents have developed several mechanisms against the effect of antibiotics. One of them is biofilm formation. In staphylococci, biofilms are associated with infections such as endocarditis, osteomyelitis, catheter-related bloodstream infections, etc. To author's best knowledge, no validated and standardized methodology evaluating candidate compound activity against staphylococcal biofilms exists. However, a variety of protocols for in vitro drug activity testing has been suggested, yet there are often fundamental differences. Based on our experience, a key methodological step that leads to credible results is to form a robust biofilm with appropriate attributes such as firm adherence to the substrate, a complex arrangement in layers, and the presence of extracellular polysaccharide matrix. At first, for the purpose of drug antibiofilm activity evaluation, the focus was put on various conditions (supplementation of cultivation media by human plasma/fetal bovine serum, shaking mode, the density of initial inoculum) that should lead to reproducible and robust in vitro staphylococcal biofilm formation in microtiter plate model. Three model staphylococcal reference strains were included in the study: Staphylococcus aureus (ATCC 29213), methicillin-resistant Staphylococcus aureus (ATCC 43300), and Staphylococcus epidermidis (ATCC 35983). The total biofilm biomass was quantified using the Christensen method with crystal violet, and results obtained from at least three independent experiments were statistically processed. Attention was also paid to the viability of the biofilm-forming staphylococcal cells and the presence of extracellular polysaccharide matrix. The conditions that led to robust biofilm biomass formation with attributes for biofilms mentioned above were then applied by introducing an alternative method analogous to the commercially available test system, the Calgary Biofilm Device. In this test system, biofilms are formed on pegs that are incorporated into the lid of the microtiter plate. This system provides several advantages (in situ detection and quantification of biofilm microbial cells that have retained their viability after drug exposure). Based on our preliminary studies, it was found that the attention to the peg surface and substrate on which the bacterial biofilms are formed should also be paid to. Therefore, further steps leading to the optimization were introduced. The surface of pegs was coated by human plasma, fetal bovine serum, and L-polylysine. Subsequently, the willingness of bacteria to adhere and form biofilm was monitored. In conclusion, suitable conditions were revealed, leading to the formation of reproducible, robust staphylococcal biofilms in vitro for the microtiter model and the system analogous to the Calgary biofilm device, as well. The robustness and typical slime texture could be detected visually. Likewise, an analysis by confocal laser scanning microscopy revealed a complex three-dimensional arrangement of biofilm forming organisms surrounded by an extracellular polysaccharide matrix.

Keywords: anti-biofilm drug activity screening, in vitro biofilm formation, microtiter plate model, the Calgary biofilm device, staphylococcal infections, substrate modification, surface coating

Procedia PDF Downloads 155
7 Evaluation of Coal Quality and Geomechanical Moduli Using Core and Geophysical Logs: Study from Middle Permian Barakar Formation of Gondwana Coalfield

Authors: Joyjit Dey, Souvik Sen

Abstract:

Middle Permian Barakar formation is the major economic coal bearing unit of vast east-west trending Damodar Valley basin of Gondwana coalfield. Primary sedimentary structures were studied from the core holes, which represent majorly four facies groups: sandstone dominated facies, sandstone-shale heterolith facies, shale facies and coal facies. Total eight major coal seams have been identified with the bottom most seam being the thickest. Laterally, continuous coal seams were deposited in the calm and quiet environment of extensive floodplain swamps. Channel sinuosity and lateral channel migration/avulsion results in lateral facies heterogeneity and coal splitting. Geophysical well logs (Gamma-Resistivity-Density logs) have been used to establish the vertical and lateral correlation of various litho units field-wide, which reveals the predominance of repetitive fining upwards cycles. Well log data being a permanent record, offers a strong foundation for generating log based property evaluation and helps in characterization of depositional units in terms of lateral and vertical heterogeneity. Low gamma, high resistivity, low density is the typical coal seam signatures in geophysical logs. Here, we have used a density cutoff of 1.6 g/cc as a primary discriminator of coal and the same has been employed to compute various coal assay parameters, which are ash, fixed carbon, moisture, volatile content, cleat porosity, vitrinite reflectance (VRo%), which were calibrated with the laboratory based measurements. The study shows ash content and VRo% increase from west to east (towards basin margin), while fixed carbon, moisture and volatile content increase towards west, depicting increased coal quality westwards. Seam wise cleat porosity decreases from east to west, this would be an effect of overburden, as overburden pressure increases westward with the deepening of basin causing more sediment packet deposited on the western side of the study area. Coal is a porous, viscoelastic material in which velocity and strain both change nonlinearly with stress, especially for stress applied perpendicular to the bedding plane. Usually, the coal seam has a high velocity contrast relative to its neighboring layers. Despite extensive discussion of the maceral and chemical properties of coal, its elastic characteristics have received comparatively little attention. The measurement of the elastic constants of coal presents many difficulties: sample-to-sample inhomogeneity and fragility and velocity dependence on stress, orientation, humidity, and chemical content. In this study, a conclusive empirical equation VS= 0.80VP-0.86 has been used to model shear velocity from compression velocity. Also the same has been used to compute various geomechanical moduli. Geomech analyses yield a Poisson ratio of 0.348 against coals. Average bulk modulus value is 3.97 GPA, while average shear modulus and Young’s modulus values are coming out as 1.34 and 3.59 GPA respectively. These middle Permian Barakar coals show an average 23.84 MPA uniaxial compressive strength (UCS) with 4.97 MPA cohesive strength and 0.46 as friction coefficient. The output values of log based proximate parameters and geomechanical moduli suggest a medium volatile Bituminous grade for the studied coal seams, which is found in the laboratory based core study as well.

Keywords: core analysis, coal characterization, geophysical log, geo-mechanical moduli

Procedia PDF Downloads 226
6 Identifying the Conservation Gaps in Poorly Studied Protected Area in the Philippines: A Study Case of Sibuyan Island

Authors: Roven Tumaneng, Angelica Kristina Monzon, Ralph Sedricke Lapuz, Jose Don De Alban, Jennica Paula Masigan, Joanne Rae Pales, Laila Monera Pornel, Dennis Tablazon, Rizza Karen Veridiano, Jackie Lou Wenceslao, Edmund Leo Rico, Neil Aldrin Mallari

Abstract:

Most protected area management plans in the Philippines, particularly the smaller and more remote islands suffer from insufficient baseline data, which should provide the bases for formulating measureable conservation targets and appropriate management interventions for these protected areas. Attempts to synthesize available data particularly on cultural and socio-economic characteristic of local peoples within and outside protected areas also suffer from the lack of comprehensive and detailed inventories, which should be considered in designing adaptive management interventions to be used for those protected areas. Mt Guiting-guiting Natural Park (MGGNP) located in Sibuyan Island is one of the poorly studied protected areas in the Philippines. In this study, we determined the highly biologically important areas of the protected area using Maximum Entropy approach (MaxEnt) from environmental predictors (i.e., topographic, bioclimatic,land cover, and soil image layers) derived from global remotely sensed data and point occurrence data of species of birds and trees recorded during field surveys on the island. A total of 23 trigger species of birds and trees was modeled and stacked to generate species richness maps for biological high conservation value areas (HCVAs). Forest habitat change was delineated using dual-polarised L-band ALOS-PALSAR mosaic data at 25 meter spatial resolution, taken at two acquisition years 2007 and 2009 to provide information on forest cover ad habitat change in the island between year 2007 and 2009. Determining the livelihood guilds were also conducted using the data gathered from171 household interviews, from which demographic and livelihood variables were extracted (i.e., age, gender, number of household members, educational attainment, years of residency, distance from forest edge, main occupation, alternative sources of food and resources during scarcity months, and sources of these alternative resources).Using Principal Component Analysis (PCA) and Kruskal-Wallis test, the diversity and patterns of forest resource use by people in the island were determined with particular focus on the economic activities that directly and indirectly affect the population of key species as well as to identify levels of forest resource use by people in different areas of the park.Results showed that there are gaps in the area occupied by the natural park, as evidenced by the mismatch of the proposed HCVAs and the existing perimeters of the park. We found out that subsistence forest gathering was the possible main driver for forest degradation out of the eight livelihood guilds that were identified in the park. Determining the high conservation areas and identifyingthe anthropogenic factors that influence the species richness and abundance of key species in the different management zone of MGGNP would provide guidance for the design of a protected area management plan and future monitoring programs. However, through intensive communication and consultation with government stakeholders and local communities our results led to setting conservation targets in local development plans and serve as a basis for the reposition of the boundaries and reconfiguration of the management zones of MGGNP.

Keywords: conservation gaps, livelihood guilds, MaxEnt, protected area

Procedia PDF Downloads 407
5 Moths of Indian Himalayas: Data Digging for Climate Change Monitoring

Authors: Angshuman Raha, Abesh Kumar Sanyal, Uttaran Bandyopadhyay, Kaushik Mallick, Kamalika Bhattacharyya, Subrata Gayen, Gaurab Nandi Das, Mohd. Ali, Kailash Chandra

Abstract:

Indian Himalayan Region (IHR), due to its sheer latitudinal and altitudinal expanse, acts as a mixing ground for different zoogeographic faunal elements. The innumerable unique and distributional restricted rare species of IHR are constantly being threatened with extinction by the ongoing climate change scenario. Many of which might have faced extinction without even being noticed or discovered. Monitoring the community dynamics of a suitable taxon is indispensable to assess the effect of this global perturbation at micro-habitat level. Lepidoptera, particularly moths are suitable for this purpose due to their huge diversity and strict herbivorous nature. The present study aimed to collate scattered historical records of moths from IHR and spatially disseminate the same in Geographic Information System (GIS) domain. The study also intended to identify moth species with significant altitudinal shifts which could be prioritised for monitoring programme to assess the effect of climate change on biodiversity. A robust database on moths recorded from IHR was prepared from voluminous secondary literature and museum collections. Historical sampling points were transformed into richness grids which were spatially overlaid on altitude, annual precipitation and vegetation layers separately to show moth richness patterns along major environmental gradients. Primary samplings were done by setting standard light traps at 11 Protected Areas representing five Indian Himalayan biogeographic provinces. To identify significant altitudinal shifts, past and present altitudinal records of the identified species from primary samplings were compared. A consolidated list of 4107 species belonging to 1726 genera of 62 families of moths was prepared from a total of 10,685 historical records from IHR. Family-wise assemblage revealed Erebidae to be the most speciose family with 913 species under 348 genera, followed by Geometridae with 879 species under 309 genera and Noctuidae with 525 species under 207 genera. Among biogeographic provinces, Central Himalaya represented maximum records with 2248 species, followed by Western and North-western Himalaya with 1799 and 877 species, respectively. Spatial analysis revealed species richness was more or less uniform (up to 150 species record per cell) across IHR. Throughout IHR, the middle elevation zones between 1000-2000m encompassed high species richness. Temperate coniferous forest associated with 1500-2000mm rainfall zone showed maximum species richness. Total 752 species of moths were identified representing 23 families from the present sampling. 13 genera were identified which were restricted to specialized habitats of alpine meadows over 3500m. Five historical localities with high richness of >150 species were selected which could be considered for repeat sampling to assess climate change influence on moth assemblage. Of the 7 species exhibiting significant altitudinal ascend of >2000m, Trachea auriplena, Diphtherocome fasciata (Noctuidae) and Actias winbrechlini (Saturniidae) showed maximum range shift of >2500m, indicating intensive monitoring of these species. Great Himalayan National Park harbours most diverse assemblage of high-altitude restricted species and should be a priority site for habitat conservation. Among the 13 range restricted genera, Arichanna, Opisthograptis, Photoscotosia (Geometridae), Phlogophora, Anaplectoides and Paraxestia (Noctuidae) were dominant and require rigorous monitoring, as they are most susceptible to climatic perturbations.

Keywords: altitudinal shifts, climate change, historical records, Indian Himalayan region, Lepidoptera

Procedia PDF Downloads 169
4 Microdiamond and Moissanite Inclusions in Garnets from Pohorje Mountains, Eastern Alps, Slovenia

Authors: Mirijam Vrabec, Marian Janak, Bojan Ambrozic, Angelja K. Surca, Nastja Rogan Smuc, Nina Zupancic, Saso Sturm

Abstract:

Natural microdiamonds and moissanite (SiC) can form during the orogenic events under ultrahigh-pressure metamorphic conditions (UHP), when parts of Earth’s crust are subducted to extreme depths. So far, such processes were identified only in few places on the Earth, and therefore, represent unique opportunity to study the evolution of the Earth’s deep interior. An important discovery of microdiamonds and moissanite was reported from Pohorje, (Slovenia), where they occurred as single or polyphase inclusions in garnets. Metasedimentary rocks from Pohorje are predominantly gneisses representing parts of the Austroalpine metamorphic units of the Eastern Alps. During Cretaceous orogeny, (ca. 95–92 Ma) continental crustal rocks were deeply subducted to the mantle depths (below 100 km) and metamorphosed at pressures exceeding 3.5 GPa and temperatures between 800–850 °C. Microstructural and phase analysis of the inclusions as well as detailed elemental analysis of host garnets were carried out combining several analytical techniques: optical microscope in plane polarized transmitted light, electron probe microanalysis (EPMA) with wavelength-dispersive x-ray spectrometry (WDS) and field-emission scanning microscope (FEG-SEM) with energy-dispersive x-ray spectroscopy (EDS). Micro-Raman analysis revealed sharp, first order diamond bands sometimes accompanied by graphite bands implying that transformation of diamond back to graphite occurred. To study the chemical and crystallographic relationship between microdiamonds and co-inclusions, advanced techniques of transmission electron microscopy (TEM) were applied, which included high-angle annular dark-field scanning transmission electron microscopy (HAADF-STEM), combined with EDS and electron energy-loss spectroscopy (EELS). To prepare electron transparent TEM lamellae selectively a dual-beam Focused Ion Beam/SEM (FIB/SEM) was employed. Detailed study of TEM lamellae, which was cross-sectioned from the highly faceted inclusion body located within the host garnet crystal matrix, revealed rich and rather complex internal structure. Namely, the negative crystal facets of the main inclusion body were typically decorated with up to 1 μm thick amorphous layer, reflecting the general garnet composition with slight variations in Fe/Ca content. Within these layers, ELNES analysis revealed the presence of a 28–30 nm thick layer of amorphous carbon. The very last section of this layer corresponds to composition of SiO2. Within the inclusion, besides diamond and moissanite alumosilicate mineral with pronounced layered structure, iron sulfides and chlorine were identified under TEM and CO2 and CH4 using Raman. Moissanite is found as single crystal or composed from numerous highly textured nano-crystals with the average size of 10 nm. Moissanite inclusions were found embedded inside the amorphous crust implying that moissanite crystalized well before the deposition of the amorphous layer. From the microstructural, crystallographic and chemical observations so far we can deduce, that polyphase inclusions in diamond bearing garnets from Pohorje most probably crystallized from reduced supercritical fluids. Based on layered interface structure of the host mineral multiphase process of crystallization is possible. The presence of microdiamonds and moissanite in rocks from Pohorje demonstrates that these parts of the Eastern Alps were subducted to extreme depths, and were subsequently exhumed back to the Earth's surface without complete breakdown of UHP mineral phases, allowing a rear and exceptional opportunity to study them in-situ.

Keywords: diamond, fluid inclusions, moissanite, TEM, UHP metamorphism.

Procedia PDF Downloads 304
3 Developing a Place-Name Gazetteer for Singapore by Mining Historical Planning Archives and Selective Crowd-Sourcing

Authors: Kevin F. Hsu, Alvin Chua, Sarah X. Lin

Abstract:

As a multilingual society, Singaporean names for different parts of the city have changed over time. Residents included Indigenous Malays, dialect-speakers from China, European settler-colonists, and Tamil-speakers from South India. Each group would name locations in their own languages. Today, as ancestral tongues are increasingly supplanted by English, contemporary Singaporeans’ understanding of once-common place names is disappearing. After demolition or redevelopment, some urban places will only exist in archival records or in human memory. United Nations conferences on the standardization of geographic names have called attention to how place names relate to identity, well-being, and a sense of belonging. The Singapore Place-Naming Project responds to these imperatives by capturing past and present place names through digitizing historical maps, mining archival records, and applying selective crowd-sourcing to trace the evolution of place names throughout the city. The project ensures that both formal and vernacular geographical names remain accessible to historians, city planners, and the public. The project is compiling a gazetteer, a geospatial archive of placenames, with streets, buildings, landmarks, and other points of interest (POI) appearing in the historic maps and planning documents of Singapore, currently held by the National Archives of Singapore, the National Library Board, university departments, and the Urban Redevelopment Authority. To create a spatial layer of information, the project links each place name to either a geo-referenced point, line segment, or polygon, along with the original source material in which the name appears. This record is supplemented by crowd-sourced contributions from civil service officers and heritage specialists, drawing from their collective memory to (1) define geospatial boundaries of historic places that appear in past documents, but maybe unfamiliar to users today, and (2) identify and record vernacular place names not captured in formal planning documents. An intuitive interface allows participants to demarcate feature classes, vernacular phrasings, time periods, and other knowledge related to historical or forgotten spaces. Participants are stratified into age bands and ethnicity to improve representativeness. Future iterations could allow additional public contributions. Names reveal meanings that communities assign to each place. While existing historical maps of Singapore allow users to toggle between present-day and historical raster files, this project goes a step further by adding layers of social understanding and planning documents. Tracking place names illuminates linguistic, cultural, commercial, and demographic shifts in Singapore, in the context of transformations of the urban environment. The project also demonstrates how a moderated, selectively crowd-sourced effort can solicit useful geospatial data at scale, sourced from different generations, and at higher granularity than traditional surveys, while mitigating negative impacts of unmoderated crowd-sourcing. Stakeholder agencies believe the project will achieve several objectives, including Supporting heritage conservation and public education; Safeguarding intangible cultural heritage; Providing historical context for street, place or development-renaming requests; Enhancing place-making with deeper historical knowledge; Facilitating emergency and social services by tagging legal addresses to vernacular place names; Encouraging public engagement with heritage by eliciting multi-stakeholder input.

Keywords: collective memory, crowd-sourced, digital heritage, geospatial, geographical names, linguistic heritage, place-naming, Singapore, Southeast Asia

Procedia PDF Downloads 129
2 Times2D: A Time-Frequency Method for Time Series Forecasting

Authors: Reza Nematirad, Anil Pahwa, Balasubramaniam Natarajan

Abstract:

Time series data consist of successive data points collected over a period of time. Accurate prediction of future values is essential for informed decision-making in several real-world applications, including electricity load demand forecasting, lifetime estimation of industrial machinery, traffic planning, weather prediction, and the stock market. Due to their critical relevance and wide application, there has been considerable interest in time series forecasting in recent years. However, the proliferation of sensors and IoT devices, real-time monitoring systems, and high-frequency trading data introduce significant intricate temporal variations, rapid changes, noise, and non-linearities, making time series forecasting more challenging. Classical methods such as Autoregressive integrated moving average (ARIMA) and Exponential Smoothing aim to extract pre-defined temporal variations, such as trends and seasonality. While these methods are effective for capturing well-defined seasonal patterns and trends, they often struggle with more complex, non-linear patterns present in real-world time series data. In recent years, deep learning has made significant contributions to time series forecasting. Recurrent Neural Networks (RNNs) and their variants, such as Long short-term memory (LSTMs) and Gated Recurrent Units (GRUs), have been widely adopted for modeling sequential data. However, they often suffer from the locality, making it difficult to capture local trends and rapid fluctuations. Convolutional Neural Networks (CNNs), particularly Temporal Convolutional Networks (TCNs), leverage convolutional layers to capture temporal dependencies by applying convolutional filters along the temporal dimension. Despite their advantages, TCNs struggle with capturing relationships between distant time points due to the locality of one-dimensional convolution kernels. Transformers have revolutionized time series forecasting with their powerful attention mechanisms, effectively capturing long-term dependencies and relationships between distant time points. However, the attention mechanism may struggle to discern dependencies directly from scattered time points due to intricate temporal patterns. Lastly, Multi-Layer Perceptrons (MLPs) have also been employed, with models like N-BEATS and LightTS demonstrating success. Despite this, MLPs often face high volatility and computational complexity challenges in long-horizon forecasting. To address intricate temporal variations in time series data, this study introduces Times2D, a novel framework that parallelly integrates 2D spectrogram and derivative heatmap techniques. The spectrogram focuses on the frequency domain, capturing periodicity, while the derivative patterns emphasize the time domain, highlighting sharp fluctuations and turning points. This 2D transformation enables the utilization of powerful computer vision techniques to capture various intricate temporal variations. To evaluate the performance of Times2D, extensive experiments were conducted on standard time series datasets and compared with various state-of-the-art algorithms, including DLinear (2023), TimesNet (2023), Non-stationary Transformer (2022), PatchTST (2023), N-HiTS (2023), Crossformer (2023), MICN (2023), LightTS (2022), FEDformer (2022), FiLM (2022), SCINet (2022a), Autoformer (2021), and Informer (2021) under the same modeling conditions. The initial results demonstrated that Times2D achieves consistent state-of-the-art performance in both short-term and long-term forecasting tasks. Furthermore, the generality of the Times2D framework allows it to be applied to various tasks such as time series imputation, clustering, classification, and anomaly detection, offering potential benefits in any domain that involves sequential data analysis.

Keywords: derivative patterns, spectrogram, time series forecasting, times2D, 2D representation

Procedia PDF Downloads 42
1 Impacts of Ibeju - Lekki New Town on Neighbouring Residents of Ibeju, Lagos Nigeria

Authors: Abolade Olajoke, Adigun Folasade Oyenike, Odunjo Oluronke Omolola Olaleye, Babajide Rotimi

Abstract:

Against the shortfall associated with unprecedented urbanization in most cities of the world, coupled with rapid expansion of outer boundaries, is the resultant birth of the development of new towns. The paper therefore examines the impacts of Ibeju - Lekki New Town on Neighbouring communities of Ibeju Lekki. Random systematic sampling was employed elicit relevant information from a total number of 269 residents at interval of five buildings in four neighbouring communities. Descriptive statistics was employed to test for the socioeconomic characteristics of respondents, problems faced by government officials during the implementation and monitoring process. Likert scale was employed to ascertain respondents view on the impact of the new town on neighbouring communities. Result from the findings shows that male (56.9%) are the most dominant occupant in the study area of which most (68.1%) fall within the most the active age group (18-39 and 40-59 years). Results further shows that 36% of the total respondents are traders and majority (32%) earn below government salary wage cap of ₦18000 thus indicating that majority of the respondents are petty traders. Results of findings from development authority reveals that the major problem encountered during monitoring and implementation is harassment of government officials (35%). Result of likert scale further show that new town has brought increase in intensity of land use within neighbouring communities (RAI 3.65), provision of job opportunity (RAI 3.57). This have consequently improve standard of living of the neighbouring community (RAI 3.27). On the contrary some (RAI 1.97) opined that attention should paid to provision of power supply and provision of recreation facilities (RAI I.63). The study recommends that government should make adequate provisions for basic facilities such power supply, adequate health care system, basic education and provision of healthy portable water. This should be given utmost priority to enhance the living condition of residents. To forestall attack from residents’ adequate security measures should be provided as backup for Government official during implementation and monitoring. Appropriate sanction to illegal occupants and demolition of illegal structures should be fully implemented, This will indubitably prevent haphazard development and also promote a liveable environment. Against the shortfall associated with unprecedented urbanization in most cities of the world, coupled with rapid expansion of outer boundaries, is the resultant birth of the development of new towns. The paper therefore examines the impacts of Ibeju - Lekki New Town on Neighbouring communities of Ibeju Lekki. Random systematic sampling was employed elicit relevant information from a total number of 269 residents at interval of five buildings in four neighbouring communities. Descriptive statistics was employed to test for the socioeconomic characteristics of respondents, problems faced by government officials during the implementation and monitoring process. Likert scale was employed to ascertain respondents view on the impact of the new town on neighbouring communities. Result from the findings shows that male (56.9%) are the most dominant occupant in the study area of which most (68.1%) fall within the most the active age group (18-39 and 40-59 years). Results further shows that 36% of the total respondents are traders and majority (32%) earn below government salary wage cap of ₦18000 thus indicating that majority of the respondents are petty traders. Results of findings from development authority reveals that the major problem encountered during monitoring and implementation is harassment of government officials (35%) Result of likert scale further show that new town has brought increase in intensity of land use within neighbouring communities (RAI 3.65), provision of job opportunity (RAI 3.57). This have consequently improve standard of living of the neighbouring community (RAI 3.27). On the contrary some (RAI 1.97) opined that attention should paid to provision of power supply and provision of recreation facilities (RAI I.63). The study recommends that government should make adequate provisions for basic facilities such power supply, adequate health care system, basic education and provision of healthy portable water. This should be given utmost priority to enhance the living condition of residents. To forestall attack from residents’ adequate security measures should be provided as backup for Government official during implementation and monitoring. Appropriate sanction to illegal occupants and demolition of illegal structures should be fully implemented, This will indubitably prevent haphazard development and also promote a liveable environment.

Keywords: new town, urbanization, infrastructure boundary

Procedia PDF Downloads 410