Search results for: continuous data
24726 Rubber Crumbs in Alkali Activated Clay Roof Tiles at Low Temperature
Authors: Aswin Kumar Krishnan, Yat Choy Wong, Reiza Mukhlis, Zipeng Zhang, Arul Arulrajah
Abstract:
The continuous increase in vehicle uptake escalates the number of rubber tyre waste which need to be managed to avoid landfilling and stockpiling. The present research focused on the sustainable use of rubber crumbs in clay roof tiles. The properties of roof tiles composed of clay, rubber crumbs, NaOH, and Na₂SiO₃ with a 10% alkaline activator were studied. Tile samples were fabricated by heating the compacted mixtures at 50°C for 72 hours, followed by a higher heating temperature of 200°C for 24 hours. The effect of rubber crumbs aggregates as a substitution for the raw clay materials was investigated by varying their concentration from 0% to 2.5%. X-ray diffraction (XRD) and scanning electron microscopy (SEM) analyses have been conducted to study the phases and microstructures of the samples. It was found that the optimum rubber crumbs concentration was at 0.5% and 1%, while cracks and larger porosity were found at higher crumbs concentrations. Water absorption and compressive strength test results demonstrated that rubber crumbs and clay satisfied the standard requirement for the roof tiles.Keywords: rubber crumbs, clay, roof tiles, alkaline activators
Procedia PDF Downloads 10424725 Design and Implementation a Virtualization Platform for Providing Smart Tourism Services
Authors: Nam Don Kim, Jungho Moon, Tae Yun Chung
Abstract:
This paper proposes an Internet of Things (IoT) based virtualization platform for providing smart tourism services. The virtualization platform provides a consistent access interface to various types of data by naming IoT devices and legacy information systems as pathnames in a virtual file system. In the other words, the IoT virtualization platform functions as a middleware which uses the metadata for underlying collected data. The proposed platform makes it easy to provide customized tourism information by using tourist locations collected by IoT devices and additionally enables to create new interactive smart tourism services focused on the tourist locations. The proposed platform is very efficient so that the provided tourism services are isolated from changes in raw data and the services can be modified or expanded without changing the underlying data structure.Keywords: internet of things (IoT), IoT platform, serviceplatform, virtual file system (VSF)
Procedia PDF Downloads 50324724 A Review on 3D Smart City Platforms Using Remotely Sensed Data to Aid Simulation and Urban Analysis
Authors: Slim Namouchi, Bruno Vallet, Imed Riadh Farah
Abstract:
3D urban models provide powerful tools for decision making, urban planning, and smart city services. The accuracy of this 3D based systems is directly related to the quality of these models. Since manual large-scale modeling, such as cities or countries is highly time intensive and very expensive process, a fully automatic 3D building generation is needed. However, 3D modeling process result depends on the input data, the proprieties of the captured objects, and the required characteristics of the reconstructed 3D model. Nowadays, producing 3D real-world model is no longer a problem. Remotely sensed data had experienced a remarkable increase in the recent years, especially data acquired using unmanned aerial vehicles (UAV). While the scanning techniques are developing, the captured data amount and the resolution are getting bigger and more precise. This paper presents a literature review, which aims to identify different methods of automatic 3D buildings extractions either from LiDAR or the combination of LiDAR and satellite or aerial images. Then, we present open source technologies, and data models (e.g., CityGML, PostGIS, Cesiumjs) used to integrate these models in geospatial base layers for smart city services.Keywords: CityGML, LiDAR, remote sensing, SIG, Smart City, 3D urban modeling
Procedia PDF Downloads 13524723 Spontaneous Message Detection of Annoying Situation in Community Networks Using Mining Algorithm
Authors: P. Senthil Kumari
Abstract:
Main concerns in data mining investigation are social controls of data mining for handling ambiguity, noise, or incompleteness on text data. We describe an innovative approach for unplanned text data detection of community networks achieved by classification mechanism. In a tangible domain claim with humble secrecy backgrounds provided by community network for evading annoying content is presented on consumer message partition. To avoid this, mining methodology provides the capability to unswervingly switch the messages and similarly recover the superiority of ordering. Here we designated learning-centered mining approaches with pre-processing technique to complete this effort. Our involvement of work compact with rule-based personalization for automatic text categorization which was appropriate in many dissimilar frameworks and offers tolerance value for permits the background of comments conferring to a variety of conditions associated with the policy or rule arrangements processed by learning algorithm. Remarkably, we find that the choice of classifier has predicted the class labels for control of the inadequate documents on community network with great value of effect.Keywords: text mining, data classification, community network, learning algorithm
Procedia PDF Downloads 50824722 Dehydration of Residues from WTP for Application in Building Materials and Reuse of Water from the Waste Treatment: A Feasible Solution to Complete Treatment Systems
Authors: Marco Correa, Flavio Araujo, Paulo Scalize, Antonio Albuquerque
Abstract:
The increasing reduction of the volumes of surface water sources which supply most municipalities, as well as the continued rise of demand for treated water, combined with the disposal of effluents from washing of decanters and filters of the water treatment plants, generates a continuous search for correct environmentally solutions to these problems. The effluents generated by the water treatment industry need to be suitably processed for return to the environment or re-use. This article shows an alternative for the dehydration of sludge from the water treatment plants (WTP) and eventual disposal of sludge drained. Using the simple design methodology, we present a case study for a drainage in tanks geotextile, full-scale, which involve five sludge drainage tanks from WTP of the Rio Verde City. Aiming to the reutilization the water drained from the sludge and enabling its reuse both at the beginning of the treatment process at the WTP and in less noble services as for watering the gardens of the local town hall. The sludge will be used to production of building materials.Keywords: re-use, residue, sustainable, water treatment plants, sludge
Procedia PDF Downloads 49024721 Expanding the Evaluation Criteria for a Wind Turbine Performance
Authors: Ivan Balachin, Geanette Polanco, Jiang Xingliang, Hu Qin
Abstract:
The problem of global warming raised up interest towards renewable energy sources. To reduce cost of wind energy is a challenge. Before building of wind park conditions such as: average wind speed, direction, time for each wind, probability of icing, must be considered in the design phase. Operation values used on the setting of control systems also will depend on mentioned variables. Here it is proposed a procedure to be include in the evaluation of the performance of a wind turbine, based on the amplitude of wind changes, the number of changes and their duration. A generic study case based on actual data is presented. Data analysing techniques were applied to model the power required for yaw system based on amplitude and data amount of wind changes. A theoretical model between time, amplitude of wind changes and angular speed of nacelle rotation was identified.Keywords: field data processing, regression determination, wind turbine performance, wind turbine placing, yaw system losses
Procedia PDF Downloads 39024720 An Exhaustive All-Subsets Examination of Trade Theory on WTO Data
Authors: Masoud Charkhabi
Abstract:
We examine trade theory with this motivation. The full set of World Trade Organization data are organized into country-year pairs, each treated as a different entity. Topological Data Analysis reveals that among the 16 region and 240 region-year pairs there exists in fact a distinguishable group of region-period pairs. The generally accepted periods of shifts from dissimilar-dissimilar to similar-similar trade in goods among regions are examined from this new perspective. The period breaks are treated as cumulative and are flexible. This type of all-subsets analysis is motivated from computer science and is made possible with Lossy Compression and Graph Theory. The results question many patterns in similar-similar to dissimilar-dissimilar trade. They also show indications of economic shifts that only later become evident in other economic metrics.Keywords: econometrics, globalization, network science, topological data, analysis, trade theory, visualization, world trade
Procedia PDF Downloads 37224719 Preparation, Characterisation and Electrical Properties of Metal/polymer-DNA Nanohybrids
Authors: Mahdi Almaky
Abstract:
Conducting polymer of N-(3-pyrrol-1-yl-propyl)-2,2`-bipyridinium hexafluoro-phosphate (PPBH) was prepared via chemical and electrochemical polymerization methods. The bulk polymer showed conductivity in the order of 10-12 S cm-1. DNA-templated polymer nano wires of PPBH (PolyPPBH-DNA) have been chemically prepared then used as templates to direct the formation of metal nanowires (Cu) in order to enhance the electrical properties of the polymer/DNA wires. The chemical structures, morphology and the electrical characterisation of the as obtained structures have been characterized through spectroscopic (FTIR, UV-vis and XPS), single-crystal X-ray diffraction and microscopic (AFM, EFM and c-AFM) techniques. The morphology of the nanomaterials has been observed by AFM; showing the nanowires are uniform and continuous. The polymer conductivity was slightly improved after metallization. The conductivity of Cu-PolyPPBH-DNA nanowires was estimated to be 7.1x10-2 S cm-1. This conductivity is slightly higher than the conductivity of PolyPPBH-DNA nano wires (2.0 x 10-2 S cm-1), but it is lower than the measurements for PPy/DNA nano wires (2.1 x 10-1 S cm-1) prepared and measured by using c-AFM probe. These results reflect the large effect of the chemical structure (N-substitution) on the electrical properties of these polymers by reducing the extended conjugation.Keywords: DNA, template, nano wires, N-Alkylatedpyrrole, copper
Procedia PDF Downloads 42224718 Discussing Embedded versus Central Machine Learning in Wireless Sensor Networks
Authors: Anne-Lena Kampen, Øivind Kure
Abstract:
Machine learning (ML) can be implemented in Wireless Sensor Networks (WSNs) as a central solution or distributed solution where the ML is embedded in the nodes. Embedding improves privacy and may reduce prediction delay. In addition, the number of transmissions is reduced. However, quality factors such as prediction accuracy, fault detection efficiency and coordinated control of the overall system suffer. Here, we discuss and highlight the trade-offs that should be considered when choosing between embedding and centralized ML, especially for multihop networks. In addition, we present estimations that demonstrate the energy trade-offs between embedded and centralized ML. Although the total network energy consumption is lower with central prediction, it makes the network more prone for partitioning due to the high forwarding load on the one-hop nodes. Moreover, the continuous improvements in the number of operations per joule for embedded devices will move the energy balance toward embedded prediction.Keywords: central machine learning, embedded machine learning, energy consumption, local machine learning, wireless sensor networks, WSN
Procedia PDF Downloads 15424717 Using Probe Person Data for Travel Mode Detection
Authors: Muhammad Awais Shafique, Eiji Hato, Hideki Yaginuma
Abstract:
Recently GPS data is used in a lot of studies to automatically reconstruct travel patterns for trip survey. The aim is to minimize the use of questionnaire surveys and travel diaries so as to reduce their negative effects. In this paper data acquired from GPS and accelerometer embedded in smart phones is utilized to predict the mode of transportation used by the phone carrier. For prediction, Support Vector Machine (SVM) and Adaptive boosting (AdaBoost) are employed. Moreover a unique method to improve the prediction results from these algorithms is also proposed. Results suggest that the prediction accuracy of AdaBoost after improvement is relatively better than the rest.Keywords: accelerometer, AdaBoost, GPS, mode prediction, support vector machine
Procedia PDF Downloads 35924716 Building Energy Modeling for Networks of Data Centers
Authors: Eric Kumar, Erica Cochran, Zhiang Zhang, Wei Liang, Ronak Mody
Abstract:
The objective of this article was to create a modelling framework that exposes the marginal costs of shifting workloads across geographically distributed data-centers. Geographical distribution of internet services helps to optimize their performance for localized end users with lowered communications times and increased availability. However, due to the geographical and temporal effects, the physical embodiments of a service's data center infrastructure can vary greatly. In this work, we first identify that the sources of variances in the physical infrastructure primarily stem from local weather conditions, specific user traffic profiles, energy sources, and the types of IT hardware available at the time of deployment. Second, we create a traffic simulator that indicates the IT load at each data-center in the set as an approximator for user traffic profiles. Third, we implement a framework that quantifies the global level energy demands using building energy models and the traffic profiles. The results of the model provide a time series of energy demands that can be used for further life cycle analysis of internet services.Keywords: data-centers, energy, life cycle, network simulation
Procedia PDF Downloads 14724715 Predicting National Football League (NFL) Match with Score-Based System
Authors: Marcho Setiawan Handok, Samuel S. Lemma, Abdoulaye Fofana, Naseef Mansoor
Abstract:
This paper is proposing a method to predict the outcome of the National Football League match with data from 2019 to 2022 and compare it with other popular models. The model uses open-source statistical data of each team, such as passing yards, rushing yards, fumbles lost, and scoring. Each statistical data has offensive and defensive. For instance, a data set of anticipated values for a specific matchup is created by comparing the offensive passing yards obtained by one team to the defensive passing yards given by the opposition. We evaluated the model’s performance by contrasting its result with those of established prediction algorithms. This research is using a neural network to predict the score of a National Football League match and then predict the winner of the game.Keywords: game prediction, NFL, football, artificial neural network
Procedia PDF Downloads 8424714 Economical and Environmental Impact of Deforestation on Charcoal Production in Gaza Province
Authors: Paulo Cumbe
Abstract:
This work analyzes the economic and environmental impact of the exploitation of forest resources on populations and their sustainability in the regions where it occurs. There is an intensive and continuous activity of charcoal production, in the Massingir and Mabalane districts, in Gaza, Mozambique, to supply the most used fuel that is used by the population of the capital city, Maputo. Charcoal production is one of the sources of income for several families. However, it causes a negative environmental impact on biodiversity. We have analyzed different studies carried out in these communities that measure the speed, the level, and the impact of deforestation involving different actors, to deepen our understanding of this issue. The results of these studies reveal that the degraded area in five years would need one hundred years to be restored, which is unsustainable from an environmental point of view it is. Populations seek new areas for the same practice to maintain their livelihood, progressing with ecosystem degradation and increasing carbon dioxide emissions into the atmosphere. It is believed that environmental education, creation, and dissemination of new forms of charcoal production that are more profitable and less aggressive to the environment and forest repopulation actions need to be carried out to guarantee the sustainable development of the populations in these regions.Keywords: deforestation, emissions, sustainability, charcoal
Procedia PDF Downloads 7024713 Nonlinear Vibration Analysis of a Functionally Graded Micro-Beam under a Step DC Voltage
Authors: Ali Raheli, Rahim Habibifar, Behzad Mohammadi-Alasti, Mahdi Abbasgholipour
Abstract:
This paper presents vibration behavior of a FGM micro-beam and its pull-in instability under a nonlinear electrostatic pressure. An exponential function has been applied to show the continuous gradation of the properties along thickness. Nonlinear integro-differential-electro-mechanical equation based on Euler–Bernoulli beam theory has been derived. The governing equation in the static analysis has been solved using Step-by-Step Linearization Method and Finite Difference Method. Fixed points or equilibrium positions and singular points have been shown in the state control space. In order to find the response to a step DC voltage, the nonlinear equation of motion has been solved using Galerkin-based reduced-order model and time histories and phase portrait for different applied voltages have been shown. The effects of electrostatic pressure on stability of FGM micro-beams having various amounts of the ceramic constituent have been investigated.Keywords: FGM, MEMS, nonlinear vibration, electrical, dynamic pull-in voltage
Procedia PDF Downloads 45724712 Assimilating Multi-Mission Satellites Data into a Hydrological Model
Authors: Mehdi Khaki, Ehsan Forootan, Joseph Awange, Michael Kuhn
Abstract:
Terrestrial water storage, as a source of freshwater, plays an important role in human lives. Hydrological models offer important tools for simulating and predicting water storages at global and regional scales. However, their comparisons with 'reality' are imperfect mainly due to a high level of uncertainty in input data and limitations in accounting for all complex water cycle processes, uncertainties of (unknown) empirical model parameters, as well as the absence of high resolution (both spatially and temporally) data. Data assimilation can mitigate this drawback by incorporating new sets of observations into models. In this effort, we use multi-mission satellite-derived remotely sensed observations to improve the performance of World-Wide Water Resources Assessment system (W3RA) hydrological model for estimating terrestrial water storages. For this purpose, we assimilate total water storage (TWS) data from the Gravity Recovery And Climate Experiment (GRACE) and surface soil moisture data from the Advanced Microwave Scanning Radiometer for the Earth Observing System (AMSR-E) into W3RA. This is done to (i) improve model estimations of water stored in ground and soil moisture, and (ii) assess the impacts of each satellite of data (from GRACE and AMSR-E) and their combination on the final terrestrial water storage estimations. These data are assimilated into W3RA using the Ensemble Square-Root Filter (EnSRF) filtering technique over Mississippi Basin (the United States) and Murray-Darling Basin (Australia) between 2002 and 2013. In order to evaluate the results, independent ground-based groundwater and soil moisture measurements within each basin are used.Keywords: data assimilation, GRACE, AMSR-E, hydrological model, EnSRF
Procedia PDF Downloads 28924711 The Optimization of Immobilization Conditions for Biohydrogen Production from Palm Industry Wastewater
Authors: A. W. Zularisam, Sveta Thakur, Lakhveer Singh, Mimi Sakinah Abdul Munaim
Abstract:
Clostridium sp. LS2 was immobilised by entrapment in polyethylene glycol (PEG) gel beads to improve the biohydrogen production rate from palm oil mill effluent (POME). We sought to explore and optimise the hydrogen production capability of the immobilised cells by studying the conditions for cell immobilisation, including PEG concentration, cell loading and curing times, as well as the effects of temperature and K2HPO4 (500–2000 mg/L), NiCl2 (0.1–5.0 mg/L), FeCl2 (100–400 mg/L) MgSO4 (50–200 mg/L) concentrations on hydrogen production rate. The results showed that by optimising the PEG concentration (10% w/v), initial biomass (2.2 g dry weight), curing time (80 min) and temperature (37 °C), as well as the concentrations of K2HPO4 (2000 mg/L), NiCl2 (1 mg/L), FeCl2 (300 mg/L) and MgSO4 (100 mg/L), a maximum hydrogen production rate of 7.3 L/L-POME/day and a yield of 0.31 L H2/g chemical oxygen demand were obtained during continuous operation. We believe that this process may be potentially expanded for sustained and large-scale hydrogen production.Keywords: hydrogen, polyethylene glycol, immobilised cell, fermentation, palm oil mill effluent
Procedia PDF Downloads 27224710 The Numerical Model of the Onset of Acoustic Oscillation in Pulse Tube Engine
Authors: Alexander I. Dovgyallo, Evgeniy A. Zinoviev, Svetlana O. Nekrasova
Abstract:
The most of works applied for the pulse tube converters contain the workflow description implemented through the use of mathematical models on stationary modes. However, the study of the thermoacoustic systems unsteady behavior in the start, stop, and acoustic load changes modes is in the particular interest. The aim of the present study was to develop a mathematical thermal excitation model of acoustic oscillations in pulse tube engine (PTE) as a small-scale scheme of pulse tube engine operating at atmospheric air. Unlike some previous works this standing wave configuration is a fully closed system. The improvements over previous mathematical models are the following: the model allows specifying any values of porosity for regenerator, takes into account the piston weight and the friction in the cylinder and piston unit, and determines the operating frequency. The numerical method is based on the relation equations between the pressure and volume velocity variables at the ends of each element of PTE which is recorded through the appropriate transformation matrix. A solution demonstrates that the PTE operation frequency is the complex value, and it depends on the piston mass and the dynamic friction due to its movement in the cylinder. On the basis of the determined frequency thermoacoustically induced heat transport and generation of acoustic power equations were solved for channel with temperature gradient on its ends. The results of numerical simulation demonstrate the features of the initialization process of oscillation and show that that generated acoustic power more than power on the steady mode in a factor of 3…4. But doesn`t mean the possibility of its further continuous utilizing due to its existence only in transient mode which lasts only for a 30-40 sec. The experiments were carried out on small-scale PTE. The results shows that the value of acoustic power is in the range of 0.7..1.05 W for the defined frequency range f = 13..18 Hz and pressure amplitudes 11..12 kPa. These experimental data are satisfactorily correlated with the numerical modeling results. The mathematical model can be straightforwardly applied for the thermoacoustic devices with variable temperatures of thermal reservoirs and variable transduction loads which are expected to occur in practical implementations of portable thermoacoustic engines.Keywords: nonlinear processes, pulse tube engine, thermal excitation, standing wave
Procedia PDF Downloads 37724709 Manage an Acute Pain Unit based on the Balanced Scorecard
Authors: Helena Costa Oliveira, Carmem Oliveira, Rita Moutinho
Abstract:
The Balanced Scorecard (BSC) is a continuous strategic monitoring model focused not only on financial issues but also on internal processes, patients/users, and learning and growth. Initially dedicated to business management, it currently serves organizations of other natures - such as hospitals. This paper presents a BSC designed for a Portuguese Acute Pain Unit (APU). This study is qualitative and based on the experience of collaborators at the APU. The management of APU is based on four perspectives – users, internal processes, learning and growth, and financial and legal. For each perspective, there were identified strategic objectives, critical factors, lead indicators and initiatives. The strategic map of the APU outlining sustained strategic relations among strategic objectives. This study contributes to the development of research in the health management area as it explores how organizational insufficiencies and inconsistencies in this particular case can be addressed, through the identification of critical factors, to clearly establish core outcomes and initiatives to set up.Keywords: acute pain unit, balanced scorecard, hospital management, organizational performance, Portugal
Procedia PDF Downloads 14824708 Reinforcement of an Electric Vehicle Battery Pack Using Honeycomb Structures
Authors: Brandon To, Yong S. Park
Abstract:
As more battery electric vehicles are being introduced into the automobile industry, continuous advancements are constantly made in the electric vehicle space. Improvements in lithium-ion battery technology allow electric vehicles to be capable of traveling long distances. The batteries are capable of being charged faster, allowing for a sufficient range in shorter amounts of time. With increased reliance on battery technology and the changes in vehicle power trains, new challenges arise from this. Resulting electric vehicle fires caused by collisions are potentially more dangerous than those of the typical internal combustion engine. To further reduce the battery failures involved with side collisions, this project intends to reinforce an existing battery pack of an electric vehicle with honeycomb structures such that intrusion into the batteries can be minimized with weight restrictions in place. Honeycomb structures of hexagonal geometry are implemented into the side extrusions of the battery pack. With the use of explicit dynamics simulations performed in ANSYS, quantitative results such as deformation, strain, and stress are used to compare the performance of the battery pack with and without the implemented honeycomb structures.Keywords: battery pack, electric vehicle, honeycomb, side impact
Procedia PDF Downloads 12124707 Airborne CO₂ Lidar Measurements for Atmospheric Carbon and Transport: America (ACT-America) Project and Active Sensing of CO₂ Emissions over Nights, Days, and Seasons 2017-2018 Field Campaigns
Authors: Joel F. Campbell, Bing Lin, Michael Obland, Susan Kooi, Tai-Fang Fan, Byron Meadows, Edward Browell, Wayne Erxleben, Doug McGregor, Jeremy Dobler, Sandip Pal, Christopher O'Dell, Ken Davis
Abstract:
The Active Sensing of CO₂ Emissions over Nights, Days, and Seasons (ASCENDS) CarbonHawk Experiment Simulator (ACES) is a NASA Langley Research Center instrument funded by NASA’s Science Mission Directorate that seeks to advance technologies critical to measuring atmospheric column carbon dioxide (CO₂ ) mixing ratios in support of the NASA ASCENDS mission. The ACES instrument, an Intensity-Modulated Continuous-Wave (IM-CW) lidar, was designed for high-altitude aircraft operations and can be directly applied to space instrumentation to meet the ASCENDS mission requirements. The ACES design demonstrates advanced technologies critical for developing an airborne simulator and spaceborne instrument with lower platform consumption of size, mass, and power, and with improved performance. The Atmospheric Carbon and Transport – America (ACT-America) is an Earth Venture Suborbital -2 (EVS-2) mission sponsored by the Earth Science Division of NASA’s Science Mission Directorate. A major objective is to enhance knowledge of the sources/sinks and transport of atmospheric CO₂ through the application of remote and in situ airborne measurements of CO₂ and other atmospheric properties on spatial and temporal scales. ACT-America consists of five campaigns to measure regional carbon and evaluate transport under various meteorological conditions in three regional areas of the Continental United States. Regional CO₂ distributions of the lower atmosphere were observed from the C-130 aircraft by the Harris Corp. Multi-Frequency Fiber Laser Lidar (MFLL) and the ACES lidar. The airborne lidars provide unique data that complement the more traditional in situ sensors. This presentation shows the applications of CO₂ lidars in support of these science needs.Keywords: CO₂ measurement, IMCW, CW lidar, laser spectroscopy
Procedia PDF Downloads 16224706 Surface-Enhanced Raman Detection in Chip-Based Chromatography via a Droplet Interface
Authors: Renata Gerhardt, Detlev Belder
Abstract:
Raman spectroscopy has attracted much attention as a structurally descriptive and label-free detection method. It is particularly suited for chemical analysis given as it is non-destructive and molecules can be identified via the fingerprint region of the spectra. In this work possibilities are investigated how to integrate Raman spectroscopy as a detection method for chip-based chromatography, making use of a droplet interface. A demanding task in lab-on-a-chip applications is the specific and sensitive detection of low concentrated analytes in small volumes. Fluorescence detection is frequently utilized but restricted to fluorescent molecules. Furthermore, no structural information is provided. Another often applied technique is mass spectrometry which enables the identification of molecules based on their mass to charge ratio. Additionally, the obtained fragmentation pattern gives insight into the chemical structure. However, it is only applicable as an end-of-the-line detection because analytes are destroyed during measurements. In contrast to mass spectrometry, Raman spectroscopy can be applied on-chip and substances can be processed further downstream after detection. A major drawback of Raman spectroscopy is the inherent weakness of the Raman signal, which is due to the small cross-sections associated with the scattering process. Enhancement techniques, such as surface enhanced Raman spectroscopy (SERS), are employed to overcome the poor sensitivity even allowing detection on a single molecule level. In SERS measurements, Raman signal intensity is improved by several orders of magnitude if the analyte is in close proximity to nanostructured metal surfaces or nanoparticles. The main gain of lab-on-a-chip technology is the building block-like ability to seamlessly integrate different functionalities, such as synthesis, separation, derivatization and detection on a single device. We intend to utilize this powerful toolbox to realize Raman detection in chip-based chromatography. By interfacing on-chip separations with a droplet generator, the separated analytes are encapsulated into numerous discrete containers. These droplets can then be injected with a silver nanoparticle solution and investigated via Raman spectroscopy. Droplet microfluidics is a sub-discipline of microfluidics which instead of a continuous flow operates with the segmented flow. Segmented flow is created by merging two immiscible phases (usually an aqueous phase and oil) thus forming small discrete volumes of one phase in the carrier phase. The study surveys different chip designs to realize coupling of chip-based chromatography with droplet microfluidics. With regards to maintaining a sufficient flow rate for chromatographic separation and ensuring stable eluent flow over the column different flow rates of eluent and oil phase are tested. Furthermore, the detection of analytes in droplets with surface enhanced Raman spectroscopy is examined. The compartmentalization of separated compounds preserves the analytical resolution since the continuous phase restricts dispersion between the droplets. The droplets are ideal vessels for the insertion of silver colloids thus making use of the surface enhancement effect and improving the sensitivity of the detection. The long-term goal of this work is the first realization of coupling chip based chromatography with droplets microfluidics to employ surface enhanced Raman spectroscopy as means of detection.Keywords: chip-based separation, chip LC, droplets, Raman spectroscopy, SERS
Procedia PDF Downloads 24524705 Development of a Nano-Alumina-Zirconia Composite Catalyst as an Active Thin Film in Biodiesel Production
Authors: N. Marzban, J. K. Heydarzadeh M. Pourmohammadbagher, M. H. Hatami, A. Samia
Abstract:
A nano-alumina-zirconia composite catalyst was synthesized by a simple aqueous sol-gel method using AlCl3.6H2O and ZrCl4 as precursors. Thermal decomposition of the precursor and subsequent formation of γ-Al2O3 and t-Zr were investigated by thermal analysis. XRD analysis showed that γ-Al2O3 and t-ZrO2 phases were formed at 700 °C. FT-IR analysis also indicated that the phase transition to γ-Al2O3 occurred in corroboration with X-ray studies. TEM analysis of the calcined powder revealed that spherical particles were in the range of 8-12 nm. The nano-alumina-zirconia composite particles were mesoporous and uniformly distributed in their crystalline phase. In order to measure the catalytic activity, esterification reaction was carried out. Biodiesel, as a renewable fuel, was formed in a continuous packed column reactor. Free fatty acid (FFA) was esterified with ethanol in a heterogeneous catalytic reactor. It was found that the synthesized γ-Al2O3/ZrO2 composite had the potential to be used as a heterogeneous base catalyst for biodiesel production processes.Keywords: nano alumina-zirconia, composite catalyst, thin film, biodiesel
Procedia PDF Downloads 23324704 Development and Power Characterization of an IoT Network for Agricultural Imaging Applications
Authors: Jacob Wahl, Jane Zhang
Abstract:
This paper describes the development and characterization of a prototype IoT network for use with agricultural imaging and monitoring applications. The sensor and gateway nodes are designed using the ESP32 SoC with integrated Bluetooth Low Energy 4.2 and Wi-Fi. A development board, the Arducam IoTai ESP32, is used for prototyping, testing, and power measurements. Google’s Firebase is used as the cloud storage site for image data collected by the sensor. The sensor node captures images using the OV2640 2MP camera module and transmits the image data to the gateway via Bluetooth Low Energy. The gateway then uploads the collected images to Firebase via a known nearby Wi-Fi network connection. This image data can then be processed and analyzed by computer vision and machine learning pipelines to assess crop growth or other needs. The sensor node achieves a wireless transmission data throughput of 220kbps while consuming 150mA of current; the sensor sleeps at 162µA. The sensor node device lifetime is estimated to be 682 days on a 6600mAh LiPo battery while acquiring five images per day based on the development board power measurements. This network can be utilized by any application that requires high data rates, low power consumption, short-range communication, and large amounts of data to be transmitted at low-frequency intervals.Keywords: Bluetooth low energy, ESP32, firebase cloud, IoT, smart farming
Procedia PDF Downloads 13924703 Treating Voxels as Words: Word-to-Vector Methods for fMRI Meta-Analyses
Authors: Matthew Baucum
Abstract:
With the increasing popularity of fMRI as an experimental method, psychology and neuroscience can greatly benefit from advanced techniques for summarizing and synthesizing large amounts of data from brain imaging studies. One promising avenue is automated meta-analyses, in which natural language processing methods are used to identify the brain regions consistently associated with certain semantic concepts (e.g. “social”, “reward’) across large corpora of studies. This study builds on this approach by demonstrating how, in fMRI meta-analyses, individual voxels can be treated as vectors in a semantic space and evaluated for their “proximity” to terms of interest. In this technique, a low-dimensional semantic space is built from brain imaging study texts, allowing words in each text to be represented as vectors (where words that frequently appear together are near each other in the semantic space). Consequently, each voxel in a brain mask can be represented as a normalized vector sum of all of the words in the studies that showed activation in that voxel. The entire brain mask can then be visualized in terms of each voxel’s proximity to a given term of interest (e.g., “vision”, “decision making”) or collection of terms (e.g., “theory of mind”, “social”, “agent”), as measured by the cosine similarity between the voxel’s vector and the term vector (or the average of multiple term vectors). Analysis can also proceed in the opposite direction, allowing word cloud visualizations of the nearest semantic neighbors for a given brain region. This approach allows for continuous, fine-grained metrics of voxel-term associations, and relies on state-of-the-art “open vocabulary” methods that go beyond mere word-counts. An analysis of over 11,000 neuroimaging studies from an existing meta-analytic fMRI database demonstrates that this technique can be used to recover known neural bases for multiple psychological functions, suggesting this method’s utility for efficient, high-level meta-analyses of localized brain function. While automated text analytic methods are no replacement for deliberate, manual meta-analyses, they seem to show promise for the efficient aggregation of large bodies of scientific knowledge, at least on a relatively general level.Keywords: FMRI, machine learning, meta-analysis, text analysis
Procedia PDF Downloads 44924702 Determining the Most Efficient Test Available in Software Testing
Authors: Qasim Zafar, Matthew Anderson, Esteban Garcia, Steven Drager
Abstract:
Software failures can present an enormous detriment to people's lives and cost millions of dollars to repair when they are unexpectedly encountered in the wild. Despite a significant portion of the software development lifecycle and resources are dedicated to testing, software failures are a relatively frequent occurrence. Nevertheless, the evaluation of testing effectiveness remains at the forefront of ensuring high-quality software and software metrics play a critical role in providing valuable insights into quantifiable objectives to assess the level of assurance and confidence in the system. As the selection of appropriate metrics can be an arduous process, the goal of this paper is to shed light on the significance of software metrics by examining a range of testing techniques and metrics as well as identifying key areas for improvement. Additionally, through this investigation, readers will gain a deeper understanding of how metrics can help to drive informed decision-making on delivering high-quality software and facilitate continuous improvement in testing practices.Keywords: software testing, software metrics, testing effectiveness, black box testing, random testing, adaptive random testing, combinatorial testing, fuzz testing, equivalence partition, boundary value analysis, white box testing
Procedia PDF Downloads 8924701 Hidden Hot Spots: Identifying and Understanding the Spatial Distribution of Crime
Authors: Lauren C. Porter, Andrew Curtis, Eric Jefferis, Susanne Mitchell
Abstract:
A wealth of research has been generated examining the variation in crime across neighborhoods. However, there is also a striking degree of crime concentration within neighborhoods. A number of studies show that a small percentage of street segments, intersections, or addresses account for a large portion of crime. Not surprisingly, a focus on these crime hot spots can be an effective strategy for reducing community level crime and related ills, such as health problems. However, research is also limited in an important respect. Studies tend to use official data to identify hot spots, such as 911 calls or calls for service. While the use of call data may be more representative of the actual level and distribution of crime than some other official measures (e.g. arrest data), call data still suffer from the 'dark figure of crime.' That is, there is most certainly a degree of error between crimes that occur versus crimes that are reported to the police. In this study, we present an alternative method of identifying crime hot spots, that does not rely on official data. In doing so, we highlight the potential utility of neighborhood-insiders to identify and understand crime dynamics within geographic spaces. Specifically, we use spatial video and geo-narratives to record the crime insights of 36 police, ex-offenders, and residents of a high crime neighborhood in northeast Ohio. Spatial mentions of crime are mapped to identify participant-identified hot spots, and these are juxtaposed with calls for service (CFS) data. While there are bound to be differences between these two sources of data, we find that one location, in particular, a corner store, emerges as a hot spot for all three groups of participants. Yet it does not emerge when we examine CFS data. A closer examination of the space around this corner store and a qualitative analysis of narrative data reveal important clues as to why this store may indeed be a hot spot, but not generate disproportionate calls to the police. In short, our results suggest that researchers who rely solely on official data to study crime hot spots may risk missing some of the most dangerous places.Keywords: crime, narrative, video, neighborhood
Procedia PDF Downloads 23824700 Advancing Urban Sustainability through Data-Driven Machine Learning Solutions
Authors: Nasim Eslamirad, Mahdi Rasoulinezhad, Francesco De Luca, Sadok Ben Yahia, Kimmo Sakari Lylykangas, Francesco Pilla
Abstract:
With the ongoing urbanization, cities face increasing environmental challenges impacting human well-being. To tackle these issues, data-driven approaches in urban analysis have gained prominence, leveraging urban data to promote sustainability. Integrating Machine Learning techniques enables researchers to analyze and predict complex environmental phenomena like Urban Heat Island occurrences in urban areas. This paper demonstrates the implementation of data-driven approach and interpretable Machine Learning algorithms with interpretability techniques to conduct comprehensive data analyses for sustainable urban design. The developed framework and algorithms are demonstrated for Tallinn, Estonia to develop sustainable urban strategies to mitigate urban heat waves. Geospatial data, preprocessed and labeled with UHI levels, are used to train various ML models, with Logistic Regression emerging as the best-performing model based on evaluation metrics to derive a mathematical equation representing the area with UHI or without UHI effects, providing insights into UHI occurrences based on buildings and urban features. The derived formula highlights the importance of building volume, height, area, and shape length to create an urban environment with UHI impact. The data-driven approach and derived equation inform mitigation strategies and sustainable urban development in Tallinn and offer valuable guidance for other locations with varying climates.Keywords: data-driven approach, machine learning transparent models, interpretable machine learning models, urban heat island effect
Procedia PDF Downloads 3924699 Optimization of Production Scheduling through the Lean and Simulation Integration in Automotive Company
Authors: Guilherme Gorgulho, Carlos Roberto Camello Lima
Abstract:
Due to the competitive market in which companies are currently engaged, the constant changes require companies to react quickly regarding the variability of demand and process. The changes are caused by customers, or by demand fluctuations or variations of products, or the need to serve customers within agreed delivery taking into account the continuous search for quality and competitive prices in products. These changes end up influencing directly or indirectly the activities of the Planning and Production Control (PPC), which does business in strategic, tactical and operational levels of production systems. One area of concern for organizations is in the short term (operational level), because this planning stage any error or divergence will cause waste and impact on the delivery of products on time to customers. Thus, this study aims to optimize the efficiency of production scheduling, using different sequencing strategies in an automotive company. Seeking to aim the proposed objective, we used the computer simulation in conjunction with lean manufacturing to build and validate the current model, and subsequently the creation of future scenarios.Keywords: computational simulation, lean manufacturing, production scheduling, sequencing strategies
Procedia PDF Downloads 27124698 Cross-Comparison between Land Surface Temperature from Polar and Geostationary Satellite over Heterogenous Landscape: A Case Study in Hong Kong
Authors: Ibrahim A. Adeniran, Rui F. Zhu, Man S. Wong
Abstract:
Owing to the insufficiency in the spatial representativeness and continuity of in situ temperature measurements from weather stations (WS), the use of temperature measurement from WS for large-range diurnal analysis in heterogenous landscapes has been limited. This has made the accurate estimation of land surface temperature (LST) from remotely sensed data more crucial. Moreover, the study of dynamic interaction between the atmosphere and the physical surface of the Earth could be enhanced at both annual and diurnal scales by using optimal LST data derived from satellite sensors. The tradeoff between the spatial and temporal resolution of LSTs from satellite’s thermal infrared sensors (TIRS) has, however, been a major challenge, especially when high spatiotemporal LST data are recommended. It is well-known from existing literature that polar satellites have the advantage of high spatial resolution, while geostationary satellites have a high temporal resolution. Hence, this study is aimed at designing a framework for the cross-comparison of LST data from polar and geostationary satellites in a heterogeneous landscape. This could help to understand the relationship between the LST estimates from the two satellites and, consequently, their integration in diurnal LST analysis. Landsat-8 satellite data will be used as the representative of the polar satellite due to the availability of its long-term series, while the Himawari-8 satellite will be used as the data source for the geostationary satellite because of its improved TIRS. For the study area, Hong Kong Special Administrative Region (HK SAR) will be selected; this is due to the heterogeneity in the landscape of the region. LST data will be retrieved from both satellites using the Split window algorithm (SWA), and the resulting data will be validated by comparing satellite-derived LST data with temperature data from automatic WS in HK SAR. The LST data from the satellite data will then be separated based on the land use classification in HK SAR using the Global Land Cover by National Mapping Organization version3 (GLCNMO 2013) data. The relationship between LST data from Landsat-8 and Himawari-8 will then be investigated based on the land-use class and over different seasons of the year in order to account for seasonal variation in their relationship. The resulting relationship will be spatially and statistically analyzed and graphically visualized for detailed interpretation. Findings from this study will reveal the relationship between the two satellite data based on the land use classification within the study area and the seasons of the year. While the information provided by this study will help in the optimal combination of LST data from Polar (Landsat-8) and geostationary (Himawari-8) satellites, it will also serve as a roadmap in the annual and diurnal urban heat (UHI) analysis in Hong Kong SAR.Keywords: automatic weather station, Himawari-8, Landsat-8, land surface temperature, land use classification, split window algorithm, urban heat island
Procedia PDF Downloads 7324697 Enhancing Organizational Performance through Adaptive Learning: A Case Study of ASML
Authors: Ramin Shadani
Abstract:
This study introduces adaptive performance as a key organizational performance dimension and explores the relationship between the dimensions of a learning organization and adaptive performance. A survey was therefore conducted using the dimensions of the Learning Organization Questionnaire (DLOQ), followed by factor analysis and structural equation modeling in order to investigate the dynamics between learning organization practices and adaptive performance. Results confirm that adaptive performance is indeed one important dimension of organizational performance. The study also shows that perceived knowledge and adaptive performance mediate the positive relationship between the practices of a learning organization with perceived financial performance. We extend existing DLOQ research by demonstrating that adaptive performance, as a nonfinancial organizational learning outcome, has a significant impact on financial performance. Our study also provides additional validation of the measures of DLOQ's performance. Indeed, organizations need to take a glance at how the activities of learning and development can provide better overall improvement in performance, especially in enhancing adaptive capability. The study has provided requisite empirical support that activities of learning and development within organizations allow much-improved intangible performance outcomes, especially through adaptive performance.Keywords: adaptive performance, continuous learning, financial performance, leadership style, organizational learning, organizational performance
Procedia PDF Downloads 30