Search results for: optimal digital signal processing
1009 Distraction from Pain: An fMRI Study on the Role of Age-Related Changes in Executive Functions
Authors: Katharina M. Rischer, Angelika Dierolf, Ana M. Gonzalez-Roldan, Pedro Montoya, Fernand Anton, Marian van der Meulen
Abstract:
Even though age has been associated with increased and prolonged episodes of pain, little is known about potential age-related changes in the ˈtop-downˈ modulation of pain, such as cognitive distraction from pain. The analgesic effects of distraction result from competition for attentional resources in the prefrontal cortex (PFC), a region that is also involved in executive functions. Given that the PFC shows pronounced age-related atrophy, distraction may be less effective in reducing pain in older compared to younger adults. The aim of this study was to investigate the influence of aging on task-related analgesia and the underpinning neural mechanisms, with a focus on the role of executive functions in distraction from pain. In a first session, 64 participants (32 young adults: 26.69 ± 4.14 years; 32 older adults: 68.28 ± 7.00 years) completed a battery of neuropsychological tests. In a second session, participants underwent a pain distraction paradigm, while fMRI images were acquired. In this paradigm, participants completed a low (0-back) and a high (2-back) load condition of a working memory task while receiving either warm or painful thermal stimuli to their lower arm. To control for age-related differences in sensitivity to pain and perceived task difficulty, stimulus intensity, and task speed were individually calibrated. Results indicate that both age groups showed significantly reduced activity in a network of regions involved in pain processing when completing the high load distraction task; however, young adults showed a larger neural distraction effect in different parts of the insula and the thalamus. Moreover, better executive functions, in particular inhibitory control abilities, were associated with a larger behavioral and neural distraction effect. These findings clearly demonstrate that top-down control of pain is affected in older age, and could explain the higher vulnerability for older adults to develop chronic pain. Moreover, our findings suggest that the assessment of executive functions may be a useful tool for predicting the efficacy of cognitive pain modulation strategies in older adults.Keywords: executive functions, cognitive pain modulation, fMRI, PFC
Procedia PDF Downloads 1451008 Development of Solid Electrolytes Based on Networked Cellulose
Authors: Boor Singh Lalia, Yarjan Abdul Samad, Raed Hashaikeh
Abstract:
Three different kinds of solid polymer electrolytes were prepared using polyethylene oxide (PEO) as a base polymer, networked cellulose (NC) as a physical support and LiClO4 as a conductive salt for the electrolytes. Networked cellulose, a modified form of cellulose, is a biodegradable and environmentally friendly additive which provides a strong fibrous networked support for structural stability of the electrolytes. Although the PEO/NC/LiClO4 electrolyte retains its structural integrity and mechanical properties at 100oC as compared to pristine PEO-based polymer electrolytes, it suffers from poor ionic conductivity. To improve the room temperature conductivity of the electrolyte, PEO is replaced by the polyethylene glycol (PEG) which is a liquid phase that provides high mobility for Li+ ions transport in the electrolyte. PEG/NC/LiClO4 shows improvement in ionic conductivity compared to PEO/NC/LiClO4 at room temperature, but it is brittle and tends to form cracks during processing. An advanced solid polymer electrolyte with optimum ionic conductivity and mechanical properties is developed by using a ternary system: TEGDME/PEO/NC+LiClO4. At room temperature, this electrolyte exhibits an ionic conductivity to the order of 10-5 S/cm, which is very high compared to that of the PEO/LiClO4 electrolyte. Pristine PEO electrolytes start melting at 65 °C and completely lose its mechanical strength. Dynamic mechanical analysis of TEGDME: PEO: NC (70:20:10 wt%) showed an improvement of storage modulus as compared to the pristine PEO in the 60–120 °C temperature range. Also, with an addition of NC, the electrolyte retains its mechanical integrity at 100 oC which is beneficial for Li-ion battery operation at high temperatures. Differential scanning calorimetry (DSC) and thermal gravimetry analysis (TGA) studies revealed that the ternary polymer electrolyte is thermally stable in the lithium ion battery operational temperature range. As-prepared polymer electrolyte was used to assemble LiFePO4/ TEGDME/PEO/NC+LiClO4/Li half cells and their electrochemical performance was studied via cyclic voltammetry and charge-discharge cycling.Keywords: solid polymer electrolyte, ionic conductivity, mechanical properties, lithium ion batteries, cyclic voltammetry
Procedia PDF Downloads 4301007 The Treatment of Nitrate Polluted Groundwater Using Bio-electrochemical Systems Inoculated with Local Groundwater Sediments
Authors: Danish Laidin, Peter Gostomski, Aaron Marshall, Carlo Carere
Abstract:
Groundwater contamination of nitrate (NO3-) is becoming more prevalent in regions of intensive and extensive agricultural activities. Household nitrate removal involves using ion exchange membranes and reverse osmosis (RO) systems, whereas industrial nitrate removal may use organic carbon substrates (e.g. methanol) for heterotrophic microbial denitrification. However, these approaches both require high capital investment and operating costs. In this study, denitrification was demonstrated using bio-electrochemical systems (BESs) inoculated from sediments and microbial enrichment cultures. The BES reactors were operated continuously as microbial electrolytic cells (MECs) with a poised potential of -0.7V and -1.1V vs Ag/AgCl. Three parallel MECs were inoculated using hydrogen-driven denitrifying enrichments, stream sediments, and biofilm harvested from a denitrifying biotrickling filter, respectively. These reactors were continuously operated for over a year as various operating conditions were investigated to determine the optimal conditions for electroactive denitrification. The mass loading rate of nitrate was varied between 10 – 70 mg NO3-/d, and the maximum observed nitrate removal rate was 22 mg NO3- /(cm2∙d) with a current of 2.1 mA. For volumetric load experiments, the dilution rate of 1 mM NO3- feed was varied between 0.01 – 0.1 hr-1 to achieve a nitrate loading rate similar to the mass loading rate experiments. Under these conditions, the maximum rate of denitrification observed was 15.8 mg NO3- /(cm2∙d) with a current of 1.7mA. Hydrogen (H2) was supplied intermittently to investigate the hydrogenotrophic potential of the denitrifying biofilm electrodes. H2 supplementation at 0.1 mL/min resulted in an increase of nitrate removal from 0.3 mg NO3- /(cm2∙d) to 3.4 mg NO3- /(cm2∙d) in the hydrogenotrophically subcultured reactor but had no impact on the reactors which exhibited direct electron transfer properties. Results from this study depict the denitrification performance of the immobilized biofilm electrodes, either by direct electron transfer or hydrogen-driven denitrification, and the contribution of the planktonic cells present in the growth medium. Other results will include the microbial community analysis via 16s rDNA amplicon sequencing, varying the effect of poising cathodic potential from 0.7V to 1.3V vs Ag/AgCl, investigating the potential of using in-situ electrochemically produced hydrogen for autotrophic denitrification and adjusting the conductivity of the feed solution to mimic groundwater conditions. These findings highlight the overall performance of sediment inoculated MECs in removing nitrate and will be used for the future development of sustainable solutions for the treatment of nitrate polluted groundwater.Keywords: bio-electrochemical systems, groundwater, electroactive denitrification, microbial electrolytic cell
Procedia PDF Downloads 661006 Modeling of Void Formation in 3D Woven Fabric During Resin Transfer Moulding
Authors: Debabrata Adhikari, Mikhail Matveev, Louise Brown, Jan Kočí, Andy Long
Abstract:
Resin transfer molding (RTM) is increasingly used for manufacturing high-quality composite structures due to its additional advantages over prepregs of low-cost out-of-autoclave processing. However, to retain the advantages, it is critical to reduce the void content during the injection. Reinforcements commonly used in RTM, such as woven fabrics, have dual-scale porosity with mesoscale pores between the yarns and the micro-scale pores within the yarns. Due to the fabric geometry and the nature of the dual-scale flow, the flow front during injection creates a complicated fingering formation which leads to void formation. Analytical modeling of void formation for woven fabrics has been widely studied elsewhere. However, there is scope for improvement to the reduction in void formation in 3D fabrics wherein the in-plane yarn layers are confined by additional through-thickness binder yarns. In the present study, the structural morphology of the tortuous pore spaces in the 3D fabric has been studied and implemented using open-source software TexGen. An analytical model for the void and the fingering formation has been implemented based on an idealized unit cell model of the 3D fabric. Since the pore spaces between the yarns are free domains, the region is treated as flow-through connected channels, whereas intra-yarn flow has been modeled using Darcy’s law with an additional term to account for capillary pressure. Later the void fraction has been characterised using the criterion of void formation by comparing the fill time for inter and intra yarn flow. Moreover, the dual-scale two-phase flow of resin with air has been simulated in the commercial CFD solver OpenFOAM/ANSYS to predict the probable location of voids and validate the analytical model. The use of an idealised unit cell model will give the insight to optimise the mesoscale geometry of the reinforcement and injection parameters to minimise the void content during the LCM process.Keywords: 3D fiber, void formation, RTM, process modelling
Procedia PDF Downloads 961005 Spectroscopy and Electron Microscopy for the Characterization of CdSxSe1-x Quantum Dots in a Glass Matrix
Authors: C. Fornacelli, P. Colomban, E. Mugnaioli, I. Memmi Turbanti
Abstract:
When semiconductor particles are reduced in scale to nanometer dimension, their optical and electro-optical properties strongly differ from those of bulk crystals of the same composition. Since sampling is often not allowed concerning cultural heritage artefacts, the potentialities of two non-invasive techniques, such as Raman and Fiber Optic Reflectance Spectroscopy (FORS), have been investigated and the results of the analysis on some original glasses of different colours (from yellow to orange and deep red) and periods (from the second decade of the 20th century to present days) are reported in the present study. In order to evaluate the potentialities of the application of non-invasive techniques to the investigation of the structure and distribution of nanoparticles dispersed in a glass matrix, Scanning Electron Microscopy (SEM) and energy-disperse spectroscopy (EDS) mapping, together with Transmission Electron Microscopy (TEM) and Electron Diffraction Tomography (EDT) have also been used. Raman spectroscopy allows a fast and non-destructive measure of the quantum dots composition and size, thanks to the evaluation of the frequencies and the broadening/asymmetry of the LO phonons bands, respectively, though the important role of the compressive strain arising from the glass matrix and the possible diffusion of zinc from the matrix to the nanocrystals should be taken into account when considering the optical-phonons frequency values. The incorporation of Zn has been assumed by an upward shifting of the LO band related to the most abundant anion (S or Se), while the role of the surface phonons as well as the confinement-induced scattering by phonons with a non-zero wavevectors on the Raman peaks broadening has been verified. The optical band gap varies from 2.42 eV (pure CdS) to 1.70 eV (CdSe). For the compositional range between 0.5≤x≤0.2, the presence of two absorption edges has been related to the contribution of both pure CdS and the CdSxSe1-x solid solution; this particular feature is probably due to the presence of unaltered cubic zinc blende structures of CdS that is not taking part to the formation of the solid solution occurring only between hexagonal CdS and CdSe. Moreover, the band edge tailing originating from the disorder due to the formation of weak bonds and characterized by the Urbach edge energy has been studied and, together with the FWHM of the Raman signal, has been assumed as a good parameter to evaluate the degree of topological disorder. SEM-EDS mapping showed a peculiar distribution of the major constituents of the glass matrix (fluxes and stabilizers), especially concerning those samples where a layered structure has been assumed thanks to the spectroscopic study. Finally, TEM-EDS and EDT were used to get high-resolution information about nanocrystals (NCs) and heterogeneous glass layers. The presence of ZnO NCs (< 4 nm) dispersed in the matrix has been verified for most of the samples, while, for those samples where a disorder due to a more complex distribution of the size and/or composition of the NCs has been assumed, the TEM clearly verified most of the assumption made by the spectroscopic techniques.Keywords: CdSxSe1-x, EDT, glass, spectroscopy, TEM-EDS
Procedia PDF Downloads 3001004 Physical and Chemical Alternative Methods of Fresh Produce Disinfection
Authors: Tuji Jemal Ahmed
Abstract:
Fresh produce is an essential component of a healthy diet. However, it can also be a potential source of pathogenic microorganisms that can cause foodborne illnesses. Traditional disinfection methods, such as washing with water and chlorine, have limitations and may not effectively remove or inactivate all microorganisms. This has led to the development of alternative/new methods of fresh produce disinfection, including physical and chemical methods. In this paper, we explore the physical and chemical new methods of fresh produce disinfection, their advantages and disadvantages, and their suitability for different types of produce. Physical methods of disinfection, such as ultraviolet (UV) radiation and high-pressure processing (HPP), are crucial in ensuring the microbiological safety of fresh produce. UV radiation uses short-wavelength UV-C light to damage the DNA and RNA of microorganisms, and HPP applies high levels of pressure to fresh produce to reduce the microbial load. These physical methods are highly effective in killing a wide range of microorganisms, including bacteria, viruses, and fungi. However, they may not penetrate deep enough into the product to kill all microorganisms and can alter the sensory characteristics of the product. Chemical methods of disinfection, such as acidic electrolyzed water (AEW), ozone, and peroxyacetic acid (PAA), are also important in ensuring the microbiological safety of fresh produce. AEW uses a low concentration of hypochlorous acid and a high concentration of hydrogen ions to inactivate microorganisms, ozone uses ozone gas to damage the cell membranes and DNA of microorganisms, and PAA uses a combination of hydrogen peroxide and acetic acid to inactivate microorganisms. These chemical methods are highly effective in killing a wide range of microorganisms, but they may cause discoloration or changes in the texture and flavor of some products and may require specialized equipment and trained personnel to produce and apply. In conclusion, the selection of the most suitable method of fresh produce disinfection should take into consideration the type of product, the level of microbial contamination, the effectiveness of the method in reducing the microbial load, and any potential negative impacts on the sensory characteristics, nutritional composition, and safety of the produce.Keywords: fresh produce, pathogenic microorganisms, foodborne illnesses, disinfection methods
Procedia PDF Downloads 741003 Downward Vertical Evacuation for Disabilities People from Tsunami Using Escape Bunker Technology
Authors: Febrian Tegar Wicaksana, Niqmatul Kurniati, Surya Nandika
Abstract:
Indonesia is one of the countries that have great number of disaster occurrence and threat because it is located in not only between three tectonic plates such as Eurasia plates, Indo-Australia plates and Pacific plates, but also in the Ring of Fire path, like earthquake, Tsunami, volcanic eruption and many more. Recently, research shows that there are potential areas that will be devastated by Tsunami in southern coast of Java. Tsunami is a series of waves in a body of water caused by the displacement of a large volume of water, generally in an ocean. When the waves enter shallow water, they may rise to several feet or, in rare cases, tens of feet, striking the coast with devastating force. The parameter for reference such as magnitude, the depth of epicentre, distance between epicentres with land, the depth of every points, when reached the shore and the growth of waves. Interaction between parameters will bring the big variance of Tsunami wave. Based on that, we can formulate preparation that needed for disaster mitigation strategies. The mitigation strategies will take the important role in an effort to reduce the number of victims and damage in the area. It will reduce the number of victim and casualties. Reducing is directed to the most difficult mobilization casualties in the tsunami disaster area like old people, sick people and disabilities people. Until now, the method that used for rescuing people from Tsunami is basic horizontal evacuation. This evacuation system is not optimal because it needs so long time and it cannot be used by people with disabilities. The writers propose to create a vertical evacuation model with an escape bunker system. This bunker system is chosen because the downward vertical evacuation is considered more efficient and faster. Especially in coastal areas without any highlands surround it. The downward evacuation system is better than upward evacuation because it can avoid the risk of erosion at the ground around the structure which can affect the building. The structure of the bunker and the evacuation process while, and even after, disaster are the main priority to be considered. The power of bunker has quake’s resistance, the durability from water stream, variety of interaction to the ground, and waterproof design. When the situation is back to normal, victim and casualties can go into the safer place. The bunker will be located near the hospital and public places, and will have wide entrance supported by large slide in it so it will ease the disabilities people. The technology of the escape bunker system is expected to reduce the number of victims who have low mobility in the Tsunami.Keywords: escape bunker, tsunami, vertical evacuation, mitigation, disaster management
Procedia PDF Downloads 4951002 Image-Based UAV Vertical Distance and Velocity Estimation Algorithm during the Vertical Landing Phase Using Low-Resolution Images
Authors: Seyed-Yaser Nabavi-Chashmi, Davood Asadi, Karim Ahmadi, Eren Demir
Abstract:
The landing phase of a UAV is very critical as there are many uncertainties in this phase, which can easily entail a hard landing or even a crash. In this paper, the estimation of relative distance and velocity to the ground, as one of the most important processes during the landing phase, is studied. Using accurate measurement sensors as an alternative approach can be very expensive for sensors like LIDAR, or with a limited operational range, for sensors like ultrasonic sensors. Additionally, absolute positioning systems like GPS or IMU cannot provide distance to the ground independently. The focus of this paper is to determine whether we can measure the relative distance and velocity of UAV and ground in the landing phase using just low-resolution images taken by a monocular camera. The Lucas-Konda feature detection technique is employed to extract the most suitable feature in a series of images taken during the UAV landing. Two different approaches based on Extended Kalman Filters (EKF) have been proposed, and their performance in estimation of the relative distance and velocity are compared. The first approach uses the kinematics of the UAV as the process and the calculated optical flow as the measurement; On the other hand, the second approach uses the feature’s projection on the camera plane (pixel position) as the measurement while employing both the kinematics of the UAV and the dynamics of variation of projected point as the process to estimate both relative distance and relative velocity. To verify the results, a sequence of low-quality images taken by a camera that is moving on a specifically developed testbed has been used to compare the performance of the proposed algorithm. The case studies show that the quality of images results in considerable noise, which reduces the performance of the first approach. On the other hand, using the projected feature position is much less sensitive to the noise and estimates the distance and velocity with relatively high accuracy. This approach also can be used to predict the future projected feature position, which can drastically decrease the computational workload, as an important criterion for real-time applications.Keywords: altitude estimation, drone, image processing, trajectory planning
Procedia PDF Downloads 1131001 Improving Predictions of Coastal Benthic Invertebrate Occurrence and Density Using a Multi-Scalar Approach
Authors: Stephanie Watson, Fabrice Stephenson, Conrad Pilditch, Carolyn Lundquist
Abstract:
Spatial data detailing both the distribution and density of functionally important marine species are needed to inform management decisions. Species distribution models (SDMs) have proven helpful in this regard; however, models often focus only on species occurrences derived from spatially expansive datasets and lack the resolution and detail required to inform regional management decisions. Boosted regression trees (BRT) were used to produce high-resolution SDMs (250 m) at two spatial scales predicting probability of occurrence, abundance (count per sample unit), density (count per km2) and uncertainty for seven coastal seafloor taxa that vary in habitat usage and distribution to examine prediction differences and implications for coastal management. We investigated if small scale regionally focussed models (82,000 km2) can provide improved predictions compared to data-rich national scale models (4.2 million km2). We explored the variability in predictions across model type (occurrence vs abundance) and model scale to determine if specific taxa models or model types are more robust to geographical variability. National scale occurrence models correlated well with broad-scale environmental predictors, resulting in higher AUC (Area under the receiver operating curve) and deviance explained scores; however, they tended to overpredict in the coastal environment and lacked spatially differentiated detail for some taxa. Regional models had lower overall performance, but for some taxa, spatial predictions were more differentiated at a localised ecological scale. National density models were often spatially refined and highlighted areas of ecological relevance producing more useful outputs than regional-scale models. The utility of a two-scale approach aids the selection of the most optimal combination of models to create a spatially informative density model, as results contrasted for specific taxa between model type and scale. However, it is vital that robust predictions of occurrence and abundance are generated as inputs for the combined density model as areas that do not spatially align between models can be discarded. This study demonstrates the variability in SDM outputs created over different geographical scales and highlights implications and opportunities for managers utilising these tools for regional conservation, particularly in data-limited environments.Keywords: Benthic ecology, spatial modelling, multi-scalar modelling, marine conservation.
Procedia PDF Downloads 791000 An Absolute Femtosecond Rangefinder for Metrological Support in Coordinate Measurements
Authors: Denis A. Sokolov, Andrey V. Mazurkevich
Abstract:
In the modern world, there is an increasing demand for highly precise measurements in various fields, such as aircraft, shipbuilding, and rocket engineering. This has resulted in the development of appropriate measuring instruments that are capable of measuring the coordinates of objects within a range of up to 100 meters, with an accuracy of up to one micron. The calibration process for such optoelectronic measuring devices (trackers and total stations) involves comparing the measurement results from these devices to a reference measurement based on a linear or spatial basis. The reference used in such measurements could be a reference base or a reference range finder with the capability to measure angle increments (EDM). The base would serve as a set of reference points for this purpose. The concept of the EDM for replicating the unit of measurement has been implemented on a mobile platform, which allows for angular changes in the direction of laser radiation in two planes. To determine the distance to an object, a high-precision interferometer with its own design is employed. The laser radiation travels to the corner reflectors, which form a spatial reference with precisely known positions. When the femtosecond pulses from the reference arm and the measuring arm coincide, an interference signal is created, repeating at the frequency of the laser pulses. The distance between reference points determined by interference signals is calculated in accordance with recommendations from the International Bureau of Weights and Measures for the indirect measurement of time of light passage according to the definition of a meter. This distance is D/2 = c/2nF, approximately 2.5 meters, where c is the speed of light in a vacuum, n is the refractive index of a medium, and F is the frequency of femtosecond pulse repetition. The achieved uncertainty of type A measurement of the distance to reflectors 64 m (N•D/2, where N is an integer) away and spaced apart relative to each other at a distance of 1 m does not exceed 5 microns. The angular uncertainty is calculated theoretically since standard high-precision ring encoders will be used and are not a focus of research in this study. The Type B uncertainty components are not taken into account either, as the components that contribute most do not depend on the selected coordinate measuring method. This technology is being explored in the context of laboratory applications under controlled environmental conditions, where it is possible to achieve an advantage in terms of accuracy. In general, the EDM tests showed high accuracy, and theoretical calculations and experimental studies on an EDM prototype have shown that the uncertainty type A of distance measurements to reflectors can be less than 1 micrometer. The results of this research will be utilized to develop a highly accurate mobile absolute range finder designed for the calibration of high-precision laser trackers and laser rangefinders, as well as other equipment, using a 64 meter laboratory comparator as a reference.Keywords: femtosecond laser, pulse correlation, interferometer, laser absolute range finder, coordinate measurement
Procedia PDF Downloads 60999 Enhancing VR Exposure Therapy for the Treatment of Phobias with the Use of Photorealistic VR Environments and Stimuli, and the Use of Tactile Feedback Suits and Responsive Systems
Authors: Vardan Melkonyan, Arman Azizyan, Astghik Boyajyan
Abstract:
Virtual reality (VR) exposure therapy is a form of cognitive-behavioral therapy that uses immersive virtual environments to expose individuals to the feared stimuli or situations that trigger their phobia. VR exposure therapy has become an increasingly popular treatment for phobias, including fear of heights, public speaking, and flying, due to its ability to provide a controlled and safe environment for individuals to confront their fears while also allowing therapists to tailor the virtual exposure to the specific needs and goals of each individual. It is also a cost-effective and accessible treatment option, as it can be delivered remotely and does not require the use of drugs. Overall, VR exposure therapy has the potential to be a valuable tool for therapists in the treatment of phobias. But current methods may be improved by incorporating advanced technology such as photorealistic VR environments, tactile feedback suits, and responsive systems. The aim of this study was to identify the most effective approach for enhancing VR exposure therapy for the treatment of phobias. Photorealistic VR environments and stimuli can greatly enhance the effectiveness of VR exposure therapy for the treatment of phobias. By creating immersive, realistic virtual environments that closely mimic the real-life situations that trigger phobia responses, patients are able to more fully engage in the therapeutic process and confront their fears in a controlled and safe manner. This can help to reduce the severity of phobia symptoms and increase treatment outcomes. The use of tactile feedback suits and responsive systems can further enhance the VR exposure therapy experience by adding a physical element to the virtual environment. These suits, which can mimic the sensations of touch, pressure, and movement, allow patients to fully immerse themselves in the virtual world and feel as if they are physically present in the situation. This can help to increase the realism of the virtual environment and make it more effective in reducing phobia symptoms. Additionally, responsive systems can be used to trigger specific events or responses within the virtual environment based on the patient's actions, providing a more interactive and personalized treatment experience. A comprehensive literature review was conducted, including studies on VR exposure therapy for phobias and the use of advanced technology to enhance the therapy. Results indicate that incorporating these enhancements may significantly increase the effectiveness of VR exposure therapy for phobias. Further research is needed to fully understand the potential of these enhancements and to determine the optimal combination and implementation.Keywords: virtual reality, mental health, phobias, fears, treatment, photorealistic, immersive, phobia
Procedia PDF Downloads 89998 Exploring the Impact of Input Sequence Lengths on Long Short-Term Memory-Based Streamflow Prediction in Flashy Catchments
Authors: Farzad Hosseini Hossein Abadi, Cristina Prieto Sierra, Cesar Álvarez Díaz
Abstract:
Predicting streamflow accurately in flashy catchments prone to floods is a major research and operational challenge in hydrological modeling. Recent advancements in deep learning, particularly Long Short-Term Memory (LSTM) networks, have shown to be promising in achieving accurate hydrological predictions at daily and hourly time scales. In this work, a multi-timescale LSTM (MTS-LSTM) network was applied to the context of regional hydrological predictions at an hourly time scale in flashy catchments. The case study includes 40 catchments allocated in the Basque Country, north of Spain. We explore the impact of hyperparameters on the performance of streamflow predictions given by regional deep learning models through systematic hyperparameter tuning - where optimal regional values for different catchments are identified. The results show that predictions are highly accurate, with Nash-Sutcliffe (NSE) and Kling-Gupta (KGE) metrics values as high as 0.98 and 0.97, respectively. A principal component analysis reveals that a hyperparameter related to the length of the input sequence contributes most significantly to the prediction performance. The findings suggest that input sequence lengths have a crucial impact on the model prediction performance. Moreover, employing catchment-scale analysis reveals distinct sequence lengths for individual basins, highlighting the necessity of customizing this hyperparameter based on each catchment’s characteristics. This aligns with well known “uniqueness of the place” paradigm. In prior research, tuning the length of the input sequence of LSTMs has received limited focus in the field of streamflow prediction. Initially it was set to 365 days to capture a full annual water cycle. Later, performing limited systematic hyper-tuning using grid search, revealed a modification to 270 days. However, despite the significance of this hyperparameter in hydrological predictions, usually studies have overlooked its tuning and fixed it to 365 days. This study, employing a simultaneous systematic hyperparameter tuning approach, emphasizes the critical role of input sequence length as an influential hyperparameter in configuring LSTMs for regional streamflow prediction. Proper tuning of this hyperparameter is essential for achieving accurate hourly predictions using deep learning models.Keywords: LSTMs, streamflow, hyperparameters, hydrology
Procedia PDF Downloads 72997 A Machine Learning Approach for Assessment of Tremor: A Neurological Movement Disorder
Authors: Rajesh Ranjan, Marimuthu Palaniswami, A. A. Hashmi
Abstract:
With the changing lifestyle and environment around us, the prevalence of the critical and incurable disease has proliferated. One such condition is the neurological disorder which is rampant among the old age population and is increasing at an unstoppable rate. Most of the neurological disorder patients suffer from some movement disorder affecting the movement of their body parts. Tremor is the most common movement disorder which is prevalent in such patients that infect the upper or lower limbs or both extremities. The tremor symptoms are commonly visible in Parkinson’s disease patient, and it can also be a pure tremor (essential tremor). The patients suffering from tremor face enormous trouble in performing the daily activity, and they always need a caretaker for assistance. In the clinics, the assessment of tremor is done through a manual clinical rating task such as Unified Parkinson’s disease rating scale which is time taking and cumbersome. Neurologists have also affirmed a challenge in differentiating a Parkinsonian tremor with the pure tremor which is essential in providing an accurate diagnosis. Therefore, there is a need to develop a monitoring and assistive tool for the tremor patient that keep on checking their health condition by coordinating them with the clinicians and caretakers for early diagnosis and assistance in performing the daily activity. In our research, we focus on developing a system for automatic classification of tremor which can accurately differentiate the pure tremor from the Parkinsonian tremor using a wearable accelerometer-based device, so that adequate diagnosis can be provided to the correct patient. In this research, a study was conducted in the neuro-clinic to assess the upper wrist movement of the patient suffering from Pure (Essential) tremor and Parkinsonian tremor using a wearable accelerometer-based device. Four tasks were designed in accordance with Unified Parkinson’s disease motor rating scale which is used to assess the rest, postural, intentional and action tremor in such patient. Various features such as time-frequency domain, wavelet-based and fast-Fourier transform based cross-correlation were extracted from the tri-axial signal which was used as input feature vector space for the different supervised and unsupervised learning tools for quantification of severity of tremor. A minimum covariance maximum correlation energy comparison index was also developed which was used as the input feature for various classification tools for distinguishing the PT and ET tremor types. An automatic system for efficient classification of tremor was developed using feature extraction methods, and superior performance was achieved using K-nearest neighbors and Support Vector Machine classifiers respectively.Keywords: machine learning approach for neurological disorder assessment, automatic classification of tremor types, feature extraction method for tremor classification, neurological movement disorder, parkinsonian tremor, essential tremor
Procedia PDF Downloads 154996 Storms Dynamics in the Black Sea in the Context of the Climate Changes
Authors: Eugen Rusu
Abstract:
The objective of the work proposed is to perform an analysis of the wave conditions in the Black Sea basin. This is especially focused on the spatial and temporal occurrences and on the dynamics of the most extreme storms in the context of the climate changes. A numerical modelling system, based on the spectral phase averaged wave model SWAN, has been implemented and validated against both in situ measurements and remotely sensed data, all along the sea. Moreover, a successive correction method for the assimilation of the satellite data has been associated with the wave modelling system. This is based on the optimal interpolation of the satellite data. Previous studies show that the process of data assimilation improves considerably the reliability of the results provided by the modelling system. This especially concerns the most sensitive cases from the point of view of the accuracy of the wave predictions, as the extreme storm situations are. Following this numerical approach, it has to be highlighted that the results provided by the wave modelling system above described are in general in line with those provided by some similar wave prediction systems implemented in enclosed or semi-enclosed sea basins. Simulations of this wave modelling system with data assimilation have been performed for the 30-year period 1987-2016. Considering this database, the next step was to analyze the intensity and the dynamics of the higher storms encountered in this period. According to the data resulted from the model simulations, the western side of the sea is considerably more energetic than the rest of the basin. In this western region, regular strong storms provide usually significant wave heights greater than 8m. This may lead to maximum wave heights even greater than 15m. Such regular strong storms may occur several times in one year, usually in the wintertime, or in late autumn, and it can be noticed that their frequency becomes higher in the last decade. As regards the case of the most extreme storms, significant wave heights greater than 10m and maximum wave heights close to 20m (and even greater) may occur. Such extreme storms, which in the past were noticed only once in four or five years, are more recent to be faced almost every year in the Black Sea, and this seems to be a consequence of the climate changes. The analysis performed included also the dynamics of the monthly and annual significant wave height maxima as well as the identification of the most probable spatial and temporal occurrences of the extreme storm events. Finally, it can be concluded that the present work provides valuable information related to the characteristics of the storm conditions and on their dynamics in the Black Sea. This environment is currently subjected to high navigation traffic and intense offshore and nearshore activities and the strong storms that systematically occur may produce accidents with very serious consequences.Keywords: Black Sea, extreme storms, SWAN simulations, waves
Procedia PDF Downloads 250995 International Students into the Irish Higher Education System: Supporting the Transition
Authors: Tom Farrelly, Yvonne Kavanagh, Tony Murphy
Abstract:
The sharp rise in international students into Ireland has provided colleges with a number of opportunities but also a number of challenges, both at an institutional and individual lecturer level and of course for the incoming student. Previously, Ireland’s population, particularly its higher education student population was largely homogenous, largely drawn from its own shores and thus reflecting the ethnic, cultural and religious demographics of the day. However, over the twenty years Ireland witnessed considerable economic growth, downturn and subsequent growth all of which has resulted in an Ireland that has changed both culturally and demographically. Propelled by Ireland’s economic success up to the late 2000s, one of the defining features of this change was an unprecedented rise in the number of migrants, both academic and economic. In 2013, Ireland’s National Forum for the Enhancement for Teaching and Learning in Higher Education (hereafter the National Forum) invited proposals for inter-institutional collaborative projects aimed at different student groups’ transitioning in or out of higher education. Clearly, both as a country and a higher education sector we want incoming students to have a productive and enjoyable time in Ireland. One of the ways that will help the sector help the students make a successful transition is by developing strategies and polices that are well informed and student driven. This abstract outlines the research undertaken by the five colleges Institutes of Technology: Carlow; Cork; Tralee & Waterford and University College Cork) in Ireland that constitute the Southern cluster aimed at helping international students transition into the Irish higher education system. The aim of the southern clusters’ project was to develop a series of online learning units that can be accessed by prospective incoming international students prior to coming to Ireland and by Irish based lecturing staff. However, in order to make the units as relevant and informed as possible there was a strong research element to the project. As part of the southern cluster’s research strategy a large-scale online survey using SurveyMonkey was undertaken across the five colleges drawn from their respective international student communities. In total, there were 573 responses from students coming from over twenty different countries. The results from the survey have provided some interesting insights into the way that international students interact with and understand the Irish higher education system. The research and results will act as a model for consistent practice applicable across institutional clusters, thereby allowing institutions to minimise costs and focus on the unique aspects of transitioning international students into their institution.Keywords: digital, international, support, transitions
Procedia PDF Downloads 284994 Impact of Financial and Nutrition Support on Blood Health, Dietary Intake, and Well-Being among Female Student-Athletes
Authors: Kaila A. Vento
Abstract:
Within the field of sports science, financial situations have been reported as a key barrier in purchasing high-quality foods. A lack of proper nutrition leads to insecurities of health, impairs training, and diminishes optimal performances. Consequently, insufficient nutrient intake, disordered eating patterns, and eating disorders may arise, leading to poor health and well-being. Athletic scholarships, nutrition resources, and meal programs are available, yet are disproportionally allocated, favoring male sports, Caucasian athletes, and higher sport levels. Direct athlete finances towards nutrition at various sport levels and the role race influences aid received has yet to be examined. Additionally, a diverse female athlete population is missing in the sports science literature, specifically in nutrition. To address this gap, the current project assesses how financial and nutrition support and nutrition knowledge impacts physical health, dietary intake, and overall quality of life of a diverse sample of female athletes at the National Collegiate Athletic Association (NCAA), National Junior Collegiate Athletic Association (NJCAA), and cub sport levels. The project will identify differences in financial support in relation to race, as well. Approximately (N = 120) female athletes will participate in a single 30-minute lab visit. At this visit, body composition (i.e., height, weight, body mass index, and fat percent), blood health indicators (fasted blood glucose and lipids), and resting blood pressure are measured. In addition, three validated questionnaires pertaining to nutrition knowledge (Sports Nutrition Knowledge Questionnaire; SNKQ), dietary intake (Rapid Eating Assessment for Participants; REAP), and quality of life (World Health Organization Quality of Life Brief; WHOQL-B) are gathered. Body composition and blood health indicators will be compared with the results of self-reported sports nutrition knowledge, dietary intake, and quality of life questionnaires. It is hypothesized that 1) financial and nutrition support and nutrition knowledge will differ between the sport levels and 2) financial and nutrition support and nutrition knowledge will have a positive association with quality of dietary intake and blood health indicators, 3) financial and nutrition support will differ significantly among racial background across the various competition levels, and 4) dietary intake will influence blood health indicators and quality of life. The findings from this study could have positive implications on athletic associations' policies on equity of financial and nutrition support to improve the health and safety of all female athletes across several sport levels.Keywords: athlete, equity, finances, health, resources
Procedia PDF Downloads 106993 Effect of Different Methods to Control the Parasitic Weed Phelipanche ramosa (L. Pomel) in Tomato Crop
Authors: Disciglio G., Lops F., Carlucci A., Gatta G., Tarantino A., Frabboni L, Tarantino E.
Abstract:
The Phelipanche ramosa is considered the most damaging obligate flowering parasitic weed on a wide species of cultivated plants. The semiarid regions of the world are considered the main center of this parasitic weed, where heavy infestation are due to the ability to produce high numbers of seeds (up to 200,000), that remain viable for extended period (more than 19 years). In this paper 13 treatments of parasitic weed control, as physical, chemical, biological and agronomic methods, including the use of the resistant plants, have been carried out. In 2014 a trial was performed on processing tomato (cv Docet), grown in pots filled with soil taken from a plot heavily infested by Phelipanche ramosa, at the Department of Agriculture, Food and Environment, University of Foggia (southern Italy). Tomato seedlings were transplanted on August 8, 2014 on a clay soil (USDA) 100 kg ha-1 of N; 60 kg ha-1 of P2O5 and 20 kg ha-1 of S. Afterwards, top dressing was performed with 70 kg ha-1 of N. The randomized block design with 3 replicates was adopted. During the growing cycle of the tomato, at 70-75-81 and 88 days after transplantation the number of parasitic shoots emerged in each pot was detected. Also values of leaf chlorophyll Meter SPAD of tomato plants were measured. All data were subjected to analysis of variance (ANOVA) using the JMP software (SAS Institute Inc., Cary, NC, USA), and for comparison of means was used Tukey's test. The results show lower values of the color index SPAD in tomato plants parasitized compared to those healthy. In addition, each treatment studied did not provide complete control against Phelipanche ramosa. However the virulence of the attacks was mitigated by some treatments: radicon product, compost activated with Fusarium, mineral fertilizer nitrogen, sulfur, enzone and resistant tomato genotype. It is assumed that these effects can be improved by combining some of these treatments each other, especially for a gradual and continuing reduction of the “seed bank” of the parasite in the soil.Keywords: control methods, Phelipanche ramose, tomato crop
Procedia PDF Downloads 614992 Multi-Labeled Aromatic Medicinal Plant Image Classification Using Deep Learning
Authors: Tsega Asresa, Getahun Tigistu, Melaku Bayih
Abstract:
Computer vision is a subfield of artificial intelligence that allows computers and systems to extract meaning from digital images and video. It is used in a wide range of fields of study, including self-driving cars, video surveillance, medical diagnosis, manufacturing, law, agriculture, quality control, health care, facial recognition, and military applications. Aromatic medicinal plants are botanical raw materials used in cosmetics, medicines, health foods, essential oils, decoration, cleaning, and other natural health products for therapeutic and Aromatic culinary purposes. These plants and their products not only serve as a valuable source of income for farmers and entrepreneurs but also going to export for valuable foreign currency exchange. In Ethiopia, there is a lack of technologies for the classification and identification of Aromatic medicinal plant parts and disease type cured by aromatic medicinal plants. Farmers, industry personnel, academicians, and pharmacists find it difficult to identify plant parts and disease types cured by plants before ingredient extraction in the laboratory. Manual plant identification is a time-consuming, labor-intensive, and lengthy process. To alleviate these challenges, few studies have been conducted in the area to address these issues. One way to overcome these problems is to develop a deep learning model for efficient identification of Aromatic medicinal plant parts with their corresponding disease type. The objective of the proposed study is to identify the aromatic medicinal plant parts and their disease type classification using computer vision technology. Therefore, this research initiated a model for the classification of aromatic medicinal plant parts and their disease type by exploring computer vision technology. Morphological characteristics are still the most important tools for the identification of plants. Leaves are the most widely used parts of plants besides roots, flowers, fruits, and latex. For this study, the researcher used RGB leaf images with a size of 128x128 x3. In this study, the researchers trained five cutting-edge models: convolutional neural network, Inception V3, Residual Neural Network, Mobile Network, and Visual Geometry Group. Those models were chosen after a comprehensive review of the best-performing models. The 80/20 percentage split is used to evaluate the model, and classification metrics are used to compare models. The pre-trained Inception V3 model outperforms well, with training and validation accuracy of 99.8% and 98.7%, respectively.Keywords: aromatic medicinal plant, computer vision, convolutional neural network, deep learning, plant classification, residual neural network
Procedia PDF Downloads 192991 Adaptive Environmental Control System Strategy for Cabin Air Quality in Commercial Aircrafts
Authors: Paolo Grasso, Sai Kalyan Yelike, Federico Benzi, Mathieu Le Cam
Abstract:
The cabin air quality (CAQ) in commercial aircraft is of prime interest, especially in the context of the COVID-19 pandemic. Current Environmental Control Systems (ECS) rely on a prescribed fresh airflow per passenger to dilute contaminants. An adaptive ECS strategy is proposed, leveraging air sensing and filtration technologies to ensure a better CAQ. This paper investigates the CAQ level achieved in commercial aircraft’s cabin during various flight scenarios. The modeling and simulation analysis is performed in a Modelica-based environment describing the dynamic behavior of the system. The model includes the following three main systems: cabin, recirculation loop and air-conditioning pack. The cabin model evaluates the thermo-hygrometric conditions and the air quality in the cabin depending on the number of passengers and crew members, the outdoor conditions and the conditions of the air supplied to the cabin. The recirculation loop includes models of the recirculation fan, ordinary and novel filtration technology, mixing chamber and outflow valve. The air-conditioning pack includes models of heat exchangers and turbomachinery needed to condition the hot pressurized air bled from the engine, as well as selected contaminants originated from the outside or bled from the engine. Different ventilation control strategies are modeled and simulated. Currently, a limited understanding of contaminant concentrations in the cabin and the lack of standardized and systematic methods to collect and record data constitute a challenge in establishing a causal relationship between CAQ and passengers' comfort. As a result, contaminants are neither measured nor filtered during flight, and the current sub-optimal way to avoid their accumulation is their dilution with the fresh air flow. However, the use of a prescribed amount of fresh air comes with a cost, making the ECS the most energy-demanding non-propulsive system within an aircraft. In such a context, this study shows that an ECS based on a reduced and adaptive fresh air flow, and relying on air sensing and filtration technologies, provides promising results in terms of CAQ control. The comparative simulation results demonstrate that the proposed adaptive ECS brings substantial improvements to the CAQ in terms of both controlling the asymptotic values of the concentration of the contaminant and in mitigating hazardous scenarios, such as fume events. Original architectures allowing for adaptive control of the inlet air flow rate based on monitored CAQ will change the requirements for filtration systems and redefine the ECS operation.Keywords: cabin air quality, commercial aircraft, environmental control system, ventilation
Procedia PDF Downloads 102990 God, The Master Programmer: The Relationship Between God and Computers
Authors: Mohammad Sabbagh
Abstract:
Anyone who reads the Torah or the Quran learns that GOD created everything that is around us, seen and unseen, in six days. Within HIS plan of creation, HE placed for us a key proof of HIS existence which is essentially computers and the ability to program them. Digital computer programming began with binary instructions, which eventually evolved to what is known as high-level programming languages. Any programmer in our modern time can attest that you are essentially giving the computer commands by words and when the program is compiled, whatever is processed as output is limited to what the computer was given as an ability and furthermore as an instruction. So one can deduce that GOD created everything around us with HIS words, programming everything around in six days, just like how we can program a virtual world on the computer. GOD did mention in the Quran that one day where GOD’s throne is, is 1000 years of what we count; therefore, one might understand that GOD spoke non-stop for 6000 years of what we count, and gave everything it’s the function, attributes, class, methods and interactions. Similar to what we do in object-oriented programming. Of course, GOD has the higher example, and what HE created is much more than OOP. So when GOD said that everything is already predetermined, it is because any input, whether physical, spiritual or by thought, is outputted by any of HIS creatures, the answer has already been programmed. Any path, any thought, any idea has already been laid out with a reaction to any decision an inputter makes. Exalted is GOD!. GOD refers to HIMSELF as The Fastest Accountant in The Quran; the Arabic word that was used is close to processor or calculator. If you create a 3D simulation of a supernova explosion to understand how GOD produces certain elements and fuses protons together to spread more of HIS blessings around HIS skies; in 2022 you are going to require one of the strongest, fastest, most capable supercomputers of the world that has a theoretical speed of 50 petaFLOPS to accomplish that. In other words, the ability to perform one quadrillion (1015) floating-point operations per second. A number a human cannot even fathom. To put in more of a perspective, GOD is calculating when the computer is going through those 50 petaFLOPS calculations per second and HE is also calculating all the physics of every atom and what is smaller than that in all the actual explosion, and it’s all in truth. When GOD said HE created the world in truth, one of the meanings a person can understand is that when certain things occur around you, whether how a car crashes or how a tree grows; there is a science and a way to understand it, and whatever programming or science you deduce from whatever event you observed, it can relate to other similar events. That is why GOD might have said in The Quran that it is the people of knowledge, scholars, or scientist that fears GOD the most! One thing that is essential for us to keep up with what the computer is doing and for us to track our progress along with any errors is we incorporate logging mechanisms and backups. GOD in The Quran said that ‘WE used to copy what you used to do’. Essentially as the world is running, think of it as an interactive movie that is being played out in front of you, in a full-immersive non-virtual reality setting. GOD is recording it, from every angle to every thought, to every action. This brings the idea of how scary the Day of Judgment will be when one might realize that it’s going to be a fully immersive video when we would be getting and reading our book.Keywords: programming, the Quran, object orientation, computers and humans, GOD
Procedia PDF Downloads 107989 Studies on Optimizing the Level of Liquid Biofertilizers in Peanut and Maize and Their Economic Analysis
Authors: Chandragouda R. Patil, K. S. Jagadeesh, S. D. Kalolgi
Abstract:
Biofertilizers containing live microbial cells can mobilize one or more nutrients to plants when applied to either seed or rhizosphere. They form an integral part of nutrient management strategies for sustainable production of agricultural crops. Annually, about 22 tons of lignite-based biofertilizers are being produced and supplied to farmers at the Institute of Organic Farming, University of Agricultural Sciences, Dharwad, Karnataka state India. Although carrier based biofertilizers are common, they have shorter shelf life, poor quality, high contamination, unpredictable field performance and high cost of solid carriers. Hence, liquid formulations are being developed to increase their efficacy and broaden field applicability. An attempt was made to develop liquid formulation of strains of Rhizobium NC-92 (Groundnut), Azospirillum ACD15 both nitrogen-fixing biofertilizers and Pseudomonas striata an efficient P-solubilizing bacteria (PSB). Different concentration of amendments such as additives (glycerol and polyethylene glycol), adjuvants (carboxyl methyl cellulose), gum arabica (GA), surfactant (polysorbate) and trehalose specifically for Azospirillum were found essential. Combinations of formulations of Rhizobium and PSB for groundnut and Azospirillum and PSB for maize were evaluated under field conditions to determine the optimum level of inoculum required. Each biofertilizer strain was inoculated at the rate of 2, 4, 8 ml per kg of seeds and the efficacy of each formulation both individually and in combinations was evaluated against the lignite-based formulation at the rate of 20 g each per kg seeds and a un-inoculated set was included to compare the inoculation effect. The field experiment had 17 treatments in three replicates and the best level of inoculum was decided based on net returns and cost: benefit ratio. In peanut, the combination of 4 ml of Rhizobium and 2 ml of PSB resulted in the highest net returns and higher cost to benefit ratio of 1:2.98 followed by treatment with a combination of 2 ml per kg each of Rhizobium and PSB with a B;C ratio of 1:2.84. The benefits in terms of net returns were to the extent of 16 percent due to inoculation with lignite based formulations while it was up to 48 percent due to the best combination of liquid biofertilizers. In maize combination of liquid formulations consisting of 4 ml of Azospirillum and 2 ml of PSB resulted in the highest net returns; about 53 percent higher than the un-inoculated control and 20 percent higher than the treatment with lignite based formulation. In both the crops inoculation with lignite based formulations significantly increased the net returns over un-inoculated control while levels higher or lesser than 4 ml of Rhizobium and Azospirillum and higher or lesser than 2 ml of PSB were not economical and hence not optimal for these two crops.Keywords: Rhizobium, Azospirillum, phosphate solubilizing bacteria, liquid formulation, benefit-cost ratio
Procedia PDF Downloads 493988 The Model of Learning Centre on OTOP Production Process Based on Sufficiency Economic Philosophy for Sustainable Life Quality
Authors: Napasri Suwanajote
Abstract:
The purposes of this research were to analyse and evaluate successful factors in OTOP production process for the developing of learning centre on OTOP production process based on Sufficiency Economic Philosophy for sustainable life quality. The research has been designed as a qualitative study to gather information from 30 OTOP producers in Bangkontee District, Samudsongkram Province. They were all interviewed on 3 main parts. Part 1 was about the production process including 1) production 2) product development 3) the community strength 4) marketing possibility and 5) product quality. Part 2 evaluated appropriate successful factors including 1) the analysis of the successful factors 2) evaluate the strategy based on Sufficiency Economic Philosophy and 3) the model of learning centre on OTOP production process based on Sufficiency Economic Philosophy for sustainable life quality. The results showed that the production did not affect the environment with potential in continuing standard quality production. They used the raw materials in the country. On the aspect of product and community strength in the past 1 year, it was found that there was no appropriate packaging showing product identity according to global market standard. They needed the training on packaging especially for food and drink products. On the aspect of product quality and product specification, it was found that the products were certified by the local OTOP standard. There should be a responsible organization to help the uncertified producers pass the standard. However, there was a problem on food contamination which was hazardous to the consumers. The producers should cooperate with the government sector or educational institutes involving with food processing to reach FDA standard. The results from small group discussion showed that the community expected high education and better standard living. Some problems reported by the community included informal debt and drugs in the community. There were 8 steps in developing the model of learning centre on OTOP production process based on Sufficiency Economic Philosophy for sustainable life quality.Keywords: production process, OTOP, sufficiency economic philosophy, marketing management
Procedia PDF Downloads 234987 Relationship between Structure of Some Nitroaromatic Pollutants and Their Degradation Kinetic Parameters in UV-VIS/TIO2 System
Authors: I. Nitoi, P. Oancea, M. Raileanu, M. Crisan, L. Constantin, I. Cristea
Abstract:
Hazardous organic compounds like nitroaromatics are frequently found in chemical and petroleum industries discharged effluents. Due to their bio-refractory character and high chemical stability cannot be efficiently removed by classical biological or physical-chemical treatment processes. In the past decades, semiconductor photocatalysis has been frequently applied for the advanced degradation of toxic pollutants. Among various semiconductors titania was a widely studied photocatalyst, due to its chemical inertness, low cost, photostability and nontoxicity. In order to improve optical absorption and photocatalytic activity of TiO2 many attempts have been made, one feasible approach consists of doping oxide semiconductor with metal. The degradation of dinitrobenzene (DNB) and dinitrotoluene (DNT) from aqueous solution under UVA-VIS irradiation using heavy metal (0.5% Fe, 1%Co, 1%Ni ) doped titania was investigated. The photodegradation experiments were carried out using a Heraeus laboratory scale UV-VIS reactor equipped with a medium-pressure mercury lamp which emits in the range: 320-500 nm. Solutions with (0.34-3.14) x 10-4 M pollutant content were photo-oxidized in the following working conditions: pH = 5-9; photocatalyst dose = 200 mg/L; irradiation time = 30 – 240 minutes. Prior to irradiation, the photocatalyst powder was added to the samples, and solutions were bubbled with air (50 L/hour), in the dark, for 30 min. Dopant type, pH, structure and initial pollutant concentration influence on the degradation efficiency were evaluated in order to set up the optimal working conditions which assure substrate advanced degradation. The kinetics of nitroaromatics degradation and organic nitrogen mineralization was assessed and pseudo-first order rate constants were calculated. Fe doped photocatalyst with lowest metal content (0.5 wt.%) showed a considerable better behaviour in respect to pollutant degradation than Co and Ni (1wt.%) doped titania catalysts. For the same working conditions, degradation efficiency was higher for DNT than DNB in accordance with their calculated adsobance constants (Kad), taking into account that degradation process occurs on catalyst surface following a Langmuir-Hinshalwood model. The presence of methyl group in the structure of DNT allows its degradation by oxidative and reductive pathways, while DNB is converted only by reductive route, which also explain the highest DNT degradation efficiency. For highest pollutant concentration tested (3 x 10-4 M), optimum working conditions (0.5 wt.% Fe doped –TiO2 loading of 200 mg/L, pH=7 and 240 min. irradiation time) assures advanced nitroaromatics degradation (ηDNB=89%, ηDNT=94%) and organic nitrogen mineralization (ηDNB=44%, ηDNT=47%).Keywords: hazardous organic compounds, irradiation, nitroaromatics, photocatalysis
Procedia PDF Downloads 317986 Design and Development of an Autonomous Beach Cleaning Vehicle
Authors: Mahdi Allaoua Seklab, Süleyman BaşTürk
Abstract:
In the quest to enhance coastal environmental health, this study introduces a fully autonomous beach cleaning machine, a breakthrough in leveraging green energy and advanced artificial intelligence for ecological preservation. Designed to operate independently, the machine is propelled by a solar-powered system, underscoring a commitment to sustainability and the use of renewable energy in autonomous robotics. The vehicle's autonomous navigation is achieved through a sophisticated integration of LIDAR and a camera system, utilizing an SSD MobileNet V2 object detection model for accurate and real-time trash identification. The SSD framework, renowned for its efficiency in detecting objects in various scenarios, is coupled with the lightweight and precise highly MobileNet V2 architecture, making it particularly suited for the computational constraints of on-board processing in mobile robotics. Training of the SSD MobileNet V2 model was conducted on Google Colab, harnessing cloud-based GPU resources to facilitate a rapid and cost-effective learning process. The model was refined with an extensive dataset of annotated beach debris, optimizing the parameters using the Adam optimizer and a cross-entropy loss function to achieve high-precision trash detection. This capability allows the machine to intelligently categorize and target waste, leading to more effective cleaning operations. This paper details the design and functionality of the beach cleaning machine, emphasizing its autonomous operational capabilities and the novel application of AI in environmental robotics. The results showcase the potential of such technology to fill existing gaps in beach maintenance, offering a scalable and eco-friendly solution to the growing problem of coastal pollution. The deployment of this machine represents a significant advancement in the field, setting a new standard for the integration of autonomous systems in the service of environmental stewardship.Keywords: autonomous beach cleaning machine, renewable energy systems, coastal management, environmental robotics
Procedia PDF Downloads 30985 Modeling Search-And-Rescue Operations by Autonomous Mobile Robots at Sea
Authors: B. Kriheli, E. Levner, T. C. E. Cheng, C. T. Ng
Abstract:
During the last decades, research interest in planning, scheduling, and control of emergency response operations, especially people rescue and evacuation from the dangerous zone of marine accidents, has increased dramatically. Until the survivors (called ‘targets’) are found and saved, it may cause loss or damage whose extent depends on the location of the targets and the search duration. The problem is to efficiently search for and detect/rescue the targets as soon as possible with the help of intelligent mobile robots so as to maximize the number of saved people and/or minimize the search cost under restrictions on the amount of saved people within the allowable response time. We consider a special situation when the autonomous mobile robots (AMR), e.g., unmanned aerial vehicles and remote-controlled robo-ships have no operator on board as they are guided and completely controlled by on-board sensors and computer programs. We construct a mathematical model for the search process in an uncertain environment and provide a new fast algorithm for scheduling the activities of the autonomous robots during the search-and rescue missions after an accident at sea. We presume that in the unknown environments, the AMR’s search-and-rescue activity is subject to two types of error: (i) a 'false-negative' detection error where a target object is not discovered (‘overlooked') by the AMR’s sensors in spite that the AMR is in a close neighborhood of the latter and (ii) a 'false-positive' detection error, also known as ‘a false alarm’, in which a clean place or area is wrongly classified by the AMR’s sensors as a correct target. As the general resource-constrained discrete search problem is NP-hard, we restrict our study to finding local-optimal strategies. A specificity of the considered operational research problem in comparison with the traditional Kadane-De Groot-Stone search models is that in our model the probability of the successful search outcome depends not only on cost/time/probability parameters assigned to each individual location but, as well, on parameters characterizing the entire history of (unsuccessful) search before selecting any next location. We provide a fast approximation algorithm for finding the AMR route adopting a greedy search strategy in which, in each step, the on-board computer computes a current search effectiveness value for each location in the zone and sequentially searches for a location with the highest search effectiveness value. Extensive experiments with random and real-life data provide strong evidence in favor of the suggested operations research model and corresponding algorithm.Keywords: disaster management, intelligent robots, scheduling algorithm, search-and-rescue at sea
Procedia PDF Downloads 174984 Exploration of the Nonlinear Viscoelastic Behavior of Yogurt Using Lissajous Curves
Authors: Hugo Espinosa-Andrews
Abstract:
Introduction: Yogurt is widely accepted worldwide due to its high nutritional value, consistency, and texture. Their rheological properties play a significant role in consumer acceptance and are related to the manufacturing process and formulation. Typically, the viscoelastic characteristics of yogurts are studied using the small amplitude oscillatory shear test; however, the initial stages of flow and oral processing are described in the nonlinear zone, in which a large amplitude oscillatory stress test is applied. The objective of this work was to analyze the nonlinear viscoelastic behavior of commercial yogurts using Lissajous curves. Methods: Two commercial yogurts were purchased in a local store in Guadalajara Jalisco Mexico: a natural Greek-style yogurt and a low-fat traditional yogurt. Viscoelastic properties were evaluated using a large amplitude oscillatory stress procedure (LAOS). A crosshatch geometry of 40 mm and a truncation of 1000 µm were used. Stress sweeps were performed at 6.28 rad/s from 1 to 250 Pa at 5°C. The nonlinear viscoelastic properties were analyzed using the Lissajous curves. Results: The yogurts showed strain-viscoelastic behavior related to deformation-dependent materials. In the low-strain region, the elastic modulus predominated over the viscous modulus, showing gel-elastic properties. The sol-gel transitions were observed at approximately 66.5 Pa for the Greek yogurt, double that detected for traditional yogurt. The viscoelastic behavior of the yogurts was characteristic of weak excess deformation: behavior indicating a stable molecular structure at rest, and moderate structure at medium shear-forces. The normalized Lissajous curves characterized viscoelastic transitions of the yogurt as the stress increased. Greater viscoelasticity deformation was observed in Greek yogurt than in traditional yogurt, which is related to the presence of a protein network with a greater degree of crosslinking. Conclusions: The yogurt composition influences the viscoelastic properties of the material. Yogurt with the higher percentage of protein has greater viscoelastic and viscous properties, which describe a product of greater consistency and creaminess.Keywords: yogurt, viscoelastic properties, LAOS, elastic modulus
Procedia PDF Downloads 27983 Experimental and Computational Fluid Dynamic Modeling of a Progressing Cavity Pump Handling Newtonian Fluids
Authors: Deisy Becerra, Edwar Perez, Nicolas Rios, Miguel Asuaje
Abstract:
Progressing Cavity Pump (PCP) is a type of positive displacement pump that is being awarded greater importance as capable artificial lift equipment in the heavy oil field. The most commonly PCP used is driven single lobe pump that consists of a single external helical rotor turning eccentrically inside a double internal helical stator. This type of pump was analyzed by the experimental and Computational Fluid Dynamic (CFD) approach from the DCAB031 model located in a closed-loop arrangement. Experimental measurements were taken to determine the pressure rise and flow rate with a flow control valve installed at the outlet of the pump. The flowrate handled was measured by a FLOMEC-OM025 oval gear flowmeter. For each flowrate considered, the pump’s rotational speed and power input were controlled using an Invertek Optidrive E3 frequency driver. Once a steady-state operation was attained, pressure rise measurements were taken with a Sper Scientific wide range digital pressure meter. In this study, water and three Newtonian oils of different viscosities were tested at different rotational speeds. The CFD model implementation was developed on Star- CCM+ using an Overset Mesh that includes the relative motion between rotor and stator, which is one of the main contributions of the present work. The simulations are capable of providing detailed information about the pressure and velocity fields inside the device in laminar and unsteady regimens. The simulations have a good agreement with the experimental data due to Mean Squared Error (MSE) in under 21%, and the Grid Convergence Index (GCI) was calculated for the validation of the mesh, obtaining a value of 2.5%. In this case, three different rotational speeds were evaluated (200, 300, 400 rpm), and it is possible to show a directly proportional relationship between the rotational speed of the rotor and the flow rate calculated. The maximum production rates for the different speeds for water were 3.8 GPM, 4.3 GPM, and 6.1 GPM; also, for the oil tested were 1.8 GPM, 2.5 GPM, 3.8 GPM, respectively. Likewise, an inversely proportional relationship between the viscosity of the fluid and pump performance was observed, since the viscous oils showed the lowest pressure increase and the lowest volumetric flow pumped, with a degradation around of 30% of the pressure rise, between performance curves. Finally, the Productivity Index (PI) remained approximately constant for the different speeds evaluated; however, between fluids exist a diminution due to the viscosity.Keywords: computational fluid dynamic, CFD, Newtonian fluids, overset mesh, PCP pressure rise
Procedia PDF Downloads 128982 A Methodology for Developing New Technology Ideas to Avoid Patent Infringement: F-Term Based Patent Analysis
Authors: Kisik Song, Sungjoo Lee
Abstract:
With the growing importance of intangible assets recently, the impact of patent infringement on the business of a company has become more evident. Accordingly, it is essential for firms to estimate the risk of patent infringement risk before developing a technology and create new technology ideas to avoid the risk. Recognizing the needs, several attempts have been made to help develop new technology opportunities and most of them have focused on identifying emerging vacant technologies from patent analysis. In these studies, the IPC (International Patent Classification) system or keywords from text-mining application to patent documents was generally used to define vacant technologies. Unlike those studies, this study adopted F-term, which classifies patent documents according to the technical features of the inventions described in them. Since the technical features are analyzed by various perspectives by F-term, F-term provides more detailed information about technologies compared to IPC while more systematic information compared to keywords. Therefore, if well utilized, it can be a useful guideline to create a new technology idea. Recognizing the potential of F-term, this paper aims to suggest a novel approach to developing new technology ideas to avoid patent infringement based on F-term. For this purpose, we firstly collected data about F-term and then applied text-mining to the descriptions about classification criteria and attributes. From the text-mining results, we could identify other technologies with similar technical features of the existing one, the patented technology. Finally, we compare the technologies and extract the technical features that are commonly used in other technologies but have not been used in the existing one. These features are presented in terms of “purpose”, “function”, “structure”, “material”, “method”, “processing and operation procedure” and “control means” and so are useful for creating new technology ideas that help avoid infringing patent rights of other companies. Theoretically, this is one of the earliest attempts to adopt F-term to patent analysis; the proposed methodology can show how to best take advantage of F-term with the wealth of technical information. In practice, the proposed methodology can be valuable in the ideation process for successful product and service innovation without infringing the patents of other companies.Keywords: patent infringement, new technology ideas, patent analysis, F-term
Procedia PDF Downloads 270981 Gravitational Water Vortex Power Plant: Experimental-Parametric Design of a Hydraulic Structure Capable of Inducing the Artificial Formation of a Gravitational Water Vortex Appropriate for Hydroelectric Generation
Authors: Henrry Vicente Rojas Asuero, Holger Manuel Benavides Muñoz
Abstract:
Approximately 80% of the energy consumed worldwide is generated from fossil sources, which are responsible for the emission of a large volume of greenhouse gases. For this reason, the global trend, at present, is the widespread use of energy produced from renewable sources. This seeks safety and diversification of energy supply, based on social cohesion, economic feasibility and environmental protection. In this scenario, small hydropower systems (P ≤ 10MW) stand out due to their high efficiency, economic competitiveness and low environmental impact. Small hydropower systems, along with wind and solar energy, are expected to represent a significant percentage of the world's energy matrix in the near term. Among the various technologies present in the state of the art, relating to small hydropower systems, is the Gravitational Water Vortex Power Plant, a recent technology that excels because of its versatility of operation, since it can operate with jumps in the range of 0.70 m-2.00 m and flow rates from 1 m3/s to 20 m3/s. Its operating system is based on the utilization of the energy of rotation contained within a large water vortex artificially induced. This paper presents the study and experimental design of an optimal hydraulic structure with the capacity to induce the artificial formation of a gravitational water vortex trough a system of easy application and high efficiency, able to operate in conditions of very low head and minimum flow. The proposed structure consists of a channel, with variable base, vortex inductor, tangential flow generator, coupled to a circular tank with a conical transition bottom hole. In the laboratory test, the angular velocity of the water vortex was related to the geometric characteristics of the inductor channel, as well as the influence of the conical transition bottom hole on the physical characteristics of the water vortex. The results show angular velocity values of greater magnitude as a function of depth, in addition the presence of the conical transition in the bottom hole of the circular tank improves the water vortex formation conditions while increasing the angular velocity values. Thus, the proposed system is a sustainable solution for the energy supply of rural areas near to watercourses.Keywords: experimental model, gravitational water vortex power plant, renewable energy, small hydropower
Procedia PDF Downloads 291980 A Literature Review on the Use of Information and Communication Technology within and between Emergency Medical Teams during a Disaster
Authors: Badryah Alshehri, Kevin Gormley, Gillian Prue, Karen McCutcheon
Abstract:
In a disaster event, sharing patient information between the pre-hospitals Emergency Medical Services (EMS) and Emergency Department (ED) hospitals is a complex process during which important information may be altered or lost due to poor communication. The aim of this study was to critically discuss the current evidence base in relation to communication between pre-EMS hospital and ED hospital professionals by the use of Information and Communication Systems (ICT). This study followed the systematic approach; six electronic databases were searched: CINAHL, Medline, Embase, PubMed, Web of Science, and IEEE Xplore Digital Library were comprehensively searched in January 2018 and a second search was completed in April 2020 to capture more recent publications. The study selection process was undertaken independently by the study authors. Both qualitative and quantitative studies were chosen that focused on factors which are positively or negatively associated with coordinated communication between pre-hospital EMS and ED teams in a disaster event. These studies were assessed for quality and the data were analysed according to the key screening themes which emerged from the literature search. Twenty-two studies were included. Eleven studies employed quantitative methods, seven studies used qualitative methods, and four studies used mixed methods. Four themes emerged on communication between EMTs (pre-hospital EMS and ED staff) in a disaster event using the ICT. (1) Disaster preparedness plans and coordination. This theme reported that disaster plans are in place in hospitals, and in some cases, there are interagency agreements with pre-hospital and relevant stakeholders. However, the findings showed that the disaster plans highlighted in these studies lacked information regarding coordinated communications within and between the pre-hospital and hospital. (2) Communication systems used in the disaster. This theme highlighted that although various communication systems are used between and within hospitals and pre-hospitals, technical issues have influenced communication between teams during disasters. (3) Integrated information management systems. This theme suggested the need for an integrated health information system which can help pre-hospital and hospital staff to record patient data and ensure the data is shared. (4) Disaster training and drills. While some studies analysed disaster drills and training, the majority of these studies were focused on hospital departments other than EMTs. These studies suggest the need for simulation disaster training and drills, including EMTs. This review demonstrates that considerable gaps remain in the understanding of the communication between the EMS and ED hospitals staff in relation to response in disasters. The review shows that although different types of ICTs are used, various issues remain which affect coordinated communication among the relevant professionals.Keywords: communication, emergency communication services, emergency medical teams, emergency physicians, emergency nursing, paramedics, information and communication technology, communication systems
Procedia PDF Downloads 86