Search results for: sound monitoring
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3884

Search results for: sound monitoring

1334 Investigation of Crack Formation in Ordinary Reinforced Concrete Beams and in Beams Strengthened with Carbon Fiber Sheet: Theory and Experiment

Authors: Anton A. Bykov, Irina O. Glot, Igor N. Shardakov, Alexey P. Shestakov

Abstract:

This paper presents the results of experimental and theoretical investigations of the mechanisms of crack formation in reinforced concrete beams subjected to quasi-static bending. The boundary-value problem has been formulated in the framework of brittle fracture mechanics and has been solved by using the finite-element method. Numerical simulation of the vibrations of an uncracked beam and a beam with cracks of different size serves to determine the pattern of changes in the spectrum of eigenfrequencies observed during crack evolution. Experiments were performed on the sequential quasistatic four-point bending of the beam leading to the formation of cracks in concrete. At each loading stage, the beam was subjected to an impulse load to induce vibrations. Two stages of cracking were detected. At the first stage the conservative process of deformation is realized. The second stage is an active cracking, which is marked by a sharp change in eingenfrequencies. The boundary of a transition from one stage to another is well registered. The vibration behavior was examined for the beams strengthened by carbon-fiber sheet before loading and at the intermediate stage of loading after the grouting of initial cracks. The obtained results show that the vibrodiagnostic approach is an effective tool for monitoring of cracking and for assessing the quality of measures aimed at strengthening concrete structures.

Keywords: crack formation, experiment, mathematical modeling, reinforced concrete, vibrodiagnostics

Procedia PDF Downloads 306
1333 Real-Time Pedestrian Detection Method Based on Improved YOLOv3

Authors: Jingting Luo, Yong Wang, Ying Wang

Abstract:

Pedestrian detection in image or video data is a very important and challenging task in security surveillance. The difficulty of this task is to locate and detect pedestrians of different scales in complex scenes accurately. To solve these problems, a deep neural network (RT-YOLOv3) is proposed to realize real-time pedestrian detection at different scales in security monitoring. RT-YOLOv3 improves the traditional YOLOv3 algorithm. Firstly, the deep residual network is added to extract vehicle features. Then six convolutional neural networks with different scales are designed and fused with the corresponding scale feature maps in the residual network to form the final feature pyramid to perform pedestrian detection tasks. This method can better characterize pedestrians. In order to further improve the accuracy and generalization ability of the model, a hybrid pedestrian data set training method is used to extract pedestrian data from the VOC data set and train with the INRIA pedestrian data set. Experiments show that the proposed RT-YOLOv3 method achieves 93.57% accuracy of mAP (mean average precision) and 46.52f/s (number of frames per second). In terms of accuracy, RT-YOLOv3 performs better than Fast R-CNN, Faster R-CNN, YOLO, SSD, YOLOv2, and YOLOv3. This method reduces the missed detection rate and false detection rate, improves the positioning accuracy, and meets the requirements of real-time detection of pedestrian objects.

Keywords: pedestrian detection, feature detection, convolutional neural network, real-time detection, YOLOv3

Procedia PDF Downloads 142
1332 Inulinase Immobilization on Functionalized Magnetic Nanoparticles Prepared with Soy Protein Isolate Conjugated Bovine Serum Albumin for High Fructose Syrup Production

Authors: Homa Torabizadeh, Mohaddeseh Mikani

Abstract:

Inulinase from Aspergillus niger was covalently immobilized on magnetic nanoparticles (MNPs/Fe3O4) covered with soy protein isolate (SPI/Fe3O4) functionalized by bovine serum albumin (BSA) nanoparticles. MNPs are promising enzyme carriers because they separate easily under external magnetic fields and have enhanced immobilized enzyme reusability. As MNPs aggregate simply, surface coating strategy was employed. SPI functionalized by BSA was a suitable candidate for nanomagnetite coating due to its superior biocompatibility and hydrophilicity. Fe3O4@SPI-BSA nanoparticles were synthesized as a novel carrier with narrow particle size distribution. Step by step fabrication monitoring of Fe3O4@SPI-BSA nanoparticles was performed using field emission scanning electron microscopy and dynamic light scattering. The results illustrated that nanomagnetite with the spherical morphology was well monodispersed with the diameter of about 35 nm. The average size of the SPI-BSA nanoparticles was 80 to 90 nm, and their zeta potential was around −34 mV. Finally, the mean diameter of fabricated Fe3O4@SPI-BSA NPs was less than 120 nm. Inulinase enzyme from Aspergillus niger was covalently immobilized through gluteraldehyde on Fe3O4@SPI-BSA nanoparticles successfully. Fourier transform infrared spectra and field emission scanning electron microscopy images provided sufficient proof for the enzyme immobilization on the nanoparticles with 80% enzyme loading.

Keywords: high fructose syrup, inulinase immobilization, functionalized magnetic nanoparticles, soy protein isolate

Procedia PDF Downloads 299
1331 Carbon Capture and Storage in Geological Formation, its Legal, Regulatory Imperatives and Opportunities in India

Authors: Kalbende Krunal Ramesh

Abstract:

The Carbon Capture and Storage Technology (CCS) provides a veritable platform to bridge the gap between the seemingly irreconcilable twin global challenges of ensuring a secure, reliable and diversified energy supply and mitigating climate change by reducing atmospheric emissions of carbon dioxide. Making its proper regulatory policy and making it flexible for the government and private company by law to regulate, also exploring the opportunity in this sector is the main aim of this paper. India's total annual emissions was 1725 Mt CO2 in 2011, which comprises of 6% of total global emission. It is very important to control the greenhouse gas emission for the environment protection. This paper discusses the various regulatory policy and technology adopted by some of the countries for successful using CCS technology. The brief geology of sedimentary basins in India is studied, ranging from the category I to category IV and deep water and potential for mature technology in CCS is reviewed. Areas not suitable for CO2 storage using presently mature technologies were over viewed. CSS and Clean development mechanism was developed for India, considering the various aspects from research and development, project appraisal, approval and validation, implementation, monitoring and verification, carbon credit issued, cap and trade system and its storage potential. The opportunities in oil and gas operations, power sector, transport sector is discussed briefly.

Keywords: carbon credit issued, cap and trade system, carbon capture and storage technology, greenhouse gas

Procedia PDF Downloads 433
1330 Experimental and Numerical Investigation of Fracture Behavior of Foamed Concrete Based on Three-Point Bending Test of Beams with Initial Notch

Authors: M. Kozłowski, M. Kadela

Abstract:

Foamed concrete is known for its low self-weight and excellent thermal and acoustic properties. For many years, it has been used worldwide for insulation to foundations and roof tiles, as backfill to retaining walls, sound insulation, etc. However, in the last years it has become a promising material also for structural purposes e.g. for stabilization of weak soils. Due to favorable properties of foamed concrete, many interests and studies were involved to analyze its strength, mechanical, thermal and acoustic properties. However, these studies do not cover the investigation of fracture energy which is the core factor governing the damage and fracture mechanisms. Only limited number of publications can be found in literature. The paper presents the results of experimental investigation and numerical campaign of foamed concrete based on three-point bending test of beams with initial notch. First part of the paper presents the results of a series of static loading tests performed to investigate the fracture properties of foamed concrete of varying density. Beam specimens with dimensions of 100×100×840 mm with a central notch were tested in three-point bending. Subsequently, remaining halves of the specimens with dimensions of 100×100×420 mm were tested again as un-notched beams in the same set-up with reduced distance between supports. The tests were performed in a hydraulic displacement controlled testing machine with a load capacity of 5 kN. Apart from measuring the loading and mid-span displacement, a crack mouth opening displacement (CMOD) was monitored. Based on the load – displacement curves of notched beams the values of fracture energy and tensile stress at failure were calculated. The flexural tensile strength was obtained on un-notched beams with dimensions of 100×100×420 mm. Moreover, cube specimens 150×150×150 mm were tested in compression to determine the compressive strength. Second part of the paper deals with numerical investigation of the fracture behavior of beams with initial notch presented in the first part of the paper. Extended Finite Element Method (XFEM) was used to simulate and analyze the damage and fracture process. The influence of meshing and variation of mechanical properties on results was investigated. Numerical models simulate correctly the behavior of beams observed during three-point bending. The numerical results show that XFEM can be used to simulate different fracture toughness of foamed concrete and fracture types. Using the XFEM and computer simulation technology allow for reliable approximation of load–bearing capacity and damage mechanisms of beams made of foamed concrete, which provides some foundations for realistic structural applications.

Keywords: foamed concrete, fracture energy, three-point bending, XFEM

Procedia PDF Downloads 300
1329 Design and Modeling of a Green Building Energy Efficient System

Authors: Berhane Gebreslassie

Abstract:

Conventional commericial buildings are among the highest unwisely consumes enormous amount of energy and as consequence produce significant amount Carbon Dioxide (CO2). Traditional/conventional buildings have been built for years without consideration being given to their impact on the global warming issues as well as their CO2 contributions. Since 1973, simulation of Green Building (GB) for Energy Efficiency started and many countries in particular the US showed a positive response to minimize the usage of energy in respect to reducing the CO2 emission. As a consequence many software companies developed their own unique building energy efficiency simulation software, interfacing interoperability with Building Information Modeling (BIM). The last decade has witnessed very rapid growing number of researches on GB energy efficiency system. However, the study also indicates that the results of current GB simulation are not yet satisfactory to meet the objectives of GB. In addition most of these previous studies are unlikely excluded the studies of ultimate building energy efficiencies simulation. The aim of this project is to meet the objectives of GB by design, modeling and simulation of building ultimate energy efficiencies system. This research project presents multi-level, L-shape office building in which every particular part of the building materials has been tested for energy efficiency. An overall of 78.62% energy is saved, approaching to NetZero energy saving. Furthermore, the building is implements with distributed energy resources like renewable energies and integrating with Smart Building Automation System (SBAS) for controlling and monitoring energy usage.

Keywords: ultimate energy saving, optimum energy saving, green building, sustainable materials and renewable energy

Procedia PDF Downloads 275
1328 XAI Implemented Prognostic Framework: Condition Monitoring and Alert System Based on RUL and Sensory Data

Authors: Faruk Ozdemir, Roy Kalawsky, Peter Hubbard

Abstract:

Accurate estimation of RUL provides a basis for effective predictive maintenance, reducing unexpected downtime for industrial equipment. However, while models such as the Random Forest have effective predictive capabilities, they are the so-called ‘black box’ models, where interpretability is at a threshold to make critical diagnostic decisions involved in industries related to aviation. The purpose of this work is to present a prognostic framework that embeds Explainable Artificial Intelligence (XAI) techniques in order to provide essential transparency in Machine Learning methods' decision-making mechanisms based on sensor data, with the objective of procuring actionable insights for the aviation industry. Sensor readings have been gathered from critical equipment such as turbofan jet engine and landing gear, and the prediction of the RUL is done by a Random Forest model. It involves steps such as data gathering, feature engineering, model training, and evaluation. These critical components’ datasets are independently trained and evaluated by the models. While suitable predictions are served, their performance metrics are reasonably good; such complex models, however obscure reasoning for the predictions made by them and may even undermine the confidence of the decision-maker or the maintenance teams. This is followed by global explanations using SHAP and local explanations using LIME in the second phase to bridge the gap in reliability within industrial contexts. These tools analyze model decisions, highlighting feature importance and explaining how each input variable affects the output. This dual approach offers a general comprehension of the overall model behavior and detailed insight into specific predictions. The proposed framework, in its third component, incorporates the techniques of causal analysis in the form of Granger causality tests in order to move beyond correlation toward causation. This will not only allow the model to predict failures but also present reasons, from the key sensor features linked to possible failure mechanisms to relevant personnel. The causality between sensor behaviors and equipment failures creates much value for maintenance teams due to better root cause identification and effective preventive measures. This step contributes to the system being more explainable. Surrogate Several simple models, including Decision Trees and Linear Models, can be used in yet another stage to approximately represent the complex Random Forest model. These simpler models act as backups, replicating important jobs of the original model's behavior. If the feature explanations obtained from the surrogate model are cross-validated with the primary model, the insights derived would be more reliable and provide an intuitive sense of how the input variables affect the predictions. We then create an iterative explainable feedback loop, where the knowledge learned from the explainability methods feeds back into the training of the models. This feeds into a cycle of continuous improvement both in model accuracy and interpretability over time. By systematically integrating new findings, the model is expected to adapt to changed conditions and further develop its prognosis capability. These components are then presented to the decision-makers through the development of a fully transparent condition monitoring and alert system. The system provides a holistic tool for maintenance operations by leveraging RUL predictions, feature importance scores, persistent sensor threshold values, and autonomous alert mechanisms. Since the system will provide explanations for the predictions given, along with active alerts, the maintenance personnel can make informed decisions on their end regarding correct interventions to extend the life of the critical machinery.

Keywords: predictive maintenance, explainable artificial intelligence, prognostic, RUL, machine learning, turbofan engines, C-MAPSS dataset

Procedia PDF Downloads 7
1327 The Implementation of the Lean Six Sigma Production Process in a Telecommunications Company in Brazil

Authors: Carlos Fontanillas

Abstract:

The implementation of the lean six sigma methodology aims to implement practices to systematically improve processes by eliminating defects, making them cheaper. The implementation of projects with the methodology uses a division into five phases: definition, measurement, analysis, implementation, and control. In this process, it is understood that the implementation of said methodology generates benefits to organizations that adhere through the improvement of their processes. In the case of a telecommunications company, it was realized that the implementation of a lean six sigma project contributed to the improvement of the presented process, generating a financial return with the avoided cost. However, such study has limitations such as a specific segment of performance and procedure, i.e., it can not be defined that return under other circumstances will be the same. It is also concluded that lean six sigma projects tend to contribute to improved processes evaluated due to their methodology that is based on statistical analysis and quality management tools and can generate a financial return. It is hoped that the present study can be used to provide a clearer view of the methodology for entrepreneurs who wish to implement process improvement actions in their companies, as well as to provide a foundation for professionals working with lean six sigma projects. After the review of the processes, the completion of the project stages and the monitoring for three months in partnership with the owner of the process to ensure the effectiveness of the actions, the project was completed with the objective reached. There was an average of 60% reduction with the issuance of undue invoices generated after the deactivation and it was possible to extend the project to other companies, which allowed a reduction well above the initially stipulated target.

Keywords: quality, process, lean six sigma, organization

Procedia PDF Downloads 129
1326 Development of pm2.5 Forecasting System in Seoul, South Korea Using Chemical Transport Modeling and ConvLSTM-DNN

Authors: Ji-Seok Koo, Hee‑Yong Kwon, Hui-Young Yun, Kyung-Hui Wang, Youn-Seo Koo

Abstract:

This paper presents a forecasting system for PM2.5 levels in Seoul, South Korea, leveraging a combination of chemical transport modeling and ConvLSTM-DNN machine learning technology. Exposure to PM2.5 has known detrimental impacts on public health, making its prediction crucial for establishing preventive measures. Existing forecasting models, like the Community Multiscale Air Quality (CMAQ) and Weather Research and Forecasting (WRF), are hindered by their reliance on uncertain input data, such as anthropogenic emissions and meteorological patterns, as well as certain intrinsic model limitations. The system we've developed specifically addresses these issues by integrating machine learning and using carefully selected input features that account for local and distant sources of PM2.5. In South Korea, the PM2.5 concentration is greatly influenced by both local emissions and long-range transport from China, and our model effectively captures these spatial and temporal dynamics. Our PM2.5 prediction system combines the strengths of advanced hybrid machine learning algorithms, convLSTM and DNN, to improve upon the limitations of the traditional CMAQ model. Data used in the system include forecasted information from CMAQ and WRF models, along with actual PM2.5 concentration and weather variable data from monitoring stations in China and South Korea. The system was implemented specifically for Seoul's PM2.5 forecasting.

Keywords: PM2.5 forecast, machine learning, convLSTM, DNN

Procedia PDF Downloads 54
1325 Improved Embroidery Based Textile Electrodes for Sustainability of Impedance Measurement Characteristics

Authors: Bulcha Belay Etana

Abstract:

Research shows that several challenges are to be resolved for textile sensors and wearable smart textiles systems to make it accurate and reproducible minimizing variability issues when tested. To achieve this, we developed stimulating embroidery electrode with three different filling textiles such as 3Dknit, microfiber, and nonwoven fabric, and tested with FTT for high recoverability on compression. Hence The impedance characteristics of wetted electrodes were caried out after 1hr of wetting under normal environmental conditions. The wetted 3D knit (W-3D knit), Wetted nonwoven (W-nonwoven), and wetted microfiber (W-microfiber) developed using Satin stitch performed better than a dry standard stitch or dry Satin stitch electrodes. Its performance was almost the same as that of the gel electrode (Ag/AgCl) as shown by the impedance result in figure 2 .The impedance characteristics of Dry and wetted 3D knit based Embroidered electrodes are better than that of the microfiber, and nonwoven filling textile. This is due to the fact that 3D knit fabric has high recoverability on compression to retain electrolyte gel than microfiber, and nonwoven. However,The non-woven fabric held the electrolyte for longer time without releasing it to the skin when needed, thus making its impedance characteristics poor as observed from the results. Whereas the dry Satin stitch performs better than the standard stitch based developed electrode. The inter electrode distance of all types of the electrode was 25mm, with the area of the electrode being 20mm by 20mm. Detail evaluation and further analysis is in progress for EMG monitoring application

Keywords: impedance, moisture retention, 3D knit fabric, microfiber, nonwoven

Procedia PDF Downloads 140
1324 Performance Comparison of Resource Allocation without Feedback in Wireless Body Area Networks by Various Pseudo Orthogonal Sequences

Authors: Ojin Kwon, Yong-Jin Yoon, Liu Xin, Zhang Hongbao

Abstract:

Wireless Body Area Network (WBAN) is a short-range wireless communication around human body for various applications such as wearable devices, entertainment, military, and especially medical devices. WBAN attracts the attention of continuous health monitoring system including diagnostic procedure, early detection of abnormal conditions, and prevention of emergency situations. Compared to cellular network, WBAN system is more difficult to control inter- and inner-cell interference due to the limited power, limited calculation capability, mobility of patient, and non-cooperation among WBANs. In this paper, we compare the performance of resource allocation scheme based on several Pseudo Orthogonal Codewords (POCs) to mitigate inter-WBAN interference. Previously, the POCs are widely exploited for a protocol sequence and optical orthogonal code. Each POCs have different properties of auto- and cross-correlation and spectral efficiency according to its construction of POCs. To identify different WBANs, several different pseudo orthogonal patterns based on POCs exploits for resource allocation of WBANs. By simulating these pseudo orthogonal resource allocations of WBANs on MATLAB, we obtain the performance of WBANs according to different POCs and can analyze and evaluate the suitability of POCs for the resource allocation in the WBANs system.

Keywords: wireless body area network, body sensor network, resource allocation without feedback, interference mitigation, pseudo orthogonal pattern

Procedia PDF Downloads 353
1323 Comparison of Two Anesthetic Methods during Interventional Neuroradiology Procedure: Propofol versus Sevoflurane Using Patient State Index

Authors: Ki Hwa Lee, Eunsu Kang, Jae Hong Park

Abstract:

Background: Interventional neuroradiology (INR) has been a rapidly growing and evolving neurosurgical part during the past few decades. Sevoflurane and propofol are both suitable anesthetics for INR procedure. Monitoring of depth of anesthesia is being used very widely. SEDLine™ monitor, a 4-channel processed EEG monitor, uses a proprietary algorithm to analyze the raw EEG signal and displays the Patient State Index (PSI) values. There are only a fewer studies examining the PSI in the neuro-anesthesia. We aimed to investigate the difference of PSI values and hemodynamic variables between sevoflurane and propofol anesthesia during INR procedure. Methods: We reviewed the medical records of patients who scheduled to undergo embolization of non-ruptured intracranial aneurysm by a single operator from May 2013 to December 2014, retrospectively. Sixty-five patients were categorized into two groups; sevoflurane (n = 33) vs propofol (n = 32) group. The PSI values, hemodynamic variables, and the use of hemodynamic drugs were analyzed. Results: Significant differences were seen between PSI values obtained during different perioperative stages in both two groups (P < 0.0001). The PSI values of propofol group were lower than that of sevoflurane group during INR procedure (P < 0.01). The patients in propofol group had more prolonged time of extubation and more phenylephrine requirement than sevoflurane group (p < 0.05). Anti-hypertensive drug was more administered to the patients during extubation in sevoflurane group (p < 0.05). Conclusions: The PSI can detect depth of anesthesia and changes of concentration of anesthetics during INR procedure. Extubation was faster in sevoflurane group, but smooth recovery was shown in propofol group.

Keywords: interventional neuroradiology, patient state index, propofol, sevoflurane

Procedia PDF Downloads 180
1322 Cognitive and Metacognitive Space in the Task Design at Postgraduate Taught Level

Authors: Mei Lin, Lana Yj Liu, Thin Ngoc Pham

Abstract:

Postgraduate taught (PGT) students’ learning strategies align with what the learning task constitutes and the environment that the task creates. Cognitively, they can discover new perspectives, challenge general assumptions, establish clear connections, and synthesise information. Metacognitively, their engagement is conducive to the development of planning, monitoring, and evaluating strategies. Given that there has been a lack of longitudinal insights into international PGT students’ experiences of the cognitive and metacognitive space created in the tasks, this paper presentation aims to fill the gaps by longitudinally exploring (1) the fundamentals of task designs to create cognitive and metacognitive space and (2) the opportunities and challenges of multicultural group discussions as a pedagogical approach for the implementation of cognitive and metacognitive space in the learning tasks. Data were collected from the two rounds of semi-structured interviews with 11 international PGT students in two programmes at a UK university -at the end of semester one and at the end of semester two. The findings show that the task designs, to create cognitive and metacognitive space, need to include four interconnected factors: clarity, relevance, motivation, and practicality. In addition, international PGT students perceived that they practised and developed their cognitive and metacognitive abilities while getting immersed in multicultural group discussions. The findings, from the learners’ point of view, make some pedagogy-related suggestions to the task designs at the master’s level, particularly how to engage students in learning during their transition into higher education in a different cultural setting.

Keywords: cognitive space, master students, metacognitive space, task design

Procedia PDF Downloads 59
1321 Coastalization and Urban Sprawl in the Mediterranean: Using High-Resolution Multi-Temporal Data to Identify Typologies of Spatial Development

Authors: Apostolos Lagarias, Anastasia Stratigea

Abstract:

Coastal urbanization is heavily affecting the Mediterranean, taking the form of linear urban sprawl along the coastal zone. This process is posing extreme pressure on ecosystems, leading to an unsustainable model of growth. The aim of this research is to analyze coastal urbanization patterns in the Mediterranean using High-resolution multi-temporal data provided by the Global Human Settlement Layer (GHSL) database. Methodology involves the estimation of a set of spatial metrics characterizing the density, aggregation/clustering and dispersion of built-up areas. As case study areas, the Spanish Coast and the Adriatic Italian Coast are examined. Coastalization profiles are examined and selected sub-areas massively affected by tourism development and suburbanization trends (Costa Blanca/Murcia, Costa del Sol, Puglia, Emilia-Romagna Coast) are analyzed and compared. Results show that there are considerable differences between the Spanish and the Italian typologies of spatial development, related to the land use structure and planning policies applied in each case. Monitoring and analyzing spatial patterns could inform integrated Mediterranean strategies for coastal areas and redirect spatial/environmental policies towards a more sustainable model of growth

Keywords: coastalization, Mediterranean, multi-temporal, urban sprawl, spatial metrics

Procedia PDF Downloads 138
1320 Risk Management in Industrial Supervision Projects

Authors: Érick Aragão Ribeiro, George André Pereira Thé, José Marques Soares

Abstract:

Several problems in industrial supervision software development projects may lead to the delay or cancellation of projects. These problems can be avoided or contained by using identification methods, analysis and control of risks. These procedures can give an overview of the possible problems that can happen in the projects and what are the immediate solutions. Therefore, we propose a risk management method applied to the teaching and development of industrial supervision software. The method is developed through a literature review and previous projects can be divided into phases of management and have basic features that are validated with experimental research carried out by mechatronics engineering students and professionals. The management is conducted through the stages of identification, analysis, planning, monitoring, control and communication of risks. Programmers use a method of prioritizing risks considering the gravity and the possibility of occurrence of the risk. The outputs of the method indicate which risks occurred or are about to happen. The first results indicate which risks occur at different stages of the project and what risks have a high probability of occurring. The results show the efficiency of the proposed method compared to other methods, showing the improvement of software quality and leading developers in their decisions. This new way of developing supervision software helps students identify design problems, evaluate software developed and propose effective solutions. We conclude that the risk management optimizes the development of the industrial process control software and provides higher quality to the product.

Keywords: supervision software, risk management, industrial supervision, project management

Procedia PDF Downloads 356
1319 The Nurse Practitioner’s Role Functions in Multi-Specialist Team When Caring for a Metastatic Colon Cancer Patient with Acute Intestinal Obstruction

Authors: Yun-Tsuen Chen, Shih-Ting Huang, Pi-Fen Cheng, Yu-Ting Su, Joffrey Hsu, Hui-Zhu Chen

Abstract:

Acute intestinal obstruction is one of the differentials of acute abdomen and requires timely alleviation of intestinal distention and abdominal pain to avoid perforation, intra-abdominal infection, and peritonitis. Investigation to identify the cause of obstruction will direct treatment planning and allow for more effective management. In this study, we present a 71-year-old female presenting with symptoms of acute intestinal obstruction for five days. After extensive history taking, physical exam, medical imaging, and pathology, the patient was diagnosed with colon cancer with lung metastasis and acute intestinal obstruction. The patient was placed on nil per os status with intravenous fluid support, intravenous antibiotics, and a decompression nasogastric tube was placed. The patient received decompression with colostomy creation surgery. After assessing the patient’s clinical condition and tumor staging, a multidisciplinary healthcare team created an individualized treatment plan, which included plans to prepare the patient for home self-care and maintain good mental health with regular monitoring in the clinic setting. This case demonstrates the importance of early diagnosis, effective treatment, and a multidisciplinary approach to the management of acute intestinal obstruction secondary to colon cancer.

Keywords: acute intestinal obstruction, colostomy surgery, metastatic colon cancer, multidisciplinary healthcare team

Procedia PDF Downloads 115
1318 High Sensitive Graphene-Based Strain Sensors for SHM of Composite Laminates

Authors: A. Rinaldi, A. Proietti, C. Aquarelli, F. Marra, A. Tamburrano, M. Ciminello, M. S. Sarto

Abstract:

A new type of high sensitive piezoresistive sensors based on graphene was developed within the SARISTU project for application on Structural Health Monitoring (SHM). The new sensor consists of a graphene-based film, obtained through the spray deposition of a colloidal suspension of Multi-Layer Graphene (MLGs) nano platelets over a substrate. MLGs are produced by liquid exfoliation of thermally expanded Graphite Intercalation Compound. An array of 8 sensors is produced by spray deposition over an aeronautical CFRC plate of dimensions 550 mm (length) × 550 mm (width) × 3 mm (thickness). Electromechanical tests were performed in order to assess the sensitivity of the new piezoresistive sensors, which are characterized by an isotropic response. In the quasi-static characterizations, the CFRC plate was clamped on one side and loaded on the opposite one. The local strain map of the plate was then obtained from displacement measurements and numerical analysis. The dynamic tests were performed lying the plate over an anti-vibration table and actuating a piezoelectric element located in the middle of the sensing array. The obtained experimental results demonstrated that the sensors possess a good repeatability and a high constant gauge factor (~200) in the applied strain range 0.001%-0.02%. Moreover, they can follow dynamics up to 400 kHz and for this reason they are good candidates for Lamb-wave analysis.

Keywords: graphene, strain sensor, spray deposition, lamb-wave analysis

Procedia PDF Downloads 431
1317 Evaluation of Turbulence Prediction over Washington, D.C.: Comparison of DCNet Observations and North American Mesoscale Model Outputs

Authors: Nebila Lichiheb, LaToya Myles, William Pendergrass, Bruce Hicks, Dawson Cagle

Abstract:

Atmospheric transport of hazardous materials in urban areas is increasingly under investigation due to the potential impact on human health and the environment. In response to health and safety concerns, several dispersion models have been developed to analyze and predict the dispersion of hazardous contaminants. The models of interest usually rely on meteorological information obtained from the meteorological models of NOAA’s National Weather Service (NWS). However, due to the complexity of the urban environment, NWS forecasts provide an inadequate basis for dispersion computation in urban areas. A dense meteorological network in Washington, DC, called DCNet, has been operated by NOAA since 2003 to support the development of urban monitoring methodologies and provide the driving meteorological observations for atmospheric transport and dispersion models. This study focuses on the comparison of wind observations from the DCNet station on the U.S. Department of Commerce Herbert C. Hoover Building against the North American Mesoscale (NAM) model outputs for the period 2017-2019. The goal is to develop a simple methodology for modifying NAM outputs so that the dispersion requirements of the city and its urban area can be satisfied. This methodology will allow us to quantify the prediction errors of the NAM model and propose adjustments of key variables controlling dispersion model calculation.

Keywords: meteorological data, Washington D.C., DCNet data, NAM model

Procedia PDF Downloads 234
1316 A Study on Compromised Periodontal Health Status among the Pregnant Woman of Jamshedpur, Jharkhand, India

Authors: Rana Praween Kumar

Abstract:

Preterm-low birth weight delivery is a major cause of infant morbidity and mortality in developing countries and has been linked to poor periodontal health during pregnancy. Gingivitis and chronic periodontitis are highly prevalent chronic inflammatory oral diseases. The detection and diagnosis of these common diseases is a fundamentally important component of oral health care. This study is intended to investigate predisposing and enabling factors as determinants of oral health indicators in pregnancy as well as the association between periodontal problems during pregnancy with age and socio economic status of the individual. A community –based prospective cohort study will be conducted in Jamshedpur, Jharkhand, India among pregnant women using completed interviews and a full mouth oral clinical examination using the CPITN (Community Periodontal Index of Treatment Need) and OHI-S (Simplified Oral Hygiene) indices with adequate sample size and informed consent to the patient following proper inclusion and exclusion criteria. Multiple logistic regression analyses will be used to identify independent determinants of periodontal problems and use of dental services during pregnancy. Analysis of covariance (ANCOVA) will be used to investigate the relationship between periodontal problems with the age and socioeconomic status. The result will help in proper monitoring of periodontal health during pregnancy encouraging the delivery of healthy child and the maintenance of proper health of the mother.

Keywords: infant, periodontal problems, pregnancy, pre-term-low birth weight delivery

Procedia PDF Downloads 163
1315 Accessibility Centres in Higher Education Institutions: Inclusiveness and Peer Tutoring Programmes

Authors: Vassilis Argyropoulos, Magda Nikolaraizi, Maria Papazafiri

Abstract:

A growing number of students with disabilities attend institutions of higher education, and according to evidenced-based data, it seems that they face many obstacles regarding their academic access and inclusion. The fact that more and more students decide to actively participate in higher education, on the one hand, empowers and strengthens inclusiveness in tertiary education, but on the other hand, it brings new challenges to their access to scientific content as well as to their interactions with other students and faculty members. For this, accessibility centres have come to the fore in many higher education institutions, in order to respond to the needs of students with disabilities. In this paper, we present a study regarding the peer tutoring program, which is a service delivered by the Accessibility Centre at the University of Thessaly in Greece. Specifically, the current paper aims to describe the experiences of tutees and tutors regarding their relationships developed throughout the peer tutoring program. Twelve tutors and eight tutees with disabilities participated in the study, whose experiences were explored through interviews and were analyzed in a qualitative way. In our study, all tutees and most of the tutors described their relationship as friendly, while a few tutors preferred a more formal relationship. Also, both tutors and tutees described some of the challenges, such as setting limits or arranging an appointment. Finally, peer tutoring programs seem very promising, but in order to be effective, there is a need for training and supporting students regarding their role as well as monitoring the progress of the peer tutoring program, ensuring its smooth operation and success for both tutors and tutees.

Keywords: disability, higher education institutions, interviews, peer tutoring, inclusiveness

Procedia PDF Downloads 53
1314 Optimization of Photocatalytic Degradation of Para-Nitrophenol in Visible Light by Nitrogen and Phosphorus Co-Doped Zinc Oxide Using Factorial Design of Experimental

Authors: Friday Godwin Okibe, Elaoyi David Paul, Oladayo Thomas Ojekunle

Abstract:

In this study, Nitrogen and Phosphorous co-doped Zinc Oxide (NPZ) was prepared through a solvent-free reaction. The NPZ was characterized by Scanning Electron Microscopy (SEM) and Fourier Transform Infrared (FTIR) spectroscopy. The photocatalytic activity of the catalyst was investigated by monitoring the degradation of para-nitrophenol (PNP) under visible light irradiation and the process was optimized using factorial design of experiment. The factors investigated were initial concentration of para-nitrophenol, catalyst loading, pH and irradiation time. The characterization results revealed a successful doping of ZnO by nitrogen and phosphorus and an improvement in the surface morphology of the catalyst. The photo-catalyst exhibited improved photocatalytic activity under visible light by 73.8%. The statistical analysis of the optimization result showed that the model terms were significant at 95% confidence level. Interactions plots revealed that irradiation time was the most significant factor affecting the degradation process. The cube plots of the interactions of the variables showed that an optimum degradation efficiency of 66.9% was achieved at 10mg/L initial PNP concentration, 0.5g catalyst loading, pH 7 and 150 minutes irradiation time.

Keywords: nitrogen and phosphorous co-doped Zno, p-nitrophenol, photocatalytic degradation, optimization, factorial design of experimental

Procedia PDF Downloads 526
1313 Synthesis of ZnO Nanoparticles with Varying Calcination Temperature for Photocatalytic Degradation of Ethylbenzene

Authors: Darlington Ashiegbu, Herman Johannes Potgieter

Abstract:

The increasing utilization of Zinc Oxide (ZnO) as a better alternative to TiO₂ has been attributed to its wide bandgap (3.37eV), lower production cost, ability to absorb over a larger range of the UV-spectrum and higher efficiency in some cases. ZnO nanoparticles were synthesized via sol-gel process and calcined at 400ᵒC, 500ᵒC, and 650ᵒC. The as-synthesized nanoparticles were characterized by X-ray diffraction (XRD), scanning electron microscopy (SEM), energy dispersive X-ray spectroscopy (EDS), and Brunauer–Emmett–Teller (BET) surface area measurement. Scanning electron micrograph revealed pseudo-spherical and rod-like morphologies and a high rate of agglomeration for the sample calcined at 650ᵒC, Brunnauer Emmett Teller (BET) surface area measurement was highest in the sample calcined at 500ᵒC, energy dispersive X-ray spectroscopy (EDS) results confirmed the purity of the samples as only Zn and O₂ were detected and X-ray diffraction (XRD) results revealed crystalline hexagonal wurtzite structure of the ZnO nanoparticles. All three samples were utilized in the degradation of ethylbenzene, and a UV-Vis spectrophotometer was utilized in monitoring degradation of ethylbenzene. The sample calcined at 500ᵒC had the highest surface area for reaction, lowest agglomeration and the highest photocatalytic activity in the degradation of ethylbenzene. This revealed temperature as a very important factor in improved and higher photocatalytic activity.

Keywords: ethylbenzene, pseudo-spherical, sol-gel, zinc oxide

Procedia PDF Downloads 163
1312 Client Hacked Server

Authors: Bagul Abhijeet

Abstract:

Background: Client-Server model is the backbone of today’s internet communication. In which normal user can not have control over particular website or server? By using the same processing model one can have unauthorized access to particular server. In this paper, we discussed about application scenario of hacking for simple website or server consist of unauthorized way to access the server database. This application emerges to autonomously take direct access of simple website or server and retrieve all essential information maintain by administrator. In this system, IP address of server given as input to retrieve user-id and password of server. This leads to breaking administrative security of server and acquires the control of server database. Whereas virus helps to escape from server security by crashing the whole server. Objective: To control malicious attack and preventing all government website, and also find out illegal work to do hackers activity. Results: After implementing different hacking as well as non-hacking techniques, this system hacks simple web sites with normal security credentials. It provides access to server database and allow attacker to perform database operations from client machine. Above Figure shows the experimental result of this application upon different servers and provides satisfactory results as required. Conclusion: In this paper, we have presented a to view to hack the server which include some hacking as well as non-hacking methods. These algorithms and methods provide efficient way to hack server database. By breaking the network security allow to introduce new and better security framework. The terms “Hacking” not only consider for its illegal activities but also it should be use for strengthen our global network.

Keywords: Hacking, Vulnerabilities, Dummy request, Virus, Server monitoring

Procedia PDF Downloads 252
1311 Strategies of Translation: Unlocking the Secret of 'Locksley Hall'

Authors: Raja Lahiani

Abstract:

'Locksley Hall' is a poem that Lord Alfred Tennyson (1809-1892) published in 1842. It is believed to be his first attempt to face as a poet some of the most painful of his experiences, as it is a study of his rising out of sickness into health, conquering his selfish sorrow by faith and hope. So far, in Victorian scholarship as in modern criticism, 'Locksley Hall' has been studied and approached as a canonical Victorian English poem. The aim of this project is to prove that some strategies of translation were used in this poem in such a way as to guarantee its assimilation into the English canon and hence efface to a large extent its Arabic roots. In its relationship with its source text, 'Locksley Hall' is at the same time mimetic and imitative. As part of the terminology used in translation studies, ‘imitation’ means almost the exact opposite of what it means in ordinary English. By adopting an imitative procedure, a translator would do something totally different from the original author, wandering far and freely from the words and sense of the original text. An imitation is thus aimed at an audience which wants the work of the particular translator rather than the work of the original poet. Hallam Tennyson, the poet’s biographer, asserts that 'Locksley Hall' is a simple invention of place, incidents, and people, though he notes that he remembers the poet claiming that Sir William Jones’ prose translation of the Mu‘allaqat (pre-Islamic poems) gave him the idea of the poem. A comparative work would prove that 'Locksley Hall' mirrors a great deal of Tennyson’s biography and hence is not a simple invention of details as asserted by his biographer. It would be challenging to prove that 'Locksley Hall' shares so many details with the Mu‘allaqat, as declared by Tennyson himself, that it needs to be studied as an imitation of the Mu‘allaqat of Imru’ al-Qays and ‘Antara in addition to its being a poem in its own right. Thus, the main aim of this work is to unveil the imitative and mimetic strategies used by Tennyson in his composition of 'Locksley Hall.' It is equally important that this project researches the acculturating assimilative tools used by the poet to root his poem in its Victorian English literary, cultural and spatiotemporal settings. This work adopts a comparative methodology. Comparison is done at different levels. The poem will be contextualized in its Victorian English literary framework. Alien details related to structure, socio-spatial setting, imagery and sound effects shall be compared to Arabic poems from the Mu‘allaqat collection. This would determine whether the poem is a translation, an adaption, an imitation or a genuine work. The ultimate objective of the project is to unveil in this canonical poem a new dimension that has for long been either marginalized or ignored. By proving that 'Locksley Hall' is an imitation of classical Arabic poetry, the project aspires to consolidate its literary value and open up new gates of accessing it.

Keywords: comparative literature, imitation, Locksley Hall, Lord Alfred Tennyson, translation, Victorian poetry

Procedia PDF Downloads 201
1310 Investigating the English Speech Processing System of EFL Japanese Older Children

Authors: Hiromi Kawai

Abstract:

This study investigates the nature of EFL older children’s L2 perceptive and productive abilities using classroom data, in order to find a pedagogical solution to the teaching of L2 sounds at an early stage of learning in a formal school setting. It is still inconclusive whether older children with only EFL formal school instruction at the initial stage of L2 learning are able to attain native-like perception and production in English within the very limited amount of exposure to the target language available. Based on the notion of the lack of study of EFL Japanese children’s acquisition of English segments, the researcher uses a model of L1 speech processing which was developed for investigating L1 English children’s speech and literacy difficulties using a psycholinguistic framework. The model is composed of input channel, output channel, and lexical representation, and examines how a child receives information from spoken or written language, remembers and stores it within the lexical representations and how the child selects and produces spoken or written words. Concerning language universality and language specificity in the language acquisitional process, the aim of finding any sound errors in L1 English children seemed to conform to the author’s intention to find abilities of English sounds in older Japanese children at the novice level of English in an EFL setting. 104 students in Grade 5 (between the ages of 10 and 11 years old) of an elementary school in Tokyo participated in this study. Four tests to measure their perceptive ability and three oral repetition tests to measure their productive ability were conducted with/without reference to lexical representation. All the test items were analyzed to calculate item facility (IF) indices, and correlational analyses and Structural Equation Modeling (SEM) were conducted to examine the relationship between the receptive ability and the productive ability. IF analysis showed that (1) the participants were better at perceiving a segment than producing a segment, (2) they had difficulty in auditory discrimination of paired consonants when one of them does not exist in the Japanese inventory, (3) they had difficulty in both perceiving and producing English vowels, and (4) their L1 loan word knowledge had an influence on their ability to perceive and produce L2 sounds. The result of the Multiple Regression Modeling showed that the two production tests could predict the participants’ auditory ability of real words in English. The result of SEM showed that the hypothesis that perceptive ability affects productive ability was supported. Based on these findings, the author discusses the possible explicit method of teaching English segments to EFL older children in a formal school setting.

Keywords: EFL older children, english segments, perception, production, speech processing system

Procedia PDF Downloads 243
1309 Stable Time Reversed Integration of the Navier-Stokes Equation Using an Adjoint Gradient Method

Authors: Jurriaan Gillissen

Abstract:

This work is concerned with stabilizing the numerical integration of the Navier-Stokes equation (NSE), backwards in time. Applications involve the detection of sources of, e.g., sound, heat, and pollutants. Stable reverse numerical integration of parabolic differential equations is also relevant for image de-blurring. While the literature addresses the reverse integration problem of the advection-diffusion equation, the problem of numerical reverse integration of the NSE has, to our knowledge, not yet been addressed. Owing to the presence of viscosity, the NSE is irreversible, i.e., when going backwards in time, the fluid behaves, as if it had a negative viscosity. As an effect, perturbations from the perfect solution, due to round off errors or discretization errors, grow exponentially in time, and reverse integration of the NSE is inherently unstable, regardless of using an implicit time integration scheme. Consequently, some sort of filtering is required, in order to achieve a stable, numerical, reversed integration. The challenge is to find a filter with a minimal adverse affect on the accuracy of the reversed integration. In the present work, we explore an adjoint gradient method (AGM) to achieve this goal, and we apply this technique to two-dimensional (2D), decaying turbulence. The AGM solves for the initial velocity field u0 at t = 0, that, when integrated forward in time, produces a final velocity field u1 at t = 1, that is as close as is feasibly possible to some specified target field v1. The initial field u0 defines a minimum of a cost-functional J, that measures the distance between u1 and v1. In the minimization procedure, the u0 is updated iteratively along the gradient of J w.r.t. u0, where the gradient is obtained by transporting J backwards in time from t = 1 to t = 0, using the adjoint NSE. The AGM thus effectively replaces the backward integration by multiple forward and backward adjoint integrations. Since the viscosity is negative in the adjoint NSE, each step of the AGM is numerically stable. Nevertheless, when applied to turbulence, the AGM develops instabilities, which limit the backward integration to small times. This is due to the exponential divergence of phase space trajectories in turbulent flow, which produces a multitude of local minima in J, when the integration time is large. As an effect, the AGM may select unphysical, noisy initial conditions. In order to improve this situation, we propose two remedies. First, we replace the integration by a sequence of smaller integrations, i.e., we divide the integration time into segments, where in each segment the target field v1 is taken as the initial field u0 from the previous segment. Second, we add an additional term (regularizer) to J, which is proportional to a high-order Laplacian of u0, and which dampens the gradients of u0. We show that suitable values for the segment size and for the regularizer, allow a stable reverse integration of 2D decaying turbulence, with accurate results for more then O(10) turbulent, integral time scales.

Keywords: time reversed integration, parabolic differential equations, adjoint gradient method, two dimensional turbulence

Procedia PDF Downloads 224
1308 A Numerical Studies for Improving the Performance of Vertical Axis Wind Turbine by a Wind Power Tower

Authors: Soo-Yong Cho, Chong-Hyun Cho, Chae-Whan Rim, Sang-Kyu Choi, Jin-Gyun Kim, Ju-Seok Nam

Abstract:

Recently, vertical axis wind turbines (VAWT) have been widely used to produce electricity even in urban. They have several merits such as low sound noise, easy installation of the generator and simple structure without yaw-control mechanism and so on. However, their blades are operated under the influence of the trailing vortices generated by the preceding blades. This phenomenon deteriorates its output power and makes difficulty predicting correctly its performance. In order to improve the performance of VAWT, wind power towers can be applied. Usually, the wind power tower can be constructed as a multi-story building to increase the frontal area of the wind stream. Hence, multiple sets of the VAWT can be installed within the wind power tower, and they can be operated at high elevation. Many different types of wind power tower can be used in the field. In this study, a wind power tower with circular column shape was applied, and the VAWT was installed at the center of the wind power tower. Seven guide walls were used as a strut between the floors of the wind power tower. These guide walls were utilized not only to increase the wind velocity within the wind power tower but also to adjust the wind direction for making a better working condition on the VAWT. Hence, some important design variables, such as the distance between the wind turbine and the guide wall, the outer diameter of the wind power tower, the direction of the guide wall against the wind direction, should be considered to enhance the output power on the VAWT. A numerical analysis was conducted to find the optimum dimension on design variables by using the computational fluid dynamics (CFD) among many prediction methods. The CFD could be an accurate prediction method compared with the stream-tube methods. In order to obtain the accurate results in the CFD, it needs the transient analysis and the full three-dimensional (3-D) computation. However, this full 3-D CFD could be hard to be a practical tool because it requires huge computation time. Therefore, the reduced computational domain is applied as a practical method. In this study, the computations were conducted in the reduced computational domain and they were compared with the experimental results in the literature. It was examined the mechanism of the difference between the experimental results and the computational results. The computed results showed this computational method could be an effective method in the design methodology using the optimization algorithm. After validation of the numerical method, the CFD on the wind power tower was conducted with the important design variables affecting the performance of VAWT. The results showed that the output power of the VAWT obtained using the wind power tower was increased compared to them obtained without the wind power tower. In addition, they showed that the increased output power on the wind turbine depended greatly on the dimension of the guide wall.

Keywords: CFD, performance, VAWT, wind power tower

Procedia PDF Downloads 387
1307 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling

Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra

Abstract:

Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.

Keywords: multi-temporal satellite image, urban growth, non-stationary, stochastic model

Procedia PDF Downloads 428
1306 Changing the Way South Africa Think about Parking Provision at Tertiary Institutions

Authors: M. C. Venter, G. Hitge, S. C. Krygsman, J. Thiart

Abstract:

For decades, South Africa has been planning transportation systems from a supply, rather than a demand side, perspective. In terms of parking, this relates to requiring the minimum parking provision that is enforced by city officials. Newer insight is starting to indicate that South Africa needs to re-think this philosophy in light of a new policy environment that desires a different outcome. Urban policies have shifted from reliance on the private car for access, to employing a wide range of alternative modes. Car dominated travel is influenced by various parameters, of which the availability and location of parking plays a significant role. The question is therefore, what is the right strategy to achieve the desired transport outcomes for SA. The focus of this paper is used to assess this issue with regard to parking provision, and specifically at a tertiary institution. A parking audit was conducted at the Stellenbosch campus of Stellenbosch University, monitoring occupancy at all 60 parking areas, every hour during business hours over a five-day period. The data from this survey was compared with the prescribed number of parking bays according to the Stellenbosch Municipality zoning scheme (requiring a minimum of 0.4 bays per student). The analysis shows that by providing 0.09 bays per student, the maximum total daily occupation of all the parking areas did not exceed an 80% occupation rate. It is concluded that the prevailing parking standards are not supportive of the new urban and transport policy environment, but that it is extremely conservative from a practical demand point of view.

Keywords: parking provision, parking requirements, travel behaviour, travel demand management

Procedia PDF Downloads 181
1305 Estimation of Soil Moisture at High Resolution through Integration of Optical and Microwave Remote Sensing and Applications in Drought Analyses

Authors: Donglian Sun, Yu Li, Paul Houser, Xiwu Zhan

Abstract:

California experienced severe drought conditions in the past years. In this study, the drought conditions in California are analyzed using soil moisture anomalies derived from integrated optical and microwave satellite observations along with auxiliary land surface data. Based on the U.S. Drought Monitor (USDM) classifications, three typical drought conditions were selected for the analysis: extreme drought conditions in 2007 and 2013, severe drought conditions in 2004 and 2009, and normal conditions in 2005 and 2006. Drought is defined as negative soil moisture anomaly. To estimate soil moisture at high spatial resolutions, three approaches are explored in this study: the universal triangle model that estimates soil moisture from Normalized Difference Vegetation Index (NDVI) and Land Surface Temperature (LST); the basic model that estimates soil moisture under different conditions with auxiliary data like precipitation, soil texture, topography, and surface types; and the refined model that uses accumulated precipitation and its lagging effects. It is found that the basic model shows better agreements with the USDM classifications than the universal triangle model, while the refined model using precipitation accumulated from the previous summer to current time demonstrated the closest agreements with the USDM patterns.

Keywords: soil moisture, high resolution, regional drought, analysis and monitoring

Procedia PDF Downloads 136