Search results for: classification method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20062

Search results for: classification method

16912 Comparative Analysis of Enzyme Activities Concerned in Decomposition of Toluene

Authors: Ayuko Itsuki, Sachiyo Aburatani

Abstract:

In recent years, pollutions of the environment by toxic substances become a serious problem. While there are many methods of environmental clean-up, the methods by microorganisms are considered to be reasonable and safety for environment. Compost is known that it catabolize the meladorous substancess in its production process, however the mechanism of its catabolizing system is not known yet. In the catabolization process, organic matters turn into inorganic by the released enzymes from lots of microorganisms which live in compost. In other words, the cooperative of activated enzymes in the compost decomposes malodorous substances. Thus, clarifying the interaction among enzymes is important for revealing the catabolizing system of meladorous substance in compost. In this study, we utilized statistical method to infer the interaction among enzymes. We developed a method which combined partial correlation with cross correlation to estimate the relevance between enzymes especially from time series data of few variables. Because of using cross correlation, we can estimate not only the associative structure but also the reaction pathway. We applied the developed method to the enzyme measured data and estimated an interaction among the enzymes in decomposition mechanism of toluene.

Keywords: enzyme activities, comparative analysis, compost, toluene

Procedia PDF Downloads 255
16911 Dimensionality Reduction in Modal Analysis for Structural Health Monitoring

Authors: Elia Favarelli, Enrico Testi, Andrea Giorgetti

Abstract:

Autonomous structural health monitoring (SHM) of many structures and bridges became a topic of paramount importance for maintenance purposes and safety reasons. This paper proposes a set of machine learning (ML) tools to perform automatic feature selection and detection of anomalies in a bridge from vibrational data and compare different feature extraction schemes to increase the accuracy and reduce the amount of data collected. As a case study, the Z-24 bridge is considered because of the extensive database of accelerometric data in both standard and damaged conditions. The proposed framework starts from the first four fundamental frequencies extracted through operational modal analysis (OMA) and clustering, followed by density-based time-domain filtering (tracking). The fundamental frequencies extracted are then fed to a dimensionality reduction block implemented through two different approaches: feature selection (intelligent multiplexer) that tries to estimate the most reliable frequencies based on the evaluation of some statistical features (i.e., mean value, variance, kurtosis), and feature extraction (auto-associative neural network (ANN)) that combine the fundamental frequencies to extract new damage sensitive features in a low dimensional feature space. Finally, one class classifier (OCC) algorithms perform anomaly detection, trained with standard condition points, and tested with normal and anomaly ones. In particular, a new anomaly detector strategy is proposed, namely one class classifier neural network two (OCCNN2), which exploit the classification capability of standard classifiers in an anomaly detection problem, finding the standard class (the boundary of the features space in normal operating conditions) through a two-step approach: coarse and fine boundary estimation. The coarse estimation uses classics OCC techniques, while the fine estimation is performed through a feedforward neural network (NN) trained that exploits the boundaries estimated in the coarse step. The detection algorithms vare then compared with known methods based on principal component analysis (PCA), kernel principal component analysis (KPCA), and auto-associative neural network (ANN). In many cases, the proposed solution increases the performance with respect to the standard OCC algorithms in terms of F1 score and accuracy. In particular, by evaluating the correct features, the anomaly can be detected with accuracy and an F1 score greater than 96% with the proposed method.

Keywords: anomaly detection, frequencies selection, modal analysis, neural network, sensor network, structural health monitoring, vibration measurement

Procedia PDF Downloads 109
16910 Optimized Simultaneous Determination of Theobromine and Caffeine in Fermented and Unfermented Cacao Beans and in Cocoa Products Using Step Gradient Solvent System in Reverse Phase HPLC

Authors: Ian Marc G. Cabugsa, Kim Ryan A. Won

Abstract:

Fast, reliable and simultaneous HPLC analysis of theobromine and caffeine in cacao and cocoa products was optimized in this study. The samples tested were raw, fermented, and roasted cacao beans as well as commercially available cocoa products. The HPLC analysis was carried out using step gradient solvent system with acetonitrile and water buffered with H3PO4 as the mobile phase. The HPLC system was optimized using 273 nm wavelength at 35 °C for the column temperature with a flow rate of 1.0 mL/min. Using this method, the theobromine percent recovery mean, Limit of Detection (LOD) and Limit of Quantification (LOQ) is 118.68(±3.38)%, 0.727 and 1.05 respectively. The percent recovery mean, LOD and LOQ for caffeine is 105.53(±3.25)%, 2.42 and 3.50 respectively. The inter-day and intra-day precision for theobromine is 4.31% and 4.48% respectively, while 7.02% and 7.03% was for caffeine respectively. Compared to the standard method in AOAC using methanol in isocratic solvent system, the results of the study produced lesser chromatogram noise with emphasis on theobromine and caffeine. The method is readily usable for cacao and cocoa substances analyses using HPLC with step gradient capability.

Keywords: cacao, caffeine, HPLC, step gradient solvent system, theobromine

Procedia PDF Downloads 266
16909 Edge Detection in Low Contrast Images

Authors: Koushlendra Kumar Singh, Manish Kumar Bajpai, Rajesh K. Pandey

Abstract:

The edges of low contrast images are not clearly distinguishable to the human eye. It is difficult to find the edges and boundaries in it. The present work encompasses a new approach for low contrast images. The Chebyshev polynomial based fractional order filter has been used for filtering operation on an image. The preprocessing has been performed by this filter on the input image. Laplacian of Gaussian method has been applied on preprocessed image for edge detection. The algorithm has been tested on two test images.

Keywords: low contrast image, fractional order differentiator, Laplacian of Gaussian (LoG) method, chebyshev polynomial

Procedia PDF Downloads 613
16908 Pharmacovigilance in Hospitals: Retrospective Study at the Pharmacovigilance Service of UHE-Oran, Algeria

Authors: Nadjet Mekaouche, Hanane Zitouni, Fatma Boudia, Habiba Fetati, A. Saleh, A. Lardjam, H. Geniaux, A. Coubret, H. Toumi

Abstract:

Medicines have undeniably played a major role in prolonging shelf life and improving quality. The absolute efficacy of the drug remains a lever for innovation, its benefit/risk balance is not always assured and it does not always have the expected effects. Prior to marketing, knowledge about adverse drug reactions is incomplete. Once on the market, phase IV drug studies begin. For years, the drug was prescribed with less care to a large number of very heterogeneous patients and often in combination with other drugs. It is at this point that previously unknown adverse effects may appear, hence the need for the implementation of a pharmacovigilance system. Pharmacovigilance represents all methods for detecting, evaluating, informing and preventing the risks of adverse drug reactions. The most severe adverse events occur frequently in hospital and that a significant proportion of adverse events result in hospitalizations. In addition, the consequences of hospital adverse events in terms of length of stay, mortality and costs are considerable. It, therefore, appears necessary to develop ‘hospital pharmacovigilance’ aimed at reducing the incidence of adverse reactions in hospitals. The most widely used monitoring method in pharmacovigilance is spontaneous notification. However, underreporting of adverse drug reactions is common in many countries and is a major obstacle to pharmacovigilance assessment. It is in this context that this study aims to describe the experience of the pharmacovigilance service at the University Hospital of Oran (EHUO). This is a retrospective study extending from 2011 to 2017, carried out on archived records of declarations collected at the level of the EHUO Pharmacovigilance Department. Reporting was collected by two methods: ‘spontaneous notification’ and ‘active pharmacovigilance’ targeting certain clinical services. We counted 217 statements. It involved 56% female patients and 46% male patients. Age ranged from 5 to 78 years with an average of 46 years. The most common adverse reaction was drug toxidermy. For the drugs in question, they were essentially according to the ATC classification of anti-infectives followed by anticancer drugs. As regards the evolution of declarations by year, a low rate of notification was noted in 2011. That is why we decided to set up an active approach at the level of some services where a resident of reference attended the staffs every week. This has resulted in an increase in the number of reports. The declarations came essentially from the services where the active approach was installed. This highlights the need for ongoing communication between all relevant health actors to stimulate reporting and secure drug treatments.

Keywords: adverse drug reactions, hospital, pharmacovigilance, spontaneous notification

Procedia PDF Downloads 152
16907 Adoption of Digital Storytelling Tool to Teach 21st Century Skills by Malaysian Pre-service Teachers

Authors: Siti Aisyah binti Jumpaan

Abstract:

21ˢᵗ century skills (PAK-21) integration has made its way into Malaysian curriculum when Ministry of Education introduce its implementation since 2016. This study was conducted to explore pre-service teachers’ readiness in integrating 21st century skills in the classroom via the digital storytelling (DST) method and to find gaps between theory and practice that can be integral towards pre-service teachers’ professional growth. Qualitative research method was used in this research involving six respondents who were selected using a purposive sampling method. Their response from interviews and lesson plan analysis were analysed using narrative analysis. The findings showed that pre-service teachers showed a moderate level of readiness in integrating 21st century skills using DST. Pre-service teachers demonstrated high level of preparedness in writing their lesson plan, but their interview revealed that they faced struggles in implementation due to several factors, such as lack of technology and failure to obtain students’ participation. This study further strengthens the need for specialised curriculum for pre-service teachers in teaching 21st century skills via DST.

Keywords: digital storytelling, 21ˢᵗ century skills, preservice teachers, teacher training

Procedia PDF Downloads 76
16906 A Novel Method for Isolation of Kaempferol and Quercetin from Podophyllum Hexandrum Rhizome

Authors: S. B. Bhandare, K. S. Laddha

Abstract:

Podphyllum hexandrum belonging to family berberidaceae has gained attention in phytochemical and pharmacological research as it shows excellent anticancer activity and has been used in treatment of skin diseases, sunburns and radioprotection. Chemically it contains lignans and flavonoids such as kaempferol, quercetin and their glycosides. Objective: To isolate and identify Kaempferol and Quercetin from Podophyllum rhizome. Method: The powdered rhizome of Podophyllum hexandrum was subjected to soxhlet extraction with methanol. This methanolic extract is used to obtain podophyllin. Podohyllin was extracted with ethyl acetate and this extract was then concentrated and subjected to column chromatography to obtain purified kaempferol and quercetin. Result: Isolated kaempferol, quercetin were light yellow and dark yellow in colour respectively. TLC of the isolated compounds was performed using chloroform: methanol (9:1) which showed single band on silica plate at Rf 0.6 and 0.4 for kaempferol and quercetin. UV spectrometric studies showed UV maxima (methanol) at 259, 360 nm and 260, 370 nm which are identical with standard kaempferol and quercetin respectively. Both IR spectra exhibited prominent absorption bands for free phenolic OH at 3277 and 3296.2 cm-1 and for conjugated C=O at 1597 and 1659.7 cm-1 respectively. The mass spectrum of kaempferol and quercetin showed (M+1) peak at m/z 287 and 303.09 respectively. 1H NMR analysis of both isolated compounds exhibited typical four-peak pattern of two doublets at δ 6.86 and δ 8.01 which was assigned to H-3’,5’ and H-2’,6’ respectively. Absence of signals less than δ 6.81 in the 1H NMR spectrum supported the aromatic nature of compound. Kaempferol and Quercetin showed 98.1% and 97% purity by HPLC at UV 370 nm. Conclusion: Easy and simple method for isolation of Kaempferol and Quercetin was developed and their structures were confirmed by UV, IR, NMR and mass studies. Method has shown good reproducibility, yield and purity.

Keywords: flavonoids, kaempferol, podophyllum rhizome, quercetin

Procedia PDF Downloads 291
16905 Hydrodynamic Study and Sizing of a Distillation Column by HYSYS Software

Authors: Derrouazin Mohammed Redhouane, Souakri Mohammed Lotfi, Henini Ghania

Abstract:

This work consists, first of all, of mastering one of the powerful process simulation tools currently used in the industrial processes, which is the HYSYS sizing software, and second, of simulating a petroleum distillation column. This study is divided into two parts; where the first one consists of a dimensioning of the column with a fast approximating method using state equations, iterative calculations, and then a precise simulation method with the HYSYS software. The second part of this study is a hydrodynamic study in order to verify by obtained results the proper functioning of the plates.

Keywords: industry process engineering, water distillation, environment, HYSYS simulation tool

Procedia PDF Downloads 110
16904 Prediction of Anticancer Potential of Curcumin Nanoparticles by Means of Quasi-Qsar Analysis Using Monte Carlo Method

Authors: Ruchika Goyal, Ashwani Kumar, Sandeep Jain

Abstract:

The experimental data for anticancer potential of curcumin nanoparticles was calculated by means of eclectic data. The optimal descriptors were examined using Monte Carlo method based CORAL SEA software. The statistical quality of the model is following: n = 14, R² = 0.6809, Q² = 0.5943, s = 0.175, MAE = 0.114, F = 26 (sub-training set), n =5, R²= 0.9529, Q² = 0.7982, s = 0.086, MAE = 0.068, F = 61, Av Rm² = 0.7601, ∆R²m = 0.0840, k = 0.9856 and kk = 1.0146 (test set) and n = 5, R² = 0.6075 (validation set). This data can be used to build predictive QSAR models for anticancer activity.

Keywords: anticancer potential, curcumin, model, nanoparticles, optimal descriptors, QSAR

Procedia PDF Downloads 304
16903 The Effect of Accounting Conservatism on Cost of Capital: A Quantile Regression Approach for MENA Countries

Authors: Maha Zouaoui Khalifa, Hakim Ben Othman, Hussaney Khaled

Abstract:

Prior empirical studies have investigated the economic consequences of accounting conservatism by examining its impact on the cost of equity capital (COEC). However, findings are not conclusive. We assume that inconsistent results of such association may be attributed to the regression models used in data analysis. To address this issue, we re-examine the effect of different dimension of accounting conservatism: unconditional conservatism (U_CONS) and conditional conservatism (C_CONS) on the COEC for a sample of listed firms from Middle Eastern and North Africa (MENA) countries, applying quantile regression (QR) approach developed by Koenker and Basset (1978). While classical ordinary least square (OLS) method is widely used in empirical accounting research, however it may produce inefficient and bias estimates in the case of departures from normality or long tail error distribution. QR method is more powerful than OLS to handle this kind of problem. It allows the coefficient on the independent variables to shift across the distribution of the dependent variable whereas OLS method only estimates the conditional mean effects of a response variable. We find as predicted that U_CONS has a significant positive effect on the COEC however, C_CONS has a negative impact. Findings suggest also that the effect of the two dimensions of accounting conservatism differs considerably across COEC quantiles. Comparing results from QR method with those of OLS, this study throws more lights on the association between accounting conservatism and COEC.

Keywords: unconditional conservatism, conditional conservatism, cost of equity capital, OLS, quantile regression, emerging markets, MENA countries

Procedia PDF Downloads 343
16902 Analysis of Patient No-Shows According to Health Conditions

Authors: Sangbok Lee

Abstract:

There has been much effort on process improvement for outpatient clinics to provide quality and acute care to patients. One of the efforts is no-show analysis or prediction. This work analyzes patient no-shows along with patient health conditions. The health conditions refer to clinical symptoms that each patient has, out of the followings; hyperlipidemia, diabetes, metastatic solid tumor, dementia, chronic obstructive pulmonary disease, hypertension, coronary artery disease, myocardial infraction, congestive heart failure, atrial fibrillation, stroke, drug dependence abuse, schizophrenia, major depression, and pain. A dataset from a regional hospital is used to find the relationship between the number of the symptoms and no-show probabilities. Additional analysis reveals how each symptom or combination of symptoms affects no-shows. In the above analyses, cross-classification of patients by age and gender is carried out. The findings from the analysis will be used to take extra care to patients with particular health conditions. They will be forced to visit clinics by being informed about their health conditions and possible consequences more clearly. Moreover, this work will be used in the preparation of making institutional guidelines for patient reminder systems.

Keywords: healthcare system, no show analysis, process improvment, statistical data analysis

Procedia PDF Downloads 218
16901 Reliability Analysis of Heat Exchanger Cycle Using Non-Parametric Method

Authors: Apurv Kulkarni, Shreyas Badave, B. Rajiv

Abstract:

Non-parametric reliability technique is useful for assessment of reliability of systems for which failure rates are not available. This is useful when detection of malfunctioning of any component is the key purpose during ongoing operation of the system. The main purpose of the Heat Exchanger Cycle discussed in this paper is to provide hot water at a constant temperature for longer periods of time. In such a cycle, certain components play a crucial role and this paper presents an effective way to predict the malfunctioning of the components by determination of system reliability. The method discussed in the paper is feasible and this is clarified with the help of various test cases.

Keywords: heat exchanger cycle, k-statistics, PID controller, system reliability

Procedia PDF Downloads 370
16900 Modeling Dynamics and Control of Transversal Vibration of an Underactuated Flexible Plate Using Controlled Lagrangian Method

Authors: Mahmood Khalghollah, Mohammad Tavallaeinejad, Mohammad Eghtesad

Abstract:

The method of Controlled Lagrangian is an energy shaping control technique for under actuated Lagrangian systems. Energy shaping control design methods are appealing as they retain the underlying nonlinear dynamics and can provide stability results that hold over larger domain than can be obtained using linear design and analysis. In the present study, controlled lagrangian is employed for designing a controller in an under actuated rotating flexible plate system. In the system of rotating flexible plate, due to its nonlinear characteristics and coupled dynamics of rigid and flexible components, controller design is a known challenge. In this paper, controller objectives are considered to be vibration reduction of flexible component and position control of the tip of the plate. To achieve the goals, a method based on both kinetic and potential energy shaping is introduced. The stability of the closed-loop system is investigated and proved around its equilibrium points. Moreover, the proposed controller is shown to be robust against disturbance and plant uncertainties.

Keywords: controlled lagrangian, underactuated system, flexible rotating plate, disturbance

Procedia PDF Downloads 434
16899 Integrated Best Worst PROMETHEE to Evaluate Public Transport Service Quality

Authors: Laila Oubahman, Duleba Szabolcs

Abstract:

Public transport stakeholders aim to increase the ridership ratio by encouraging citizens to use common transportation modes. For this sight, improving service quality is a crucial option to reach the quality desired by users and reduce the gap between desired and perceived quality. Multi-criteria decision aid has been applied in literature in recent decades because it provides efficient models to assess the most impacting criteria on the overall assessment. In this paper, the PROMETHEE method is combined with the best-worst approach to construct a consensual model that avoids rank reversal to support stakeholders in ameliorating service quality.

Keywords: best-worst method, MCDA, PROMETHEE, public transport

Procedia PDF Downloads 182
16898 An Enhanced Approach in Validating Analytical Methods Using Tolerance-Based Design of Experiments (DoE)

Authors: Gule Teri

Abstract:

The effective validation of analytical methods forms a crucial component of pharmaceutical manufacturing. However, traditional validation techniques can occasionally fail to fully account for inherent variations within datasets, which may result in inconsistent outcomes. This deficiency in validation accuracy is particularly noticeable when quantifying low concentrations of active pharmaceutical ingredients (APIs), excipients, or impurities, introducing a risk to the reliability of the results and, subsequently, the safety and effectiveness of the pharmaceutical products. In response to this challenge, we introduce an enhanced, tolerance-based Design of Experiments (DoE) approach for the validation of analytical methods. This approach distinctly measures variability with reference to tolerance or design margins, enhancing the precision and trustworthiness of the results. This method provides a systematic, statistically grounded validation technique that improves the truthfulness of results. It offers an essential tool for industry professionals aiming to guarantee the accuracy of their measurements, particularly for low-concentration components. By incorporating this innovative method, pharmaceutical manufacturers can substantially advance their validation processes, subsequently improving the overall quality and safety of their products. This paper delves deeper into the development, application, and advantages of this tolerance-based DoE approach and demonstrates its effectiveness using High-Performance Liquid Chromatography (HPLC) data for verification. This paper also discusses the potential implications and future applications of this method in enhancing pharmaceutical manufacturing practices and outcomes.

Keywords: tolerance-based design, design of experiments, analytical method validation, quality control, biopharmaceutical manufacturing

Procedia PDF Downloads 57
16897 Review of Research on Waste Plastic Modified Asphalt

Authors: Song Xinze, Cai Kejian

Abstract:

To further explore the application of waste plastics in asphalt pavement, this paper begins with the classification and characteristics of waste plastics. It then provides a state-of-the-art review of the preparation methods and processes of waste plastic modifiers, waste plastic-modified asphalt, and waste plastic-modified asphalt mixtures. The paper also analyzes the factors influencing the compatibility between waste plastics and asphalt and summarizes the performance evaluation indicators for waste plastic-modified asphalt and its mixtures. It explores the research approaches and findings of domestic and international scholars and presents examples of waste plastics applications in pavement engineering. The author believes that there is a basic consensus that waste plastics can improve the high-temperature performance of asphalt. The use of cracking processes to solve the storage stability of waste plastic polymer-modified asphalt is the key to promoting its application. Additionally, the author anticipates that future research will concentrate on optimizing the recycling, processing, screening, and preparation of waste plastics, along with developing composite plastic modifiers to improve their compatibility and long-term performance in asphalt pavements.

Keywords: waste plastics, asphalt pavement, asphalt performance, asphalt modification

Procedia PDF Downloads 19
16896 Combination of Topology and Rough Set for Analysis of Power System Control

Authors: M. Kamel El-Sayed

Abstract:

In this research, we have linked the concept of rough set and topological structure to the creation of a new topological structure that assists in the analysis of the information systems of some electrical engineering issues. We used non-specific information whose boundaries do not have an empty set in the top topological structure is rough set. It is characterized by the fact that it does not contain a large number of elements and facilitates the establishment of rules. We used this structure in reducing the specifications of electrical information systems. We have provided a detailed example of this method illustrating the steps used. This method opens the door to obtaining multiple topologies, each of which uses one of the non-defined groups (rough set) in the overall information system.

Keywords: electrical engineering, information system, rough set, rough topology, topology

Procedia PDF Downloads 442
16895 The Effects of Xiang Sha Liu Jun Zi Tang to Diarrhea and Growth Performance of Piglets

Authors: Siao-Wei Jiang, Boy-Young Hsieh, Ching-Liang Hsieh, Cheng-Yung Lin

Abstract:

The problems of multiple drug resistance in the pig farming industry have been emphasized in recent years. Diarrhea syndrome is common in weaning piglets and often treated with antibiotics as a feed additive, leading to the rapid spread of antibiotic resistance and posing high health risks to humans. The study aimed to alleviate diarrhea syndrome with traditional herbal medicine, Xiang Sha Liu Jun Zi Tang, whose effects enhanced digestive function. Piglets at 4 weeks old with stool classified to Bristol stool classification type 6 or type 7 were randomly divided into the control group, group A (1% of Xiang Sha Liu Jun Zi Tang) and group B (0.1% Colistin). The piglets were administrated for 7 days, and their weight, feed intake, and stool score were recorded daily before and after the trial. The results showed that the diarrhea index score in group A and group B improved significantly compared to the control group, indicating that Xiang Sha Liu Jun Zi Tang may have the same effect on alleviating diarrhea syndrome as Colistin, and it may be another replacement for antibiotics.

Keywords: pig, diarrhea, herbal medicine, Xiang Sha Liu Jun Zi Tang

Procedia PDF Downloads 35
16894 Iterative Solver for Solving Large-Scale Frictional Contact Problems

Authors: Thierno Diop, Michel Fortin, Jean Deteix

Abstract:

Since the precise formulation of the elastic part is irrelevant for the description of the algorithm, we shall consider a generic case. In practice, however, we will have to deal with a non linear material (for instance a Mooney-Rivlin model). We are interested in solving a finite element approximation of the problem, leading to large-scale non linear discrete problems and, after linearization, to large linear systems and ultimately to calculations needing iterative methods. This also implies that penalty method, and therefore augmented Lagrangian method, are to be banned because of their negative effect on the condition number of the underlying discrete systems and thus on the convergence of iterative methods. This is in rupture to the mainstream of methods for contact in which augmented Lagrangian is the principal tool. We shall first present the problem and its discretization; this will lead us to describe a general solution algorithm relying on a preconditioner for saddle-point problems which we shall describe in some detail as it is not entirely standard. We will propose an iterative approach for solving three-dimensional frictional contact problems between elastic bodies, including contact with a rigid body, contact between two or more bodies and also self-contact.

Keywords: frictional contact, three-dimensional, large-scale, iterative method

Procedia PDF Downloads 190
16893 Classify Land Use/Cover Change and Its Impact on Soil Erosion Using GIS from 2005 to 2015 in Nzhelele Valley Limpopo Province, South Africa

Authors: Blessing Mavhuru, Nthaduleni Nethengwe, Hector Chikoore, Onyango Beneah Daniel Odhiambo

Abstract:

The main objective of this study was to classify land use/cover and how it has changed in Nzhelele Valley Limpopo Province, South Africa. The study aimed to identify and analyse the types of land use/cover in the years 2005, 2010, and 2015 with a view to assess the impact on soil erosion over time. Using GIS, the changes within land use/cover were assessed through the classification of satellite images. The study area was classified into four major land cover/use classes, which are vegetation, gravel road, built up land and agricultural fields. Over the period 2005-2015 the resultant land use/cover demonstrated (i) a significant increase (12%) for vegetation cover, (ii) a significant decrease in agriculture (16%) land use/cover, (iii) increase in built-up land (1%), as well as (iv) an increase in gravel roads (3%). This study envisages assisting policy makers in decision making on land use management for Nzhelele Valley.

Keywords: land use, land cover, change, soil erosion

Procedia PDF Downloads 237
16892 Ge₁₋ₓSnₓ Alloys with Tuneable Energy Band Gap on GaAs (100) Substrate Manufactured by a Modified Magnetron Co-Sputtering

Authors: Li Qian, Jinchao Tong, Daohua Zhang, Weijun Fan, Fei Suo

Abstract:

Photonic applications based on group IV semiconductors have always been an interest but also a challenge for the research community. We report manufacturing group IV Ge₁₋ₓSnₓ alloys with tuneable energy band gap on (100) GaAs substrate by a modified radio frequency magnetron co-sputtering. Images were taken by atomic force microscope, and scanning electron microscope clearly demonstrates a smooth surface profile, and Ge₁₋ₓSnₓ nano clusters are with the size of several tens of nanometers. Transmittance spectra were measured by Fourier Transform Infrared Spectroscopy that showed changing energy gaps with the variation in elementary composition. Calculation results by 8-band k.p method are consistent with measured gaps. Our deposition system realized direct growth of Ge₁₋ₓSnₓ thin film on GaAs (100) substrate by sputtering. This simple deposition method was modified to be able to grow high-quality photonic materials with tuneable energy gaps. This work provides an alternative and successful method for fabricating Group IV photonic semiconductor materials.

Keywords: GeSn, crystal growth, sputtering, photonic

Procedia PDF Downloads 133
16891 Sentinel-2 Based Burn Area Severity Assessment Tool in Google Earth Engine

Authors: D. Madhushanka, Y. Liu, H. C. Fernando

Abstract:

Fires are one of the foremost factors of land surface disturbance in diverse ecosystems, causing soil erosion and land-cover changes and atmospheric effects affecting people's lives and properties. Generally, the severity of the fire is calculated as the Normalized Burn Ratio (NBR) index. This is performed manually by comparing two images obtained afterward. Then by using the bitemporal difference of the preprocessed satellite images, the dNBR is calculated. The burnt area is then classified as either unburnt (dNBR<0.1) or burnt (dNBR>= 0.1). Furthermore, Wildfire Severity Assessment (WSA) classifies burnt areas and unburnt areas using classification levels proposed by USGS and comprises seven classes. This procedure generates a burn severity report for the area chosen by the user manually. This study is carried out with the objective of producing an automated tool for the above-mentioned process, namely the World Wildfire Severity Assessment Tool (WWSAT). It is implemented in Google Earth Engine (GEE), which is a free cloud-computing platform for satellite data processing, with several data catalogs at different resolutions (notably Landsat, Sentinel-2, and MODIS) and planetary-scale analysis capabilities. Sentinel-2 MSI is chosen to obtain regular processes related to burnt area severity mapping using a medium spatial resolution sensor (15m). This tool uses machine learning classification techniques to identify burnt areas using NBR and to classify their severity over the user-selected extent and period automatically. Cloud coverage is one of the biggest concerns when fire severity mapping is performed. In WWSAT based on GEE, we present a fully automatic workflow to aggregate cloud-free Sentinel-2 images for both pre-fire and post-fire image compositing. The parallel processing capabilities and preloaded geospatial datasets of GEE facilitated the production of this tool. This tool consists of a Graphical User Interface (GUI) to make it user-friendly. The advantage of this tool is the ability to obtain burn area severity over a large extent and more extended temporal periods. Two case studies were carried out to demonstrate the performance of this tool. The Blue Mountain national park forest affected by the Australian fire season between 2019 and 2020 is used to describe the workflow of the WWSAT. This site detected more than 7809 km2, using Sentinel-2 data, giving an error below 6.5% when compared with the area detected on the field. Furthermore, 86.77% of the detected area was recognized as fully burnt out, of which high severity (17.29%), moderate-high severity (19.63%), moderate-low severity (22.35%), and low severity (27.51%). The Arapaho and Roosevelt National Forest Park, California, the USA, which is affected by the Cameron peak fire in 2020, is chosen for the second case study. It was found that around 983 km2 had burned out, of which high severity (2.73%), moderate-high severity (1.57%), moderate-low severity (1.18%), and low severity (5.45%). These spots also can be detected through the visual inspection made possible by cloud-free images generated by WWSAT. This tool is cost-effective in calculating the burnt area since satellite images are free and the cost of field surveys is avoided.

Keywords: burnt area, burnt severity, fires, google earth engine (GEE), sentinel-2

Procedia PDF Downloads 211
16890 Ray Tracing Modified 3D Image Method Simulation of Picocellular Propagation Channel Environment

Authors: Fathi Alwafie

Abstract:

In this paper we present the simulation of the propagation characteristics of the picocellular propagation channel environment. The first aim has been to find a correct description of the environment for received wave. The result of the first investigations is that the environment of the indoor wave significantly changes as we change the electric parameters of material constructions. A modified 3D ray tracing image method tool has been utilized for the coverage prediction. A detailed analysis of the dependence of the indoor wave on the wide-band characteristics of the channel: Root Mean Square (RMS) delay spread characteristics and mean excess delay, is also investigated.

Keywords: propagation, ray tracing, network, mobile computing

Procedia PDF Downloads 387
16889 OCR/ICR Text Recognition Using ABBYY FineReader as an Example Text

Authors: A. R. Bagirzade, A. Sh. Najafova, S. M. Yessirkepova, E. S. Albert

Abstract:

This article describes a text recognition method based on Optical Character Recognition (OCR). The features of the OCR method were examined using the ABBYY FineReader program. It describes automatic text recognition in images. OCR is necessary because optical input devices can only transmit raster graphics as a result. Text recognition describes the task of recognizing letters shown as such, to identify and assign them an assigned numerical value in accordance with the usual text encoding (ASCII, Unicode). The peculiarity of this study conducted by the authors using the example of the ABBYY FineReader, was confirmed and shown in practice, the improvement of digital text recognition platforms developed by Electronic Publication.

Keywords: ABBYY FineReader system, algorithm symbol recognition, OCR/ICR techniques, recognition technologies

Procedia PDF Downloads 153
16888 Fuzzy Multi-Criteria Decision-Making Based on Ignatian Discernment Process

Authors: Pathinathan Theresanathan, Ajay Minj

Abstract:

Ignatian Discernment Process (IDP) is an intense decision-making tool to decide on life-issues. Decisions are influenced by various factors outside of the decision maker and inclination within. This paper develops IDP in the context of Fuzzy Multi-criteria Decision Making (FMCDM) process. Extended VIKOR method is a decision-making method which encompasses even conflict situations and accommodates weightage to various issues. Various aspects of IDP, namely three ways of decision making and tactics of inner desires, are observed, analyzed and articulated within the frame work of fuzzy rules. The decision-making situations are broadly categorized into two types. The issues outside of the decision maker influence the person. The inner feeling also plays vital role in coming to a conclusion. IDP integrates both the categories using Extended VIKOR method. Case studies are carried out and analyzed with FMCDM process. Finally, IDP is verified with an illustrative case study and results are interpreted. A confused person who could not come to a conclusion is able to take decision on a concrete way of life through IDP. The proposed IDP model recommends an integrated and committed approach to value-based decision making.

Keywords: AHP, FMCDM, IDP, ignatian discernment, MCDM, VIKOR

Procedia PDF Downloads 240
16887 Segmentation Using Multi-Thresholded Sobel Images: Application to the Separation of Stuck Pollen Grains

Authors: Endrick Barnacin, Jean-Luc Henry, Jimmy Nagau, Jack Molinie

Abstract:

Being able to identify biological particles such as spores, viruses, or pollens is important for health care professionals, as it allows for appropriate therapeutic management of patients. Optical microscopy is a technology widely used for the analysis of these types of microorganisms, because, compared to other types of microscopy, it is not expensive. The analysis of an optical microscope slide is a tedious and time-consuming task when done manually. However, using machine learning and computer vision, this process can be automated. The first step of an automated microscope slide image analysis process is segmentation. During this step, the biological particles are localized and extracted. Very often, the use of an automatic thresholding method is sufficient to locate and extract the particles. However, in some cases, the particles are not extracted individually because they are stuck to other biological elements. In this paper, we propose a stuck particles separation method based on the use of the Sobel operator and thresholding. We illustrate it by applying it to the separation of 813 images of adjacent pollen grains. The method correctly separated 95.4% of these images.

Keywords: image segmentation, stuck particles separation, Sobel operator, thresholding

Procedia PDF Downloads 118
16886 Study on Seismic Assessment of Earthquake-Damaged Reinforced Concrete Buildings

Authors: Fu-Pei Hsiao, Fung-Chung Tu, Chien-Kuo Chiu

Abstract:

In this work, to develop a method for detailed assesses of post-earthquake seismic performance for RC buildings in Taiwan, experimental data for several column specimens with various failure modes (flexural failure, flexural-shear failure, and shear failure) are used to derive reduction factors of seismic capacity for specified damage states. According to the damage states of RC columns and their corresponding seismic reduction factors suggested by experimental data, this work applies the detailed seismic performance assessment method to identify the seismic capacity of earthquake-damaged RC buildings. Additionally, a post-earthquake emergent assessment procedure is proposed that can provide the data needed for decision about earthquake-damaged buildings in a region with high seismic hazard. Finally, three actual earthquake-damaged school buildings in Taiwan are used as a case study to demonstrate application of the proposed assessment method.

Keywords: seismic assessment, seismic reduction factor, residual seismic ratio, post-earthquake, reinforced concrete, building

Procedia PDF Downloads 388
16885 Application of Lean Manufacturing Tools in Hot Asphalt Production

Authors: S. Bayona, J. Nunez, D. Paez, C. Diaz

Abstract:

The application of Lean manufacturing tools continues to be an effective solution for increasing productivity, reducing costs and eliminating waste in the manufacture of goods and services. This article analyzes the production process of a hot asphalt manufacturing company from an administrative and technical perspective. Three main phases were analyzed, the first phase was related to the determination of the risk priority number of the main operations in asphalt mix production process by an FMEA (Failure Mode Effects Analysis), in the second phase the Value Stream Mapping (VSM) of the production line was performed and in the third phase a SWOT (Strengths, Weaknesses Opportunities, Threats) matrix was constructed. Among the most valued failure modes were the lack training of workers in occupational safety and health issues, the lack of signaling and classification of granulated material, and the overweight of vehicles loaded. The analysis of the results in the three phases agree on the importance of training operational workers, improve communication with external actors in order to minimize delays in material orders and strengthen control suppliers.

Keywords: asphalt, lean manufacturing, productivity, process

Procedia PDF Downloads 103
16884 Quantification of Hydrogen Sulfide and Methyl Mercaptan in Air Samples from a Waste Management Facilities

Authors: R. F. Vieira, S. A. Figueiredo, O. M. Freitas, V. F. Domingues, C. Delerue-Matos

Abstract:

The presence of sulphur compounds like hydrogen sulphide and mercaptans is one of the reasons for waste-water treatment and waste management being associated with odour emissions. In this context having a quantifying method for these compounds helps in the optimization of treatment with the goal of their elimination, namely biofiltration processes. The aim of this study was the development of a method for quantification of odorous gases in waste treatment plants air samples. A method based on head space solid phase microextraction (HS-SPME) coupled with gas chromatography - flame photometric detector (GC-FPD) was used to analyse H2S and Metil Mercaptan (MM). The extraction was carried out with a 75-μm Carboxen-polydimethylsiloxane fiber coating at 22 ºC for 20 min, and analysed by a GC 2010 Plus A from Shimadzu with a sulphur filter detector: splitless mode (0.3 min), the column temperature program was from 60 ºC, increased by 15 ºC/min to 100 ºC (2 min). The injector temperature was held at 250 ºC, and the detector at 260 ºC. For calibration curve a gas diluter equipment (digital Hovagas G2 - Multi Component Gas Mixer) was used to do the standards. This unit had two input connections, one for a stream of the dilute gas and another for a stream of nitrogen and an output connected to a glass bulb. A 40 ppm H2S and a 50 ppm MM cylinders were used. The equipment was programmed to the selected concentration, and it automatically carried out the dilution to the glass bulb. The mixture was left flowing through the glass bulb for 5 min and then the extremities were closed. This method allowed the calibration between 1-20 ppm for H2S and 0.02-0.1 ppm and 1-3.5 ppm for MM. Several quantifications of air samples from inlet and outlet of a biofilter operating in a waste management facility in the north of Portugal allowed the evaluation the biofilters performance.

Keywords: biofiltration, hydrogen sulphide, mercaptans, quantification

Procedia PDF Downloads 459
16883 Behavior of Beam-Column Nodes Reinforced Concrete in Earthquake Zones

Authors: Zaidour Mohamed, Ghalem Ali Jr., Achit Henni Mohamed

Abstract:

This project is destined to study pole junctions of reinforced concrete beams subjected to seismic loads. A literature review was made to clarify the work done by researchers in the last three decades and especially the results of the last two years that were studied for the determination of the method of calculating the transverse reinforcement in the different nodes of a structure. For implementation efforts in the columns and beams of a building R + 4 in zone 3 were calculated using the finite element method through software. These results are the basis of our work which led to the calculation of the transverse reinforcement of the nodes of the structure in question.

Keywords: beam–column joints, cyclic loading, shearing force, damaged joint

Procedia PDF Downloads 532