Search results for: accuracy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3525

Search results for: accuracy

1065 Easymodel: Web-based Bioinformatics Software for Protein Modeling Based on Modeller

Authors: Alireza Dantism

Abstract:

Presently, describing the function of a protein sequence is one of the most common problems in biology. Usually, this problem can be facilitated by studying the three-dimensional structure of proteins. In the absence of a protein structure, comparative modeling often provides a useful three-dimensional model of the protein that is dependent on at least one known protein structure. Comparative modeling predicts the three-dimensional structure of a given protein sequence (target) mainly based on its alignment with one or more proteins of known structure (templates). Comparative modeling consists of four main steps 1. Similarity between the target sequence and at least one known template structure 2. Alignment of target sequence and template(s) 3. Build a model based on alignment with the selected template(s). 4. Prediction of model errors 5. Optimization of the built model There are many computer programs and web servers that automate the comparative modeling process. One of the most important advantages of these servers is that it makes comparative modeling available to both experts and non-experts, and they can easily do their own modeling without the need for programming knowledge, but some other experts prefer using programming knowledge and do their modeling manually because by doing this they can maximize the accuracy of their modeling. In this study, a web-based tool has been designed to predict the tertiary structure of proteins using PHP and Python programming languages. This tool is called EasyModel. EasyModel can receive, according to the user's inputs, the desired unknown sequence (which we know as the target) in this study, the protein sequence file (template), etc., which also has a percentage of similarity with the primary sequence, and its third structure Predict the unknown sequence and present the results in the form of graphs and constructed protein files.

Keywords: structural bioinformatics, protein tertiary structure prediction, modeling, comparative modeling, modeller

Procedia PDF Downloads 59
1064 Using Predictive Analytics to Identify First-Year Engineering Students at Risk of Failing

Authors: Beng Yew Low, Cher Liang Cha, Cheng Yong Teoh

Abstract:

Due to a lack of continual assessment or grade related data, identifying first-year engineering students in a polytechnic education at risk of failing is challenging. Our experience over the years tells us that there is no strong correlation between having good entry grades in Mathematics and the Sciences and excelling in hardcore engineering subjects. Hence, identifying students at risk of failure cannot be on the basis of entry grades in Mathematics and the Sciences alone. These factors compound the difficulty of early identification and intervention. This paper describes the development of a predictive analytics model in the early detection of students at risk of failing and evaluates its effectiveness. Data from continual assessments conducted in term one, supplemented by data of student psychological profiles such as interests and study habits, were used. Three classification techniques, namely Logistic Regression, K Nearest Neighbour, and Random Forest, were used in our predictive model. Based on our findings, Random Forest was determined to be the strongest predictor with an Area Under the Curve (AUC) value of 0.994. Correspondingly, the Accuracy, Precision, Recall, and F-Score were also highest among these three classifiers. Using this Random Forest Classification technique, students at risk of failure could be identified at the end of term one. They could then be assigned to a Learning Support Programme at the beginning of term two. This paper gathers the results of our findings. It also proposes further improvements that can be made to the model.

Keywords: continual assessment, predictive analytics, random forest, student psychological profile

Procedia PDF Downloads 98
1063 Intercultural Trainings for Future Global Managers: Evaluating the Effect on the Global Mind-Set

Authors: Nina Dziatzko, Christopher Stehr, Franziska Struve

Abstract:

Intercultural competence as an explicit required skill nearly never appears in job advertisements in international or even global contexts. But especially those who have to deal with different nationalities and cultures in their everyday business need to have several intercultural competencies and further a global mind-set. This way the question arises how potential future global managers can be trained to learn these competencies. In this regard, it might be helpful to see if different types of intercultural trainings have different effects on those skills. This paper outlines lessons learned based on the evaluation of two different intercultural trainings for management students. The main differences between the observed intercultural trainings are the amount of theoretical input in relation to hands-on experiences, the number of trainers as well as the used methods to teach implicit cultural rules. Both groups contain management students with the willingness and perspective to work abroad or to work in international context. The research is carried out with a pre-training-survey and a post-training-survey which consists of questions referring the international context of the students and a self-estimation of 19 identified intercultural and global mind-set skills, such as: cosmopolitanism, empathy, differentiation and adaptability. Whereas there is no clear result which training gets overall a significant higher increase of skills, there is a clear difference between the focus of competencies trained by each of the intercultural trainings. This way this research provides a guideline for both academicals institutions as well as companies for the decision between different types of intercultural trainings, if the to be trained required skills are defined. Therefore the efficiency and the accuracy of fit of the education of future global managers get optimized.

Keywords: global mind-set, intercultural competencies, intercultural training, learning experiences

Procedia PDF Downloads 247
1062 Acoustic Modeling of a Data Center with a Hot Aisle Containment System

Authors: Arshad Alfoqaha, Seth Bard, Dustin Demetriou

Abstract:

A new multi-physics acoustic modeling approach using ANSYS Mechanical FEA and FLUENT CFD methods is developed for modeling servers mounted to racks, such as IBM Z and IBM Power Systems, in data centers. This new approach allows users to determine the thermal and acoustic conditions that people are exposed to within the data center. The sound pressure level (SPL) exposure for a human working inside a hot aisle containment system inside the data center is studied. The SPL is analyzed at the noise source, at the human body, on the rack walls, on the containment walls, and on the ceiling and flooring plenum walls. In the acoustic CFD simulation, it is assumed that a four-inch diameter sphere with monopole acoustic radiation, placed in the middle of each rack, provides a single-source representation of all noise sources within the rack. Ffowcs Williams & Hawkings (FWH) acoustic model is employed. The target frequency is 1000 Hz, and the total simulation time for the transient analysis is 1.4 seconds, with a very small time step of 3e-5 seconds and 10 iterations to ensure convergence and accuracy. A User Defined Function (UDF) is developed to accurately simulate the acoustic noise source, and a Dynamic Mesh is applied to ensure acoustic wave propagation. Initial validation of the acoustic CFD simulation using a closed-form solution for the spherical propagation of an acoustic point source is performed.

Keywords: data centers, FLUENT, acoustics, sound pressure level, SPL, hot aisle containment, IBM

Procedia PDF Downloads 145
1061 Image Processing of Scanning Electron Microscope Micrograph of Ferrite and Pearlite Steel for Recognition of Micro-Constituents

Authors: Subir Gupta, Subhas Ganguly

Abstract:

In this paper, we demonstrate the new area of application of image processing in metallurgical images to develop the more opportunity for structure-property correlation based approaches of alloy design. The present exercise focuses on the development of image processing tools suitable for phrase segmentation, grain boundary detection and recognition of micro-constituents in SEM micrographs of ferrite and pearlite steels. A comprehensive data of micrographs have been experimentally developed encompassing the variation of ferrite and pearlite volume fractions and taking images at different magnification (500X, 1000X, 15000X, 2000X, 3000X and 5000X) under scanning electron microscope. The variation in the volume fraction has been achieved using four different plain carbon steel containing 0.1, 0.22, 0.35 and 0.48 wt% C heat treated under annealing and normalizing treatments. The obtained data pool of micrographs arbitrarily divided into two parts to developing training and testing sets of micrographs. The statistical recognition features for ferrite and pearlite constituents have been developed by learning from training set of micrographs. The obtained features for microstructure pattern recognition are applied to test set of micrographs. The analysis of the result shows that the developed strategy can successfully detect the micro constitutes across the wide range of magnification and variation of volume fractions of the constituents in the structure with an accuracy of about +/- 5%.

Keywords: SEM micrograph, metallurgical image processing, ferrite pearlite steel, microstructure

Procedia PDF Downloads 171
1060 Estimation of PM10 Concentration Using Ground Measurements and Landsat 8 OLI Satellite Image

Authors: Salah Abdul Hameed Saleh, Ghada Hasan

Abstract:

The aim of this work is to produce an empirical model for the determination of particulate matter (PM10) concentration in the atmosphere using visible bands of Landsat 8 OLI satellite image over Kirkuk city- IRAQ. The suggested algorithm is established on the aerosol optical reflectance model. The reflectance model is a function of the optical properties of the atmosphere, which can be related to its concentrations. The concentration of PM10 measurements was collected using Particle Mass Profiler and Counter in a Single Handheld Unit (Aerocet 531) meter simultaneously by the Landsat 8 OLI satellite image date. The PM10 measurement locations were defined by a handheld global positioning system (GPS). The obtained reflectance values for visible bands (Coastal aerosol, Blue, Green and blue bands) of landsat 8 OLI image were correlated with in-suite measured PM10. The feasibility of the proposed algorithms was investigated based on the correlation coefficient (R) and root-mean-square error (RMSE) compared with the PM10 ground measurement data. A choice of our proposed multispectral model was founded on the highest value correlation coefficient (R) and lowest value of the root mean square error (RMSE) with PM10 ground data. The outcomes of this research showed that visible bands of Landsat 8 OLI were capable of calculating PM10 concentration with an acceptable level of accuracy.

Keywords: air pollution, PM10 concentration, Lansat8 OLI image, reflectance, multispectral algorithms, Kirkuk area

Procedia PDF Downloads 420
1059 Analytical Development of a Failure Limit and Iso-Uplift Curves for Eccentrically Loaded Shallow Foundations

Authors: N. Abbas, S. Lagomarsino, S. Cattari

Abstract:

Examining existing experimental results for shallow rigid foundations subjected to vertical centric load (N), accompanied or not with a bending moment (M), two main non-linear mechanisms governing the cyclic ‎response of the soil-foundation system can be distinguished: foundation uplift and soil yielding. A soil-foundation failure limit, is defined as a domain of resistance in the two dimensional (2D) load space (N, M) inside of which lie all the admissible combinations of loads; these latter correspond to a pure elastic, non-linear elastic or plastic behavior of the soil-foundation system, while the points lying on the failure limit correspond to a combination of loads leading to a failure of the soil-foundation system. In this study, the proposed resistance domain is constructed analytically based on mechanics. Original elastic limit, uplift initiation ‎limit and iso-uplift limits are constructed inside this domain. These limits give a prediction ‎of the mechanisms activated for each combination of loads applied to the ‎foundation. A comparison of the proposed failure limit with experimental tests existing in the literature shows interesting results. Also, the developed uplift initiation limit and iso-uplift curves are confronted with others already proposed in the literature and widely used due to the absence of other alternatives, and remarkable differences are noted, showing evident errors in the past proposals and relevant accuracy for those given in the present work.

Keywords: foundation uplift, iso-uplift curves, resistance domain, soil yield

Procedia PDF Downloads 356
1058 Analysis of Veterinary Drug Residues and Pesticide Residues in Beehive Products

Authors: Alba Luna Jimenez, Maria Dolores Hernando

Abstract:

The administration of veterinary treatments at higher doses than the recommended Varroa mite control in beehive matrices has the potential to generate residues in the honeybee colony and in the derived products for consumption. Honeybee colonies can also be indirectly exposed to residues of plant protection products when foraging in crops, wildflowers near the crops, or in urban gardens just after spraying. The study evaluates the presence of both types of residues, veterinary treatments, and pesticides in beeswax, bee bread, and honey. The study was carried out in apiaries located in agricultural zones and forest areas in Andalusia, Spain. Up to nineteen residues were identified above LOQ using gas chromatography-triple quadrupole-mass spectrometry analysis (GC-MS/MS). Samples were extracted by a modified QuEChERs method. Chlorfenvinphos was detected in beeswax and bee bread despite its use is not authorized for Varroa mite control. Residues of fluvalinate-tau, authorized as veterinary treatment, were detected in most of the samples of beeswax and bee bread, presumably due to overdose or also to its potential for accumulation associated with its marked liposolubility. Residues of plant protection products were also detected in samples of beeswax and bee bread. Pesticide residues were detected above the LOQ that was established at 5 µg.kg⁻¹, which is the minimum concentration that can be quantified with acceptable accuracy and precision, as described in the European guidelines for pesticide residue analysis SANTE/11945/2015. No residues of phytosanitary treatments used in agriculture were detected in honey.

Keywords: honeybee colony, mass spectrometry analysis, pesticide residues, Varroa destructor, veterinary treatment

Procedia PDF Downloads 126
1057 Comparative Mesh Sensitivity Study of Different Reynolds Averaged Navier Stokes Turbulence Models in OpenFOAM

Authors: Zhuoneng Li, Zeeshan A. Rana, Karl W. Jenkins

Abstract:

In industry, to validate a case, often a multitude of simulation are required and in order to demonstrate confidence in the process where users tend to use a coarser mesh. Therefore, it is imperative to establish the coarsest mesh that could be used while keeping reasonable simulation accuracy. To date, the two most reliable, affordable and broadly used advanced simulations are the hybrid RANS (Reynolds Averaged Navier Stokes)/LES (Large Eddy Simulation) and wall modelled LES. The potentials in these two simulations will still be developed in the next decades mainly because the unaffordable computational cost of a DNS (Direct Numerical Simulation). In the wall modelled LES, the turbulence model is applied as a sub-grid scale model in the most inner layer near the wall. The RANS turbulence models cover the entire boundary layer region in a hybrid RANS/LES (Detached Eddy Simulation) and its variants, therefore, the RANS still has a very important role in the state of art simulations. This research focuses on the turbulence model mesh sensitivity analysis where various turbulence models such as the S-A (Spalart-Allmaras), SSG (Speziale-Sarkar-Gatski), K-Omega transitional SST (Shear Stress Transport), K-kl-Omega, γ-Reθ transitional model, v2f are evaluated within the OpenFOAM. The simulations are conducted on a fully developed turbulent flow over a flat plate where the skin friction coefficient as well as velocity profiles are obtained to compare against experimental values and DNS results. A concrete conclusion is made to clarify the mesh sensitivity for different turbulence models.

Keywords: mesh sensitivity, turbulence models, OpenFOAM, RANS

Procedia PDF Downloads 229
1056 Multi-Temporal Urban Land Cover Mapping Using Spectral Indices

Authors: Mst Ilme Faridatul, Bo Wu

Abstract:

Multi-temporal urban land cover mapping is of paramount importance for monitoring urban sprawl and managing the ecological environment. For diversified urban activities, it is challenging to map land covers in a complex urban environment. Spectral indices have proved to be effective for mapping urban land covers. To improve multi-temporal urban land cover classification and mapping, we evaluate the performance of three spectral indices, e.g. modified normalized difference bare-land index (MNDBI), tasseled cap water and vegetation index (TCWVI) and shadow index (ShDI). The MNDBI is developed to evaluate its performance of enhancing urban impervious areas by separating bare lands. A tasseled cap index, TCWVI is developed to evaluate its competence to detect vegetation and water simultaneously. The ShDI is developed to maximize the spectral difference between shadows of skyscrapers and water and enhance water detection. First, this paper presents a comparative analysis of three spectral indices using Landsat Enhanced Thematic Mapper (ETM), Thematic Mapper (TM) and Operational Land Imager (OLI) data. Second, optimized thresholds of the spectral indices are imputed to classify land covers, and finally, their performance of enhancing multi-temporal urban land cover mapping is assessed. The results indicate that the spectral indices are competent to enhance multi-temporal urban land cover mapping and achieves an overall classification accuracy of 93-96%.

Keywords: land cover, mapping, multi-temporal, spectral indices

Procedia PDF Downloads 117
1055 Challenges and Insights by Electrical Characterization of Large Area Graphene Layers

Authors: Marcus Klein, Martina GrießBach, Richard Kupke

Abstract:

The current advances in the research and manufacturing of large area graphene layers are promising towards the introduction of this exciting material in the display industry and other applications that benefit from excellent electrical and optical characteristics. New production technologies in the fabrication of flexible displays, touch screens or printed electronics apply graphene layers on non-metal substrates and bring new challenges to the required metrology. Traditional measurement concepts of layer thickness, sheet resistance, and layer uniformity, are difficult to apply to graphene production processes and are often harmful to the product layer. New non-contact sensor concepts are required to adapt to the challenges and even the foreseeable inline production of large area graphene. Dedicated non-contact measurement sensors are a pioneering method to leverage these issues in a large variety of applications, while significantly lowering the costs of development and process setup. Transferred and printed graphene layers can be characterized with high accuracy in a huge measurement range using a very high resolution. Large area graphene mappings are applied for process optimization and for efficient quality control for transfer, doping, annealing and stacking processes. Examples of doped, defected and excellent Graphene are presented as quality images and implications for manufacturers are explained.

Keywords: graphene, doping and defect testing, non-contact sheet resistance measurement, inline metrology

Procedia PDF Downloads 278
1054 Prevalence of Anxiety and Depression: A Descriptive Cross-Sectional Study among Individuals with Substance-Related Disorders in Argentina

Authors: Badino Manuel, Farias María Alejandra

Abstract:

Anxiety and depression are considered the main mental health issues found in people with substance-related disorders. Furthermore, substance-related disorders, anxiety-related and depressive disorders are among the leading causes of disability and are associated with increased mortality. The co-occurrence of substance-related disorders and these mental health conditions affect the accuracy in diagnosis, treatment plan, and recovery process. The aim is to describe the prevalence of anxiety and depression in patients with substance-related disorders in a mental health service in Córdoba, Argentina. A descriptive cross-sectional study was conducted among patients with substance-related disorders (N=305). Anxiety and depression were assessed using the Patient Health Questionnaire-4 (PHQ-4) during the period from December 2021 to March 2022. For a total of 305 participants, 71,8% were male, 25,6% female and 2,6% non-binary. As regards marital status, 51,5% were single, 21,6% as a couple, 5,9% married, 15,4% separated and 5,6% divorced. In relation to education status, 26,2% finished university, 56,1% high school, 16,4% only primary school and 1,3% no formal schooling. Regarding age, 10,8% were young, 84,3% were adults, and 4,9% were elderly. In-person treatment represented 64,6% of service users, and 35,4% were conducted through teleconsultation. 15,7% of service users scored 3 or higher for anxiety, and 32,1% scored 3 or higher for depression in the PHQ-4. 13,1% obtained a score of 3 or higher for both anxiety and depression. It is recommended to identify anxiety and depression among patients with substance-related disorders to improve the quality of diagnosis, treatment, and recovery. It is suggested to apply PHQ-4, PHQ-9 within the protocol of care for these patients.

Keywords: addiction, anxiety, depression, mental health

Procedia PDF Downloads 69
1053 In and Out-Of-Sample Performance of Non Simmetric Models in International Price Differential Forecasting in a Commodity Country Framework

Authors: Nicola Rubino

Abstract:

This paper presents an analysis of a group of commodity exporting countries' nominal exchange rate movements in relationship to the US dollar. Using a series of Unrestricted Self-exciting Threshold Autoregressive models (SETAR), we model and evaluate sixteen national CPI price differentials relative to the US dollar CPI. Out-of-sample forecast accuracy is evaluated through calculation of mean absolute error measures on the basis of two-hundred and fifty-three months rolling window forecasts and extended to three additional models, namely a logistic smooth transition regression (LSTAR), an additive non linear autoregressive model (AAR) and a simple linear Neural Network model (NNET). Our preliminary results confirm presence of some form of TAR non linearity in the majority of the countries analyzed, with a relatively higher goodness of fit, with respect to the linear AR(1) benchmark, in five countries out of sixteen considered. Although no model appears to statistically prevail over the other, our final out-of-sample forecast exercise shows that SETAR models tend to have quite poor relative forecasting performance, especially when compared to alternative non-linear specifications. Finally, by analyzing the implied half-lives of the > coefficients, our results confirms the presence, in the spirit of arbitrage band adjustment, of band convergence with an inner unit root behaviour in five of the sixteen countries analyzed.

Keywords: transition regression model, real exchange rate, nonlinearities, price differentials, PPP, commodity points

Procedia PDF Downloads 254
1052 A Comparative Analysis of the Indoor Thermal Environment of a Room with and without Transitional Space or Threshold in Traditional Row Houses Adjacent to a Narrow Alley 'Rupchan Lane' in Old Dhaka, Bangladesh

Authors: Fatema Tasmia, Brishti Majumder, Atiqur Rahman

Abstract:

Attaining appropriate thermal comfort conditions in a place where the climate is hot and humid can be perplexing. Especially, when it resides at a congested place like old Dhaka Bangladesh, the provision of giving cross ventilation and building with proper orientation is quite difficult. This paper aims to investigate the indoor thermal environment of a room with and without transitional space or threshold in traditional row houses adjacent to a narrow alley of old Dhaka through field measurements. Transitional spaces are the part of buildings which are used for semi-outdoor household activities, social gathering and it is also proved to provide an indoor thermal effect. The field study was conducted by collecting thermal data (temperature, humidity and airflow) respectively, among the outdoor narrow alley, transitional space and adjacent indoor. This east-west elongated alley has an average width of 2.13 meter (varies from 1.5 to 2.6 meter) holding row houses on both sides. Among different aspects of thermal environment, the study of this paper is based on the analysis of temperature of corresponding cases. Other aspects and their variables were considered as constant (especially material) for accuracy and avoidance of confusion. This study focuses on the outcome that can ultimately contribute to the configuration of row houses with transitional spaces and in its relation to the adjacent outdoor space while achieving thermal comfort.

Keywords: alley, Old-Dhaka, row houses, temperature, thermal comfort, threshold, transitional space

Procedia PDF Downloads 152
1051 Valence and Arousal-Based Sentiment Analysis: A Comparative Study

Authors: Usama Shahid, Muhammad Zunnurain Hussain

Abstract:

This research paper presents a comprehensive analysis of a sentiment analysis approach that employs valence and arousal as its foundational pillars, in comparison to traditional techniques. Sentiment analysis is an indispensable task in natural language processing that involves the extraction of opinions and emotions from textual data. The valence and arousal dimensions, representing the intensity and positivity/negativity of emotions, respectively, enable the creation of four quadrants, each representing a specific emotional state. The study seeks to determine the impact of utilizing these quadrants to identify distinct emotional states on the accuracy and efficiency of sentiment analysis, in comparison to traditional techniques. The results reveal that the valence and arousal-based approach outperforms other approaches, particularly in identifying nuanced emotions that may be missed by conventional methods. The study's findings are crucial for applications such as social media monitoring and market research, where the accurate classification of emotions and opinions is paramount. Overall, this research highlights the potential of using valence and arousal as a framework for sentiment analysis and offers invaluable insights into the benefits of incorporating specific types of emotions into the analysis. These findings have significant implications for researchers and practitioners in the field of natural language processing, as they provide a basis for the development of more accurate and effective sentiment analysis tools.

Keywords: sentiment analysis, valence and arousal, emotional states, natural language processing, machine learning, text analysis, sentiment classification, opinion mining

Procedia PDF Downloads 53
1050 Comprehensive Feature Extraction for Optimized Condition Assessment of Fuel Pumps

Authors: Ugochukwu Ejike Akpudo, Jank-Wook Hur

Abstract:

The increasing demand for improved productivity, maintainability, and reliability has prompted rapidly increasing research studies on the emerging condition-based maintenance concept- Prognostics and health management (PHM). Varieties of fuel pumps serve critical functions in several hydraulic systems; hence, their failure can have daunting effects on productivity, safety, etc. The need for condition monitoring and assessment of these pumps cannot be overemphasized, and this has led to the uproar in research studies on standard feature extraction techniques for optimized condition assessment of fuel pumps. By extracting time-based, frequency-based and the more robust time-frequency based features from these vibrational signals, a more comprehensive feature assessment (and selection) can be achieved for a more accurate and reliable condition assessment of these pumps. With the aid of emerging deep classification and regression algorithms like the locally linear embedding (LLE), we propose a method for comprehensive condition assessment of electromagnetic fuel pumps (EMFPs). Results show that the LLE as a comprehensive feature extraction technique yields better feature fusion/dimensionality reduction results for condition assessment of EMFPs against the use of single features. Also, unlike other feature fusion techniques, its capabilities as a fault classification technique were explored, and the results show an acceptable accuracy level using standard performance metrics for evaluation.

Keywords: electromagnetic fuel pumps, comprehensive feature extraction, condition assessment, locally linear embedding, feature fusion

Procedia PDF Downloads 89
1049 Revisiting the Historical Narratives of the Old Churches in Albay, Bikol Region, Philippines

Authors: Ruby Ann L. Ayo

Abstract:

As cultural heritage reflects the historical origin of a certain group of people, it reveals their customs, traits, beliefs, practices and even values they hold on for years. One of the tangible examples of cultural heritage is the physical structures including the old churches. The study looked-into the existing historical narratives of the century Old Catholic churches in the Province of Albay, Bikol Region, Philippines: NuestraSeñora de Salvacion in Joroan, Tiwi, Albay; the Our Lady of the Gate in Daraga, Albay; the San Juan de Bautista in Tabaco City and the St. John the Baptist in Camalig, Albay. The historical narratives were analysed in terms of validity and reliability of the secondary documents with reference to the elements of history revealing consistency and adequacy of historical facts. The contents were examined using a modified Checklist of Historical Documents. The historical narratives were likewise submitted to the content expert for validation as regards historical authenticity and accuracy. The contents of the narratives were scrutinized according to the following codes: (1.1) the Patron Saints;(1.2) factors that paved to their constructions; (1.3) the people responsible for their constructions; (1.4) the misconceptions about their constructions; and (1.5) their contributions to Bikol heritage. Based on the codes, themes were identified as: (2.1) Marian Devotees and Christ-centered Patron Saints; (2.2) geographical, socio-political and cultural factors; (2.3) church and government officials; (2.4) misconceptions on the dates of constructions and original sites; and (2.5) popular pilgrim sites and well-admired architectural designs.

Keywords: historical narratives, old churches, cultural heritage, historical validity and reliability, elements of history

Procedia PDF Downloads 264
1048 A Boundary-Fitted Nested Grid Model for Modeling Tsunami Propagation of 2004 Indonesian Tsunami along Southern Thailand

Authors: Fazlul Karim, Esa Al-Islam

Abstract:

Many problems in oceanography and environmental sciences require the solution of shallow water equations on physical domains having curvilinear coastlines and abrupt changes of ocean depth near the shore. Finite-difference technique for the shallow water equations representing the boundary as stair step may give inaccurate results near the coastline where results are of greatest interest for various applications. This suggests the use of methods which are capable of incorporating the irregular boundary in coastal belts. At the same time, large velocity gradient is expected near the beach and islands as water depth vary abruptly near the coast. A nested numerical scheme with fine resolution is the best resort to enhance the numerical accuracy with the least grid numbers for the region of interests where the velocity changes rapidly and which is unnecessary for the away of the region. This paper describes the development of a boundary fitted nested grid (BFNG) model to compute tsunami propagation of 2004 Indonesian tsunami in Southern Thailand coastal waters. In this paper, we develop a numerical model employing the shallow water nested model and an orthogonal boundary fitted grid to investigate the tsunami impact on the Southern Thailand due to the Indonesian tsunami of 2004. Comparisons of water surface elevation obtained from numerical simulations and field measurements are made.

Keywords: Indonesian tsunami of 2004, Boundary-fitted nested grid model, Southern Thailand, finite difference method

Procedia PDF Downloads 404
1047 Participation in IAEA Proficiency Test to Analyse Cobalt, Strontium and Caesium in Seawater Using Direct Counting and Radiochemical Techniques

Authors: S. Visetpotjanakit, C. Khrautongkieo

Abstract:

Radiation monitoring in the environment and foodstuffs is one of the main responsibilities of Office of Atoms for Peace (OAP) as the nuclear regulatory body of Thailand. The main goal of the OAP is to assure the safety of the Thai people and environment from any radiological incidents. Various radioanalytical methods have been developed to monitor radiation and radionuclides in the environmental and foodstuff samples. To validate our analytical performance, several proficiency test exercises from the International Atomic Energy Agency (IAEA) have been performed. Here, the results of a proficiency test exercise referred to as the Proficiency Test for Tritium, Cobalt, Strontium and Caesium Isotopes in Seawater 2017 (IAEA-RML-2017-01) are presented. All radionuclides excepting ³H were analysed using various radioanalytical methods, i.e. direct gamma-ray counting for determining ⁶⁰Co, ¹³⁴Cs and ¹³⁷Cs and developed radiochemical techniques for analysing ¹³⁴Cs, ¹³⁷Cs using AMP pre-concentration technique and 90Sr using di-(2-ethylhexyl) phosphoric acid (HDEHP) liquid extraction technique. The analysis results were submitted to IAEA. All results passed IAEA criteria, i.e. accuracy, precision and trueness and obtained ‘Accepted’ statuses. These confirm the data quality from the OAP environmental radiation laboratory to monitor radiation in the environment.

Keywords: international atomic energy agency, proficiency test, radiation monitoring, seawater

Procedia PDF Downloads 148
1046 Enhancing Temporal Extrapolation of Wind Speed Using a Hybrid Technique: A Case Study in West Coast of Denmark

Authors: B. Elshafei, X. Mao

Abstract:

The demand for renewable energy is significantly increasing, major investments are being supplied to the wind power generation industry as a leading source of clean energy. The wind energy sector is entirely dependable and driven by the prediction of wind speed, which by the nature of wind is very stochastic and widely random. This s0tudy employs deep multi-fidelity Gaussian process regression, used to predict wind speeds for medium term time horizons. Data of the RUNE experiment in the west coast of Denmark were provided by the Technical University of Denmark, which represent the wind speed across the study area from the period between December 2015 and March 2016. The study aims to investigate the effect of pre-processing the data by denoising the signal using empirical wavelet transform (EWT) and engaging the vector components of wind speed to increase the number of input data layers for data fusion using deep multi-fidelity Gaussian process regression (GPR). The outcomes were compared using root mean square error (RMSE) and the results demonstrated a significant increase in the accuracy of predictions which demonstrated that using vector components of the wind speed as additional predictors exhibits more accurate predictions than strategies that ignore them, reflecting the importance of the inclusion of all sub data and pre-processing signals for wind speed forecasting models.

Keywords: data fusion, Gaussian process regression, signal denoise, temporal extrapolation

Procedia PDF Downloads 108
1045 On the Development of Medical Additive Manufacturing in Egypt

Authors: Khalid Abdelghany

Abstract:

Additive Manufacturing (AM) is the manufacturing technology that is used to fabricate fast products direct from CAD models in very short time and with minimum operation steps. Jointly with the advancement in medical computer modeling, AM proved to be a very efficient tool to help physicians, orthopedic surgeons and dentists design and fabricate patient-tailored surgical guides, templates and customized implants from the patient’s CT / MRI images. AM jointly with computer-assisted designing/computer-assisted manufacturing (CAD/CAM) technology have enabled medical practitioners to tailor physical models in a patient-and purpose-specific fashion and helped to design and manufacture of templates, appliances and devices with a high range of accuracy using biocompatible materials. In developing countries, there are some technical and financial limitations of implementing such advanced tools as an essential portion of medical applications. CMRDI institute in Egypt has been working in the field of Medical Additive Manufacturing since 2003 and has assisted in the recovery of hundreds of poor patients using these advanced tools. This paper focuses on the surgical and dental use of 3D printing technology in Egypt as a developing country. The presented case studies have been designed and processed using the software tools and additive manufacturing machines in CMRDI through cooperative engineering and medical works. Results showed that the implementation of the additive manufacturing tools in developed countries is successful and could be economical comparing to long treatment plans.

Keywords: additive manufacturing, dental and orthopeadic stents, patient specific surgical tools, titanium implants

Procedia PDF Downloads 282
1044 Research on the United Navigation Mechanism of Land, Sea and Air Targets under Multi-Sources Information Fusion

Authors: Rui Liu, Klaus Greve

Abstract:

The navigation information is a kind of dynamic geographic information, and the navigation information system is a kind of special geographic information system. At present, there are many researches on the application of centralized management and cross-integration application of basic geographic information. However, the idea of information integration and sharing is not deeply applied into the research of navigation information service. And the imperfection of navigation target coordination and navigation information sharing mechanism under certain navigation tasks has greatly affected the reliability and scientificity of navigation service such as path planning. Considering this, the project intends to study the multi-source information fusion and multi-objective united navigation information interaction mechanism: first of all, investigate the actual needs of navigation users in different areas, and establish the preliminary navigation information classification and importance level model; and then analyze the characteristics of the remote sensing and GIS vector data, and design the fusion algorithm from the aspect of improving the positioning accuracy and extracting the navigation environment data. At last, the project intends to analyze the feature of navigation information of the land, sea and air navigation targets, and design the united navigation data standard and navigation information sharing model under certain navigation tasks, and establish a test navigation system for united navigation simulation experiment. The aim of this study is to explore the theory of united navigation service and optimize the navigation information service model, which will lay the theory and technology foundation for the united navigation of land, sea and air targets.

Keywords: information fusion, united navigation, dynamic path planning, navigation information visualization

Procedia PDF Downloads 251
1043 The Application and Relevance of Costing Techniques in Service-Oriented Business Organizations a Review of the Activity-Based Costing (ABC) Technique

Authors: Udeh Nneka Evelyn

Abstract:

The shortcoming of traditional costing system in terms of validity, accuracy, consistency, and Relevance increased the need for modern management accounting system. Activity –Based Costing (ABC) can be used as a modern tool for planning, Control and decision making for management. Past studies on ABC system have focused on manufacturing firms thereby making the studies on service firms scanty to some extent. This paper reviewed the application and relevance of activity-based costing technique in service oriented business organizations by employing a qualitative research method which relied heavily on literature review of past and current relevant articles focusing on ABC. Findings suggest that ABC is not only appropriate for use in a manufacturing environment; it is also most appropriate for service organizations such as financial institutions, the healthcare industry and government organization. In fact, some banking and financial institutions have been applying the concept for years under other names. One of them is unit costing, which is used to calculate the cost of banking services by determining the cost and consumption of each unit of output of functions required to deliver the service. ABC in very basic terms may provide very good payback for businesses. Some of the benefits that relate directly to the financial services industry are: identification the most profitable customers: more accurate product and service pricing: increase product profitability: Well organized process costs.

Keywords: business, costing, organizations, planning, techniques

Procedia PDF Downloads 217
1042 Parallel Self Organizing Neural Network Based Estimation of Archie’s Parameters and Water Saturation in Sandstone Reservoir

Authors: G. M. Hamada, A. A. Al-Gathe, A. M. Al-Khudafi

Abstract:

Determination of water saturation in sandstone is a vital question to determine the initial oil or gas in place in reservoir rocks. Water saturation determination using electrical measurements is mainly on Archie’s formula. Consequently accuracy of Archie’s formula parameters affects water saturation values rigorously. Determination of Archie’s parameters a, m, and n is proceeded by three conventional techniques, Core Archie-Parameter Estimation (CAPE) and 3-D. This work introduces the hybrid system of parallel self-organizing neural network (PSONN) targeting accepted values of Archie’s parameters and, consequently, reliable water saturation values. This work focuses on Archie’s parameters determination techniques; conventional technique, CAPE technique, and 3-D technique, and then the calculation of water saturation using current. Using the same data, a hybrid parallel self-organizing neural network (PSONN) algorithm is used to estimate Archie’s parameters and predict water saturation. Results have shown that estimated Arche’s parameters m, a, and n are highly accepted with statistical analysis, indicating that the PSONN model has a lower statistical error and higher correlation coefficient. This study was conducted using a high number of measurement points for 144 core plugs from a sandstone reservoir. PSONN algorithm can provide reliable water saturation values, and it can supplement or even replace the conventional techniques to determine Archie’s parameters and thereby calculate water saturation profiles.

Keywords: water saturation, Archie’s parameters, artificial intelligence, PSONN, sandstone reservoir

Procedia PDF Downloads 105
1041 Unlocking the Puzzle of Borrowing Adult Data for Designing Hybrid Pediatric Clinical Trials

Authors: Rajesh Kumar G

Abstract:

A challenging aspect of any clinical trial is to carefully plan the study design to meet the study objective in optimum way and to validate the assumptions made during protocol designing. And when it is a pediatric study, there is the added challenge of stringent guidelines and difficulty in recruiting the necessary subjects. Unlike adult trials, there is not much historical data available for pediatrics, which is required to validate assumptions for planning pediatric trials. Typically, pediatric studies are initiated as soon as approval is obtained for a drug to be marketed for adults, so with the adult study historical information and with the available pediatric pilot study data or simulated pediatric data, the pediatric study can be well planned. Generalizing the historical adult study for new pediatric study is a tedious task; however, it is possible by integrating various statistical techniques and utilizing the advantage of hybrid study design, which will help to achieve the study objective in a smoother way even with the presence of many constraints. This research paper will explain how well the hybrid study design can be planned along with integrated technique (SEV) to plan the pediatric study; In brief the SEV technique (Simulation, Estimation (using borrowed adult data and applying Bayesian methods)) incorporates the use of simulating the planned study data and getting the desired estimates to Validate the assumptions.This method of validation can be used to improve the accuracy of data analysis, ensuring that results are as valid and reliable as possible, which allow us to make informed decisions well ahead of study initiation. With professional precision, this technique based on the collected data allows to gain insight into best practices when using data from historical study and simulated data alike.

Keywords: adaptive design, simulation, borrowing data, bayesian model

Procedia PDF Downloads 36
1040 Satellite Technology Usage for Greenhouse Gas Emissions Monitoring and Verification: Policy Considerations for an International System

Authors: Timiebi Aganaba-Jeanty

Abstract:

Accurate and transparent monitoring, reporting and verification of Greenhouse Gas (GHG) emissions and removals is a requirement of the United Nations Framework Convention on Climate Change (UNFCCC). Several countries are obligated to prepare and submit an annual national greenhouse gas inventory covering anthropogenic emissions by sources and removals by sinks, subject to a review conducted by an international team of experts. However, the process is not without flaws. The self-reporting varies enormously in thoroughness, frequency and accuracy including inconsistency in the way such reporting occurs. The world’s space agencies are calling for a new generation of satellites that would be precise enough to map greenhouse gas emissions from individual nations. The plan is delicate politically because the global system could verify or cast doubt on emission reports from the member states of the UNFCCC. A level playing field is required and an idea that an international system should be perceived as an instrument to facilitate fairness and equality rather than to spy on or punish. This change of perspective is required to get buy in for an international verification system. The research proposes the viability of a satellite system that provides independent access to data regarding greenhouse gas emissions and the policy and governance implications of its potential use as a monitoring and verification system for the Paris Agreement. It assesses the foundations of the reporting monitoring and verification system as proposed in Paris and analyzes this in light of a proposed satellite system. The use of remote sensing technology has been debated for verification purposes and as evidence in courts but this is not without controversy. Lessons can be learned from its use in this context.

Keywords: greenhouse gas emissions, reporting, monitoring and verification, satellite, UNFCCC

Procedia PDF Downloads 263
1039 Grain Selection in Spiral Grain Selectors during Casting Single-Crystal Turbine Blades

Authors: M. Javahar, H. B. Dong

Abstract:

Single crystal components manufactured using Ni-base Superalloys are routinely used in the hot sections of aero engines and industrial gas turbines due to their outstanding high temperature strength, toughness and resistance to degradation in corrosive and oxidative environments. To control the quality of the single crystal turbine blades, particular attention has been paid to grain selection, which is used to obtain the single crystal morphology from a plethora of columnar grains. For this purpose, different designs of grain selectors are employed and the most common type is the spiral grain selector. A typical spiral grain selector includes a starter block and a spiral (helix) located above. It has been found that the grains with orientation well aligned to the thermal gradient survive in the starter block by competitive grain growth while the selection of the single crystal grain occurs in the spiral part. In the present study, 2D spiral selectors with different geometries were designed and produced using a state-of-the-art Bridgeman Directional Solidification casting furnace to investigate the competitive growth during grain selection in 2d grain selectors. The principal advantage of using a 2-D selector is to facilitate the wax injection process in investment casting by enabling significant degree of automation. The automation within the process can be derived by producing 2D grain selector wax patterns parts using a split die (metal mold model) coupled with wax injection stage. This will not only produce the part with high accuracy but also at an acceptable production rate.

Keywords: grain selector, single crystal, directional solidification, CMSX-4 superalloys, investment casting

Procedia PDF Downloads 553
1038 Importance of Developing a Decision Support System for Diagnosis of Glaucoma

Authors: Murat Durucu

Abstract:

Glaucoma is a condition of irreversible blindness, early diagnosis and appropriate interventions to make the patients able to see longer time. In this study, it addressed that the importance of developing a decision support system for glaucoma diagnosis. Glaucoma occurs when pressure happens around the eyes it causes some damage to the optic nerves and deterioration of vision. There are different levels ranging blindness of glaucoma disease. The diagnosis at an early stage allows a chance for therapies that slows the progression of the disease. In recent years, imaging technology from Heidelberg Retinal Tomography (HRT), Stereoscopic Disc Photo (SDP) and Optical Coherence Tomography (OCT) have been used for the diagnosis of glaucoma. This better accuracy and faster imaging techniques in response technique of OCT have become the most common method used by experts. Although OCT images or HRT precision and quickness, especially in the early stages, there are still difficulties and mistakes are occurred in diagnosis of glaucoma. It is difficult to obtain objective results on diagnosis and placement process of the doctor's. It seems very important to develop an objective decision support system for diagnosis and level the glaucoma disease for patients. By using OCT images and pattern recognition systems, it is possible to develop a support system for doctors to make their decisions on glaucoma. Thus, in this recent study, we develop an evaluation and support system to the usage of doctors. Pattern recognition system based computer software would help the doctors to make an objective evaluation for their patients. It is intended that after development and evaluation processes of the software, the system is planning to be serve for the usage of doctors in different hospitals.

Keywords: decision support system, glaucoma, image processing, pattern recognition

Procedia PDF Downloads 259
1037 Feature Analysis of Predictive Maintenance Models

Authors: Zhaoan Wang

Abstract:

Research in predictive maintenance modeling has improved in the recent years to predict failures and needed maintenance with high accuracy, saving cost and improving manufacturing efficiency. However, classic prediction models provide little valuable insight towards the most important features contributing to the failure. By analyzing and quantifying feature importance in predictive maintenance models, cost saving can be optimized based on business goals. First, multiple classifiers are evaluated with cross-validation to predict the multi-class of failures. Second, predictive performance with features provided by different feature selection algorithms are further analyzed. Third, features selected by different algorithms are ranked and combined based on their predictive power. Finally, linear explainer SHAP (SHapley Additive exPlanations) is applied to interpret classifier behavior and provide further insight towards the specific roles of features in both local predictions and global model behavior. The results of the experiments suggest that certain features play dominant roles in predictive models while others have significantly less impact on the overall performance. Moreover, for multi-class prediction of machine failures, the most important features vary with type of machine failures. The results may lead to improved productivity and cost saving by prioritizing sensor deployment, data collection, and data processing of more important features over less importance features.

Keywords: automated supply chain, intelligent manufacturing, predictive maintenance machine learning, feature engineering, model interpretation

Procedia PDF Downloads 105
1036 Determining of the Performance of Data Mining Algorithm Determining the Influential Factors and Prediction of Ischemic Stroke: A Comparative Study in the Southeast of Iran

Authors: Y. Mehdipour, S. Ebrahimi, A. Jahanpour, F. Seyedzaei, B. Sabayan, A. Karimi, H. Amirifard

Abstract:

Ischemic stroke is one of the common reasons for disability and mortality. The fourth leading cause of death in the world and the third in some other sources. Only 1/3 of the patients with ischemic stroke fully recover, 1/3 of them end in permanent disability and 1/3 face death. Thus, the use of predictive models to predict stroke has a vital role in reducing the complications and costs related to this disease. Thus, the aim of this study was to specify the effective factors and predict ischemic stroke with the help of DM methods. The present study was a descriptive-analytic study. The population was 213 cases from among patients referring to Ali ibn Abi Talib (AS) Hospital in Zahedan. Data collection tool was a checklist with the validity and reliability confirmed. This study used DM algorithms of decision tree for modeling. Data analysis was performed using SPSS-19 and SPSS Modeler 14.2. The results of the comparison of algorithms showed that CHAID algorithm with 95.7% accuracy has the best performance. Moreover, based on the model created, factors such as anemia, diabetes mellitus, hyperlipidemia, transient ischemic attacks, coronary artery disease, and atherosclerosis are the most effective factors in stroke. Decision tree algorithms, especially CHAID algorithm, have acceptable precision and predictive ability to determine the factors affecting ischemic stroke. Thus, by creating predictive models through this algorithm, will play a significant role in decreasing the mortality and disability caused by ischemic stroke.

Keywords: data mining, ischemic stroke, decision tree, Bayesian network

Procedia PDF Downloads 147