Search results for: validation techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7979

Search results for: validation techniques

7649 A Study of Evolving Cloud Computing Data Security: A Machine Learning Perspective

Authors: Shinoy Vengaramkode Bhaskaran

Abstract:

The advancement of cloud computing led to a variety of security issues for both consumers and industries. Whereas machine learning (ML) is one approach to securing Cloud-based systems. Various methods have been employed to prevent or detect attacks and security vulnerabilities on the Cloud using ML techniques. In this paper, we present an ML perspective on the methodologies and techniques of cloud security. Initially, an investigative study on cloud computing is conducted with a primary emphasis on the gaps with two research questions that are impeding the adoption of cloud technology, as well as the challenges associated with threat solutions. Next, some ideas are generated based on machine learning methods to mitigate certain types of attacks that are frequently discussed through the application of ML techniques. Finally, we review different machine learning algorithms and their adoption in cloud computing.

Keywords: artificial intelligence, machine learning, cloud computing infrastructure as a service, support vector machine, platform as a service

Procedia PDF Downloads 6
7648 Validation of an Educative Manual for Patients with Breast Cancer Submitted to Radiation Therapy

Authors: Flavia Oliveira de A. M. Cruz, Edison Tostes Faria, Paula Elaine D. Reis

Abstract:

When the breast is submitted to radiation therapy (RT), the most common effects are pain, skin changes, mobility restrictions, local sensory alteration, and fatigue. These effects, if not managed properly, may reduce the quality of life of cancer patients and may lead to the treatment discontinuation. Therefore, promoting knowledge and guidelines for symptom management remain a high priority for patients and a challenge for health professionals, due to the need to handle side effects in a population with a life-threatening disease. Printed materials are important strategies for supporting educative activities since they help the individual to assimilate and understand the amount of information transmitted. Nurses' behavior can be systematized through the use of an educative manual, which may be effective in promoting information regarding the treatment, self-care and how to control the effects of RT at home. In view of the importance of guaranteeing the validity of the material before its use, the objective of this research was to validate the content and appearance of an educative manual for breast cancer patients undergoing RT. The Theory of Psychometrics was used for the validation process in this descriptive methodological research. A minimum agreement rate (AR) of 80% was considered to guarantee the validity of the material. The data were collected from October to December 2017, by means of two assessments tools, constructed in the form of a Likert scale, with five levels of understanding. These instruments addressed different aspects of the evaluation, in view of two different groups of participants; 17 experts in the theme area of the educative manual, and 12 women that received RT previously to treat breast cancer. The manual was titled 'Orientation Manual: radiation therapy in breast', and was focused on breast cancer patients attended at the Department of Oncology of the Brasília University Hospital (UNACON/HUB). The research project was submitted to the Research Ethics Committee at the School of Health Sciences of the University of Brasília (CAAE: 24592213.1.0000.0030). Only two items of the assessment tool for the experts, one related to the manual's ability to promote behavioral and attitude changes and the other related to the extent of its use for other health services, obtained AR < 80% and were reformulated based on the participants' suggestions and in the literature. All other items were considered appropriate and/or complete appropriate in the three blocks proposed for the experts: objectives - 89%, structure and form - 93%, and relevance - 93%; and good and/or very good in the five blocks of analysis proposed for patients: objectives - 100%, organization - 100%, writing style - 100%, appearance - 100%, and motivation. The appearance and content validation of the educative manual proposed were attended to. The educative manual was considered relevant and pertinent and may contribute to the understanding of the therapeutic process by breast cancer patients during RT, as well as support clinical practice through the nursing consultation.

Keywords: oncology nursing, nursing care, validation studies, educational technology

Procedia PDF Downloads 127
7647 Data Mining Techniques for Anti-Money Laundering

Authors: M. Sai Veerendra

Abstract:

Today, money laundering (ML) poses a serious threat not only to financial institutions but also to the nation. This criminal activity is becoming more and more sophisticated and seems to have moved from the cliché of drug trafficking to financing terrorism and surely not forgetting personal gain. Most of the financial institutions internationally have been implementing anti-money laundering solutions (AML) to fight investment fraud activities. However, traditional investigative techniques consume numerous man-hours. Recently, data mining approaches have been developed and are considered as well-suited techniques for detecting ML activities. Within the scope of a collaboration project on developing a new data mining solution for AML Units in an international investment bank in Ireland, we survey recent data mining approaches for AML. In this paper, we present not only these approaches but also give an overview on the important factors in building data mining solutions for AML activities.

Keywords: data mining, clustering, money laundering, anti-money laundering solutions

Procedia PDF Downloads 539
7646 Multi-Scaled Non-Local Means Filter for Medical Images Denoising: Empirical Mode Decomposition vs. Wavelet Transform

Authors: Hana Rabbouch

Abstract:

In recent years, there has been considerable growth of denoising techniques mainly devoted to medical imaging. This important evolution is not only due to the progress of computing techniques, but also to the emergence of multi-resolution analysis (MRA) on both mathematical and algorithmic bases. In this paper, a comparative study is conducted between the two best-known MRA-based decomposition techniques: the Empirical Mode Decomposition (EMD) and the Discrete Wavelet Transform (DWT). The comparison is carried out in a framework of multi-scale denoising, where a Non-Local Means (NLM) filter is performed scale-by-scale to a sample of benchmark medical images. The results prove the effectiveness of the multiscaled denoising, especially when the NLM filtering is coupled with the EMD.

Keywords: medical imaging, non local means, denoising, multiscaled analysis, empirical mode decomposition, wavelets

Procedia PDF Downloads 142
7645 Review of Malaria Diagnosis Techniques

Authors: Lubabatu Sada Sodangu

Abstract:

Malaria is a major cause of death in tropical and subtropical nations. Malaria cases are continually rising as a result of a number of factors, despite the fact that the condition is now treatable using effective methods. In this situation, quick and effective diagnostic methods are essential for the management and control of malaria. Malaria diagnosis using conventional methods is still troublesome, hence new technologies have been created and implemented to get around the drawbacks. The review describes the currently known malaria diagnostic techniques, their strengths and shortcomings.

Keywords: malaria, technique, diagnosis, Africa

Procedia PDF Downloads 55
7644 Review of Malaria Diagnosis Techniques

Authors: Lubabatu Sada Sodangi

Abstract:

Malaria is a major cause of death in tropical and subtropical nations. Malaria cases are continually rising as a result of a number of factors, despite the fact that the condition is now treatable using effective methods. In this situation, quick and effective diagnostic methods are essential for the management and control of malaria. Malaria diagnosis using conventional methods is still troublesome; hence, new technologies have been created and implemented to get around the drawbacks. The review describes the currently known malaria diagnostic techniques, their strengths, and shortcomings.

Keywords: malaria, technique, diagnosis, Africa

Procedia PDF Downloads 60
7643 Application of Learning Media Based Augmented Reality on Molecular Geometry Concept

Authors: F. S. Irwansyah, I. Farida, Y. Maulana

Abstract:

Studying chemistry requires the ability to understand three levels of understanding in the form of macroscopic, submicroscopic and symbolic, but the lack of emphasis on the submicroscopic level leads to the understanding of chemical concepts becoming incomplete, due to the limitations of the tools capable of providing visualization of submicroscopic concepts. The purpose of this study describes the stages of making augmented reality learning media on the concept of molecular geometry and analyze the feasibility test result of augmented reality learning media on the concept of molecular geometry. This research uses Research and Development (R & D) method which produces a product of AR learning media on molecular geometry concept and test the effectiveness of the product. Research stages include concept analysis and learning indicators, design development, validation, feasibility, and limited testing. The stages of validation and limited trial are aimed to get feedback in the form of assessment, suggestion and improvement on learning aspect, material substance aspect, visual communication aspect and software engineering aspects and media feasibility in terms of media creation purpose to be used in learning. The results of the overall feasibility test obtained r-calculation 0,7-0,9 with the interpretation of high feasibility value, whereas the result of limited trial got the percentage of eligibility with the average value equal to 70,83-92,5%. This percentage indicates that AR's learning media product on the concept of molecular geometry, deserves to be used as a learning resource.

Keywords: android, augmented reality, chemical learning, geometry

Procedia PDF Downloads 207
7642 Cost Reduction Techniques for Provision of Shelter to Homeless

Authors: Mukul Anand

Abstract:

Quality oriented affordable shelter for all has always been the key issue in the housing sector of our country. Homelessness is the acute form of housing need. It is a paradox that in spite of innumerable government initiated programmes for affordable housing, certain section of society is still devoid of shelter. About nineteen million (18.78 million) households grapple with housing shortage in Urban India in 2012. In Indian scenario there is major mismatch between the people for whom the houses are being built and those who need them. The prime force faced by public authorities in facilitation of quality housing for all is high cost of construction. The present paper will comprehend executable techniques for dilution of cost factor in housing the homeless. The key actors responsible for delivery of cheap housing stock such as capacity building, resource optimization, innovative low cost building material and indigenous skeleton housing system will also be incorporated in developing these techniques. Time performance, which is an important angle of above actors, will also be explored so as to increase the effectiveness of low cost housing. Along with this best practices will be taken up as case studies where both conventional techniques of housing and innovative low cost housing techniques would be cited. Transportation consists of approximately 30% of total construction budget. Thus use of alternative local solutions depending upon the region would be covered so as to highlight major components of low cost housing. Government is laid back regarding base line information on use of innovative low cost method and technique of resource optimization. Therefore, the paper would be an attempt to bring to light simpler solutions for achieving low cost housing.

Keywords: construction, cost, housing, optimization, shelter

Procedia PDF Downloads 447
7641 Educating Empathy: Combining Active Listening and Moral Discovery to Facilitate Prosocial Connection

Authors: Erika Price, Lisa Johnson

Abstract:

Cognitive and dispositional empathy is decreasing among students worldwide, particularly those at university. This paper looks at the effects of encouraging empathetic positioning in divisive topics by teaching listening skills and moral discovery to university students. Two groups of university students were given the assignment to interview individuals they disagreed with on social issues (e.g. abortion, gun control, legalization of drugs, involvement in Ukraine, etc.). One group completed the assignment with no other instruction. The second group completed the assignment after receiving instruction in active listening and Jonathan Haidt’s theory of moral foundations in politics. Results show that when students are given both active listening techniques and awareness of moral foundations, they are significantly more likely to have socially positive interactions with those they disagree with on issues as compared to those who listen passively to ideological opponents. As students interacted with those they disagreed with, they evidenced prosocial behaviors of acknowledgement, validation, and even commonalities with their opponents’ viewpoints, signifying a heartening trend of empathetic connection that is waning in students. The research suggests that empathy is a skill that can be nurtured by active listening but that it is more fully cultivated when paired with the concept of moral foundations underpinning political ideologies. These findings shed light on how to create more effective pedagogies for social and emotional learning, as well as inclusion.

Keywords: empathy, listening skills, moral discovery, pedagogy, prosocial behavior

Procedia PDF Downloads 69
7640 Artificial Neural Networks Application on Nusselt Number and Pressure Drop Prediction in Triangular Corrugated Plate Heat Exchanger

Authors: Hany Elsaid Fawaz Abdallah

Abstract:

This study presents a new artificial neural network(ANN) model to predict the Nusselt Number and pressure drop for the turbulent flow in a triangular corrugated plate heat exchanger for forced air and turbulent water flow. An experimental investigation was performed to create a new dataset for the Nusselt Number and pressure drop values in the following range of dimensionless parameters: The plate corrugation angles (from 0° to 60°), the Reynolds number (from 10000 to 40000), pitch to height ratio (from 1 to 4), and Prandtl number (from 0.7 to 200). Based on the ANN performance graph, the three-layer structure with {12-8-6} hidden neurons has been chosen. The training procedure includes back-propagation with the biases and weight adjustment, the evaluation of the loss function for the training and validation dataset and feed-forward propagation of the input parameters. The linear function was used at the output layer as the activation function, while for the hidden layers, the rectified linear unit activation function was utilized. In order to accelerate the ANN training, the loss function minimization may be achieved by the adaptive moment estimation algorithm (ADAM). The ‘‘MinMax’’ normalization approach was utilized to avoid the increase in the training time due to drastic differences in the loss function gradients with respect to the values of weights. Since the test dataset is not being used for the ANN training, a cross-validation technique is applied to the ANN network using the new data. Such procedure was repeated until loss function convergence was achieved or for 4000 epochs with a batch size of 200 points. The program code was written in Python 3.0 using open-source ANN libraries such as Scikit learn, TensorFlow and Keras libraries. The mean average percent error values of 9.4% for the Nusselt number and 8.2% for pressure drop for the ANN model have been achieved. Therefore, higher accuracy compared to the generalized correlations was achieved. The performance validation of the obtained model was based on a comparison of predicted data with the experimental results yielding excellent accuracy.

Keywords: artificial neural networks, corrugated channel, heat transfer enhancement, Nusselt number, pressure drop, generalized correlations

Procedia PDF Downloads 89
7639 Current Practices of Permitted Daily Exposure (PDE) Calculation and Selection

Authors: Annie Ramanbhai Mecwan

Abstract:

Cleaning validation in a pharmaceutical manufacturing facility is documented evidence that a cleaning process has effectively removed contaminants, residues from previous drug products and cleaning agents below a pre-defined threshold from the reusable tools and parts of equipment. In shared manufacturing facilities more than one drug product is prepared. After cleaning of reusable tools and parts of equipment after one drug product manufacturing, there are chances that some residues of drug substance from previously manufactured drug products may be retained on the equipment and can carried forward to the next drug product and thus cause cross-contamination. Health-based limits through the derivation of a safe threshold value called permitted daily exposure (PDE) for the residues of drug substances should be employed to identify the risks posed at these manufacturing facilities. The PDE represents a substance-specific dose that is unlikely to cause an adverse effect if an individual is exposed to or below this dose every day for a lifetime. There are different practices to calculate PDE. Data for all APIs in the public domain are considered to calculate PDE value though, company to company may vary the final PDE value based on different toxicologist’s perspective or their subjective evaluation. Hence, Regulatory agencies should take responsibility for publishing PDE values for all APIs as it is done for elemental PDEs. This will harmonize the PDE values all over the world and prevent the unnecessary load on manufacturers for cleaning validation

Keywords: active pharmaceutical ingredient, good manufacturing practice, NOAEL, no observed adverse effect level, permitted daily exposure

Procedia PDF Downloads 91
7638 Measurement of Solids Concentration in Hydrocyclone Using ERT: Validation Against CFD

Authors: Vakamalla Teja Reddy, Narasimha Mangadoddy

Abstract:

Hydrocyclones are used to separate particles into different size fractions in the mineral processing, chemical and metallurgical industries. High speed video imaging, Laser Doppler Anemometry (LDA), X-ray and Gamma ray tomography are previously used to measure the two-phase flow characteristics in the cyclone. However, investigation of solids flow characteristics inside the cyclone is often impeded by the nature of the process due to slurry opaqueness and solid metal wall vessels. In this work, a dual-plane high speed Electrical resistance tomography (ERT) is used to measure hydrocyclone internal flow dynamics in situ. Experiments are carried out in 3 inch hydrocyclone for feed solid concentrations varying in the range of 0-50%. ERT data analysis through the optimized FEM mesh size and reconstruction algorithms on air-core and solid concentration tomograms is assessed. Results are presented in terms of the air-core diameter and solids volume fraction contours using Maxwell’s equation for various hydrocyclone operational parameters. It is confirmed by ERT that the air core occupied area and wall solids conductivity levels decreases with increasing the feed solids concentration. Algebraic slip mixture based multi-phase computational fluid dynamics (CFD) model is used to predict the air-core size and the solid concentrations in the hydrocyclone. Validation of air-core size and mean solid volume fractions by ERT measurements with the CFD simulations is attempted.

Keywords: air-core, electrical resistance tomography, hydrocyclone, multi-phase CFD

Procedia PDF Downloads 379
7637 Comparative Analysis of Classification Methods in Determining Non-Active Student Characteristics in Indonesia Open University

Authors: Dewi Juliah Ratnaningsih, Imas Sukaesih Sitanggang

Abstract:

Classification is one of data mining techniques that aims to discover a model from training data that distinguishes records into the appropriate category or class. Data mining classification methods can be applied in education, for example, to determine the classification of non-active students in Indonesia Open University. This paper presents a comparison of three methods of classification: Naïve Bayes, Bagging, and C.45. The criteria used to evaluate the performance of three methods of classification are stratified cross-validation, confusion matrix, the value of the area under the ROC Curve (AUC), Recall, Precision, and F-measure. The data used for this paper are from the non-active Indonesia Open University students in registration period of 2004.1 to 2012.2. Target analysis requires that non-active students were divided into 3 groups: C1, C2, and C3. Data analyzed are as many as 4173 students. Results of the study show: (1) Bagging method gave a high degree of classification accuracy than Naïve Bayes and C.45, (2) the Bagging classification accuracy rate is 82.99 %, while the Naïve Bayes and C.45 are 80.04 % and 82.74 % respectively, (3) the result of Bagging classification tree method has a large number of nodes, so it is quite difficult in decision making, (4) classification of non-active Indonesia Open University student characteristics uses algorithms C.45, (5) based on the algorithm C.45, there are 5 interesting rules which can describe the characteristics of non-active Indonesia Open University students.

Keywords: comparative analysis, data mining, clasiffication, Bagging, Naïve Bayes, C.45, non-active students, Indonesia Open University

Procedia PDF Downloads 316
7636 Estimation of Ribb Dam Catchment Sediment Yield and Reservoir Effective Life Using Soil and Water Assessment Tool Model and Empirical Methods

Authors: Getalem E. Haylia

Abstract:

The Ribb dam is one of the irrigation projects in the Upper Blue Nile basin, Ethiopia, to irrigate the Fogera plain. Reservoir sedimentation is a major problem because it reduces the useful reservoir capacity by the accumulation of sediments coming from the watersheds. Estimates of sediment yield are needed for studies of reservoir sedimentation and planning of soil and water conservation measures. The objective of this study was to simulate the Ribb dam catchment sediment yield using SWAT model and to estimate Ribb reservoir effective life according to trap efficiency methods. The Ribb dam catchment is found in North Western part of Ethiopia highlands, and it belongs to the upper Blue Nile and Lake Tana basins. Soil and Water Assessment Tool (SWAT) was selected to simulate flow and sediment yield in the Ribb dam catchment. The model sensitivity, calibration, and validation analysis at Ambo Bahir site were performed with Sequential Uncertainty Fitting (SUFI-2). The flow data at this site was obtained by transforming the Lower Ribb gauge station (2002-2013) flow data using Area Ratio Method. The sediment load was derived based on the sediment concentration yield curve of Ambo site. Stream flow results showed that the Nash-Sutcliffe efficiency coefficient (NSE) was 0.81 and the coefficient of determination (R²) was 0.86 in calibration period (2004-2010) and, 0.74 and 0.77 in validation period (2011-2013), respectively. Using the same periods, the NS and R² for the sediment load calibration were 0.85 and 0.79 and, for the validation, it became 0.83 and 0.78, respectively. The simulated average daily flow rate and sediment yield generated from Ribb dam watershed were 3.38 m³/s and 1772.96 tons/km²/yr, respectively. The effective life of Ribb reservoir was estimated using the developed empirical methods of the Brune (1953), Churchill (1948) and Brown (1958) methods and found to be 30, 38 and 29 years respectively. To conclude, massive sediment comes from the steep slope agricultural areas, and approximately 98-100% of this incoming annual sediment loads have been trapped by the Ribb reservoir. In Ribb catchment, as well as reservoir systematic and thorough consideration of technical, social, environmental, and catchment managements and practices should be made to lengthen the useful life of Ribb reservoir.

Keywords: catchment, reservoir effective life, reservoir sedimentation, Ribb, sediment yield, SWAT model

Procedia PDF Downloads 188
7635 Fake News Detection for Korean News Using Machine Learning Techniques

Authors: Tae-Uk Yun, Pullip Chung, Kee-Young Kwahk, Hyunchul Ahn

Abstract:

Fake news is defined as the news articles that are intentionally and verifiably false, and could mislead readers. Spread of fake news may provoke anxiety, chaos, fear, or irrational decisions of the public. Thus, detecting fake news and preventing its spread has become very important issue in our society. However, due to the huge amount of fake news produced every day, it is almost impossible to identify it by a human. Under this context, researchers have tried to develop automated fake news detection using machine learning techniques over the past years. But, there have been no prior studies proposed an automated fake news detection method for Korean news to our best knowledge. In this study, we aim to detect Korean fake news using text mining and machine learning techniques. Our proposed method consists of two steps. In the first step, the news contents to be analyzed is convert to quantified values using various text mining techniques (topic modeling, TF-IDF, and so on). After that, in step 2, classifiers are trained using the values produced in step 1. As the classifiers, machine learning techniques such as logistic regression, backpropagation network, support vector machine, and deep neural network can be applied. To validate the effectiveness of the proposed method, we collected about 200 short Korean news from Seoul National University’s FactCheck. which provides with detailed analysis reports from 20 media outlets and links to source documents for each case. Using this dataset, we will identify which text features are important as well as which classifiers are effective in detecting Korean fake news.

Keywords: fake news detection, Korean news, machine learning, text mining

Procedia PDF Downloads 276
7634 A Comparative Analysis of Various Companding Techniques Used to Reduce PAPR in VLC Systems

Authors: Arushi Singh, Anjana Jain, Prakash Vyavahare

Abstract:

Recently, Li-Fi(light-fiedelity) has been launched based on VLC(visible light communication) technique, 100 times faster than WiFi. Now 5G mobile communication system is proposed to use VLC-OFDM as the transmission technique. The VLC system focused on visible rays, is considered for efficient spectrum use and easy intensity modulation through LEDs. The reason of high speed in VLC is LED, as they flicker incredibly fast(order of MHz). Another advantage of employing LED is-it acts as low pass filter results no out-of-band emission. The VLC system falls under the category of ‘green technology’ for utilizing LEDs. In present scenario, OFDM is used for high data-rates, interference immunity and high spectral efficiency. Inspite of the advantages OFDM suffers from large PAPR, ICI among carriers and frequency offset errors. Since, the data transmission technique used in VLC system is OFDM, the system suffers the drawbacks of OFDM as well as VLC, the non-linearity dues to non-linear characteristics of LED and PAPR of OFDM due to which the high power amplifier enters in non-linear region. The proposed paper focuses on reduction of PAPR in VLC-OFDM systems. Many techniques are applied to reduce PAPR such as-clipping-introduces distortion in the carrier; selective mapping technique-suffers wastage of bandwidth; partial transmit sequence-very complex due to exponentially increased number of sub-blocks. The paper discusses three companding techniques namely- µ-law, A-law and advance A-law companding technique. The analysis shows that the advance A-law companding techniques reduces the PAPR of the signal by adjusting the companding parameter within the range. VLC-OFDM systems are the future of the wireless communication but non-linearity in VLC-OFDM is a severe issue. The proposed paper discusses the techniques to reduce PAPR, one of the non-linearities of the system. The companding techniques mentioned in this paper provides better results without increasing the complexity of the system.

Keywords: non-linear companding techniques, peak to average power ratio (PAPR), visible light communication (VLC), VLC-OFDM

Procedia PDF Downloads 286
7633 Awareness and Utilization of E-Learning Technologies in Teaching and Learning of Human Kinetics and Health Education Courses in Nigeria Universities

Authors: Ibrahim Laro ABUBAKAR

Abstract:

The study examined the Availability and Utilization of E-Learning Technologies in Teaching of Human Kinetics and Health Education courses in Nigerian Universities, specifically, Universities in Kwara State. Two purposes were formulated to guide the study from which two research questions and two hypotheses were raised. The descriptive research design was used in the research. Three Hundred respondents (100 Lecturers and 200 Students) made up the population for the study. There was no sampling, as the population of the study was not much. A structured questionnaire tagged ‘Availability and Utilization of E-Learning Technologies in Teaching and Learning Questionnaire’ (AUETTLQ) was used for data collection. The questionnaire was subjected to face and content validation, and it was equally pilot tested. The validation yielded a reliability coefficient of 0.78. The data collected from the study were statistically analyzed using frequencies and percentage count for personal data of the respondents, mean and standard deviation to answer the research questions. The null hypotheses were tested at 0.05 level of significance using the independent t-test. One among other findings of this study showed that lecturers and Student are aware of synchronous e-learning technologies in teaching and learning of Human Kinetics and Health Education but often utilize the synchronous e-learning technologies. It was recommended among others that lecturers and Students should be sensitized through seminars and workshops on the need to maximally utilize available e-learning technologies in teaching and learning of Human Kinetics and Health Education courses in Universities.

Keywords: awareness, utilization, E-Learning, technologies, human kinetics synchronous

Procedia PDF Downloads 120
7632 Lessons Learned from Interlaboratory Noise Modelling in Scope of Environmental Impact Assessments in Slovenia

Authors: S. Cencek, A. Markun

Abstract:

Noise assessment methods are regularly used in scope of Environmental Impact Assessments for planned projects to assess (predict) the expected noise emissions of these projects. Different noise assessment methods could be used. In recent years, we had an opportunity to collaborate in some noise assessment procedures where noise assessments of different laboratories have been performed simultaneously. We identified some significant differences in noise assessment results between laboratories in Slovenia. We estimate that despite good input Georeferenced Data to set up acoustic model exists in Slovenia; there is no clear consensus on methods for predictive noise methods for planned projects. We analyzed input data, methods and results of predictive noise methods for two planned industrial projects, both were done independently by two laboratories. We also analyzed the data, methods and results of two interlaboratory collaborative noise models for two existing noise sources (railway and motorway). In cases of predictive noise modelling, the validations of acoustic models were performed by noise measurements of surrounding existing noise sources, but in varying durations. The acoustic characteristics of existing buildings were also not described identically. The planned noise sources were described and digitized differently. Differences in noise assessment results between different laboratories have ranged up to 10 dBA, which considerably exceeds the acceptable uncertainty ranged between 3 to 6 dBA. Contrary to predictive noise modelling, in cases of collaborative noise modelling for two existing noise sources the possibility to perform the validation noise measurements of existing noise sources greatly increased the comparability of noise modelling results. In both cases of collaborative noise modelling for existing motorway and railway, the modelling results of different laboratories were comparable. Differences in noise modeling results between different laboratories were below 5 dBA, which was acceptable uncertainty set up by interlaboratory noise modelling organizer. The lessons learned from the study were: 1) Predictive noise calculation using formulae from International standard SIST ISO 9613-2: 1997 is not an appropriate method to predict noise emissions of planned projects since due to complexity of procedure they are not used strictly, 2) The noise measurements are important tools to minimize noise assessment errors of planned projects and should be in cases of predictive noise modelling performed at least for validation of acoustic model, 3) National guidelines should be made on the appropriate data, methods, noise source digitalization, validation of acoustic model etc. in order to unify the predictive noise models and their results in scope of Environmental Impact Assessments for planned projects.

Keywords: environmental noise assessment, predictive noise modelling, spatial planning, noise measurements, national guidelines

Procedia PDF Downloads 234
7631 Machine Learning for Classifying Risks of Death and Length of Stay of Patients in Intensive Unit Care Beds

Authors: Itamir de Morais Barroca Filho, Cephas A. S. Barreto, Ramon Malaquias, Cezar Miranda Paula de Souza, Arthur Costa Gorgônio, João C. Xavier-Júnior, Mateus Firmino, Fellipe Matheus Costa Barbosa

Abstract:

Information and Communication Technologies (ICT) in healthcare are crucial for efficiently delivering medical healthcare services to patients. These ICTs are also known as e-health and comprise technologies such as electronic record systems, telemedicine systems, and personalized devices for diagnosis. The focus of e-health is to improve the quality of health information, strengthen national health systems, and ensure accessible, high-quality health care for all. All the data gathered by these technologies make it possible to help clinical staff with automated decisions using machine learning. In this context, we collected patient data, such as heart rate, oxygen saturation (SpO2), blood pressure, respiration, and others. With this data, we were able to develop machine learning models for patients’ risk of death and estimate the length of stay in ICU beds. Thus, this paper presents the methodology for applying machine learning techniques to develop these models. As a result, although we implemented these models on an IoT healthcare platform, helping clinical staff in healthcare in an ICU, it is essential to create a robust clinical validation process and monitoring of the proposed models.

Keywords: ICT, e-health, machine learning, ICU, healthcare

Procedia PDF Downloads 114
7630 Gradient Boosted Trees on Spark Platform for Supervised Learning in Health Care Big Data

Authors: Gayathri Nagarajan, L. D. Dhinesh Babu

Abstract:

Health care is one of the prominent industries that generate voluminous data thereby finding the need of machine learning techniques with big data solutions for efficient processing and prediction. Missing data, incomplete data, real time streaming data, sensitive data, privacy, heterogeneity are few of the common challenges to be addressed for efficient processing and mining of health care data. In comparison with other applications, accuracy and fast processing are of higher importance for health care applications as they are related to the human life directly. Though there are many machine learning techniques and big data solutions used for efficient processing and prediction in health care data, different techniques and different frameworks are proved to be effective for different applications largely depending on the characteristics of the datasets. In this paper, we present a framework that uses ensemble machine learning technique gradient boosted trees for data classification in health care big data. The framework is built on Spark platform which is fast in comparison with other traditional frameworks. Unlike other works that focus on a single technique, our work presents a comparison of six different machine learning techniques along with gradient boosted trees on datasets of different characteristics. Five benchmark health care datasets are considered for experimentation, and the results of different machine learning techniques are discussed in comparison with gradient boosted trees. The metric chosen for comparison is misclassification error rate and the run time of the algorithms. The goal of this paper is to i) Compare the performance of gradient boosted trees with other machine learning techniques in Spark platform specifically for health care big data and ii) Discuss the results from the experiments conducted on datasets of different characteristics thereby drawing inference and conclusion. The experimental results show that the accuracy is largely dependent on the characteristics of the datasets for other machine learning techniques whereas gradient boosting trees yields reasonably stable results in terms of accuracy without largely depending on the dataset characteristics.

Keywords: big data analytics, ensemble machine learning, gradient boosted trees, Spark platform

Procedia PDF Downloads 241
7629 Programming without Code: An Approach and Environment to Conditions-On-Data Programming

Authors: Philippe Larvet

Abstract:

This paper presents the concept of an object-based programming language where tests (if... then... else) and control structures (while, repeat, for...) disappear and are replaced by conditions on data. According to the object paradigm, by using this concept, data are still embedded inside objects, as variable-value couples, but object methods are expressed into the form of logical propositions (‘conditions on data’ or COD).For instance : variable1 = value1 AND variable2 > value2 => variable3 = value3. Implementing this approach, a central inference engine turns and examines objects one after another, collecting all CODs of each object. CODs are considered as rules in a rule-based system: the left part of each proposition (left side of the ‘=>‘ sign) is the premise and the right part is the conclusion. So, premises are evaluated and conclusions are fired. Conclusions modify the variable-value couples of the object and the engine goes to examine the next object. The paper develops the principles of writing CODs instead of complex algorithms. Through samples, the paper also presents several hints for implementing a simple mechanism able to process this ‘COD language’. The proposed approach can be used within the context of simulation, process control, industrial systems validation, etc. By writing simple and rigorous conditions on data, instead of using classical and long-to-learn languages, engineers and specialists can easily simulate and validate the functioning of complex systems.

Keywords: conditions on data, logical proposition, programming without code, object-oriented programming, system simulation, system validation

Procedia PDF Downloads 222
7628 Analysis of Spamming Threats and Some Possible Solutions for Online Social Networking Sites (OSNS)

Authors: Dilip Singh Sisodia, Shrish Verma

Abstract:

Spamming is the most common issue seen nowadays in the Internet especially in Online Social Networking Sites (like Facebook, Twitter, and Google+ etc.). Spam messages keep wasting Internet bandwidth and the storage space of servers. On social network sites; spammers often disguise themselves by creating fake accounts and hijacking user’s accounts for personal gains. They behave like normal user and they continue to change their spamming strategy. To prevent this, most modern spam-filtering solutions are deployed on the receiver side; they are good at filtering spam for end users. In this paper we are presenting some spamming techniques their behaviour and possible solutions. We have analyzed how Spammers enters into online social networking sites (OSNSs) and how they target it and the techniques they use for it. The five discussed techniques of spamming techniques which are clickjacking, social engineered attacks, cross site scripting, URL shortening, and drive by download. We have used elgg framework for demonstration of some of spamming threats and respective implementation of solutions.

Keywords: online social networking sites, spam, attacks, internet, clickjacking / likejacking, drive-by-download, URL shortening, networking, socially engineered attacks, elgg framework

Procedia PDF Downloads 348
7627 Optimizing E-commerce Retention: A Detailed Study of Machine Learning Techniques for Churn Prediction

Authors: Saurabh Kumar

Abstract:

In the fiercely competitive landscape of e-commerce, understanding and mitigating customer churn has become paramount for sustainable business growth. This paper presents a thorough investigation into the application of machine learning techniques for churn prediction in e-commerce, aiming to provide actionable insights for businesses seeking to enhance customer retention strategies. We conduct a comparative study of various machine learning algorithms, including traditional statistical methods and ensemble techniques, leveraging a rich dataset sourced from Kaggle. Through rigorous evaluation, we assess the predictive performance, interpretability, and scalability of each method, elucidating their respective strengths and limitations in capturing the intricate dynamics of customer churn. We identified the XGBoost classifier to be the best performing. Our findings not only offer practical guidelines for selecting suitable modeling approaches but also contribute to the broader understanding of customer behavior in the e-commerce domain. Ultimately, this research equips businesses with the knowledge and tools necessary to proactively identify and address churn, thereby fostering long-term customer relationships and sustaining competitive advantage.

Keywords: customer churn, e-commerce, machine learning techniques, predictive performance, sustainable business growth

Procedia PDF Downloads 31
7626 Clinical Validation of an Automated Natural Language Processing Algorithm for Finding COVID-19 Symptoms and Complications in Patient Notes

Authors: Karolina Wieczorek, Sophie Wiliams

Abstract:

Introduction: Patient data is often collected in Electronic Health Record Systems (EHR) for purposes such as providing care as well as reporting data. This information can be re-used to validate data models in clinical trials or in epidemiological studies. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. Mentioning a disease in a discharge letter does not necessarily mean that a patient suffers from this disease. Many of them discuss a diagnostic process, different tests, or discuss whether a patient has a certain disease. The COVID-19 dataset in this study used natural language processing (NLP), an automated algorithm which extracts information related to COVID-19 symptoms, complications, and medications prescribed within the hospital. Free-text patient clinical patient notes are rich sources of information which contain patient data not captured in a structured form, hence the use of named entity recognition (NER) to capture additional information. Methods: Patient data (discharge summary letters) were exported and screened by an algorithm to pick up relevant terms related to COVID-19. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. A list of 124 Systematized Nomenclature of Medicine (SNOMED) Clinical Terms has been provided in Excel with corresponding IDs. Two independent medical student researchers were provided with a dictionary of SNOMED list of terms to refer to when screening the notes. They worked on two separate datasets called "A” and "B”, respectively. Notes were screened to check if the correct term had been picked-up by the algorithm to ensure that negated terms were not picked up. Results: Its implementation in the hospital began on March 31, 2020, and the first EHR-derived extract was generated for use in an audit study on June 04, 2020. The dataset has contributed to large, priority clinical trials (including International Severe Acute Respiratory and Emerging Infection Consortium (ISARIC) by bulk upload to REDcap research databases) and local research and audit studies. Successful sharing of EHR-extracted datasets requires communicating the provenance and quality, including completeness and accuracy of this data. The results of the validation of the algorithm were the following: precision (0.907), recall (0.416), and F-score test (0.570). Percentage enhancement with NLP extracted terms compared to regular data extraction alone was low (0.3%) for relatively well-documented data such as previous medical history but higher (16.6%, 29.53%, 30.3%, 45.1%) for complications, presenting illness, chronic procedures, acute procedures respectively. Conclusions: This automated NLP algorithm is shown to be useful in facilitating patient data analysis and has the potential to be used in more large-scale clinical trials to assess potential study exclusion criteria for participants in the development of vaccines.

Keywords: automated, algorithm, NLP, COVID-19

Procedia PDF Downloads 102
7625 A Model of Empowerment Evaluation of Knowledge Management in Private Banks Using Fuzzy Inference System

Authors: Nazanin Pilevari, Kamyar Mahmoodi

Abstract:

The purpose of this research is to provide a model based on fuzzy inference system for evaluating empowerment of Knowledge management. The first prototype of the research was developed based on the study of literature. In the next step, experts were provided with these models and after implementing consensus-based reform, the views of Fuzzy Delphi experts and techniques, components and Index research model were finalized. Culture, structure, IT and leadership were considered as dimensions of empowerment. Then, In order to collect and extract data for fuzzy inference system based on knowledge and Experience, the experts were interviewed. The values obtained from designed fuzzy inference system, made review and assessment of the organization's empowerment of Knowledge management possible. After the design and validation of systems to measure indexes ,empowerment of Knowledge management and inputs into fuzzy inference) in the AYANDEH Bank, a questionnaire was used. In the case of this bank, the system output indicates that the status of empowerment of Knowledge management, culture, organizational structure and leadership are at the moderate level and information technology empowerment are relatively high. Based on these results, the status of knowledge management empowerment in AYANDE Bank, was moderate. Eventually, some suggestions for improving the current situation of banks were provided. According to studies of research history, the use of powerful tools in Fuzzy Inference System for assessment of Knowledge management and knowledge management empowerment such an assessment in the field of banking, are the innovation of this Research.

Keywords: knowledge management, knowledge management empowerment, fuzzy inference system, fuzzy Delphi

Procedia PDF Downloads 361
7624 Multidrug Therapies For HIV: Hybrid On-Off, Hysteresis On-Off Control and Simple STI

Authors: Magno Enrique Mendoza Meza

Abstract:

This paper deals with the comparison of three control techniques: the hysteresis on-off control (HyOOC), the hybrid on-off control (HOOC) and the simple Structured Treatment Interruptions (sSTI). These techniques are applied to the mathematical model developed by Kirschner and Webb. To compare these techniques we use a cost functional that minimize the wild-type virus population and the mutant virus population, but the main objective is to minimize the systemic cost of treatment and maximize levels of healthy CD4+ T cells. HyOOC, HOOC, and sSTI are applied to the drug therapies using a reverse transcriptase and protease inhibitors; simulations show that these controls maintain the uninfected cells in a small, bounded neighborhood of a pre-specified level. The controller HyOOC and HOOC are designed by appropriate choice of virtual equilibrium points.

Keywords: virus dynamics, on-off control, hysteresis, multi-drug therapies

Procedia PDF Downloads 395
7623 Contrast Enhancement of Masses in Mammograms Using Multiscale Morphology

Authors: Amit Kamra, V. K. Jain, Pragya

Abstract:

Mammography is widely used technique for breast cancer screening. There are various other techniques for breast cancer screening but mammography is the most reliable and effective technique. The images obtained through mammography are of low contrast which causes problem for the radiologists to interpret. Hence, a high quality image is mandatory for the processing of the image for extracting any kind of information from it. Many contrast enhancement algorithms have been developed over the years. In the present work, an efficient morphology based technique is proposed for contrast enhancement of masses in mammographic images. The proposed method is based on Multiscale Morphology and it takes into consideration the scale of the structuring element. The proposed method is compared with other state-of-the-art techniques. The experimental results show that the proposed method is better both qualitatively and quantitatively than the other standard contrast enhancement techniques.

Keywords: enhancement, mammography, multi-scale, mathematical morphology

Procedia PDF Downloads 427
7622 Validation of SWAT Model for Prediction of Water Yield and Water Balance: Case Study of Upstream Catchment of Jebba Dam in Nigeria

Authors: Adeniyi G. Adeogun, Bolaji F. Sule, Adebayo W. Salami, Michael O. Daramola

Abstract:

Estimation of water yield and water balance in a river catchment is critical to the sustainable management of water resources at watershed level in any country. Therefore, in the present study, Soil and Water Assessment Tool (SWAT) interfaced with Geographical Information System (GIS) was applied as a tool to predict water balance and water yield of a catchment area in Nigeria. The catchment area, which was 12,992km2, is located upstream Jebba hydropower dam in North central part of Nigeria. In this study, data on the observed flow were collected and compared with simulated flow using SWAT. The correlation between the two data sets was evaluated using statistical measures, such as, Nasch-Sucliffe Efficiency (NSE) and coefficient of determination (R2). The model output shows a good agreement between the observed flow and simulated flow as indicated by NSE and R2, which were greater than 0.7 for both calibration and validation period. A total of 42,733 mm of water was predicted by the calibrated model as the water yield potential of the basin for a simulation period 1985 to 2010. This interesting performance obtained with SWAT model suggests that SWAT model could be a promising tool to predict water balance and water yield in sustainable management of water resources. In addition, SWAT could be applied to other water resources in other basins in Nigeria as a decision support tool for sustainable water management in Nigeria.

Keywords: GIS, modeling, sensitivity analysis, SWAT, water yield, watershed level

Procedia PDF Downloads 441
7621 Innovative Teaching Learning Techniques and Learning Difficulties of Adult Learners in Literacy Education Programmes in Calabar Metropolis, Cross River State, Nigeria

Authors: Simon Ibor Akpama

Abstract:

The study investigated the extent to which innovative teaching-learning techniques can influence and attenuate learning difficulties among adult learners participating in different literacy education programmes in Calabar Metropolis, Cross River State, Nigeria. A quasi-experimental design was adopted to collect data from a sample size of 150 participants of the programme. The sample was drawn using the simple random sampling method. As an experimental study, the 150 participants were divided into two equal groups –the first was the experimental group while the second was the control. A pre-test was administered to the two groups which were later exposed to a post-test after treatment. Two instruments were used for data collection. The first was the guide for the Literacy Learning Difficulties Inventory (LLDI). Three hypotheses were postulated and tested as .05 level of significance using Analysis of Covariance (ANOVA) test statistics. Results of the analysis firstly showed that the two groups (treatment and control) did not differ in the pre-test regarding their literacy learning difficulties. Secondly, the result showed that for each hypothesis, innovative teaching-learning techniques significantly influenced adult learners’ (participants) literacy learning difficulties. Based on these findings, the study recommends the use of innovative teaching-learning techniques in adult literacy education centres to mitigate the learning difficulties of adult learners in literacy education programmes in Calabar Metropolis.

Keywords: teaching, learning, techniques, innovative, difficulties, programme

Procedia PDF Downloads 123
7620 Performance Analysis of Shunt Active Power Filter for Various Reference Current Generation Techniques

Authors: Vishal V. Choudhari, Gaurao A. Dongre, S. P. Diwan

Abstract:

A number of reference current generation have been developed for analysis of shunt active power filter to mitigate the load compensation. Depending upon the type of load the technique has to be chosen. In this paper, six reference current generation techniques viz. instantaneous reactive power theory(IRP), Synchronous reference frame theory(SRF), Perfect harmonic cancellation(PHC), Unity power factor method(UPF), Self-tuning filter method(STF), Predictive filtering method(PFM) are compared for different operating conditions. The harmonics are introduced because of non-linear loads in the system. These harmonics are eliminated using above techniques. The results and performance of system simulated on MATLAB/Simulink platform. The system is experimentally implemented using DS1104 card of dSPACE system.

Keywords: SAPF, power quality, THD, IRP, SRF, dSPACE module DS1104

Procedia PDF Downloads 592