Search results for: harmony search algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3799

Search results for: harmony search algorithms

649 Predicting Radioactive Waste Glass Viscosity, Density and Dissolution with Machine Learning

Authors: Joseph Lillington, Tom Gout, Mike Harrison, Ian Farnan

Abstract:

The vitrification of high-level nuclear waste within borosilicate glass and its incorporation within a multi-barrier repository deep underground is widely accepted as the preferred disposal method. However, for this to happen, any safety case will require validation that the initially localized radionuclides will not be considerably released into the near/far-field. Therefore, accurate mechanistic models are necessary to predict glass dissolution, and these should be robust to a variety of incorporated waste species and leaching test conditions, particularly given substantial variations across international waste-streams. Here, machine learning is used to predict glass material properties (viscosity, density) and glass leaching model parameters from large-scale industrial data. A variety of different machine learning algorithms have been compared to assess performance. Density was predicted solely from composition, whereas viscosity additionally considered temperature. To predict suitable glass leaching model parameters, a large simulated dataset was created by coupling MATLAB and the chemical reactive-transport code HYTEC, considering the state-of-the-art GRAAL model (glass reactivity in allowance of the alteration layer). The trained models were then subsequently applied to the large-scale industrial, experimental data to identify potentially appropriate model parameters. Results indicate that ensemble methods can accurately predict viscosity as a function of temperature and composition across all three industrial datasets. Glass density prediction shows reliable learning performance with predictions primarily being within the experimental uncertainty of the test data. Furthermore, machine learning can predict glass dissolution model parameters behavior, demonstrating potential value in GRAAL model development and in assessing suitable model parameters for large-scale industrial glass dissolution data.

Keywords: machine learning, predictive modelling, pattern recognition, radioactive waste glass

Procedia PDF Downloads 96
648 Comparative Study of Skeletonization and Radial Distance Methods for Automated Finger Enumeration

Authors: Mohammad Hossain Mohammadi, Saif Al Ameri, Sana Ziaei, Jinane Mounsef

Abstract:

Automated enumeration of the number of hand fingers is widely used in several motion gaming and distance control applications, and is discussed in several published papers as a starting block for hand recognition systems. The automated finger enumeration technique should not only be accurate, but also must have a fast response for a moving-picture input. The high performance of video in motion games or distance control will inhibit the program’s overall speed, for image processing software such as Matlab need to produce results at high computation speeds. Since an automated finger enumeration with minimum error and processing time is desired, a comparative study between two finger enumeration techniques is presented and analyzed in this paper. In the pre-processing stage, various image processing functions were applied on a real-time video input to obtain the final cleaned auto-cropped image of the hand to be used for the two techniques. The first technique uses the known morphological tool of skeletonization to count the number of skeleton’s endpoints for fingers. The second technique uses a radial distance method to enumerate the number of fingers in order to obtain a one dimensional hand representation. For both discussed methods, the different steps of the algorithms are explained. Then, a comparative study analyzes the accuracy and speed of both techniques. Through experimental testing in different background conditions, it was observed that the radial distance method was more accurate and responsive to a real-time video input compared to the skeletonization method. All test results were generated in Matlab and were based on displaying a human hand for three different orientations on top of a plain color background. Finally, the limitations surrounding the enumeration techniques are presented.

Keywords: comparative study, hand recognition, fingertip detection, skeletonization, radial distance, Matlab

Procedia PDF Downloads 363
647 Consumer Behavior and the Demand for Sustainable Buildings in an Emerging Market: The Example of Brazil

Authors: Vinícius L. L. Morrone, David Douek, Helder M. F. Pereira, Bernadete L. M. Grandolpho

Abstract:

This work aimed to identify the relationships between the level of consumer environmental awareness and their search for sustainable properties, as well as to understand the main sustainability structures considered by these consumers during the decision process. Additionally, the paper looked up to the influence environmental awareness and financial status have over the disposition of buyers to pay more for sustainable properties. To achieve these objectives, 318 questionnaires were answered electronically, after being sent to the Green Building Brazil email basis, as to other Real Estate developers client basis. From all the questionnaires answered, 71 were discarded, leaving a total amount of 247 admitted questionnaires to be analyzed. The responses were evaluated based on the theory of consumer decision making, especially on the influence factors of this process. The data were processed using a PLS model, using the R software. The results have shown that the level of consumer environmental awareness effectively affects the consumer’s will of acquiring a sustainable property or, at least, a property with some environmental friendly structures. The consumer’s environmental awareness also positively impacts the importance consumers give to individual environmental friendly structures. Also, as a consumer value to those individual structures raises, it is also observed a raise in his will to buy a sustainable property. Additionally, the impact of consumer’s environmental awareness and financial status over the willingness to pay more for a property with those attributes. The results indicate that there was no relationship between consumers' environmental awareness and their willingness to pay more for a sustainable property. On the other hand, the financial status and the family income of the consumers showed a positive relation with the willingness to pay more for a sustainable property. This indicates that consumers with better financial conditions, which according to the analysis do not necessarily have a greater environmental awareness, are those who are willing to pay more for a sustainable property. Thus, this study indicates that, even if the environmental awareness impact positively the demand for sustainable structures and properties, this impact is not price reflected, due to the price elasticity of the consumption, especially for a category of lower income consumers. This paper adds to the literature in the way it projects some guidelines to the consumer’s decision process in the Real Estate market in emerging economies, as well as it presents some drivers to pricing decisions.

Keywords: consumer behavior, environmental awareness, real estate pricing, sustainable buildings

Procedia PDF Downloads 167
646 Toward Indoor and Outdoor Surveillance using an Improved Fast Background Subtraction Algorithm

Authors: El Harraj Abdeslam, Raissouni Naoufal

Abstract:

The detection of moving objects from a video image sequences is very important for object tracking, activity recognition, and behavior understanding in video surveillance. The most used approach for moving objects detection / tracking is background subtraction algorithms. Many approaches have been suggested for background subtraction. But, these are illumination change sensitive and the solutions proposed to bypass this problem are time consuming. In this paper, we propose a robust yet computationally efficient background subtraction approach and, mainly, focus on the ability to detect moving objects on dynamic scenes, for possible applications in complex and restricted access areas monitoring, where moving and motionless persons must be reliably detected. It consists of three main phases, establishing illumination changes in variance, background/foreground modeling and morphological analysis for noise removing. We handle illumination changes using Contrast Limited Histogram Equalization (CLAHE), which limits the intensity of each pixel to user determined maximum. Thus, it mitigates the degradation due to scene illumination changes and improves the visibility of the video signal. Initially, the background and foreground images are extracted from the video sequence. Then, the background and foreground images are separately enhanced by applying CLAHE. In order to form multi-modal backgrounds we model each channel of a pixel as a mixture of K Gaussians (K=5) using Gaussian Mixture Model (GMM). Finally, we post process the resulting binary foreground mask using morphological erosion and dilation transformations to remove possible noise. For experimental test, we used a standard dataset to challenge the efficiency and accuracy of the proposed method on a diverse set of dynamic scenes.

Keywords: video surveillance, background subtraction, contrast limited histogram equalization, illumination invariance, object tracking, object detection, behavior understanding, dynamic scenes

Procedia PDF Downloads 237
645 Low-Surface Roughness and High Optical Quality CdS Thin Film Deposited on Heated Substrate Using Room-Temperature Chemical Solution

Authors: A. Elsayed, M. H. Dewaidar, M. Ghali, M. Elkemary

Abstract:

The high production cost of the conventional solar cells requires the search for economic methods suitable for solar energy conversion. Cadmium Sulfide (CdS) is one of the most important semiconductors used in photovoltaics, especially in large area solar cells; and can be prepared in a thin film form by a wide variety of deposition techniques. The preparation techniques include vacuum evaporation, sputtering and molecular beam epitaxy. Other techniques, based on chemical solutions, are also used for depositing CdS films with dramatically low-cost compared to other vacuum-based methods. Although this technique is widely used during the last decades, due to simplicity and low-deposition temperature (~100°C), there is still a strong need for more information on the growth process and its relation with the quality of the deposited films. Here, we report on deposition of high-quality CdS thin films; with low-surface roughness ( < 3.0 nm) and sharp optical absorption edge; on low-temperature glass substrates (70°C) using a new method based on the room-temperature chemical solution. In this method, a mixture solution of cadmium acetate and thiourea at room temperature was used under special growth conditions for deposition of CdS films. X-ray diffraction (XRD) measurements were used to examine the crystal structure properties of the deposited CdS films. In addition, UV-VIS transmittance and low-temperature (4K) photoluminescence (PL) measurements were performed for quantifying optical properties of the deposited films. The deposited films show high optical quality as confirmed by observation of both, sharp edge in the transmittance spectra and strong PL intensity at room temperature. Furthermore, we found a strong effect of the growth conditions on the optical band gap of the deposited films; where remarkable red-shift in the absorption edge with temperature is clearly seen in both transmission and PL spectra. Such tuning of both optical band gap of the deposited CdS films can be utilized for tuning the electronic bands' alignments between CdS and other light-harvesting materials, like CuInGaSe or CdTe, for potential improvement in the efficiency of solar cells devices based on these heterostructures.

Keywords: chemical deposition, CdS, optical properties, surface, thin film

Procedia PDF Downloads 148
644 Diabetes Care in Detention Settings: A Systematic Review

Authors: A. Papachristou, A. Ntikoudi, L. Makris, V. Saridakis

Abstract:

Introduction: More than 10 million people are imprisoned or detained worldwide. Figures from 2011-12 show that prison inmates are more likely than the general population to suffer from chronic or infectious diseases, while most inmates are overweight or obese, and more than a quarter have high blood pressure. In 2011/12, the proportion of prisoners reporting diabetes or hyperglycemia was 899 per 10,000 prisoners, almost double the 2004 figure (483 per 10,000). It is important to ensure that this population has access to the same standard of care as people outside prisons, as access to services should be need-based. Diabetes is a public health problem associated with increased morbidity and mortality worldwide. According to the International Diabetes Federation (IDF) in 2017, approximately 425 million people worldwide had diabetes. This number is expected to increase to 629 million by 2045. Poor management of diabetes in prisons can lead to poor blood sugar control and increase the risk of complications. Aim: The aim of this review was to systematically evaluate all the available literature on diabetes care in custodial settings. Methods: An extensive literature search was conducted through electronic databases (PubMed, Scopus and CINAHL) with the terms ‘custody’, ‘diabetes Mellitus, ‘detention centers and ‘chronic disease’. Articles published in English until September 2022, were included; no other criteria on publication dates were set. Results: Most of the studies mentioned a diabetes prevalence of approximately 10%, among other common chronic. Hypertension, obesity, smoking, sedentary lifestyle were the most common comorbidities associated with diabetes. Conclusion: Good glycemic control is fundamental to managing diabetes, and while many prisoners enter prison poorly, access to regular medication and meals, as well as exercise, offers the potential for improvement. Not being able to get help as quickly as in the past can be extremely stressful, and some prisoners may deliberately raise their blood sugar levels to avoid the risk of developing hypoglycemia, especially if they know they have had previous episodes of nocturnal hypoglycemia. Thus, appropriate training and resources are critical to providing quality care to incarcerated people with diabetes.

Keywords: custody, diabetes mellitus, detention centers, chronic disease

Procedia PDF Downloads 77
643 Health Information Needs and Utilization of Information and Communication Technologies by Medical Professionals in a Northern City of India

Authors: Sonika Raj, Amarjeet Singh, Vijay Lakshmi Sharma

Abstract:

Introduction: In 21st century, due to revolution in Information and Communication Technologies (ICTs), there has been phenomenal development in quality and quantity of knowledge in the field of medical science. So, the access to relevant information to physicians is critical to the delivery of effective healthcare services to patients. The study was conducted to assess the information needs and attitudes of the medical professionals; to determine the sources and channels of information used by them; to ascertain the current usage of ICTs and the barriers faced by them in utilization of ICTs in health information access. Methodology: This descriptive cross-sectional study was carried in 2015 on hundred medical professionals working in public and private sectors of Chandigarh. The study used both quantitative and qualitative method for data collection. A semi structured questionnaire and interview schedule was used to collect data on information seeking needs, access to ICTs and barriers to healthcare information access. Five Data analysis was done using SPSS-16 and qualitative data was analyzed using thematic approach. Results: The most preferred sources to access healthcare information were internet (85%), trainings (61%) and communication with colleagues (57%). They wanted information on new drug therapy and latest developments in respective fields. All had access to computer with but almost half assessed their computer knowledge as average and only 3% had received training regarding usage. Educational status (p=0.004), place of work (p=0.004), number of years in job (p=0.004) and sector of job (p=0.04) of doctors were found to be significantly associated with their active search for information. The major themes that emerged from in-views were need; types and sources of healthcare information; exchange of information among different levels of healthcare providers; usage of ICTs to obtain and share information; barriers to access of healthcare information and quality of health information materials and involvement in their development process Conclusion and Recommendations: The medical professionals need information in their in their due course of work. However, information needs of medical professionals were not being adequately met. There should be training of professional regarding internet skills and the course on bioinformatics should be incorporated in the curricula of medical students. The policy framework must be formulated that will encourage and promote the use of ICTs as tools for health information access and dissemination.

Keywords: health information, ICTs, medical professionals, qualitative

Procedia PDF Downloads 329
642 Infrared Spectroscopy in Tandem with Machine Learning for Simultaneous Rapid Identification of Bacteria Isolated Directly from Patients' Urine Samples and Determination of Their Susceptibility to Antibiotics

Authors: Mahmoud Huleihel, George Abu-Aqil, Manal Suleiman, Klaris Riesenberg, Itshak Lapidot, Ahmad Salman

Abstract:

Urinary tract infections (UTIs) are considered to be the most common bacterial infections worldwide, which are caused mainly by Escherichia (E.) coli (about 80%). Klebsiella pneumoniae (about 10%) and Pseudomonas aeruginosa (about 6%). Although antibiotics are considered as the most effective treatment for bacterial infectious diseases, unfortunately, most of the bacteria already have developed resistance to the majority of the commonly available antibiotics. Therefore, it is crucial to identify the infecting bacteria and to determine its susceptibility to antibiotics for prescribing effective treatment. Classical methods are time consuming, require ~48 hours for determining bacterial susceptibility. Thus, it is highly urgent to develop a new method that can significantly reduce the time required for determining both infecting bacterium at the species level and diagnose its susceptibility to antibiotics. Fourier-Transform Infrared (FTIR) spectroscopy is well known as a sensitive and rapid method, which can detect minor molecular changes in bacterial genome associated with the development of resistance to antibiotics. The main goal of this study is to examine the potential of FTIR spectroscopy, in tandem with machine learning algorithms, to identify the infected bacteria at the species level and to determine E. coli susceptibility to different antibiotics directly from patients' urine in about 30minutes. For this goal, 1600 different E. coli isolates were isolated for different patients' urine sample, measured by FTIR, and analyzed using different machine learning algorithm like Random Forest, XGBoost, and CNN. We achieved 98% success in isolate level identification and 89% accuracy in susceptibility determination.

Keywords: urinary tract infections (UTIs), E. coli, Klebsiella pneumonia, Pseudomonas aeruginosa, bacterial, susceptibility to antibiotics, infrared microscopy, machine learning

Procedia PDF Downloads 146
641 Using Arts in ESL Classroom

Authors: Nazia Shehzad

Abstract:

Language and art can supplement and correlate each other. Through the ages art has been a means of visual expression used to convey a wide series of incarnated ideas. Art can take the perceiver into different times and into different worlds. It can also be used to introduce different levels of vocabulary to the learners of a second language. Learning a second language for most students is a very difficult and strenuous experience. They are not only trying to accommodate to a new language but are also trying to adjust to themselves and a new environment. They are anxious about almost everything, but they are especially self-conscious about their performance in the classroom. By relocating the focus from the student to an object, everyone participates, thus waiving a certain degree of self-consciousness. The experience, a student has with art in the classroom has to be gratifying for both the student and the teacher. If the atmosphere in the classroom is too grave it will not serve any useful purpose. Art is an excellent way to teach English and encourage collaboration and interaction between students of all ages. As making art involves many different processes, it is wonderful for classification and following/giving instructions. It is also an effective way to achieve and implement language of characterization and comparison and vocabulary acquirement for the elements of design (shape, size, color, texture, tone etc.) is so much more entertaining if done in a practical and hands-on way. Expressing ideas and feelings through art is also of immeasurable value where students are at the beginning stages of English language acquisition and for many of my Saudi students it was a form of therapy. It is also a way to respect, search, examine and share the cultural traditions of different cultures, and of the students themselves. Art not only provides a field for ideas to keep aimless, meandering minds of students' busy but is also a productive tool to analyze English language in a new order. As an ESL teacher, using art is a highly compelling way to bridge the gap between student and teacher. It’s difficult to keep students concentrated, especially when they speak a different language. To get students to actually learn and explore something in your foreign language lesson, artwork is your best friend. Many teachers feel that through amalgamation of the arts into their academic lessons students are able to learn more profoundly because they use diverse ways of thinking and problem solving. Teachers observe that drawing often retains students who might otherwise be dispassionate and can help students move ahead simple recall when they are asked to make connections and come up with an exclusive interpretation through an artwork or drawing. Students use observation skills when they are drawing, and this can help to persuade students who might otherwise remain silent or need more time to process information.

Keywords: amalgamation of arts, expressing ideas and feelings through arts, effective way to achieve and implement language, language and art can supplement and correlate each other

Procedia PDF Downloads 343
640 In Search of Commonalities in the Determinants of Child Sex Ratios in India and People's of Republic of China

Authors: Suddhasil Siddhanta, Debasish Nandy

Abstract:

Child sex ratios pattern in the Asian Population is highly masculine mainly due to birth masculinity and gender bias in child mortality. The vast and the growing literature of female deficit in world population points out the diffusion of child sex ratio pattern in many Asian as well as neighboring European countries. However, little attention has been given to understand the common factors in different demographics in explaining child sex ratio pattern. Such a scholarship is extremely important as level of gender inequity is different in different country set up. Our paper tries to explain the major structural commonalities in the child masculinity pattern in two demographic billionaires - India and China. The analysis reveals that apart from geographical diffusion of sex selection technology, patrilocal social structure, as proxied by households with more than one generation in China and proportion of population aged 65 years and above in India, can explain significant variation of missing girl child in these two countries. Even after controlling for individual capacity building factors like educational attainment, or work force participation, the measure of social stratification is coming out to be the major determinant of child sex ratio variation. Other socio economic factors that perform much well are the agency building factors of the females, like changing pattern of marriage customs which is proxied by divorce and remarriage ratio for china and percentage of female marrying at or after the age of 20 years in India and the female workforce participation. Proportion of minorities in socio-religious composition of the population and gender bias in scholastic attainment in both these counties are also found to be significant in modeling child sex ratio variations. All these significant common factors associated with child sex ratio point toward the one single most important factor: the historical evolution of patriarchy and its contemporary perpetuation in both the countries. It seems that prohibition of sex selection might not be sufficient to combat the peculiar skewness of excessive maleness in child population in both these countries. Demand sided policies is therefore utmost important to root out the gender bias in child sex ratios.

Keywords: child sex ratios, gender bias, structural factors, prosperity, patrilocality

Procedia PDF Downloads 143
639 Structural Design Optimization of Reinforced Thin-Walled Vessels under External Pressure Using Simulation and Machine Learning Classification Algorithm

Authors: Lydia Novozhilova, Vladimir Urazhdin

Abstract:

An optimization problem for reinforced thin-walled vessels under uniform external pressure is considered. The conventional approaches to optimization generally start with pre-defined geometric parameters of the vessels, and then employ analytic or numeric calculations and/or experimental testing to verify functionality, such as stability under the projected conditions. The proposed approach consists of two steps. First, the feasibility domain will be identified in the multidimensional parameter space. Every point in the feasibility domain defines a design satisfying both geometric and functional constraints. Second, an objective function defined in this domain is formulated and optimized. The broader applicability of the suggested methodology is maximized by implementing the Support Vector Machines (SVM) classification algorithm of machine learning for identification of the feasible design region. Training data for SVM classifier is obtained using the Simulation package of SOLIDWORKS®. Based on the data, the SVM algorithm produces a curvilinear boundary separating admissible and not admissible sets of design parameters with maximal margins. Then optimization of the vessel parameters in the feasibility domain is performed using the standard algorithms for the constrained optimization. As an example, optimization of a ring-stiffened closed cylindrical thin-walled vessel with semi-spherical caps under high external pressure is implemented. As a functional constraint, von Mises stress criterion is used but any other stability constraint admitting mathematical formulation can be incorporated into the proposed approach. Suggested methodology has a good potential for reducing design time for finding optimal parameters of thin-walled vessels under uniform external pressure.

Keywords: design parameters, feasibility domain, von Mises stress criterion, Support Vector Machine (SVM) classifier

Procedia PDF Downloads 307
638 Development of a Computer Aided Diagnosis Tool for Brain Tumor Extraction and Classification

Authors: Fathi Kallel, Abdulelah Alabd Uljabbar, Abdulrahman Aldukhail, Abdulaziz Alomran

Abstract:

The brain is an important organ in our body since it is responsible about the majority actions such as vision, memory, etc. However, different diseases such as Alzheimer and tumors could affect the brain and conduct to a partial or full disorder. Regular diagnosis are necessary as a preventive measure and could help doctors to early detect a possible trouble and therefore taking the appropriate treatment, especially in the case of brain tumors. Different imaging modalities are proposed for diagnosis of brain tumor. The powerful and most used modality is the Magnetic Resonance Imaging (MRI). MRI images are analyzed by doctor in order to locate eventual tumor in the brain and describe the appropriate and needed treatment. Diverse image processing methods are also proposed for helping doctors in identifying and analyzing the tumor. In fact, a large Computer Aided Diagnostic (CAD) tools including developed image processing algorithms are proposed and exploited by doctors as a second opinion to analyze and identify the brain tumors. In this paper, we proposed a new advanced CAD for brain tumor identification, classification and feature extraction. Our proposed CAD includes three main parts. Firstly, we load the brain MRI. Secondly, a robust technique for brain tumor extraction is proposed. This technique is based on both Discrete Wavelet Transform (DWT) and Principal Component Analysis (PCA). DWT is characterized by its multiresolution analytic property, that’s why it was applied on MRI images with different decomposition levels for feature extraction. Nevertheless, this technique suffers from a main drawback since it necessitates a huge storage and is computationally expensive. To decrease the dimensions of the feature vector and the computing time, PCA technique is considered. In the last stage, according to different extracted features, the brain tumor is classified into either benign or malignant tumor using Support Vector Machine (SVM) algorithm. A CAD tool for brain tumor detection and classification, including all above-mentioned stages, is designed and developed using MATLAB guide user interface.

Keywords: MRI, brain tumor, CAD, feature extraction, DWT, PCA, classification, SVM

Procedia PDF Downloads 225
637 Dynamic Reliability for a Complex System and Process: Application on Offshore Platform in Mozambique

Authors: Raed KOUTA, José-Alcebiades-Ernesto HLUNGUANE, Eric Châtele

Abstract:

The search for and exploitation of new fossil energy resources is taking place in the context of the gradual depletion of existing deposits. Despite the adoption of international targets to combat global warming, the demand for fuels continues to grow, contradicting the movement towards an energy-efficient society. The increase in the share of offshore in global hydrocarbon production tends to compensate for the depletion of terrestrial reserves, thus constituting a major challenge for the players in the sector. Through the economic potential it represents, and the energy independence it provides, offshore exploitation is also a challenge for States such as Mozambique, which have large maritime areas and whose environmental wealth must be considered. The exploitation of new reserves on economically viable terms depends on available technologies. The development of deep and ultra-deep offshore requires significant research and development efforts. Progress has also been made in managing the multiple risks inherent in this activity. Our study proposes a reliability approach to develop products and processes designed to live at sea. Indeed, the context of an offshore platform requires highly reliable solutions to overcome the difficulties of access to the system for regular maintenance and quick repairs and which must resist deterioration and degradation processes. One of the characteristics of failures that we consider is the actual conditions of use that are considered 'extreme.' These conditions depend on time and the interactions between the different causes. These are the two factors that give the degradation process its dynamic character, hence the need to develop dynamic reliability models. Our work highlights mathematical models that can explicitly manage interactions between components and process variables. These models are accompanied by numerical resolution methods that help to structure a dynamic reliability approach in a physical and probabilistic context. The application developed makes it possible to evaluate the reliability, availability, and maintainability of a floating storage and unloading platform for liquefied natural gas production.

Keywords: dynamic reliability, offshore plateform, stochastic process, uncertainties

Procedia PDF Downloads 102
636 Indian Premier League (IPL) Score Prediction: Comparative Analysis of Machine Learning Models

Authors: Rohini Hariharan, Yazhini R, Bhamidipati Naga Shrikarti

Abstract:

In the realm of cricket, particularly within the context of the Indian Premier League (IPL), the ability to predict team scores accurately holds significant importance for both cricket enthusiasts and stakeholders alike. This paper presents a comprehensive study on IPL score prediction utilizing various machine learning algorithms, including Support Vector Machines (SVM), XGBoost, Multiple Regression, Linear Regression, K-nearest neighbors (KNN), and Random Forest. Through meticulous data preprocessing, feature engineering, and model selection, we aimed to develop a robust predictive framework capable of forecasting team scores with high precision. Our experimentation involved the analysis of historical IPL match data encompassing diverse match and player statistics. Leveraging this data, we employed state-of-the-art machine learning techniques to train and evaluate the performance of each model. Notably, Multiple Regression emerged as the top-performing algorithm, achieving an impressive accuracy of 77.19% and a precision of 54.05% (within a threshold of +/- 10 runs). This research contributes to the advancement of sports analytics by demonstrating the efficacy of machine learning in predicting IPL team scores. The findings underscore the potential of advanced predictive modeling techniques to provide valuable insights for cricket enthusiasts, team management, and betting agencies. Additionally, this study serves as a benchmark for future research endeavors aimed at enhancing the accuracy and interpretability of IPL score prediction models.

Keywords: indian premier league (IPL), cricket, score prediction, machine learning, support vector machines (SVM), xgboost, multiple regression, linear regression, k-nearest neighbors (KNN), random forest, sports analytics

Procedia PDF Downloads 27
635 Radar Track-based Classification of Birds and UAVs

Authors: Altilio Rosa, Chirico Francesco, Foglia Goffredo

Abstract:

In recent years, the number of Unmanned Aerial Vehicles (UAVs) has significantly increased. The rapid development of commercial and recreational drones makes them an important part of our society. Despite the growing list of their applications, these vehicles pose a huge threat to civil and military installations: detection, classification and neutralization of such flying objects become an urgent need. Radar is an effective remote sensing tool for detecting and tracking flying objects, but scenarios characterized by the presence of a high number of tracks related to flying birds make especially challenging the drone detection task: operator PPI is cluttered with a huge number of potential threats and his reaction time can be severely affected. Flying birds compared to UAVs show similar velocity, RADAR cross-section and, in general, similar characteristics. Building from the absence of a single feature that is able to distinguish UAVs and birds, this paper uses a multiple features approach where an original feature selection technique is developed to feed binary classifiers trained to distinguish birds and UAVs. RADAR tracks acquired on the field and related to different UAVs and birds performing various trajectories were used to extract specifically designed target movement-related features based on velocity, trajectory and signal strength. An optimization strategy based on a genetic algorithm is also introduced to select the optimal subset of features and to estimate the performance of several classification algorithms (Neural network, SVM, Logistic regression…) both in terms of the number of selected features and misclassification error. Results show that the proposed methods are able to reduce the dimension of the data space and to remove almost all non-drone false targets with a suitable classification accuracy (higher than 95%).

Keywords: birds, classification, machine learning, UAVs

Procedia PDF Downloads 195
634 Level Set Based Extraction and Update of Lake Contours Using Multi-Temporal Satellite Images

Authors: Yindi Zhao, Yun Zhang, Silu Xia, Lixin Wu

Abstract:

The contours and areas of water surfaces, especially lakes, often change due to natural disasters and construction activities. It is an effective way to extract and update water contours from satellite images using image processing algorithms. However, to produce optimal water surface contours that are close to true boundaries is still a challenging task. This paper compares the performances of three different level set models, including the Chan-Vese (CV) model, the signed pressure force (SPF) model, and the region-scalable fitting (RSF) energy model for extracting lake contours. After experiment testing, it is indicated that the RSF model, in which a region-scalable fitting (RSF) energy functional is defined and incorporated into a variational level set formulation, is superior to CV and SPF, and it can get desirable contour lines when there are “holes” in the regions of waters, such as the islands in the lake. Therefore, the RSF model is applied to extracting lake contours from Landsat satellite images. Four temporal Landsat satellite images of the years of 2000, 2005, 2010, and 2014 are used in our study. All of them were acquired in May, with the same path/row (121/036) covering Xuzhou City, Jiangsu Province, China. Firstly, the near infrared (NIR) band is selected for water extraction. Image registration is conducted on NIR bands of different temporal images for information update, and linear stretching is also done in order to distinguish water from other land cover types. Then for the first temporal image acquired in 2000, lake contours are extracted via the RSF model with initialization of user-defined rectangles. Afterwards, using the lake contours extracted the previous temporal image as the initialized values, lake contours are updated for the current temporal image by means of the RSF model. Meanwhile, the changed and unchanged lakes are also detected. The results show that great changes have taken place in two lakes, i.e. Dalong Lake and Panan Lake, and RSF can actually extract and effectively update lake contours using multi-temporal satellite image.

Keywords: level set model, multi-temporal image, lake contour extraction, contour update

Procedia PDF Downloads 342
633 The Role and Effects of Communication on Occupational Safety: A Review

Authors: Pieter A. Cornelissen, Joris J. Van Hoof

Abstract:

The interest in improving occupational safety started almost simultaneously with the beginning of the Industrial Revolution. Yet, it was not until the late 1970’s before the role of communication was considered in scientific research regarding occupational safety. In recent years the importance of communication as a means to improve occupational safety has increased. Not only as communication might have a direct effect on safety performance and safety outcomes, but also as it can be viewed as a major component of other important safety-related elements (e.g., training, safety meetings, leadership). And while safety communication is an increasingly important topic in research, its operationalization is often vague and differs among studies. This is not only problematic when comparing results, but also in applying these results to practice and the work floor. By means of an in-depth analysis—building on an existing dataset—this review aims to overcome these problems. The initial database search yielded 25.527 articles, which was reduced to a research corpus of 176 articles. Focusing on the 37 articles of this corpus that addressed communication (related to safety outcomes and safety performance), the current study will provide a comprehensive overview of the role and effects of safety communication and outlines the conditions under which communication contributes to a safer work environment. The study shows that in literature a distinction is commonly made between safety communication (i.e., the exchange or dissemination of safety-related information) and feedback (i.e. a reactive form of communication). And although there is a consensus among researchers that both communication and feedback positively affect safety performance, there is a debate about the directness of this relationship. Whereas some researchers assume a direct relationship between safety communication and safety performance, others state that this relationship is mediated by safety climate. One of the key findings is that despite the strongly present view that safety communication is a formal and top-down safety management tool, researchers stress the importance of open communication that encourages and allows employees to express their worries, experiences, views, and share information. This raises questions with regard to other directions (e.g., bottom-up, horizontal) and forms of communication (e.g., informal). The current review proposes a framework to overcome the often vague and different operationalizations of safety communication. The proposed framework can be used to characterize safety communication in terms of stakeholders, direction, and characteristics of communication (e.g., medium usage).

Keywords: communication, feedback, occupational safety, review

Procedia PDF Downloads 278
632 Climate Species Lists: A Combination of Methods for Urban Areas

Authors: Andrea Gion Saluz, Tal Hertig, Axel Heinrich, Stefan Stevanovic

Abstract:

Higher temperatures, seasonal changes in precipitation, and extreme weather events are increasingly affecting trees. To counteract the increasing challenges of urban trees, strategies are increasingly being sought to preserve existing tree populations on the one hand and to prepare for the coming years on the other. One such strategy lies in strategic climate tree species selection. The search is on for species or varieties that can cope with the new climatic conditions. Many efforts in German-speaking countries deal with this in detail, such as the tree lists of the German Conference of Garden Authorities (GALK), the project Stadtgrün 2021, or the instruments of the Climate Species Matrix by Prof. Dr. Roloff. In this context, different methods for a correct species selection are offered. One possibility is to select certain physiological attributes that indicate the climate resilience of a species. To calculate the dissimilarity of the present climate of different geographic regions in relation to the future climate of any city, a weighted (standardized) Euclidean distance (SED) for seasonal climate values is calculated for each region of the Earth. The calculation was performed in the QGIS geographic information system, using global raster datasets on monthly climate values in the 1981-2010 standard period. Data from a European forest inventory were used to identify tree species growing in the calculated analogue climate regions. The inventory used is the compilation of georeferenced point data at a 1 km grid resolution on the occurrence of tree species in 21 European countries. In this project, the results of the methodological application are shown for the city of Zurich for the year 2060. In the first step, analog climate regions based on projected climate values for the measuring station Kirche Fluntern (ZH) were searched for. In a further step, the methods mentioned above were applied to generate tree species lists for the city of Zurich. These lists were then qualitatively evaluated with respect to the suitability of the different tree species for the Zurich area to generate a cleaned and thus usable list of possible future tree species.

Keywords: climate change, climate region, climate tree, urban tree

Procedia PDF Downloads 81
631 Determining the Presence of Brucella abortus Antibodies by the Indirect Elisa Method in Bovine Bulk Milk and Risk Factors in the Peri-Urban Zones of Bamenda Cameroon

Authors: Cha-ah C. N., Awah N. J., Mouiche M. M. M.

Abstract:

Brucellosis is a neglected zoonotic disease of animals and man caused by bacteria of genus Brucella. Though eradicated in some parts of the world, it remains endemic in sub-Saharan Africa including Cameroon. The aim of this study was to contribute to the epidemiology of brucellosis in the North-West region of Cameroon by detecting the presence of anti-Brucella antibodies in bovine bulk milk as this serves as a route of transmission from animals to man. A cross sectional study was conducted to determine the prevalence of Brucella abortus antibodies in bovine bulk milk in the peri-urban zones of Bamenda. One hundred bulk milk samples were collected from 100 herds and tested by milk I-ELISA test. The conducted study revealed the presence of anti-Brucella abortus antibodies in bovine bulk milk. The study revealed that bovine brucellosis is widespread in animal production systems in this area. The animal infection pressure in these systems has remained strong due to movement of livestock in search of pasture, co-existence of animal husbandry, communal sharing of grazing land, concentration of animals around water points, abortions in production systems, locality of production systems and failure to quarantine upon introduction of new animals. The circulation of Brucella abortus antibodies in cattle farms recorded in the study revealed potential public health implication and suggest economic importance of brucellosis to the cattle industry in the Northwest region of Cameroon. The risk for re-emergence and transmission of brucellosis is evident as a result of the co-existence of animal husbandry activities and social-cultural activities that promote brucellosis transmission. Well-designed countrywide, evidence-based studies of brucellosis are needed. These could help to generate reliable frequency and potential impact estimates, to identify Brucella reservoirs, and to propose control strategies of proven efficacy.

Keywords: brucellosis, bulk milk, northwest region Cameroon, prevalence

Procedia PDF Downloads 126
630 A Review of Lexical Retrieval Intervention in Primary Progressive Aphasia and Alzheimer's Disease: Mechanisms of Change, Cognition, and Generalisation

Authors: Ashleigh Beales, Anne Whitworth, Jade Cartwright

Abstract:

Background: While significant benefits of lexical retrieval intervention are evident within the Primary Progressive Aphasia (PPA) and Alzheimer’s disease (AD) literature, an understanding of the mechanisms that underlie change or improvement is limited. Change mechanisms have been explored in the non-progressive post-stroke literature that may offer insight into how interventions affect change with progressive language disorders. The potential influences of cognitive factors may also play a role here, interacting with the aims of intervention. Exploring how such processes have been applied is likely to grow our understanding of how interventions have, or have not, been effective, and how and why generalisation is likely, or not, to occur. Aims: This review of the literature aimed to (1) investigate the proposed mechanisms of change which underpin lexical interventions, mapping the PPA and AD lexical retrieval literature to theoretical accounts of mechanisms that underlie change within the broader intervention literature, (2) identify whether and which nonlinguistic cognitive functions have been engaged in intervention with these populations and any proposed influence, and (3) explore evidence of linguistic generalisation, with particular reference to change mechanisms employed in interventions. Main contribution: A search of Medline, PsycINFO, and CINAHL identified 36 articles that reported data for individuals with PPA or AD following lexical retrieval intervention. A review of the mechanisms of change identified 10 studies that used stimulation, 21 studies utilised relearning, three studies drew on reorganisation, and two studies used cognitive-relay. Significant treatment gains, predominantly based on linguistic performance measures, were reported for all client groups for each of the proposed mechanisms. Reorganisation and cognitive-relay change mechanisms were only targeted in PPA. Eighteen studies incorporated nonlinguistic cognitive functions in intervention; these were limited to autobiographical memory (16 studies), episodic memory (three studies), or both (one study). Linguistic generalisation outcomes were inconsistently reported in PPA and AD studies. Conclusion: This review highlights that individuals with PPA and AD may benefit from lexical retrieval intervention, irrespective of the mechanism of change. Thorough application of a theory of intervention is required to gain a greater understanding of the change mechanisms, as well as the interplay of nonlinguistic cognitive functions.

Keywords: Alzheimer's disease, lexical retrieval, mechanisms of change, primary progressive aphasia

Procedia PDF Downloads 176
629 Comparison Of Virtual Non-Contrast To True Non-Contrast Images Using Dual Layer Spectral Computed Tomography

Authors: O’Day Luke

Abstract:

Purpose: To validate virtual non-contrast reconstructions generated from dual-layer spectral computed tomography (DL-CT) data as an alternative for the acquisition of a dedicated true non-contrast dataset during multiphase contrast studies. Material and methods: Thirty-three patients underwent a routine multiphase clinical CT examination, using Dual-Layer Spectral CT, from March to August 2021. True non-contrast (TNC) and virtual non-contrast (VNC) datasets, generated from both portal venous and arterial phase imaging were evaluated. For every patient in both true and virtual non-contrast datasets, a region-of-interest (ROI) was defined in aorta, liver, fluid (i.e. gallbladder, urinary bladder), kidney, muscle, fat and spongious bone, resulting in 693 ROIs. Differences in attenuation for VNC and TNV images were compared, both separately and combined. Consistency between VNC reconstructions obtained from the arterial and portal venous phase was evaluated. Results: Comparison of CT density (HU) on the VNC and TNC images showed a high correlation. The mean difference between TNC and VNC images (excluding bone results) was 5.5 ± 9.1 HU and > 90% of all comparisons showed a difference of less than 15 HU. For all tissues but spongious bone, the mean absolute difference between TNC and VNC images was below 10 HU. VNC images derived from the arterial and the portal venous phase showed a good correlation in most tissue types. The aortic attenuation was somewhat dependent however on which dataset was used for reconstruction. Bone evaluation with VNC datasets continues to be a problem, as spectral CT algorithms are currently poor in differentiating bone and iodine. Conclusion: Given the increasing availability of DL-CT and proven accuracy of virtual non-contrast processing, VNC is a promising tool for generating additional data during routine contrast-enhanced studies. This study shows the utility of virtual non-contrast scans as an alternative for true non-contrast studies during multiphase CT, with potential for dose reduction, without loss of diagnostic information.

Keywords: dual-layer spectral computed tomography, virtual non-contrast, true non-contrast, clinical comparison

Procedia PDF Downloads 123
628 Integration of Educational Data Mining Models to a Web-Based Support System for Predicting High School Student Performance

Authors: Sokkhey Phauk, Takeo Okazaki

Abstract:

The challenging task in educational institutions is to maximize the high performance of students and minimize the failure rate of poor-performing students. An effective method to leverage this task is to know student learning patterns with highly influencing factors and get an early prediction of student learning outcomes at the timely stage for setting up policies for improvement. Educational data mining (EDM) is an emerging disciplinary field of data mining, statistics, and machine learning concerned with extracting useful knowledge and information for the sake of improvement and development in the education environment. The study is of this work is to propose techniques in EDM and integrate it into a web-based system for predicting poor-performing students. A comparative study of prediction models is conducted. Subsequently, high performing models are developed to get higher performance. The hybrid random forest (Hybrid RF) produces the most successful classification. For the context of intervention and improving the learning outcomes, a feature selection method MICHI, which is the combination of mutual information (MI) and chi-square (CHI) algorithms based on the ranked feature scores, is introduced to select a dominant feature set that improves the performance of prediction and uses the obtained dominant set as information for intervention. By using the proposed techniques of EDM, an academic performance prediction system (APPS) is subsequently developed for educational stockholders to get an early prediction of student learning outcomes for timely intervention. Experimental outcomes and evaluation surveys report the effectiveness and usefulness of the developed system. The system is used to help educational stakeholders and related individuals for intervening and improving student performance.

Keywords: academic performance prediction system, educational data mining, dominant factors, feature selection method, prediction model, student performance

Procedia PDF Downloads 89
627 Joint Training Offer Selection and Course Timetabling Problems: Models and Algorithms

Authors: Gianpaolo Ghiani, Emanuela Guerriero, Emanuele Manni, Alessandro Romano

Abstract:

In this article, we deal with a variant of the classical course timetabling problem that has a practical application in many areas of education. In particular, in this paper we are interested in high schools remedial courses. The purpose of such courses is to provide under-prepared students with the skills necessary to succeed in their studies. In particular, a student might be under prepared in an entire course, or only in a part of it. The limited availability of funds, as well as the limited amount of time and teachers at disposal, often requires schools to choose which courses and/or which teaching units to activate. Thus, schools need to model the training offer and the related timetabling, with the goal of ensuring the highest possible teaching quality, by meeting the above-mentioned financial, time and resources constraints. Moreover, there are some prerequisites between the teaching units that must be satisfied. We first present a Mixed-Integer Programming (MIP) model to solve this problem to optimality. However, the presence of many peculiar constraints contributes inevitably in increasing the complexity of the mathematical model. Thus, solving it through a general purpose solver may be performed for small instances only, while solving real-life-sized instances of such model requires specific techniques or heuristic approaches. For this purpose, we also propose a heuristic approach, in which we make use of a fast constructive procedure to obtain a feasible solution. To assess our exact and heuristic approaches we perform extensive computational results on both real-life instances (obtained from a high school in Lecce, Italy) and randomly generated instances. Our tests show that the MIP model is never solved to optimality, with an average optimality gap of 57%. On the other hand, the heuristic algorithm is much faster (in about the 50% of the considered instances it converges in approximately half of the time limit) and in many cases allows achieving an improvement on the objective function value obtained by the MIP model. Such an improvement ranges between 18% and 66%.

Keywords: heuristic, MIP model, remedial course, school, timetabling

Procedia PDF Downloads 585
626 ROSgeoregistration: Aerial Multi-Spectral Image Simulator for the Robot Operating System

Authors: Andrew R. Willis, Kevin Brink, Kathleen Dipple

Abstract:

This article describes a software package called ROS-georegistration intended for use with the robot operating system (ROS) and the Gazebo 3D simulation environment. ROSgeoregistration provides tools for the simulation, test, and deployment of aerial georegistration algorithms and is available at github.com/uncc-visionlab/rosgeoregistration. A model creation package is provided which downloads multi-spectral images from the Google Earth Engine database and, if necessary, incorporates these images into a single, possibly very large, reference image. Additionally a Gazebo plugin which uses the real-time sensor pose and image formation model to generate simulated imagery using the specified reference image is provided along with related plugins for UAV relevant data. The novelty of this work is threefold: (1) this is the first system to link the massive multi-spectral imaging database of Google’s Earth Engine to the Gazebo simulator, (2) this is the first example of a system that can simulate geospatially and radiometrically accurate imagery from multiple sensor views of the same terrain region, and (3) integration with other UAS tools creates a new holistic UAS simulation environment to support UAS system and subsystem development where real-world testing would generally be prohibitive. Sensed imagery and ground truth registration information is published to client applications which can receive imagery synchronously with telemetry from other payload sensors, e.g., IMU, GPS/GNSS, barometer, and windspeed sensor data. To highlight functionality, we demonstrate ROSgeoregistration for simulating Electro-Optical (EO) and Synthetic Aperture Radar (SAR) image sensors and an example use case for developing and evaluating image-based UAS position feedback, i.e., pose for image-based Guidance Navigation and Control (GNC) applications.

Keywords: EO-to-EO, EO-to-SAR, flight simulation, georegistration, image generation, robot operating system, vision-based navigation

Procedia PDF Downloads 87
625 Kernel-Based Double Nearest Proportion Feature Extraction for Hyperspectral Image Classification

Authors: Hung-Sheng Lin, Cheng-Hsuan Li

Abstract:

Over the past few years, kernel-based algorithms have been widely used to extend some linear feature extraction methods such as principal component analysis (PCA), linear discriminate analysis (LDA), and nonparametric weighted feature extraction (NWFE) to their nonlinear versions, kernel principal component analysis (KPCA), generalized discriminate analysis (GDA), and kernel nonparametric weighted feature extraction (KNWFE), respectively. These nonlinear feature extraction methods can detect nonlinear directions with the largest nonlinear variance or the largest class separability based on the given kernel function. Moreover, they have been applied to improve the target detection or the image classification of hyperspectral images. The double nearest proportion feature extraction (DNP) can effectively reduce the overlap effect and have good performance in hyperspectral image classification. The DNP structure is an extension of the k-nearest neighbor technique. For each sample, there are two corresponding nearest proportions of samples, the self-class nearest proportion and the other-class nearest proportion. The term “nearest proportion” used here consider both the local information and other more global information. With these settings, the effect of the overlap between the sample distributions can be reduced. Usually, the maximum likelihood estimator and the related unbiased estimator are not ideal estimators in high dimensional inference problems, particularly in small data-size situation. Hence, an improved estimator by shrinkage estimation (regularization) is proposed. Based on the DNP structure, LDA is included as a special case. In this paper, the kernel method is applied to extend DNP to kernel-based DNP (KDNP). In addition to the advantages of DNP, KDNP surpasses DNP in the experimental results. According to the experiments on the real hyperspectral image data sets, the classification performance of KDNP is better than that of PCA, LDA, NWFE, and their kernel versions, KPCA, GDA, and KNWFE.

Keywords: feature extraction, kernel method, double nearest proportion feature extraction, kernel double nearest feature extraction

Procedia PDF Downloads 323
624 GBKMeans: A Genetic Based K-Means Applied to the Capacitated Planning of Reading Units

Authors: Anderson S. Fonseca, Italo F. S. Da Silva, Robert D. A. Santos, Mayara G. Da Silva, Pedro H. C. Vieira, Antonio M. S. Sobrinho, Victor H. B. Lemos, Petterson S. Diniz, Anselmo C. Paiva, Eliana M. G. Monteiro

Abstract:

In Brazil, the National Electric Energy Agency (ANEEL) establishes that electrical energy companies are responsible for measuring and billing their customers. Among these regulations, it’s defined that a company must bill your customers within 27-33 days. If a relocation or a change of period is required, the consumer must be notified in writing, in advance of a billing period. To make it easier to organize a workday’s measurements, these companies create a reading plan. These plans consist of grouping customers into reading groups, which are visited by an employee responsible for measuring consumption and billing. The creation process of a plan efficiently and optimally is a capacitated clustering problem with constraints related to homogeneity and compactness, that is, the employee’s working load and the geographical position of the consuming unit. This process is a work done manually by several experts who have experience in the geographic formation of the region, which takes a large number of days to complete the final planning, and because it’s human activity, there is no guarantee of finding the best optimization for planning. In this paper, the GBKMeans method presents a technique based on K-Means and genetic algorithms for creating a capacitated cluster that respects the constraints established in an efficient and balanced manner, that minimizes the cost of relocating consumer units and the time required for final planning creation. The results obtained by the presented method are compared with the current planning of a real city, showing an improvement of 54.71% in the standard deviation of working load and 11.97% in the compactness of the groups.

Keywords: capacitated clustering, k-means, genetic algorithm, districting problems

Procedia PDF Downloads 177
623 Clinical Advice Services: Using Lean Chassis to Optimize Nurse-Driven Telephonic Triage of After-Hour Calls from Patients

Authors: Eric Lee G. Escobedo-Wu, Nidhi Rohatgi, Fouzel Dhebar

Abstract:

It is challenging for patients to navigate through healthcare systems after-hours. This leads to delays in care, patient/provider dissatisfaction, inappropriate resource utilization, readmissions, and higher costs. It is important to provide patients and providers with effective clinical decision-making tools to allow seamless connectivity and coordinated care. In August 2015, patient-centric Stanford Health Care established Clinical Advice Services (CAS) to provide clinical decision support after-hours. CAS is founded on key Lean principles: Value stream mapping, empathy mapping, waste walk, takt time calculations, standard work, plan-do-check-act cycles, and active daily management. At CAS, Clinical Assistants take the initial call and manage all non-clinical calls (e.g., appointments, directions, general information). If the patient has a clinical symptom, the CAS nurses take the call and utilize standardized clinical algorithms to triage the patient to home, clinic, urgent care, emergency department, or 911. Nurses may also contact the on-call physician based on the clinical algorithm for further direction and consultation. Since August 2015, CAS has managed 228,990 calls from 26 clinical specialties. Reporting is built into the electronic health record for analysis and data collection. 65.3% of the after-hours calls are clinically related. Average clinical algorithm adherence rate has been 92%. An average of 9% of calls was escalated by CAS nurses to the physician on call. An average of 5% of patients was triaged to the Emergency Department by CAS. Key learnings indicate that a seamless connectivity vision, cascading, multidisciplinary ownership of the problem, and synergistic enterprise improvements have contributed to this success while striving for continuous improvement.

Keywords: after hours phone calls, clinical advice services, nurse triage, Stanford Health Care

Procedia PDF Downloads 153
622 Applying Multiple Kinect on the Development of a Rapid 3D Mannequin Scan Platform

Authors: Shih-Wen Hsiao, Yi-Cheng Tsao

Abstract:

In the field of reverse engineering and creative industries, applying 3D scanning process to obtain geometric forms of the objects is a mature and common technique. For instance, organic objects such as faces and nonorganic objects such as products could be scanned to acquire the geometric information for further application. However, although the data resolution of 3D scanning device is increasing and there are more and more abundant complementary applications, the penetration rate of 3D scanning for the public is still limited by the relative high price of the devices. On the other hand, Kinect, released by Microsoft, is known for its powerful functions, considerably low price, and complete technology and database support. Therefore, related studies can be done with the applying of Kinect under acceptable cost and data precision. Due to the fact that Kinect utilizes optical mechanism to extracting depth information, limitations are found due to the reason of the straight path of the light. Thus, various angles are required sequentially to obtain the complete 3D information of the object when applying a single Kinect for 3D scanning. The integration process which combines the 3D data from different angles by certain algorithms is also required. This sequential scanning process costs much time and the complex integration process often encounter some technical problems. Therefore, this paper aimed to apply multiple Kinects simultaneously on the field of developing a rapid 3D mannequin scan platform and proposed suggestions on the number and angles of Kinects. In the content, a method of establishing the coordination based on the relation between mannequin and the specifications of Kinect is proposed, and a suggestion of angles and number of Kinects is also described. An experiment of applying multiple Kinect on the scanning of 3D mannequin is constructed by Microsoft API, and the results show that the time required for scanning and technical threshold can be reduced in the industries of fashion and garment design.

Keywords: 3D scan, depth sensor, fashion and garment design, mannequin, multiple Kinect sensor

Procedia PDF Downloads 347
621 Improved Intracellular Protein Degradation System for Rapid Screening and Quantitative Study of Essential Fungal Proteins in Biopharmaceutical Development

Authors: Patarasuda Chaisupa, R. Clay Wright

Abstract:

The selection of appropriate biomolecular targets is a crucial aspect of biopharmaceutical development. The Auxin-Inducible Degron Degradation (AID) technology has demonstrated remarkable potential in efficiently and rapidly degrading target proteins, thereby enabling the identification and acquisition of drug targets. The AID system also offers a viable method to deplete specific proteins, particularly in cases where the degradation pathway has not been exploited or when the adaptation of proteins, including the cell environment, occurs to compensate for the mutation or gene knockout. In this study, we have engineered an improved AID system tailored to deplete proteins of interest. This AID construct combines the auxin-responsive E3 ubiquitin ligase binding domain, AFB2, and the substrate degron, IAA17, fused to the target genes. Essential genes of fungi with the lowest percent amino acid similarity to human and plant orthologs, according to the Basic Local Alignment Search Tool (BLAST), were cloned into the AID construct in S. cerevisiae (AID-tagged strains) using a modular yeast cloning toolkit for multipart assembly and direct genetic modification. Each E3 ubiquitin ligase and IAA17 degron was fused to a fluorescence protein, allowing for real-time monitoring of protein levels in response to different auxin doses via cytometry. Our AID system exhibited high sensitivity, with an EC50 value of 0.040 µM (SE = 0.016) for AFB2, enabling the specific promotion of IAA17::target protein degradation. Furthermore, we demonstrate how this improved AID system enhances quantitative functional studies of various proteins in fungi. The advancements made in auxin-inducible protein degradation in this study offer a powerful approach to investigating critical target protein viability in fungi, screening protein targets for drugs, and regulating intracellular protein abundance, thus revolutionizing the study of protein function underlying a diverse range of biological processes.

Keywords: synthetic biology, bioengineering, molecular biology, biotechnology

Procedia PDF Downloads 68
620 A Comparative Analysis of Clustering Approaches for Understanding Patterns in Health Insurance Uptake: Evidence from Sociodemographic Kenyan Data

Authors: Nelson Kimeli Kemboi Yego, Juma Kasozi, Joseph Nkruzinza, Francis Kipkogei

Abstract:

The study investigated the low uptake of health insurance in Kenya despite efforts to achieve universal health coverage through various health insurance schemes. Unsupervised machine learning techniques were employed to identify patterns in health insurance uptake based on sociodemographic factors among Kenyan households. The aim was to identify key demographic groups that are underinsured and to provide insights for the development of effective policies and outreach programs. Using the 2021 FinAccess Survey, the study clustered Kenyan households based on their health insurance uptake and sociodemographic features to reveal patterns in health insurance uptake across the country. The effectiveness of k-prototypes clustering, hierarchical clustering, and agglomerative hierarchical clustering in clustering based on sociodemographic factors was compared. The k-prototypes approach was found to be the most effective at uncovering distinct and well-separated clusters in the Kenyan sociodemographic data related to health insurance uptake based on silhouette, Calinski-Harabasz, Davies-Bouldin, and Rand indices. Hence, it was utilized in uncovering the patterns in uptake. The results of the analysis indicate that inclusivity in health insurance is greatly related to affordability. The findings suggest that targeted policy interventions and outreach programs are necessary to increase health insurance uptake in Kenya, with the ultimate goal of achieving universal health coverage. The study provides important insights for policymakers and stakeholders in the health insurance sector to address the low uptake of health insurance and to ensure that healthcare services are accessible and affordable to all Kenyans, regardless of their socio-demographic status. The study highlights the potential of unsupervised machine learning techniques to provide insights into complex health policy issues and improve decision-making in the health sector.

Keywords: health insurance, unsupervised learning, clustering algorithms, machine learning

Procedia PDF Downloads 106