Search results for: derivative and convoluted derivative methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15634

Search results for: derivative and convoluted derivative methods

11194 Goal-Setting in a Peer Leader HIV Prevention Intervention to Improve Preexposure Prophylaxis Access among Black Men Who Have Sex with Men

Authors: Tim J. Walsh, Lindsay E. Young, John A. Schneider

Abstract:

Background: The disproportionate rate of HIV infection among Black men who have sex with men (BMSM) in the United States suggest the importance of Preexposure Prophylaxis (PrEP) interventions for this population. As such, there is an urgent need for innovative outreach strategies that extend beyond the traditional patient-provider relationship to reach at-risk populations. Training members of the BMSM community as peer change agents (PCAs) is one such strategy. An important piece of this training is goal-setting. Goal-setting not only encourages PCAs to define the parameters of the intervention according to their lived experience, it also helps them plan courses of action. Therefore, the aims of this mixed methods study are: (1) Characterize the goals that BMSM set at the end of their PrEP training and (2) Assess the relationship between goal types and PCA engagement. Methods: Between March 2016 and July 2016, preliminary data were collected from 68 BMSM, ages 18-33, in Chicago as part of an ongoing PrEP intervention. Once enrolled, PCAs participate in a half-day training in which they learn about PrEP, practice initiating conversations about PrEP, and identify strategies for supporting at-risk peers through the PrEP adoption process. Training culminates with a goal-setting exercise, whereby participants establish a goal related to their role as a PCA. Goals were coded for features that either emerged from the data itself or existed in extant goal-setting literature. The main outcomes were (1) number of PrEP conversations PCAs self-report during booster conversations two weeks following the intervention and (2) number of peers PCAs recruit into the study that completed the PrEP workshop. Results: PCA goals (N=68) were characterized in terms of four features: Specificity, target population, personalization, and purpose defined. To date, PCAs report a collective 52 PrEP conversations. 56, 25, and 6% of PrEP conversations occurred with friends, family, and sexual partners, respectively. PCAs with specific goals had more PrEP conversations with at-risk peers compared to those with vague goals (58% vs. 42%); PCAs with personalized goals had more PrEP conversations compared to those with de-personalized goals (60% vs. 53%); and PCAs with goals that defined a purpose had more PrEP conversations compared to those who did not define a purpose (75% vs. 52%). 100% of PCAs with goals that defined a purpose recruited peers into the study compared to 45 percent of PCAs with goals that did not define a purpose. Conclusion: Our preliminary analysis demonstrates that BMSM are motivated to set and work toward a diverse set of goals to support peers in PrEP adoption. PCAs with goals involving a clearly defined purpose had more PrEP conversations and greater peer recruitment than those with goals lacking a defined purpose. This may indicate that PCAs who define their purpose at the outset of their participation will be more engaged in the study than those who do not. Goal-setting may be considered as a component of future HIV prevention interventions to advance intervention goals and as an indicator of PCAs understanding of the intervention.

Keywords: HIV prevention, MSM, peer change agent, preexposure prophylaxis

Procedia PDF Downloads 196
11193 Exploring the Gap between Coverage, Access, Utilization of Long Lasting Insecticidal Nets (LLINs) among the People of Malaria Endemic Districts in Bangladesh

Authors: Fouzia Khanam, Tridib Chowdhury, Belal Hossain, Sajedur Rahman, Mahfuzar Rahman

Abstract:

Introduction: Over the last decades, the world has achieved a noticeable success in preventing malaria. Nevertheless, malaria, a vector-borne infectious disease, remains a major public health burden globally as well as in Bangladesh. To achieve the goal of eliminating malaria, BRAC, a leading organization of Bangladesh in collaboration with government, is distributing free LLIN to the 13 endemic districts of the country. The study was conducted with the aim of assessing the gap between coverage, access, and utilization of LLIN among the people of the 13 malaria endemic districts of Bangladesh. Methods: This baseline study employed a community cross-sectional design triangulated with qualitative methods to measure households’ ownership, access and use of 13 endemic districts. A multistage cluster random sampling was employed for the quantitative part and for qualitative part a purposive sampling strategy was done. Thus present analysis included 2640 households encompassing a total of 14475 populations. Data were collected using a pre-tested structured questionnaire through one on one face-to-face interview with respondents. All analyses were performed using STATA (Version 13.0). For the qualitative part participant observation, in-depth interview, focus group discussion, key informant interview and informal interview was done to gather the contextual data. Findings: According to our study, 99.8% of households possessed at least one-bed net in both study areas. 77.4% households possessed at least two LLIN and 43.2% households had access to LLIN for all the members. So the gap between coverage and access is 34%. 91.8% people in the 13 districts and 95.1% in Chittagong Hill Tracts areas reported having had slept under a bed net the night before interviewed. And despite the relatively low access, in 77.8% of households, all the members were used the LLIN the previous night. This higher utilization compared to access might be due to the increased awareness among the community people regarding LLIN uses. However, among those people with sufficient access to LLIN, 6% of them still did not use the LLIN which reflects the behavioral failure that needs to be addressed. The major reasons for not using LLIN, identified by both qualitative and quantitative findings, were insufficient access, sleeping or living outside the home, migration, perceived low efficacy of LLIN, fear of physical side effects or feeling uncomfortable. Conclusion: Given that LLIN access and use was a bit short of the targets, it conveys important messages to the malaria control program. Targeting specific population segments and groups for achieving expected LLIN coverage is very crucial. And also, addressing behavior failure by well-designed behavioral change interventions is mandatory.

Keywords: long lasting insecticide net, malaria, malaria control programme, World Health Organisation

Procedia PDF Downloads 188
11192 Customization of Moodle Open Source LMS for Tanzania Secondary Schools’ Use

Authors: Ellen A. Kalinga

Abstract:

Moodle is an open source learning management system that enables creation of a powerful and flexible learning environment. Many organizations, especially learning institutions have customized Moodle open source LMS for their own use. In general open source LMSs are of great interest due to many advantages they offer in terms of cost, usage and freedom to customize to fit a particular context. Tanzania Secondary School e-Learning (TanSSe-L) system is the learning management system for Tanzania secondary schools. TanSSe-L system was developed using a number of methods, one of them being customization of Moodle Open Source LMS. This paper presents few areas on the way Moodle OS LMS was customized to produce a functional TanSSe-L system fitted to the requirements and specifications of Tanzania secondary schools’ context.

Keywords: LMS, Moodle, e-learning, Tanzania, secondary school

Procedia PDF Downloads 393
11191 The Use of Geographic Information System Technologies for Geotechnical Monitoring of Pipeline Systems

Authors: A. G. Akhundov

Abstract:

Issues of obtaining unbiased data on the status of pipeline systems of oil- and oil product transportation become especially important when laying and operating pipelines under severe nature and climatic conditions. The essential attention is paid here to researching exogenous processes and their impact on linear facilities of the pipeline system. Reliable operation of pipelines under severe nature and climatic conditions, timely planning and implementation of compensating measures are only possible if operation conditions of pipeline systems are regularly monitored, and changes of permafrost soil and hydrological operation conditions are accounted for. One of the main reasons for emergency situations to appear is the geodynamic factor. Emergency situations are proved by the experience to occur within areas characterized by certain conditions of the environment and to develop according to similar scenarios depending on active processes. The analysis of natural and technical systems of main pipelines at different stages of monitoring gives a possibility of making a forecast of the change dynamics. The integration of GIS technologies, traditional means of geotechnical monitoring (in-line inspection, geodetic methods, field observations), and remote methods (aero-visual inspection, aero photo shooting, air and ground laser scanning) provides the most efficient solution of the problem. The united environment of geo information system (GIS) is a comfortable way to implement the monitoring system on the main pipelines since it provides means to describe a complex natural and technical system and every element thereof with any set of parameters. Such GIS enables a comfortable simulation of main pipelines (both in 2D and 3D), the analysis of situations and selection of recommendations to prevent negative natural or man-made processes and to mitigate their consequences. The specifics of such systems include: a multi-dimensions simulation of facilities in the pipeline system, math modelling of the processes to be observed, and the use of efficient numeric algorithms and software packets for forecasting and analyzing. We see one of the most interesting possibilities of using the monitoring results as generating of up-to-date 3D models of a facility and the surrounding area on the basis of aero laser scanning, data of aerophotoshooting, and data of in-line inspection and instrument measurements. The resulting 3D model shall be the basis of the information system providing means to store and process data of geotechnical observations with references to the facilities of the main pipeline; to plan compensating measures, and to control their implementation. The use of GISs for geotechnical monitoring of pipeline systems is aimed at improving the reliability of their operation, reducing the probability of negative events (accidents and disasters), and at mitigation of consequences thereof if they still are to occur.

Keywords: databases, 3D GIS, geotechnical monitoring, pipelines, laser scaning

Procedia PDF Downloads 189
11190 Machine Learning Techniques in Seismic Risk Assessment of Structures

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this work is to evaluate the advantages and disadvantages of various machine learning techniques in two key steps of seismic hazard and risk assessment of different types of structures. The first step is the development of ground-motion models, which are used for forecasting ground-motion intensity measures (IM) given source characteristics, source-to-site distance, and local site condition for future events. IMs such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available. Second, it is investigated how machine learning techniques could be beneficial for developing probabilistic seismic demand models (PSDMs), which provide the relationship between the structural demand responses (e.g., component deformations, accelerations, internal forces, etc.) and the ground motion IMs. In the risk framework, such models are used to develop fragility curves estimating exceeding probability of damage for pre-defined limit states, and therefore, control the reliability of the predictions in the risk assessment. In this study, machine learning algorithms like artificial neural network, random forest, and support vector machine are adopted and trained on the demand parameters to derive PSDMs for them. It is observed that such models can provide more accurate estimates of prediction in relatively shorter about of time compared to conventional methods. Moreover, they can be used for sensitivity analysis of fragility curves with respect to many modeling parameters without necessarily requiring more intense numerical response-history analysis.

Keywords: artificial neural network, machine learning, random forest, seismic risk analysis, seismic hazard analysis, support vector machine

Procedia PDF Downloads 106
11189 Preparation of Papers - Developing a Leukemia Diagnostic System Based on Hybrid Deep Learning Architectures in Actual Clinical Environments

Authors: Skyler Kim

Abstract:

An early diagnosis of leukemia has always been a challenge to doctors and hematologists. On a worldwide basis, it was reported that there were approximately 350,000 new cases in 2012, and diagnosing leukemia was time-consuming and inefficient because of an endemic shortage of flow cytometry equipment in current clinical practice. As the number of medical diagnosis tools increased and a large volume of high-quality data was produced, there was an urgent need for more advanced data analysis methods. One of these methods was the AI approach. This approach has become a major trend in recent years, and several research groups have been working on developing these diagnostic models. However, designing and implementing a leukemia diagnostic system in real clinical environments based on a deep learning approach with larger sets remains complex. Leukemia is a major hematological malignancy that results in mortality and morbidity throughout different ages. We decided to select acute lymphocytic leukemia to develop our diagnostic system since acute lymphocytic leukemia is the most common type of leukemia, accounting for 74% of all children diagnosed with leukemia. The results from this development work can be applied to all other types of leukemia. To develop our model, the Kaggle dataset was used, which consists of 15135 total images, 8491 of these are images of abnormal cells, and 5398 images are normal. In this paper, we design and implement a leukemia diagnostic system in a real clinical environment based on deep learning approaches with larger sets. The proposed diagnostic system has the function of detecting and classifying leukemia. Different from other AI approaches, we explore hybrid architectures to improve the current performance. First, we developed two independent convolutional neural network models: VGG19 and ResNet50. Then, using both VGG19 and ResNet50, we developed a hybrid deep learning architecture employing transfer learning techniques to extract features from each input image. In our approach, fusing the features from specific abstraction layers can be deemed as auxiliary features and lead to further improvement of the classification accuracy. In this approach, features extracted from the lower levels are combined into higher dimension feature maps to help improve the discriminative capability of intermediate features and also overcome the problem of network gradient vanishing or exploding. By comparing VGG19 and ResNet50 and the proposed hybrid model, we concluded that the hybrid model had a significant advantage in accuracy. The detailed results of each model’s performance and their pros and cons will be presented in the conference.

Keywords: acute lymphoblastic leukemia, hybrid model, leukemia diagnostic system, machine learning

Procedia PDF Downloads 187
11188 Assessment of Aminopolyether on 18F-FDG Samples

Authors: Renata L. C. Leão, João E. Nascimento, Natalia C. E. S. Nascimento, Elaine S. Vasconcelos, Mércia L. Oliveira

Abstract:

The quality control procedures of a radiopharmaceutical include the assessment of its chemical purity. The method suggested by international pharmacopeias consists of a thin layer chromatographic run. In this paper, the method proposed by the United States Pharmacopeia (USP) is compared to a direct method to determine the final concentration of aminopolyether in Fludeoxyglucose (18F-FDG) preparations. The approach (no chromatographic run) was achieved by placing the thin-layer chromatography (TLC) plate directly on an iodine vapor chamber. Both methods were validated and they showed adequate results to determine the concentration of aminopolyether in 18F-FDG preparations. However, the direct method is more sensitive, faster and simpler when compared to the reference method (with chromatographic run), and it may be chosen for use in routine quality control of 18F-FDG.

Keywords: chemical purity, Kryptofix 222, thin layer chromatography, validation

Procedia PDF Downloads 201
11187 Words Spotting in the Images Handwritten Historical Documents

Authors: Issam Ben Jami

Abstract:

Information retrieval in digital libraries is very important because most famous historical documents occupy a significant value. The word spotting in historical documents is a very difficult notion, because automatic recognition of such documents is naturally cursive, it represents a wide variability in the level scale and translation words in the same documents. We first present a system for the automatic recognition, based on the extraction of interest points words from the image model. The extraction phase of the key points is chosen from the representation of the image as a synthetic description of the shape recognition in a multidimensional space. As a result, we use advanced methods that can find and describe interesting points invariant to scale, rotation and lighting which are linked to local configurations of pixels. We test this approach on documents of the 15th century. Our experiments give important results.

Keywords: feature matching, historical documents, pattern recognition, word spotting

Procedia PDF Downloads 274
11186 Self-Supervised Learning for Hate-Speech Identification

Authors: Shrabani Ghosh

Abstract:

Automatic offensive language detection in social media has become a stirring task in today's NLP. Manual Offensive language detection is tedious and laborious work where automatic methods based on machine learning are only alternatives. Previous works have done sentiment analysis over social media in different ways such as supervised, semi-supervised, and unsupervised manner. Domain adaptation in a semi-supervised way has also been explored in NLP, where the source domain and the target domain are different. In domain adaptation, the source domain usually has a large amount of labeled data, while only a limited amount of labeled data is available in the target domain. Pretrained transformers like BERT, RoBERTa models are fine-tuned to perform text classification in an unsupervised manner to perform further pre-train masked language modeling (MLM) tasks. In previous work, hate speech detection has been explored in Gab.ai, which is a free speech platform described as a platform of extremist in varying degrees in online social media. In domain adaptation process, Twitter data is used as the source domain, and Gab data is used as the target domain. The performance of domain adaptation also depends on the cross-domain similarity. Different distance measure methods such as L2 distance, cosine distance, Maximum Mean Discrepancy (MMD), Fisher Linear Discriminant (FLD), and CORAL have been used to estimate domain similarity. Certainly, in-domain distances are small, and between-domain distances are expected to be large. The previous work finding shows that pretrain masked language model (MLM) fine-tuned with a mixture of posts of source and target domain gives higher accuracy. However, in-domain performance of the hate classifier on Twitter data accuracy is 71.78%, and out-of-domain performance of the hate classifier on Gab data goes down to 56.53%. Recently self-supervised learning got a lot of attention as it is more applicable when labeled data are scarce. Few works have already been explored to apply self-supervised learning on NLP tasks such as sentiment classification. Self-supervised language representation model ALBERTA focuses on modeling inter-sentence coherence and helps downstream tasks with multi-sentence inputs. Self-supervised attention learning approach shows better performance as it exploits extracted context word in the training process. In this work, a self-supervised attention mechanism has been proposed to detect hate speech on Gab.ai. This framework initially classifies the Gab dataset in an attention-based self-supervised manner. On the next step, a semi-supervised classifier trained on the combination of labeled data from the first step and unlabeled data. The performance of the proposed framework will be compared with the results described earlier and also with optimized outcomes obtained from different optimization techniques.

Keywords: attention learning, language model, offensive language detection, self-supervised learning

Procedia PDF Downloads 106
11185 Austrian Secondary School Teachers’ Perspectives on Character Education and Life Skills: First Quantitative Insights from a Mixed Methods Study

Authors: Evelyn Kropfreiter, Roland Bernhard

Abstract:

There has been an increased interest in school-based whole-child development in the Austrian education system in the last few years. Although there is a consensus among academics that teachers' beliefs are an essential component of their professional competence, there are hardly any studies in the German-speaking world examining teachers' beliefs about school-based character education. To close this gap, we are conducting a mixed methods study combining qualitative interviews and a questionnaire in Austria (doctoral thesis at the University of Salzburg). In this paper, we present preliminary insights into the quantitative strand of the project. In contrast to German-speaking countries, the Anglo-Saxon world has a long tradition of explicit character education in schools. There has been a rising interest in approaches focusing on a neo-Aristotelian form of character education in England. The Jubilee Centre strongly influences the "renaissance" of papers on neo-Aristotelian character education for Character and Virtues, founded in 2012. The quantitative questionnaire study (n = 264) is an online survey of teachers and school principals conducted in four different federal states in spring 2023. Most respondents (n = 264) from lower secondary schools (AHS-Unterstufe and Mittelschule) believe that character education in schools for 10-14-year-olds is more important for society than good exam results. Many teachers state that they consider themselves prepared to promote their students' personal development and life skills through their education and to attend further training courses. However, there are many obstacles in the education system to ensure that a comprehensive education reaches the students. Many teachers state that they consider themselves prepared to promote their students' character strengths and life skills through their education and to attend further training courses. However, there are many obstacles in the education system to ensure that a comprehensive education reaches the students. Among the most cited difficulties, teachers mention the time factor associated with an overcrowded curriculum and a strong focus on performance, which often leaves them needing more time to keep an eye on nurturing the whole person. The fact that character education is not a separate subject, and its implementation needs to be monitored also makes it challenging to implement it in everyday school life. Austrian teachers prioritize moral virtues such as compassion and honesty as character strengths in everyday school life and resilience and commitment in the next place. Our results are like those reported in other studies on teacher's beliefs about character education. They indicate that Austrian teachers want to teach character in their schools but see systemic constraints such as the curriculum, in which personality roles play a subordinate role, and the focus on performance testing in the school system and the associated lack of time as obstacles to fostering more character development in students.

Keywords: character education, life skills, teachers' beliefs, virtues

Procedia PDF Downloads 81
11184 Assessing Diagnostic and Evaluation Tools for Use in Urban Immunisation Programming: A Critical Narrative Review and Proposed Framework

Authors: Tim Crocker-Buque, Sandra Mounier-Jack, Natasha Howard

Abstract:

Background: Due to both the increasing scale and speed of urbanisation, urban areas in low and middle-income countries (LMICs) host increasingly large populations of under-immunized children, with the additional associated risks of rapid disease transmission in high-density living environments. Multiple interdependent factors are associated with these coverage disparities in urban areas and most evidence comes from relatively few countries, e.g., predominantly India, Kenya, Nigeria, and some from Pakistan, Iran, and Brazil. This study aimed to identify, describe, and assess the main tools used to measure or improve coverage of immunisation services in poor urban areas. Methods: Authors used a qualitative review design, including academic and non-academic literature, to identify tools used to improve coverage of public health interventions in urban areas. Authors selected and extracted sources that provided good examples of specific tools, or categories of tools, used in a context relevant to urban immunization. Diagnostic (e.g., for data collection, analysis, and insight generation) and programme tools (e.g., for investigating or improving ongoing programmes) and interventions (e.g., multi-component or stand-alone with evidence) were selected for inclusion to provide a range of type and availability of relevant tools. These were then prioritised using a decision-analysis framework and a tool selection guide for programme managers developed. Results: Authors reviewed tools used in urban immunisation contexts and tools designed for (i) non-immunization and/or non-health interventions in urban areas, and (ii) immunisation in rural contexts that had relevance for urban areas (e.g., Reaching every District/Child/ Zone). Many approaches combined several tools and methods, which authors categorised as diagnostic, programme, and intervention. The most common diagnostic tools were cross-sectional surveys, key informant interviews, focus group discussions, secondary analysis of routine data, and geographical mapping of outcomes, resources, and services. Programme tools involved multiple stages of data collection, analysis, insight generation, and intervention planning and included guidance documents from WHO (World Health Organisation), UNICEF (United Nations Children's Fund), USAID (United States Agency for International Development), and governments, and articles reporting on diagnostics, interventions, and/or evaluations to improve urban immunisation. Interventions involved service improvement, education, reminder/recall, incentives, outreach, mass-media, or were multi-component. The main gaps in existing tools were an assessment of macro/policy-level factors, exploration of effective immunization communication channels, and measuring in/out-migration. The proposed framework uses a problem tree approach to suggest tools to address five common challenges (i.e. identifying populations, understanding communities, issues with service access and use, improving services, improving coverage) based on context and available data. Conclusion: This study identified many tools relevant to evaluating urban LMIC immunisation programmes, including significant crossover between tools. This was encouraging in terms of supporting the identification of common areas, but problematic as data volumes, instructions, and activities could overwhelm managers and tools are not always suitably applied to suitable contexts. Further research is needed on how best to combine tools and methods to suit local contexts. Authors’ initial framework can be tested and developed further.

Keywords: health equity, immunisation, low and middle-income countries, poverty, urban health

Procedia PDF Downloads 139
11183 High Capacity Reversible Watermarking through Interpolated Error Shifting

Authors: Hae-Yeoun Lee

Abstract:

Reversible watermarking that not only protects the copyright but also preserve the original quality of the digital content have been intensively studied. In particular, the demand for reversible watermarking has increased. In this paper, we propose a reversible watermarking scheme based on interpolation-error shifting and error precompensation. The intensity of a pixel is interpolated from the intensities of neighbouring pixels, and the difference histogram between the interpolated and the original intensities is obtained and modified to embed the watermark message. By restoring the difference histogram, the embedded watermark is extracted and the original image is recovered by compensating for the interpolation error. The overflow and underflow are prevented by error precompensation. To show the performance of the method, the proposed algorithm is compared with other methods using various test images.

Keywords: reversible watermarking, high capacity, high quality, interpolated error shifting, error precompensation

Procedia PDF Downloads 322
11182 How To Get Students’ Attentions?: Little Tricks From 15 English Teachers In Labuan

Authors: Suriani Oxley

Abstract:

All teachers aim to conduct a successful and an effective teaching. Teacher will use a variety of teaching techniques and methods to ensure that students achieve the learning objectives but often the teaching and learning processes are interrupted by a number of things such as noisy students, students not paying attention, the students play and so on. Such disturbances must be addressed to ensure that students can concentrate on their learning activities. This qualitative study observed and captured a video of numerous tricks that teachers in Labuan have implemented in helping the students to pay attentions in the classroom. The tricks are such as Name Calling, Non-Verbal Clues, Body Language, Ask Question, Offer Assistance, Echo Clapping, Call and Response & Cues and Clues. All of these tricks are simple but yet interesting language learning strategies that helped students to focus on their learning activities.

Keywords: paying attention, observation, tricks, learning strategies, classroom

Procedia PDF Downloads 566
11181 A Survey on Ambient Intelligence in Agricultural Technology

Authors: C. Angel, S. Asha

Abstract:

Despite the advances made in various new technologies, application of these technologies for agriculture still remains a formidable task, as it involves integration of diverse domains for monitoring the different process involved in agricultural management. Advances in ambient intelligence technology represents one of the most powerful technology for increasing the yield of agricultural crops and to mitigate the impact of water scarcity, climatic change and methods for managing pests, weeds, and diseases. This paper proposes a GPS-assisted, machine to machine solutions that combine information collected by multiple sensors for the automated management of paddy crops. To maintain the economic viability of paddy cultivation, the various techniques used in agriculture are discussed and a novel system which uses ambient intelligence technique is proposed in this paper. The ambient intelligence based agricultural system gives a great scope.

Keywords: ambient intelligence, agricultural technology, smart agriculture, precise farming

Procedia PDF Downloads 606
11180 Impact of Electronic Guest Relationship Management (e-GRM) on Brand Loyalty: The Case of Croatian Hotels

Authors: Marina Laškarin, Vlado Galičić

Abstract:

Quick adoption of e-business and emerging influence of “Electronic Word of Mouth e-WOM” communication on guests made leading hotel brands successful examples of electronic guest relationship management. Main reasons behind such success are well established procedures in collection, analysis and usage of highly valuable data available on the Internet, generated through some form of e-GRM programme. E-GRM is more than just a technology solution. It’s a system which balance respective guest demands, hotel technological capabilities and organizational culture of employees, discharging the universal approach in guest relations “same for all”. The purpose of this research derives from the necessity of determining the importance of monitoring and applying e-WOM communication as one of the methods used in managing guest relations. This paper analyses and compares different hotelier’s opinions on e-WOM communication.

Keywords: brand loyalty, e-WOM communication, GRM programmes, organizational culture

Procedia PDF Downloads 290
11179 Ensemble Machine Learning Approach for Estimating Missing Data from CO₂ Time Series

Authors: Atbin Mahabbati, Jason Beringer, Matthias Leopold

Abstract:

To address the global challenges of climate and environmental changes, there is a need for quantifying and reducing uncertainties in environmental data, including observations of carbon, water, and energy. Global eddy covariance flux tower networks (FLUXNET), and their regional counterparts (i.e., OzFlux, AmeriFlux, China Flux, etc.) were established in the late 1990s and early 2000s to address the demand. Despite the capability of eddy covariance in validating process modelling analyses, field surveys and remote sensing assessments, there are some serious concerns regarding the challenges associated with the technique, e.g. data gaps and uncertainties. To address these concerns, this research has developed an ensemble model to fill the data gaps of CO₂ flux to avoid the limitations of using a single algorithm, and therefore, provide less error and decline the uncertainties associated with the gap-filling process. In this study, the data of five towers in the OzFlux Network (Alice Springs Mulga, Calperum, Gingin, Howard Springs and Tumbarumba) during 2013 were used to develop an ensemble machine learning model, using five feedforward neural networks (FFNN) with different structures combined with an eXtreme Gradient Boosting (XGB) algorithm. The former methods, FFNN, provided the primary estimations in the first layer, while the later, XGB, used the outputs of the first layer as its input to provide the final estimations of CO₂ flux. The introduced model showed slight superiority over each single FFNN and the XGB, while each of these two methods was used individually, overall RMSE: 2.64, 2.91, and 3.54 g C m⁻² yr⁻¹ respectively (3.54 provided by the best FFNN). The most significant improvement happened to the estimation of the extreme diurnal values (during midday and sunrise), as well as nocturnal estimations, which is generally considered as one of the most challenging parts of CO₂ flux gap-filling. The towers, as well as seasonality, showed different levels of sensitivity to improvements provided by the ensemble model. For instance, Tumbarumba showed more sensitivity compared to Calperum, where the differences between the Ensemble model on the one hand and the FFNNs and XGB, on the other hand, were the least of all 5 sites. Besides, the performance difference between the ensemble model and its components individually were more significant during the warm season (Jan, Feb, Mar, Oct, Nov, and Dec) compared to the cold season (Apr, May, Jun, Jul, Aug, and Sep) due to the higher amount of photosynthesis of plants, which led to a larger range of CO₂ exchange. In conclusion, the introduced ensemble model slightly improved the accuracy of CO₂ flux gap-filling and robustness of the model. Therefore, using ensemble machine learning models is potentially capable of improving data estimation and regression outcome when it seems to be no more room for improvement while using a single algorithm.

Keywords: carbon flux, Eddy covariance, extreme gradient boosting, gap-filling comparison, hybrid model, OzFlux network

Procedia PDF Downloads 140
11178 Fluorescence-Based Biosensor for Dopamine Detection Using Quantum Dots

Authors: Sylwia Krawiec, Joanna Cabaj, Karol Malecha

Abstract:

Nowadays, progress in the field of the analytical methods is of great interest for reliable biological research and medical diagnostics. Classical techniques of chemical analysis, despite many advantages, do not permit to obtain immediate results or automatization of measurements. Chemical sensors have displaced the conventional analytical methods - sensors combine precision, sensitivity, fast response and the possibility of continuous-monitoring. Biosensor is a chemical sensor, which except of conventer also possess a biologically active material, which is the basis for the detection of specific chemicals in the sample. Each biosensor device mainly consists of two elements: a sensitive element, where is recognition of receptor-analyte, and a transducer element which receives the signal and converts it into a measurable signal. Through these two elements biosensors can be divided in two categories: due to the recognition element (e.g immunosensor) and due to the transducer (e.g optical sensor). Working of optical sensor is based on measurements of quantitative changes of parameters characterizing light radiation. The most often analyzed parameters include: amplitude (intensity), frequency or polarization. Changes in the optical properties one of the compound which reacts with biological material coated on the sensor is analyzed by a direct method, in an indirect method indicators are used, which changes the optical properties due to the transformation of the testing species. The most commonly used dyes in this method are: small molecules with an aromatic ring, like rhodamine, fluorescent proteins, for example green fluorescent protein (GFP), or nanoparticles such as quantum dots (QDs). Quantum dots have, in comparison with organic dyes, much better photoluminescent properties, better bioavailability and chemical inertness. These are semiconductor nanocrystals size of 2-10 nm. This very limited number of atoms and the ‘nano’-size gives QDs these highly fluorescent properties. Rapid and sensitive detection of dopamine is extremely important in modern medicine. Dopamine is very important neurotransmitter, which mainly occurs in the brain and central nervous system of mammals. Dopamine is responsible for the transmission information of moving through the nervous system and plays an important role in processes of learning or memory. Detection of dopamine is significant for diseases associated with the central nervous system such as Parkinson or schizophrenia. In developed optical biosensor for detection of dopamine, are used graphene quantum dots (GQDs). In such sensor dopamine molecules coats the GQD surface - in result occurs quenching of fluorescence due to Resonance Energy Transfer (FRET). Changes in fluorescence correspond to specific concentrations of the neurotransmitter in tested sample, so it is possible to accurately determine the concentration of dopamine in the sample.

Keywords: biosensor, dopamine, fluorescence, quantum dots

Procedia PDF Downloads 364
11177 CsPbBr₃@MOF-5-Based Single Drop Microextraction for in-situ Fluorescence Colorimetric Detection of Dechlorination Reaction

Authors: Yanxue Shang, Jingbin Zeng

Abstract:

Chlorobenzene homologues (CBHs) are a category of environmental pollutants that can not be ignored. They can stay in the environment for a long period and are potentially carcinogenic. The traditional degradation method of CBHs is dechlorination followed by sample preparation and analysis. This is not only time-consuming and laborious, but the detection and analysis processes are used in conjunction with large-scale instruments. Therefore, this can not achieve rapid and low-cost detection. Compared with traditional sensing methods, colorimetric sensing is simpler and more convenient. In recent years, chromaticity sensors based on fluorescence have attracted more and more attention. Compared with sensing methods based on changes in fluorescence intensity, changes in color gradients are easier to recognize by the naked eye. Accordingly, this work proposes to use single drop microextraction (SDME) technology to solve the above problems. After the dechlorination reaction was completed, the organic droplet extracts Cl⁻ and realizes fluorescence colorimetric sensing at the same time. This method was integrated sample processing and visual in-situ detection, simplifying the detection process. As a fluorescence colorimetric sensor material, CsPbBr₃ was encapsulated in MOF-5 to construct CsPbBr₃@MOF-5 fluorescence colorimetric composite. Then the fluorescence colorimetric sensor was constructed by dispersing the composite in SDME organic droplets. When the Br⁻ in CsPbBr₃ exchanges with Cl⁻ produced by the dechlorination reactions, it is converted into CsPbCl₃. The fluorescence color of the single droplet of SDME will change from green to blue emission, thereby realizing visual observation. Therein, SDME can enhance the concentration and enrichment of Cl⁻ and instead of sample pretreatment. The fluorescence color change of CsPbBr₃@MOF-5 can replace the detection process of large-scale instruments to achieve real-time rapid detection. Due to the absorption ability of MOF-5, it can not only improve the stability of CsPbBr₃, but induce the adsorption of Cl⁻. Simultaneously, accelerate the exchange of Br- and Cl⁻ in CsPbBr₃ and the detection process of Cl⁻. The absorption process was verified by density functional theory (DFT) calculations. This method exhibits exceptional linearity for Cl⁻ in the range of 10⁻² - 10⁻⁶ M (10000 μM - 1 μM) with a limit of detection of 10⁻⁷ M. Whereafter, the dechlorination reactions of different kinds of CBHs were also carried out with this method, and all had satisfactory detection ability. Also verified the accuracy by gas chromatography (GC), and it was found that the SDME we developed in this work had high credibility. In summary, the in-situ visualization method of dechlorination reaction detection was a combination of sample processing and fluorescence colorimetric sensing. Thus, the strategy researched herein represents a promising method for the visual detection of dechlorination reactions and can be extended for applications in environments, chemical industries, and foods.

Keywords: chlorobenzene homologues, colorimetric sensor, metal halide perovskite, metal-organic frameworks, single drop microextraction

Procedia PDF Downloads 143
11176 The Biocompatibility and Osteogenic Potential of Experimental Calcium Silicate Based Root Canal Sealer, Capseal

Authors: Seok Woo Chang

Abstract:

Aim: Capseal I and Capseal II are calcium silicate and calcium phosphate based experimental root canal sealer. The aim of this study was to evaluate the biocompatibility and mineralization potential of Capseal I and Capseal II. Materials and Methods: The biocompatibility and mineralization-related gene expression (alkaline phosphatase (ALP), bone sialoprotein (BSP), and osteocalcin (OCN)) of Capseal I and Capseal II were compared using methylthiazol tetrazolium assay and reverse transcription-polymerization chain reaction analysis, respectively. The results were analyzed by Kruskal-Wallis test. P-value of < 0.05 was considered significant. Result: Both Capseal I and Capseal II were favorable in biocompatibility and influenced the messenger RNA expression of ALP and BSP. Conclusion: Within the limitation of this study, Capseal is biocompatible and have mineralization promoting potential, and thus could be a promising root canal sealer.

Keywords: biocompatibility, mineralization-related gene expression, Capseal I, Capseal II

Procedia PDF Downloads 279
11175 Protein-Thiocyanate Composite as a Sensor for Iron III Cations

Authors: Hosam El-Sayed, Amira Abou El-Kheir, Salwa Mowafi, Marwa Abou Taleb

Abstract:

Two proteinic biopolymers; namely keratin and sericin, were extracted from their respective natural resources by simple appropriate methods. The said proteins were dissolved in the appropriate solvents followed by regeneration in a form of film polyvinyl alcohol. Proteinium thiocyanate (PTC) composite was prepared by reaction of a regenerated film with potassium thiocyanate in acid medium. In another experiment, the said acidified proteins were reacted with potassium thiocyante before dissolution and regeneration in a form of PTC composite. The possibility of using PTC composite for determination of the concentration of iron III ions in domestic as well as industrial water was examined. The concentration of iron III cations in water was determined spectrophotometrically by measuring the intensity of blood red colour of iron III thiocyanate obtained by interaction of PTC with iron III cation in the tested water sample.

Keywords: iron III cations, protein, sensor, thiocyanate, water

Procedia PDF Downloads 429
11174 Risk Management in Healthcare Sector in Turkey: A Dental Hospital Case Study

Authors: Pırıl Tekin, Rızvan Erol

Abstract:

Risk management has become very important and popular in developing countries in recent years. Especially making patient and employee health and safety issues compulsory in the hospitals, raised the number of studies in Turkey. Also risk management become more important for hospital senior management from clinics to the laboratories. Because quality is really important to be chosen for both patients to consult and employees to prefer to work. And also risk management studies can lead to hospital management team about future works and methods. By this point of view, this study is the risk assessment carried out in the biggest dental hospital in the south part of Turkey. This study was conducted as a research case study, covering two different health care place; A Clinic and A Laboratory. It shows that the problems in this dental hospital and how it can solve all.

Keywords: risk management, healthcare, dental hospital, quality management

Procedia PDF Downloads 377
11173 Choking among Babies, Toddlers and Children with Special Needs: A Review of Mechanisms, Implications, Incidence, and Recommendations of Professional Prevention Guidelines

Authors: Ella Abaev, Shany Segal, Miri Gabay

Abstract:

Background: Choking is a blockage of airways that prevents efficient breathing and air flow to the lungs. Choking may be partial or full and is an emergency situation. Complete or prolonged choking leads to apnea, lack of oxygen in the tissues of the body and brain, and can cause death. There are three mechanisms of choking: obstruction of internal respiratory tracts by food or object aspiration, any material that blocks or covers external air passages, external pressure on the neck or trapping between objects. Children's airways are narrower than that of adults and therefore the risk of choking is greater, due to the aspiration of food and other foreign bodies into the lungs. In the Child Development Center at Safra Children’s Hospital, Tel Hashomer in Israel are treated infants, toddlers, and children aged 0-18 years with various developmental disabilities. Due to the increase in reports of ‘almost an event’ of choking in the past year and the serious consequences of choking event, it was decided to give an emphasis to the issue. Incidence and methods: The number of reports of ‘almost an event’ or a choking event was examined at the center during the years 2013-2018 and a thorough research work was conducted on the subject in order to build a prevention program. Findings: Between 2013 and 2018 the center reported about ten cases of ‘almost choking events’. In the middle of 2018 alone three cases of ‘almost an event’ were reported. Objective: Providing knowledge leads to awareness raise, change of perception, change in behavior and prevention. The center employs more than 130 staff members from various sectors so that it is the work of multi-professional teams to promote the quality and safety of the treatment. The familiarity of the staff with risk factors, prevention guidelines, identification of choking signs, and treatment are most important and significant in determining the outcome of a choking event. Conclusions and recommendations: After in-depth research work was carried out in cooperation with the Risk Management Unit on the subject of choking, which include a description of the definitions, mechanisms, risk factors, treatment methods and extensive recommendations for prevention (e.g. using treatment and stimulation accessories with standards association stamps and adjustment of the type of food and the way it is served to match to the child's age and the ability to swallow). The expected stages of development and emphasis on the population of children with special needs were taken into account. The research findings will be published by the staff and parents of the patients, professional publications, and lectures and there is an expectation to decrease the number of choking events in the next years.

Keywords: children with special needs, choking, educational system, prevention guidelines

Procedia PDF Downloads 179
11172 Umbrella Reinforcement Learning – A Tool for Hard Problems

Authors: Egor E. Nuzhin, Nikolay V. Brilliantov

Abstract:

We propose an approach for addressing Reinforcement Learning (RL) problems. It combines the ideas of umbrella sampling, borrowed from Monte Carlo technique of computational physics and chemistry, with optimal control methods, and is realized on the base of neural networks. This results in a powerful algorithm, designed to solve hard RL problems – the problems, with long-time delayed reward, state-traps sticking and a lack of terminal states. It outperforms the prominent algorithms, such as PPO, RND, iLQR and VI, which are among the most efficient for the hard problems. The new algorithm deals with a continuous ensemble of agents and expected return, that includes the ensemble entropy. This results in a quick and efficient search of the optimal policy in terms of ”exploration-exploitation trade-off” in the state-action space.

Keywords: umbrella sampling, reinforcement learning, policy gradient, dynamic programming

Procedia PDF Downloads 21
11171 Memetic Algorithm for Solving the One-To-One Shortest Path Problem

Authors: Omar Dib, Alexandre Caminada, Marie-Ange Manier

Abstract:

The purpose of this study is to introduce a novel approach to solve the one-to-one shortest path problem. A directed connected graph is assumed in which all edges’ weights are positive. Our method is based on a memetic algorithm in which we combine a genetic algorithm (GA) and a variable neighborhood search method (VNS). We compare our approximate method with two exact algorithms Dijkstra and Integer Programming (IP). We made experimentations using random generated, complete and real graph instances. In most case studies, numerical results show that our method outperforms exact methods with 5% average gap to the optimality. Our algorithm’s average speed is 20-times faster than Dijkstra and more than 1000-times compared to IP. The details of the experimental results are also discussed and presented in the paper.

Keywords: shortest path problem, Dijkstra’s algorithm, integer programming, memetic algorithm

Procedia PDF Downloads 467
11170 Evaluation of Salt Content in Bread and the Amount Intake by Hypertensive Patients in the Algiers Region

Authors: S.lanasri, A.Boudjerrane, R.Belgherbi, O.Hadjoudj

Abstract:

Introduction: Bread is the most popular food in Algeria. The aim of this study was to examine the consumption of salt from bread by hypertensive patients. Materials and methods: sixty breads were collected from different artisans Algiers bakeries, each sample was mixed in harm distilled water until homogeneous and filtered. Analysis of the salt content was carried out according to the Mohr method titration. We calculated the amount of salt in bread consumed by 100 hypertensive patients using a questionnaire about the average amount of bread per day. Results: The salt content values from bread were 3.4g ± 0.37 NaCl / 100g.The average amount of salt consumed per day by patients from only bread was 3.82 g ± 3.8 with a maximum of 17 g per day. Only 38.18% of patients consume bread without salt even then 95% knew that excess salt intake can complicate hypertension. Conclusion: This study showed that bread is a major contributor to salt intake by Algerian hypertensive patients.

Keywords: salt, bread, hypertensive patients, Algiers

Procedia PDF Downloads 151
11169 A Similarity/Dissimilarity Measure to Biological Sequence Alignment

Authors: Muhammad A. Khan, Waseem Shahzad

Abstract:

Analysis of protein sequences is carried out for the purpose to discover their structural and ancestry relationship. Sequence similarity determines similar protein structures, similar function, and homology detection. Biological sequences composed of amino acid residues or nucleotides provide significant information through sequence alignment. In this paper, we present a new similarity/dissimilarity measure to sequence alignment based on the primary structure of a protein. The approach finds the distance between the two given sequences using the novel sequence alignment algorithm and a mathematical model. The algorithm runs at a time complexity of O(n²). A distance matrix is generated to construct a phylogenetic tree of different species. The new similarity/dissimilarity measure outperforms other existing methods.

Keywords: alignment, distance, homology, mathematical model, phylogenetic tree

Procedia PDF Downloads 178
11168 Embracing the Uniqueness and Potential of Each Child: Moving Theory to Practice

Authors: Joy Chadwick

Abstract:

This Study of Teaching and Learning (SoTL) research focused on the experiences of teacher candidates involved in an inclusive education methods course within a four-year direct entry Bachelor of Education program. The placement of this course within the final fourteen-week practicum semester is designed to facilitate deeper theory-practice connections between effective inclusive pedagogical knowledge and the real life of classroom teaching. The course focuses on supporting teacher candidates to understand that effective instruction within an inclusive classroom context must be intentional, responsive, and relational. Diversity is situated not as exceptional but rather as expected. This interpretive qualitative study involved the analysis of twenty-nine teacher candidate reflective journals and six individual teacher candidate semi-structured interviews. The journal entries were completed at the start of the semester and at the end of the semester with the intent of having teacher candidates reflect on their beliefs of what it means to be an effective inclusive educator and how the course and practicum experiences impacted their understanding and approaches to teaching in inclusive classrooms. The semi-structured interviews provided further depth and context to the journal data. The journals and interview transcripts were coded and themed using NVivo software. The findings suggest that instructional frameworks such as universal design for learning (UDL), differentiated instruction (DI), response to intervention (RTI), social emotional learning (SEL), and self-regulation supported teacher candidate’s abilities to meet the needs of their students more effectively. Course content that focused on specific exceptionalities also supported teacher candidates to be proactive rather than reactive when responding to student learning challenges. Teacher candidates also articulated the importance of reframing their perspective about students in challenging moments and that seeing the individual worth of each child was integral to their approach to teaching. A persisting question for teacher educators exists as to what pedagogical knowledge and understanding is most relevant in supporting future teachers to be effective at planning for and embracing the diversity of student needs within classrooms today. This research directs us to consider the critical importance of addressing personal attributes and mindsets of teacher candidates regarding children as well as considering instructional frameworks when designing coursework. Further, the alignment of an inclusive education course during a teaching practicum allows for an iterative approach to learning. The practical application of course concepts while teaching in a practicum allows for a deeper understanding of instructional frameworks, thus enhancing the confidence of teacher candidates. Research findings have implications for teacher education programs as connected to inclusive education methods courses, practicum experiences, and overall teacher education program design.

Keywords: inclusion, inclusive education, pre-service teacher education, practicum experiences, teacher education

Procedia PDF Downloads 68
11167 Visualizing Class Metrics and Object Calls for Software Systems

Authors: Mohammad Alnabhan, Awni Hammouri, Mustafa Hammad, Anas Al-Badareen, Omamah Al-Thnebat

Abstract:

Software visualization is one of the main techniques used to simplify the presentation of software systems and enhance their understandability. It is used to present the software system in a visual manner using simple, clear and meaningful symbols. This study proposes a new 2D software visualization approach. In this approach, each class is represented by rectangle, the name of the class placed above the rectangle, the size of class (Line of Code) represented by the height of the rectangle. The methods and the attributes are represented by circles and triangles respectively. The relationships among classes correspond to arrows. The proposed visualization approach was evaluated in terms of applicability and efficiency. Results have confirmed successful implementation of the proposed approach, and its ability to provide a simple and effective graphical presentation of extracted software components and properties.

Keywords: software visualization, software metrics, calling relationships, 2D graphs

Procedia PDF Downloads 205
11166 Valorization of the Algerian Plaster and Dune Sand in the Building Sector

Authors: S. Dorbani, F. Kharchi, F. Salem, K. Arroudj, N. Chioukh

Abstract:

The need for thermal comfort of buildings, with the aim of saving energy, has always generated a big interest during the development of methods, to improve the mode of construction. In the present paper, which is concerned by the valorization of locally abundant materials, mixtures of plaster and dune sand have been studied. To point out the thermal performances of these mixtures, a comparative study has been established between this product and the two materials most commonly used in construction, the concrete and hollow brick. The results showed that optimal mixture is made with 1/3 plaster and 2/3 dune sand. This mortar achieved significant increases in the mechanical strengths, which allow it to be used as a carrier element for buildings, of up to two levels. The element obtained offers an acceptable thermal insulation, with a decrease the outer-wall construction thickness.

Keywords: local materials, mortar, plaster, dune sand, compaction, mechanical performance, thermal performance

Procedia PDF Downloads 483
11165 Solution of S3 Problem of Deformation Mechanics for a Definite Condition and Resulting Modifications of Important Failure Theories

Authors: Ranajay Bhowmick

Abstract:

Analysis of stresses for an infinitesimal tetrahedron leads to a situation where we obtain a cubic equation consisting of three stress invariants. This cubic equation, when solved for a definite condition, gives the principal stresses directly without requiring any cumbersome and time-consuming trial and error methods or iterative numerical procedures. Since the failure criterion of different materials are generally expressed as functions of principal stresses, an attempt has been made in this study to incorporate the solutions of the cubic equation in the form of principal stresses, obtained for a definite condition, into some of the established failure theories to determine their modified descriptions. It has been observed that the failure theories can be represented using the quadratic stress invariant and the orientation of the principal plane.

Keywords: cubic equation, stress invariant, trigonometric, explicit solution, principal stress, failure criterion

Procedia PDF Downloads 137