Search results for: balance ability
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5608

Search results for: balance ability

898 Forced-Choice Measurement Models of Behavioural, Social, and Emotional Skills: Theory, Research, and Development

Authors: Richard Roberts, Anna Kravtcova

Abstract:

Introduction: The realisation that personality can change over the course of a lifetime has led to a new companion model to the Big Five, the behavioural, emotional, and social skills approach (BESSA). BESSA hypothesizes that this set of skills represents how the individual is thinking, feeling, and behaving when the situation calls for it, as opposed to traits, which represent how someone tends to think, feel, and behave averaged across situations. The five major skill domains share parallels with the Big Five Factor (BFF) model creativity and innovation (openness), self-management (conscientiousness), social engagement (extraversion), cooperation (agreeableness), and emotional resilience (emotional stability) skills. We point to noteworthy limitations in the current operationalisation of BESSA skills (i.e., via Likert-type items) and offer up a different measurement approach: forced choice. Method: In this forced-choice paradigm, individuals were given three skill items (e.g., managing my time) and asked to select one response they believed they were “worst at” and “best at”. The Thurstonian IRT models allow these to be placed on a normative scale. Two multivariate studies (N = 1178) were conducted with a 22-item forced-choice version of the BESSA, a published measure of the BFF, and various criteria. Findings: Confirmatory factor analysis of the forced-choice assessment showed acceptable model fit (RMSEA<0.06), while reliability estimates were reasonable (around 0.70 for each construct). Convergent validity evidence was as predicted (correlations between 0.40 and 0.60 for corresponding BFF and BESSA constructs). Notable was the extent the forced-choice BESSA assessment improved upon test-criterion relationships over and above the BFF. For example, typical regression models find BFF personality accounting for 25% of the variance in life satisfaction scores; both studies showed incremental gains over the BFF exceeding 6% (i.e., BFF and BESSA together accounted for over 31% of the variance in both studies). Discussion: Forced-choice measurement models offer up the promise of creating equated test forms that may unequivocally measure skill gains and are less prone to fakability and reference bias effects. Implications for practitioners are discussed, especially those interested in selection, succession planning, and training and development. We also discuss how the forced choice method can be applied to other constructs like emotional immunity, cross-cultural competence, and self-estimates of cognitive ability.

Keywords: Big Five, forced-choice method, BFF, methods of measurements

Procedia PDF Downloads 74
897 Innovations in the Implementation of Preventive Strategies and Measuring Their Effectiveness Towards the Prevention of Harmful Incidents to People with Mental Disabilities who Receive Home and Community Based Services

Authors: Carlos V. Gonzalez

Abstract:

Background: Providers of in-home and community based services strive for the elimination of preventable harm to the people under their care as well as to the employees who support them. Traditional models of safety and protection from harm have assumed that the absence of incidents of harm is a good indicator of safe practices. However, this model creates an illusion of safety that is easily shaken by sudden and inadvertent harmful events. As an alternative, we have developed and implemented an evidence-based resilient model of safety known as C.O.P.E. (Caring, Observing, Predicting and Evaluating). Within this model, safety is not defined by the absence of harmful incidents, but by the presence of continuous monitoring, anticipation, learning, and rapid response to events that may lead to harm. Objective: The objective was to evaluate the effectiveness of the C.O.P.E. model for the reduction of harm to individuals with mental disabilities who receive home and community based services. Methods: Over the course of 2 years we counted the number of incidents of harm and near misses. We trained employees on strategies to eliminate incidents before they fully escalated. We trained employees to track different levels of patient status within a scale from 0 to 10. Additionally, we provided direct support professionals and supervisors with customized smart phone applications to track and notify the team of changes in that status every 30 minutes. Finally, the information that we collected was saved in a private computer network that analyzes and graphs the outcome of each incident. Result and conclusions: The use of the COPE model resulted in: A reduction in incidents of harm. A reduction the use of restraints and other physical interventions. An increase in Direct Support Professional’s ability to detect and respond to health problems. Improvement in employee alertness by decreasing sleeping on duty. Improvement in caring and positive interaction between Direct Support Professionals and the person who is supported. Developing a method to globally measure and assess the effectiveness of prevention from harm plans. Future applications of the COPE model for the reduction of harm to people who receive home and community based services are discussed.

Keywords: harm, patients, resilience, safety, mental illness, disability

Procedia PDF Downloads 428
896 Accurate Mass Segmentation Using U-Net Deep Learning Architecture for Improved Cancer Detection

Authors: Ali Hamza

Abstract:

Accurate segmentation of breast ultrasound images is of paramount importance in enhancing the diagnostic capabilities of breast cancer detection. This study presents an approach utilizing the U-Net architecture for segmenting breast ultrasound images aimed at improving the accuracy and reliability of mass identification within the breast tissue. The proposed method encompasses a multi-stage process. Initially, preprocessing techniques are employed to refine image quality and diminish noise interference. Subsequently, the U-Net architecture, a deep learning convolutional neural network (CNN), is employed for pixel-wise segmentation of regions of interest corresponding to potential breast masses. The U-Net's distinctive architecture, characterized by a contracting and expansive pathway, enables accurate boundary delineation and detailed feature extraction. To evaluate the effectiveness of the proposed approach, an extensive dataset of breast ultrasound images is employed, encompassing diverse cases. Quantitative performance metrics such as the Dice coefficient, Jaccard index, sensitivity, specificity, and Hausdorff distance are employed to comprehensively assess the segmentation accuracy. Comparative analyses against traditional segmentation methods showcase the superiority of the U-Net architecture in capturing intricate details and accurately segmenting breast masses. The outcomes of this study emphasize the potential of the U-Net-based segmentation approach in bolstering breast ultrasound image analysis. The method's ability to reliably pinpoint mass boundaries holds promise for aiding radiologists in precise diagnosis and treatment planning. However, further validation and integration within clinical workflows are necessary to ascertain their practical clinical utility and facilitate seamless adoption by healthcare professionals. In conclusion, leveraging the U-Net architecture for breast ultrasound image segmentation showcases a robust framework that can significantly enhance diagnostic accuracy and advance the field of breast cancer detection. This approach represents a pivotal step towards empowering medical professionals with a more potent tool for early and accurate breast cancer diagnosis.

Keywords: mage segmentation, U-Net, deep learning, breast cancer detection, diagnostic accuracy, mass identification, convolutional neural network

Procedia PDF Downloads 60
895 A Review of Digital Twins to Reduce Emission in the Construction Industry

Authors: Zichao Zhang, Yifan Zhao, Samuel Court

Abstract:

The carbon emission problem of the traditional construction industry has long been a pressing issue. With the growing emphasis on environmental protection and advancement of science and technology, the organic integration of digital technology and emission reduction has gradually become a mainstream solution. Among various sophisticated digital technologies, digital twins, which involve creating virtual replicas of physical systems or objects, have gained enormous attention in recent years as tools to improve productivity, optimize management and reduce carbon emissions. However, the relatively high implementation costs including finances, time, and manpower associated with digital twins have limited their widespread adoption. As a result, most of the current applications are primarily concentrated within a few industries. In addition, the creation of digital twins relies on a large amount of data and requires designers to possess exceptional skills in information collection, organization, and analysis. Unfortunately, these capabilities are often lacking in the traditional construction industry. Furthermore, as a relatively new concept, digital twins have different expressions and usage methods across different industries. This lack of standardized practices poses a challenge in creating a high-quality digital twin framework for construction. This paper firstly reviews the current academic studies and industrial practices focused on reducing greenhouse gas emissions in the construction industry using digital twins. Additionally, it identifies the challenges that may be encountered during the design and implementation of a digital twin framework specific to this industry and proposes potential directions for future research. This study shows that digital twins possess substantial potential and significance in enhancing the working environment within the traditional construction industry, particularly in their ability to support decision-making processes. It proves that digital twins can improve the work efficiency and energy utilization of related machinery while helping this industry save energy and reduce emissions. This work will help scholars in this field to better understand the relationship between digital twins and energy conservation and emission reduction, and it also serves as a conceptual reference for practitioners to implement related technologies.

Keywords: digital twins, emission reduction, construction industry, energy saving, life cycle, sustainability

Procedia PDF Downloads 65
894 Configuring Resilience and Environmental Sustainability to Achieve Superior Performance under Differing Conditions of Transportation Disruptions

Authors: Henry Ataburo, Dominic Essuman, Emmanuel Kwabena Anin

Abstract:

Recent trends of catastrophic events, such as the Covid-19 pandemic, the Suez Canal blockage, the Russia-Ukraine conflict, the Israel-Hamas conflict, and the climate change crisis, continue to devastate supply chains and the broader society. Prior authors have advocated for a simultaneous pursuit of resilience and sustainability as crucial for navigating these challenges. Nevertheless, the relationship between resilience and sustainability is a rather complex one: resilience and sustainability are considered unrelated, substitutes, or complements. Scholars also suggest that different firms prioritize resilience and sustainability differently for varied strategic reasons. However, we know little about whether, how, and when these choices produce different typologies of firms to explain differences in financial and market performance outcomes. This research draws inferences from the systems configuration approach to organizational fit to contend that a taxonomy of firms may emerge based on how firms configure resilience and environmental sustainability. The study further examines the effects of these taxonomies on financial and market performance in differing transportation disruption conditions. Resilience is operationalized as a firm’s ability to adjust current operations, structure, knowledge, and resources in response to disruptions, whereas environmental sustainability is operationalized as the extent to which a firm deploys resources judiciously and keeps the ecological impact of its operations to the barest minimum. Using primary data from 199 firms in Ghana and cluster analysis as an analytical tool, the study identifies four clusters of firms based on how they prioritize resilience and sustainability: Cluster 1 - "strong, moderate resilience, high sustainability firms," Cluster 2 - "sigh resilience, high sustainability firms," Cluster 3 - "high resilience, strong, moderate sustainability firms," and Cluster 4 - "weak, moderate resilience, strong, moderate sustainability firms". In addition, ANOVA and regression analysis revealed the following findings: Only clusters 1 and 2 were significantly associated with both market and financial performance. Under high transportation disruption conditions, cluster 1 firms excel better in market performance, whereas cluster 2 firms excel better in financial performance. Conversely, under low transportation disruption conditions, cluster 1 firms excel better in financial performance, whereas cluster 2 firms excel better in market performance. The study provides theoretical and empirical evidence of how resilience and environmental sustainability can be configured to achieve specific performance objectives under different disruption conditions.

Keywords: resilience, environmental sustainability, developing economy, transportation disruption

Procedia PDF Downloads 46
893 New Advanced Medical Software Technology Challenges and Evolution of the Regulatory Framework in Expert Software, Artificial Intelligence, and Machine Learning

Authors: Umamaheswari Shanmugam, Silvia Ronchi, Radu Vornicu

Abstract:

Software, artificial intelligence, and machine learning can improve healthcare through innovative and advanced technologies that are able to use the large amount and variety of data generated during healthcare services every day. As we read the news, over 500 machine learning or other artificial intelligence medical devices have now received FDA clearance or approval, the first ones even preceding the year 2000. One of the big advantages of these new technologies is the ability to get experience and knowledge from real-world use and to continuously improve their performance. Healthcare systems and institutions can have a great benefit because the use of advanced technologies improves the same time efficiency and efficacy of healthcare. Software-defined as a medical device, is stand-alone software that is intended to be used for patients for one or more of these specific medical intended uses: - diagnosis, prevention, monitoring, prediction, prognosis, treatment or alleviation of a disease, any other health conditions, replacing or modifying any part of a physiological or pathological process–manage the received information from in vitro specimens derived from the human samples (body) and without principal main action of its principal intended use by pharmacological, immunological or metabolic definition. Software qualified as medical devices must comply with the general safety and performance requirements applicable to medical devices. These requirements are necessary to ensure high performance and quality and also to protect patients’ safety. The evolution and the continuous improvement of software used in healthcare must take into consideration the increase in regulatory requirements, which are becoming more complex in each market. The gap between these advanced technologies and the new regulations is the biggest challenge for medical device manufacturers. Regulatory requirements can be considered a market barrier, as they can delay or obstacle the device approval, but they are necessary to ensure performance, quality, and safety, and at the same time, they can be a business opportunity if the manufacturer is able to define in advance the appropriate regulatory strategy. The abstract will provide an overview of the current regulatory framework, the evolution of the international requirements, and the standards applicable to medical device software in the potential market all over the world.

Keywords: artificial intelligence, machine learning, SaMD, regulatory, clinical evaluation, classification, international requirements, MDR, 510k, PMA, IMDRF, cyber security, health care systems.

Procedia PDF Downloads 71
892 Guidelines for the Management Process Development of Research Journals in Order to Develop Suan Sunandha Rajabhat University to International Standards

Authors: Araya Yordchim, Rosjana Chandhasa, Suwaree Yordchim

Abstract:

This research aims to study guidelines on the development of management process for research journals in order to develop Suan Sunandha Rajabhat University to international standards. This research investigated affecting elements ranging from the format of the article, evaluation form for research article quality, the process of creating a scholarly journal, satisfaction level of those with knowledge and competency to conduct research, arisen problems, and solutions. Drawing upon the sample size of 40 persons who had knowledge and competency in conducting research and creating scholarly journal articles at an international level, the data for this research were collected using questionnaires as a tool. Through the usage of computer software, data were analyzed by using the statistics in the forms of frequency, percentage, mean, standard deviation, and multiple regression analysis. The majority of participants were civil servants with a doctorate degree, followed by civil servants with a master's degree. Among them, the suitability of the article format was rated at a good level while the evaluation form for research articles quality was assessed at a good level. Based on participants' viewpoints, the process of creating scholarly journals was at a good level, while the satisfaction of those who had knowledge and competency in conducting research was at a satisfactory level. The problems encountered were the difficulty in accessing the website. The solution to the problem was to develop a website with user-friendly accessibility, including setting up a Google scholar profile for the purpose of references counting and the articles being used for reference in real-time. Research article format influenced the level of satisfaction of those who had the knowledge and competency to conduct research with statistical significance at the 0.01 level. The research article quality assessment form (preface section, research article writing section, preparation for research article manuscripts section, and the original article evaluation form for the author) affected the satisfaction of those with knowledge and competency to conduct research with the statistical significance at the level of 0.01. The process of establishing journals had an impact on the satisfaction of those with knowledge and ability to conduct research with statistical significance at the level of .05

Keywords: guidelines, development of management, research journals, international standards

Procedia PDF Downloads 111
891 Rhizospheric Oxygen Release of Hydroponically Grown Wetland Macrophytes as Passive Source for Cathodic Reduction in Microbial Fuel Cell

Authors: Chabungbam Niranjit Khuman, Makarand Madhao Ghangrekar, Arunabha Mitra

Abstract:

The cost of aeration is one of the limiting factors in the upscaling of microbial fuel cells (MFC) for field-scale applications. Wetland macrophytes have the ability to release oxygen into the water to maintain aerobic conditions in their root zone. In this experiment, the efficacy of rhizospheric oxygen release of wetland macrophytes as a source of oxygen in the cathodic chamber of MFC was conducted. The experiment was conducted in an MFC consisting of a three-liter anodic chamber made of ceramic cylinder and a 27 L cathodic chamber. Untreated carbon felts were used as electrodes (i.e., anode and cathode) and connected to an external load of 100 Ω using stainless steel wire. Wetland macrophytes (Canna indica) were grown in the cathodic chamber of the MFC in a hydroponic fashion using a styrofoam sheet (termed as macrophytes assisted-microbial fuel cell, M-MFC). The catholyte (i.e., water) in the M-MFC had negligible contact with atmospheric air due to the styrofoam sheet used for maintaining the hydroponic condition. There was no mixing of the catholyte in the M-MFC. Sucrose based synthetic wastewater having chemical oxygen demand (COD) of 3000 mg/L was fed into the anodic chamber of the MFC in fed-batch mode with a liquid retention time of four days. The C. indica thrived well throughout the duration of the experiment without much care. The average dissolved oxygen (DO) concentration and pH value in the M-MFC were 3.25 mg/L and 7.07, respectively, in the catholyte. Since the catholyte was not in contact with air, the DO in the catholyte might be considered as solely liberated from the rhizospheric oxygen release of C. indica. The maximum COD removal efficiency of M-MFC observed during the experiment was 76.9%. The inadequacy of terminal electron acceptor in the cathodic chamber in M-MFC might have hampered the electron transfer, which in turn, led to slower specific microbial activity, thereby resulting in lower COD removal efficiency than the traditional MFC with aerated catholyte. The average operating voltage (OV) and open-circuit voltage (OCV) of 294 mV and 594 mV, respectively, were observed in M-MFC. The maximum power density observed during polarization was 381 mW/m³, and the maximum sustainable power density observed during the experiment was 397 mW/m³ in M-MFC. The maximum normalized energy recovery and coulombic efficiency of 38.09 Wh/m³ and 1.27%, respectively, were observed. Therefore, it was evidenced that rhizospheric oxygen release of wetland macrophytes (C. indica) was capable of sustaining the cathodic reaction in MFC for field-scale applications.

Keywords: hydroponic, microbial fuel cell, rhizospheric oxygen release, wetland macrophytes

Procedia PDF Downloads 113
890 Structuring Highly Iterative Product Development Projects by Using Agile-Indicators

Authors: Guenther Schuh, Michael Riesener, Frederic Diels

Abstract:

Nowadays, manufacturing companies are faced with the challenge of meeting heterogeneous customer requirements in short product life cycles with a variety of product functions. So far, some of the functional requirements remain unknown until late stages of the product development. A way to handle these uncertainties is the highly iterative product development (HIP) approach. By structuring the development project as a highly iterative process, this method provides customer oriented and marketable products. There are first approaches for combined, hybrid models comprising deterministic-normative methods like the Stage-Gate process and empirical-adaptive development methods like SCRUM on a project management level. However, almost unconsidered is the question, which development scopes can preferably be realized with either empirical-adaptive or deterministic-normative approaches. In this context, a development scope constitutes a self-contained section of the overall development objective. Therefore, this paper focuses on a methodology that deals with the uncertainty of requirements within the early development stages and the corresponding selection of the most appropriate development approach. For this purpose, internal influencing factors like a company’s technology ability, the prototype manufacturability and the potential solution space as well as external factors like the market accuracy, relevance and volatility will be analyzed and combined into an Agile-Indicator. The Agile-Indicator is derived in three steps. First of all, it is necessary to rate each internal and external factor in terms of the importance for the overall development task. Secondly, each requirement has to be evaluated for every single internal and external factor appropriate to their suitability for empirical-adaptive development. Finally, the total sums of internal and external side are composed in the Agile-Indicator. Thus, the Agile-Indicator constitutes a company-specific and application-related criterion, on which the allocation of empirical-adaptive and deterministic-normative development scopes can be made. In a last step, this indicator will be used for a specific clustering of development scopes by application of the fuzzy c-means (FCM) clustering algorithm. The FCM-method determines sub-clusters within functional clusters based on the empirical-adaptive environmental impact of the Agile-Indicator. By means of the methodology presented in this paper, it is possible to classify requirements, which are uncertainly carried out by the market, into empirical-adaptive or deterministic-normative development scopes.

Keywords: agile, highly iterative development, agile-indicator, product development

Procedia PDF Downloads 224
889 Synthesis of Porphyrin-Functionalized Beads for Flow Cytometry

Authors: William E. Bauta, Jennifer Rebeles, Reggie Jacob

Abstract:

Porphyrins are noteworthy in biomedical science for their cancer tissue accumulation and photophysical properties. The preferential accumulation of some porphyrins in cancerous tissue has been known for many years. This, combined with their characteristic photophysical and photochemical properties, including their strong fluorescence and their ability to generate reactive oxygen species in vivo upon laser irradiation, has led to much research into the application of porphyrins as cancer diagnostic and therapeutic agents. Porphyrins have been used as dyes to detect cancer cells both in vivo and, less commonly, in vitro. In one example, human sputum samples from lung cancer patients and patients without the disease were dissociated and stained with the porphyrin TCPP (5,10,15,20-tetrakis-(4-carboxyphenyl)-porphine). Cells were analyzed by flow cytometry. Cancer samples were identified by their higher TCPP fluorescence intensity relative to the no-cancer controls. However, quantitative analysis of fluorescence in cell suspensions stained with multiple fluorophores requires particles stained with each of the individual fluorophores as controls. Fluorescent control particles must be compatible in size with flow cytometer fluidics and have favorable hydrodynamic properties in suspension. They must also display fluorescence comparable to the cells of interest and be stable upon storage amine-functionalized spherical polystyrene beads in the 5 to 20-micron diameter range that was reacted with TCPP and EDC in aqueous pH six buffer overnight to form amide bonds. Beads were isolated by centrifugation and tested by flow cytometry. The 10-micron amine-functionalized beads displayed the best combination of fluorescence intensity and hydrodynamic properties, such as lack of clumping and remaining in suspension during the experiment. These beads were further optimized by varying the stoichiometry of EDC and TCPP relative to the amine. The reaction was accompanied by the formation of a TCPP-related particulate, which was removed, after bead centrifugation, using a microfiltration process. The resultant TCPP-functionalized beads were compatible with flow cytometry conditions and displayed a fluorescence comparable to that of stained cells, which allowed their use as fluorescence standards. The beads were stable in refrigerated storage in the dark for more than eight months. This work demonstrates the first preparation of porphyrin-functionalized flow cytometry control beads.

Keywords: tetraaryl porphyrin, polystyrene beads, flow cytometry, peptide coupling

Procedia PDF Downloads 71
888 Transcending Boundaries: Integrating Urban Vibrancy with Contemporary Interior Design through Vivid Wall Pieces

Authors: B. C. Biermann

Abstract:

This in-depth exploration investigates the transformative integration of urban vibrancy into contemporary interior design through the strategic incorporation of vivid wall pieces. Bridging the gap between public dynamism and private tranquility, this study delves into the nuanced methodologies, creative processes, and profound impacts of this innovative approach. Drawing inspiration from street art's dynamic language and the timeless allure of natural beauty, these artworks serve as conduits, orchestrating a dialogue that challenges traditional boundaries and redefines the relationship between external chaos and internal sanctuaries. The fusion of urban vibrancy with contemporary interior design represents a paradigm shift, where the inherent dynamism of public spaces harmoniously converges with the curated tranquility of private environments. This paper aims to explore the underlying principles, creative processes, and transformative impacts of integrating vivid wall pieces as instruments for bringing the "outside in." Employing an innovative and meticulous methodology, street art elements are synthesized with the refined aesthetics of contemporary design. This delicate balance necessitates a nuanced understanding of both artistic realms, ensuring a synthesis that captures the essence of urban energy while seamlessly blending with the sophistication of modern interior design. The creative process involves a strategic selection of street art motifs, colors, and textures that resonate with the organic beauty found in natural landscapes, creating a symbiotic relationship between the grittiness of the streets and the elegance of interior spaces. This groundbreaking approach defies traditional boundaries by integrating dynamic street art into interior spaces, blurring the demarcation between external chaos and internal tranquility. Vivid wall pieces serve as dynamic focal points, transforming physical spaces and challenging conventional perceptions of where art belongs. This redefinition asserts that boundaries are fluid and meant to be transcended. Case studies illustrate the profound impact of integrating vivid wall pieces on the aesthetic appeal of interior spaces. Urban vibrancy revitalizes the atmosphere, infusing it with palpable energy that resonates with the vivacity of public spaces. The curated tranquility of private interiors coexists harmoniously with the dynamic visual language of street art, fostering a unique and evolving relationship between inhabitants and their living spaces. Emphasizing harmonious coexistence, the paper underscores the potential for a seamless dialogue between public urban spaces and private interiors. The integration of vivid wall pieces acts as a bridge rather than a dichotomy, merging the dynamism of street art with the curated elegance of contemporary design. This unique visual tapestry transcends traditional categorizations, fostering a symbiotic relationship between contrasting worlds. In conclusion, this paper posits that the integration of vivid wall pieces represents a transformative tool for contemporary interior design, challenging and redefining conventional boundaries. By strategically bringing the "outside in," this approach transforms interior spaces and heralds a paradigm shift in the relationship between urban aesthetics and contemporary living. The ongoing narrative between urban vibrancy and interior design creates spaces that reflect the dynamic and ever-evolving nature of the surrounding environment.

Keywords: Art Integration, Contemporary Interior Design, Interior Space Transformation, Vivid Wall Pieces

Procedia PDF Downloads 55
887 The Plight of the Rohingyas: Design Guidelines to Accommodate Displaced People in Bangladesh

Authors: Nazia Roushan, Maria Kipti

Abstract:

The sensitive issue of a large-scale entry of Rohingya refugees to Bangladesh has arisen again since August of 2017. Incited by ethnic and religious conflict, the Rohingyas—an ethnic group concentrated in the north-west state of Rakhine in Myanmar—have been fleeing to what is now Bangladesh from as early as the late 1700s in four main exoduses. This long-standing persecution has recently escalated, and accommodating the recent wave of exodus has been especially challenging due to the sheer volume of a million refugees concentrated in refugee camps in two small administrative units (upazilas) in the south-east of the country: the host area. This drastic change in the host area’s social fabric is putting a lot of strain on the country’s economic, demographic and environmental stability, and security. Although Bangladesh’s long-term experience with disaster management has enabled it to respond rapidly to the crisis, the government is failing to cope with this enormous problem and has taken insufficient steps towards improving the living conditions to inhibit the inflow of more refugees. On top of that, the absence of a comprehensive national refugee policy, and the density of the structures of the camps are constricting the upgrading of the shelters to international standards. As of December 2016, the combined number of internally displaced persons (IDPs) due to conflict and violence (stock), and new displacements due to disasters (flow) in Bangladesh had exceeded 1 million. These numbers have increased dramatically in the last few months. Moreover, by 2050, Bangladesh will have as much as 25 million climate refugees just from its coastal districts. To enhance the resilience of the vulnerable, it is crucial to methodically factorize further interventions between Disaster Risk Reduction for Resilience (DRR) and the concept of Building Back Better (BBB) in the rehabilitation-reconstruction period. Considering these points, this paper provides a palette of options for design guidelines related to the living spaces and infrastructures for refugees. This will encourage the development of national standards for refugee camps, and the national and local level rehabilitation-reconstruction practices. Unhygienic living conditions, vulnerability, and the general lack of control over life are pervasive throughout the camps. This paper, therefore, proposes site-specific strategic and physical planning and design for shelters for refugees in Bangladesh that will lead to sustainable living environments through the following: a) site survey of existing two registered and one makeshift unregistered refugee camps to document and study their physical conditions, b) questionnaires and semi-structured focus group discussions carried out among the refugees and stakeholders to understand what the lived experiences and needs are; and c) combining the findings with international minimum standards for shelter and settlement from International Federation of Red Cross and Red Crescent (IFRC), Médecins Sans Frontières (MSF), United Nations High Commissioner for Refugees (UNHCR). These proposals include temporary shelter solutions that balance between lived spaces and regimented, repetitive plans using readily available and cheap materials, erosion control and slope stabilization strategies, and most importantly, coping mechanisms for the refugees to be self-reliant and resilient.

Keywords: architecture, Bangladesh, refugee camp, resilience, Rohingya

Procedia PDF Downloads 214
886 Towards Modern Approaches of Intelligence Measurement for Clinical and Educational Practices

Authors: Alena Kulikova, Tatjana Kanonire

Abstract:

Intelligence research is one of the oldest fields of psychology. Many factors have made a research on intelligence, defined as reasoning and problem solving [1, 2], a very acute and urgent problem. Thus, it has been repeatedly shown that intelligence is a predictor of academic, professional, and social achievement in adulthood (for example, [3]); Moreover, intelligence predicts these achievements better than any other trait or ability [4]. The individual level, a comprehensive assessment of intelligence is a necessary criterion for the diagnosis of various mental conditions. For example, it is a necessary condition for psychological, medical and pedagogical commissions when deciding on educational needs and the most appropriate educational programs for school children. Assessment of intelligence is crucial in clinical psychodiagnostic and needs high-quality intelligence measurement tools. Therefore, it is not surprising that the development of intelligence tests is an essential part of psychological science and practice. Many modern intelligence tests have a long history and have been used for decades, for example, the Stanford-Binet test or the Wechsler test. However, the vast majority of these tests are based on the classic linear test structure, in which all respondents receive all tasks (see, for example, a critical review by [5]). This understanding of the testing procedure is a legacy of the pre-computer era, in which blank testing was the only diagnostic procedure available [6] and has some significant limitations that affect the reliability of the data obtained [7] and increased time costs. Another problem with measuring IQ is that classical line-structured tests do not fully allow to measure respondent's intellectual progress [8], which is undoubtedly a critical limitation. Advances in modern psychometrics allow for avoiding the limitations of existing tools. However, as in any rapidly developing industry, at the moment, psychometrics does not offer ready-made and straightforward solutions and requires additional research. In our presentation we would like to discuss the strengths and weaknesses of the current approaches to intelligence measurement and highlight “points of growth” for creating a test in accordance with modern psychometrics. Whether it is possible to create the instrument that will use all achievements of modern psychometric and remain valid and practically oriented. What would be the possible limitations for such an instrument? The theoretical framework and study design to create and validate the original Russian comprehensive computer test for measuring the intellectual development in school-age children will be presented.

Keywords: Intelligence, psychometrics, psychological measurement, computerized adaptive testing, multistage testing

Procedia PDF Downloads 64
885 Effect of Automatic Self Transcending Meditation on Perceived Stress and Sleep Quality in Adults

Authors: Divya Kanchibhotla, Shashank Kulkarni, Shweta Singh

Abstract:

Chronic stress and sleep quality reduces mental health and increases the risk of developing depression and anxiety as well. There is increasing evidence for the utility of meditation as an adjunct clinical intervention for conditions like depression and anxiety. The present study is an attempt to explore the impact of Sahaj Samadhi Meditation (SSM), a category of Automatic Self Transcending Meditation (ASTM), on perceived stress and sleep quality in adults. The study design was a single group pre-post assessment. Perceived Stress Scale (PSS) and the Pittsburgh Sleep Quality Index (PSQI) were used in this study. Fifty-two participants filled PSS, and 60 participants filled PSQI at the beginning of the program (day 0), after two weeks (day 16) and at two months (day 60). Significant pre-post differences for the perceived stress level on Day 0 - Day 16 (p < 0.01; Cohen's d = 0.46) and Day 0 - Day 60 (p < 0.01; Cohen's d = 0.76) clearly demonstrated that by practicing SSM, participants experienced reduction in the perceived stress. The effect size of the intervention observed on the 16th day of assessment was small to medium, but on the 60th day, a medium to large effect size of the intervention was observed. In addition to this, significant pre-post differences for the sleep quality on Day 0 - Day 16 and Day 0 - Day 60 (p < 0.05) clearly demonstrated that by practicing SSM, participants experienced improvement in the sleep quality. Compared with Day 0 assessment, participants demonstrated significant improvement in the quality of sleep on Day 16 and Day 60. The effect size of the intervention observed on the 16th day of assessment was small, but on the 60th day, a small to medium effect size of the intervention was observed. In the current study we found out that after practicing SSM for two months, participants reported a reduction in the perceived stress, they felt that they are more confident about their ability to handle personal problems, were able to cope with all the things that they had to do, felt that they were on top of the things, and felt less angered. Participants also reported that their overall sleep quality improved; they took less time to fall asleep; they had less disturbances in sleep and less daytime dysfunction due to sleep deprivation. The present study provides clear evidence of the efficacy and safety of non-pharmacological interventions such as SSM in reducing stress and improving sleep quality. Thus, ASTM may be considered a useful intervention to reduce psychological distress in healthy, non-clinical populations, and it can be an alternative remedy for treating poor sleep among individuals and decreasing the use of harmful sedatives.

Keywords: automatic self transcending meditation, Sahaj Samadhi meditation, sleep, stress

Procedia PDF Downloads 113
884 Leadership and Corporate Social Responsibility: The Role of Spiritual Intelligence

Authors: Meghan E. Murray, Carri R. Tolmie

Abstract:

This study aims to identify potential factors and widely applicable best practices that can contribute to improving corporate social responsibility (CSR) and corporate performance for firms by exploring the relationship between transformational leadership, spiritual intelligence, and emotional intelligence. Corporate social responsibility is when companies are cognizant of the impact of their actions on the economy, their communities, the environment, and the world as a whole while executing business practices accordingly. The prevalence of CSR has continuously strengthened over the past few years and is now a common practice in the business world, with such efforts coinciding with what stakeholders and the public now expect from corporations. Because of this, it is extremely important to be able to pinpoint factors and best practices that can improve CSR within corporations. One potential factor that may lead to improved CSR is spiritual intelligence (SQ), or the ability to recognize and live with a purpose larger than oneself. Spiritual intelligence is a measurable skill, just like emotional intelligence (EQ), and can be improved through purposeful and targeted coaching. This research project consists of two studies. Study 1 is a case study comparison of a benefit corporation and a non-benefit corporation. This study will examine the role of SQ and EQ as moderators in the relationship between the transformational leadership of employees within each company and the perception of each firm’s CSR and corporate performance. Project methodology includes creating and administering a survey comprised of multiple pre-established scales on transformational leadership, spiritual intelligence, emotional intelligence, CSR, and corporate performance. Multiple regression analysis will be used to extract significant findings from the collected data. Study 2 will dive deeper into spiritual intelligence itself by analyzing pre-existing data and identifying key relationships that may provide value to companies and their stakeholders. This will be done by performing multiple regression analysis on anonymized data provided by Deep Change, a company that has created an advanced, proprietary system to measure spiritual intelligence. Based on the results of both studies, this research aims to uncover best practices, including the unique contribution of spiritual intelligence, that can be utilized by organizations to help enhance their corporate social responsibility. If it is found that high spiritual and emotional intelligence can positively impact CSR effort, then corporations will have a tangible way to enhance their CSR: providing targeted employees with training and coaching to increase their SQ and EQ.

Keywords: corporate social responsibility, CSR, corporate performance, emotional intelligence, EQ, spiritual intelligence, SQ, transformational leadership

Procedia PDF Downloads 107
883 Handling, Exporting and Archiving Automated Mineralogy Data Using TESCAN TIMA

Authors: Marek Dosbaba

Abstract:

Within the mining sector, SEM-based Automated Mineralogy (AM) has been the standard application for quickly and efficiently handling mineral processing tasks. Over the last decade, the trend has been to analyze larger numbers of samples, often with a higher level of detail. This has necessitated a shift from interactive sample analysis performed by an operator using a SEM, to an increased reliance on offline processing to analyze and report the data. In response to this trend, TESCAN TIMA Mineral Analyzer is designed to quickly create a virtual copy of the studied samples, thereby preserving all the necessary information. Depending on the selected data acquisition mode, TESCAN TIMA can perform hyperspectral mapping and save an X-ray spectrum for each pixel or segment, respectively. This approach allows the user to browse through elemental distribution maps of all elements detectable by means of energy dispersive spectroscopy. Re-evaluation of the existing data for the presence of previously unconsidered elements is possible without the need to repeat the analysis. Additional tiers of data such as a secondary electron or cathodoluminescence images can also be recorded. To take full advantage of these information-rich datasets, TIMA utilizes a new archiving tool introduced by TESCAN. The dataset size can be reduced for long-term storage and all information can be recovered on-demand in case of renewed interest. TESCAN TIMA is optimized for network storage of its datasets because of the larger data storage capacity of servers compared to local drives, which also allows multiple users to access the data remotely. This goes hand in hand with the support of remote control for the entire data acquisition process. TESCAN also brings a newly extended open-source data format that allows other applications to extract, process and report AM data. This offers the ability to link TIMA data to large databases feeding plant performance dashboards or geometallurgical models. The traditional tabular particle-by-particle or grain-by-grain export process is preserved and can be customized with scripts to include user-defined particle/grain properties.

Keywords: Tescan, electron microscopy, mineralogy, SEM, automated mineralogy, database, TESCAN TIMA, open format, archiving, big data

Procedia PDF Downloads 89
882 Resilience and Urban Transformation: A Review of Recent Interventions in Europe and Turkey

Authors: Bilge Ozel

Abstract:

Cities are high-complex living organisms and are subjects to continuous transformations produced by the stress that derives from changing conditions. Today the metropolises are seen like “development engines” of the countries and accordingly they become the centre of better living conditions that encourages demographic growth which constitutes the main reason of the changes. Indeed, the potential for economic advancement of the cities directly represents the economic status of their countries. The term of “resilience”, which sees the changes as natural processes and represents the flexibility and adaptability of the systems in the face of changing conditions, becomes a key concept for the development of urban transformation policies. The term of “resilience” derives from the Latin word ‘resilire’, which means ‘bounce’, ‘jump back’, refers to the ability of a system to withstand shocks and still maintain the basic characteristics. A resilient system does not only survive the potential risks and threats but also takes advantage of the positive outcomes of the perturbations and ensures adaptation to the new external conditions. When this understanding is taken into the urban context - or rather “urban resilience” - it delineates the capacity of cities to anticipate upcoming shocks and changes without undergoing major alterations in its functional, physical, socio-economic systems. Undoubtedly, the issue of coordinating the urban systems in a “resilient” form is a multidisciplinary and complex process as the cities are multi-layered and dynamic structures. The concept of “urban transformation” is first launched in Europe just after World War II. It has been applied through different methods such as renovation, revitalization, improvement and gentrification. These methods have been in continuous advancement by acquiring new meanings and trends over years. With the effects of neoliberal policies in the 1980s, the concept of urban transformation has been associated with economic objectives. Subsequently this understanding has been improved over time and had new orientations such as providing more social justice and environmental sustainability. The aim of this research is to identify the most applied urban transformation methods in Turkey and its main reasons of being selected. Moreover, investigating the lacking and limiting points of the urban transformation policies in the context of “urban resilience” in a comparative way with European interventions. The emblematic examples, which symbolize the breaking points of the recent evolution of urban transformation concepts in Europe and Turkey, are chosen and reviewed in a critical way.

Keywords: resilience, urban dynamics, urban resilience, urban transformation

Procedia PDF Downloads 247
881 Computational Simulations and Assessment of the Application of Non-Circular TAVI Devices

Authors: Jonathon Bailey, Neil Bressloff, Nick Curzen

Abstract:

Transcatheter Aortic Valve Implantation (TAVI) devices are stent-like frames with prosthetic leaflets on the inside, which are percutaneously implanted. The device in a crimped state is fed through the arteries to the aortic root, where the device frame is opened through either self-expansion or balloon expansion, which reveals the prosthetic valve within. The frequency at which TAVI is being used to treat aortic stenosis is rapidly increasing. In time, TAVI is likely to become the favoured treatment over Surgical Valve Replacement (SVR). Mortality after TAVI has been associated with severe Paravalvular Aortic Regurgitation (PAR). PAR occurs when the frame of the TAVI device does not make an effective seal against the internal surface of the aortic root, allowing blood to flow backwards about the valve. PAR is common in patients and has been reported to some degree in as much as 76% of cases. Severe PAR (grade 3 or 4) has been reported in approximately 17% of TAVI patients resulting in post-procedural mortality increases from 6.7% to 16.5%. TAVI devices, like SVR devices, are circular in cross-section as the aortic root is often considered to be approximately circular in shape. In reality, however, the aortic root is often non-circular. The ascending aorta, aortic sino tubular junction, aortic annulus and left ventricular outflow tract have an average ellipticity ratio of 1.07, 1.09, 1.29, and 1.49 respectively. An elliptical aortic root does not severely affect SVR, as the leaflets are completely removed during the surgical procedure. However, an elliptical aortic root can inhibit the ability of the circular Balloon-Expandable (BE) TAVI devices to conform to the interior of the aortic root wall, which increases the risk of PAR. Self-Expanding (SE) TAVI devices are considered better at conforming to elliptical aortic roots, however the valve leaflets were not designed for elliptical function, furthermore the incidence of PAR is greater in SE devices than BE devices (19.8% vs. 12.2% respectively). If a patient’s aortic root is too severely elliptical, they will not be suitable for TAVI, narrowing the treatment options to SVR. It therefore follows that in order to increase the population who can undergo TAVI, and reduce the risk associated with TAVI, non-circular devices should be developed. Computational simulations were employed to further advance our understanding of non-circular TAVI devices. Radial stiffness of the TAVI devices in multiple directions, frame bending stiffness and resistance to balloon induced expansion are all computationally simulated. Finally, a simulation has been developed that demonstrates the expansion of TAVI devices into a non-circular patient specific aortic root model in order to assess the alterations in deployment dynamics, PAR and the stresses induced in the aortic root.

Keywords: tavi, tavr, fea, par, fem

Procedia PDF Downloads 427
880 Quantifying Automation in the Architectural Design Process via a Framework Based on Task Breakdown Systems and Recursive Analysis: An Exploratory Study

Authors: D. M. Samartsev, A. G. Copping

Abstract:

As with all industries, architects are using increasing amounts of automation within practice, with approaches such as generative design and use of AI becoming more commonplace. However, the discourse on the rate at which the architectural design process is being automated is often personal and lacking in objective figures and measurements. This results in confusion between people and barriers to effective discourse on the subject, in turn limiting the ability of architects, policy makers, and members of the public in making informed decisions in the area of design automation. This paper proposes the use of a framework to quantify the progress of automation within the design process. The use of a reductionist analysis of the design process allows it to be quantified in a manner that enables direct comparison across different times, as well as locations and projects. The methodology is informed by the design of this framework – taking on the aspects of a systematic review but compressed in time to allow for an initial set of data to verify the validity of the framework. The use of such a framework of quantification enables various practical uses such as predicting the future of the architectural industry with regards to which tasks will be automated, as well as making more informed decisions on the subject of automation on multiple levels ranging from individual decisions to policy making from governing bodies such as the RIBA. This is achieved by analyzing the design process as a generic task that needs to be performed, then using principles of work breakdown systems to split the task of designing an entire building into smaller tasks, which can then be recursively split further as required. Each task is then assigned a series of milestones that allow for the objective analysis of its automation progress. By combining these two approaches it is possible to create a data structure that describes how much various parts of the architectural design process are automated. The data gathered in the paper serves the dual purposes of providing the framework with validation, as well as giving insights into the current situation of automation within the architectural design process. The framework can be interrogated in many ways and preliminary analysis shows that almost 40% of the architectural design process has been automated in some practical fashion at the time of writing, with the rate at which progress is made slowly increasing over the years, with the majority of tasks in the design process reaching a new milestone in automation in less than 6 years. Additionally, a further 15% of the design process is currently being automated in some way, with various products in development but not yet released to the industry. Lastly, various limitations of the framework are examined in this paper as well as further areas of study.

Keywords: analysis, architecture, automation, design process, technology

Procedia PDF Downloads 81
879 Towards End-To-End Disease Prediction from Raw Metagenomic Data

Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker

Abstract:

Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.

Keywords: deep learning, disease prediction, end-to-end machine learning, metagenomics, multiple instance learning, precision medicine

Procedia PDF Downloads 103
878 Performance of HVOF Sprayed Ni-20CR and Cr3C2-NiCr Coatings on Fe-Based Superalloy in an Actual Industrial Environment of a Coal Fired Boiler

Authors: Tejinder Singh Sidhu

Abstract:

Hot corrosion has been recognized as a severe problem in steam-powered electricity generation plants and industrial waste incinerators as it consumes the material at an unpredictably rapid rate. Consequently, the load-carrying ability of the components reduces quickly, eventually leading to catastrophic failure. The inability to either totally prevent hot corrosion or at least detect it at an early stage has resulted in several accidents, leading to loss of life and/or destruction of infrastructures. A number of countermeasures are currently in use or under investigation to combat hot corrosion, such as using inhibitors, controlling the process parameters, designing a suitable industrial alloy, and depositing protective coatings. However, the protection system to be selected for a particular application must be practical, reliable, and economically viable. Due to the continuously rising cost of the materials as well as increased material requirements, the coating techniques have been given much more importance in recent times. Coatings can add value to products up to 10 times the cost of the coating. Among the different coating techniques, thermal spraying has grown into a well-accepted industrial technology for applying overlay coatings onto the surfaces of engineering components to allow them to function under extreme conditions of wear, erosion-corrosion, high-temperature oxidation, and hot corrosion. In this study, the hot corrosion performances of Ni-20Cr and Cr₃C₂-NiCr coatings developed by High Velocity Oxy-Fuel (HVOF) process have been studied. The coatings were developed on a Fe-based superalloy, and experiments were performed in an actual industrial environment of a coal-fired boiler. The cyclic study was carried out around the platen superheater zone where the temperature was around 1000°C. The study was conducted for 10 cycles, and one cycle was consisting of 100 hours of heating followed by 1 hour of cooling at ambient temperature. Both the coatings deposited on Fe-based superalloy imparted better hot corrosion resistance than the uncoated one. The Ni-20Cr coated superalloy performed better than the Cr₃C₂-NiCr coated in the actual working conditions of the coal fired boiler. It is found that the formation of chromium oxide at the boundaries of Ni-rich splats of the coating blocks the inward permeation of oxygen and other corrosive species to the substrate.

Keywords: hot corrosion, coating, HVOF, oxidation

Procedia PDF Downloads 63
877 Investigating the Status of Black Women in Brazil: Beyond Housekeepers and Samba Dancers

Authors: Sandra Maria Cerqueira Da Silva

Abstract:

The construction of the material world involves a series of social power relations. These relations, in a way, can dictate, shape, judge and drive the profiles of so-called ‘ideal’ individuals. Gender relations, as power relations, are defined based on hierarchies, obediences and inequalities, and male domination seems, with few exceptions, to be rooted in every society around the world. The profile of the Brazilian woman, beyond patriarchal and market determinations, is strongly subjected to media products. Women are, numerically, the majority in Brazilian society. The social indicators point to slight advances in terms of years of study and professional qualification, as well as access to the job market; yet, differences in opportunity and conditions — often explained though the ‘unquestionable’ cultural rancidness argument — still hinder women’s ability to reach and keep job positions. These unequalities are also visible in everyday interactions and in gender relations, and they become greater once race is added to the analysis. For a black woman, her racial origins may play a part in determining the construction of her gender roles. In these terms, there is need to investigate the racial character of the sexual differences within a larger social proccess of naturalization and justification of cultural hierarchies. Thus, the goal of this study is to identify and discuss the media-built image of black women in Brazil. Furthermore, it is necessary to seek views different than those of the ruling classes. The study uses a qualitative approach based on the feminist standpoint, which intends to hold women’s experiences as central. The body of the research — images taken from the Internet — was treated through critical content analysis. The results show that in Brazil the profile of black women, beyond the machist and sexist generalizations, objectifies them or sees them as servants, always at the disposal of non-blacks. It is necessary to overcome the history of this nation, always considering the contribution of these women to the growth and development of places and societies. This can be done through the acknowledgement and highlighting of the few black women who were able to overcome the many barriers in their path and reach leadership position in the country. There are still many important challenges in the way of finding affirmative policies and reaching a more equal society in terms of gender and race; a serious and firm political commitment seems sine qua non.

Keywords: black woman, feminist standpoint, markings, objectification

Procedia PDF Downloads 251
876 Enhancing Thai In-Service Science Teachers' Technological Pedagogical Content Knowledge Integrating Local Context and Sufficiency Economy into Science Teaching

Authors: Siriwan Chatmaneerungcharoen

Abstract:

An emerging body of ‘21st century skills’-such as adaptability, complex communication skills, technology skills and the ability to solve non-routine problems--are valuable across a wide range of jobs in the national economy. Within the Thai context, a focus on the Philosophy of Sufficiency Economy is integrated into Science Education. Thai science education has advocated infusing 21st century skills and Philosophy of Sufficiency Economy into the school curriculum and several educational levels have launched such efforts. Therefore, developing science teachers to have proper knowledge is the most important factor to success of the goals. The purposes of this study were to develop 40 Cooperative Science teachers’ Technological Pedagogical Content Knowledge (TPACK) and to develop Professional Development Model integrated with Co-teaching Model and Coaching System (Co-TPACK). TPACK is essential to career development for teachers. Forty volunteer In-service teachers who were science cooperative teachers participated in this study for 2 years. Data sources throughout the research project consisted of teacher refection, classroom observations, Semi-structure interviews, Situation interview, questionnaires and document analysis. Interpretivist framework was used to analyze the data. Findings indicate that at the beginning, the teachers understood only the meaning of Philosophy of Sufficiency Economy but they did not know how to integrate the Philosophy of Sufficiency Economy into their science classrooms. Mostly, they preferred to use lecture based teaching and experimental teaching styles. While the Co- TPACK was progressing, the teachers had blended their teaching styles and learning evaluation methods. Co-TPACK consists of 3 cycles (Student Teachers’ Preparation Cycle, Cooperative Science Teachers Cycle, Collaboration cycle (Co-teaching, Co-planning, and Co-Evaluating and Coaching System)).The Co-TPACK enhances the 40 cooperative science teachers, student teachers and university supervisor to exchange their knowledge and experience on teaching science. There are many channels that they used for communication including online. They have used more Phuket context-integrated lessons, technology-integrated teaching and Learning that can explicit Philosophy of Sufficiency Economy. Their sustained development is shown in their lesson plans and teaching practices.

Keywords: technological pedagogical content knowledge, philosophy of sufficiency economy, professional development, coaching system

Procedia PDF Downloads 436
875 The Extension of the Kano Model by the Concept of Over-Service

Authors: Lou-Hon Sun, Yu-Ming Chiu, Chen-Wei Tao, Chia-Yun Tsai

Abstract:

It is common practice for many companies to ask employees to provide heart-touching service for customers and to emphasize the attitude of 'customer first'. However, services may not necessarily gain praise, and may actually be considered excessive, if customers do not appreciate such behaviors. In reality, many restaurant businesses try to provide as much service as possible without taking into account whether over-provision may lead to negative customer reception. A survey of 894 people in Britain revealed that 49 percent of respondents consider over-attentive waiters the most annoying aspect of dining out. It can be seen that merely aiming to exceed customers’ expectations without actually addressing their needs, only further distances and dissociates the standard of services from the goals of customer satisfaction itself. Over-service is defined, as 'service provided that exceeds customer expectations, or simply that customers deemed redundant, resulting in negative perception'. It was found that customers’ reactions and complaints concerning over-service are not as intense as those against service failures caused by the inability to meet expectations; consequently, it is more difficult for managers to become aware of the existence of over-service. Thus the ability to manage over-service behaviors is a significant topic for consideration. The Kano model classifies customer preferences into five categories: attractive quality attribute, one-dimensional quality attribute, must-be quality attribute, indifferent quality attribute and reverse quality attributes. The model is still very popular for researchers to explore the quality aspects and customer satisfaction. Nevertheless, several studies indicated that Kano’s model could not fully capture the nature of service quality. The concept of over-service can be used to restructure the model and provide a better understanding of the service quality construct. In this research, the structure of Kano's two-dimensional questionnaire will be used to classify the factors into different dimensions. The same questions will be used in the second questionnaire for identifying the over-service experienced of the respondents. The finding of these two questionnaires will be used to analyze the relevance between service quality classification and over-service behaviors. The subjects of this research are customers of fine dining chain restaurants. Three hundred questionnaires will be issued based on the stratified random sampling method. Items for measurement will be derived from DINESERV scale. The tangible dimension of the questionnaire will be eliminated due to this research is focused on the employee behaviors. Quality attributes of the Kano model are often regarded as an instrument for improving customer satisfaction. The concept of over-service can be used to restructure the model and provide a better understanding of service quality construct. The extension of the Kano model will not only develop a better understanding of customer needs and expectations but also enhance the management of service quality.

Keywords: consumer satisfaction, DINESERV, kano model, over-service

Procedia PDF Downloads 139
874 Screening for Non-hallucinogenic Neuroplastogens as Drug Candidates for the Treatment of Anxiety, Depression, and Posttraumatic Stress Disorder

Authors: Jillian M. Hagel, Joseph E. Tucker, Peter J. Facchini

Abstract:

With the aim of establishing a holistic approach for the treatment of central nervous system (CNS) disorders, we are pursuing a drug development program rapidly progressing through discovery and characterization phases. The drug candidates identified in this program are referred to as neuroplastogens owing to their ability to mediate neuroplasticity, which can be beneficial to patients suffering from anxiety, depression, or posttraumatic stress disorder. These and other related neuropsychiatric conditions are associated with the onset of neuronal atrophy, which is defined as a reduction in the number and/or productivity of neurons. The stimulation of neuroplasticity results in an increase in the connectivity between neurons and promotes the restoration of healthy brain function. We have synthesized a substantial catalogue of proprietary indolethylamine derivatives based on the general structures of serotonin (5-hydroxytryptamine) and psychedelic molecules such as N,N-dimethyltryptamine (DMT) and psilocin (4-hydroxy-DMT) that function as neuroplastogens. A primary objective in our screening protocol is the identification of derivatives associated with a significant reduction in hallucination, which will allow administration of the drug at a dose that induces neuroplasticity and triggers other efficacious outcomes in the treatment of targeted CNS disorders but which does not cause a psychedelic response in the patient. Both neuroplasticity and hallucination are associated with engagement of the 5HT2A receptor, requiring drug candidates differentially coupled to these two outcomes at a molecular level. We use novel and proprietary artificial intelligence algorithms to predict the mode of binding to the 5HT2A receptor, which has been shown to correlate with the hallucinogenic response. Hallucination is tested using the mouse head-twitch response model, whereas mouse marble-burying and sucrose preference assays are used to evaluate anxiolytic and anti-depressive potential. Neuroplasticity is assays using dendritic outgrowth assays and cell-based ELISA analysis. Pharmacokinetics and additional receptor-binding analyses also contribute the selection of lead candidates. A summary of the program is presented.

Keywords: neuroplastogen, non-hallucinogenic, drug development, anxiety, depression, PTSD, indolethylamine derivatives, psychedelic-inspired, 5-HT2A receptor, computational chemistry, head-twitch response behavioural model, neurite outgrowth assay

Procedia PDF Downloads 105
873 Micro Plasma an Emerging Technology to Eradicate Pesticides from Food Surface

Authors: Muhammad Saiful Islam Khan, Yun Ji Kim

Abstract:

Organophosphorus pesticides (OPPs) have been widely used to replace more persistent organochlorine pesticides because OPPs are more soluble in water and decompose rapidly in aquatic systems. Extensive uses of OPPs in modern agriculture are the major cause of the contamination of surface water. Regardless of the advantages gained by the application of pesticides in modern agriculture, they are a threat to the public health environment. With the aim of reducing possible health threats, several physical and chemical treatment processes have been studied to eliminate biological and chemical poisons from food stuff. In the present study, a micro-plasma device was used to reduce pesticides from the surface of food stuff. Pesticide free food items chosen in this study were perilla leaf, tomato, broccoli and blueberry. To evaluate the removal efficiency of pesticides, different washing methods were followed such as soaking with water, washing with bubbling water, washing with plasma-treated water and washing with chlorine water. 2 mL of 2000 ppm pesticide samples, namely, diazinone and chlorpyrifos were individuality inoculated on food surface and was air dried for 2 hours before treated with plasma. Plasma treated water was used in two different manners one is plasma treated water with bubbling the other one is aerosolized plasma treated water. The removal efficiency of pesticides from food surface was studied using HPLC. Washing with plasma treated water, aerosolized plasma treated water and chlorine water shows minimum 72% to maximum 87 % reduction for 4 min treatment irrespective to the types of food items and the types of pesticides sample, in case of soaking and bubbling the reduction is 8% to 48%. Washing with plasma treated water, aerosolized plasma treated water and chlorine water shows somewhat similar reduction ability which is significantly higher comparing to the soaking and bubbling washing system. The temperature effect of the washing systems was also evaluated; three different temperatures were set for the experiment, such as 22°C, 10°C and 4°C. Decreasing temperature from 22°C to 10°C shows a higher reduction in the case of washing with plasma and aerosolized plasma treated water, whereas an opposite trend was observed for the washing with chlorine water. Further temperature reduction from 10°C to 4°C does not show any significant reduction of pesticides, except for the washing with chlorine water. Chlorine water treatment shows lesser pesticide reduction with the decrease in temperature. The color changes of the treated sample were measured immediately and after one week to evaluate if there is any effect of washing with plasma treated water and with chlorine water. No significant color changes were observed for either of the washing systems, except for broccoli washing with chlorine water.

Keywords: chlorpyrifos, diazinone, pesticides, micro plasma

Procedia PDF Downloads 165
872 Application of Zeolite Nanoparticles in Biomedical Optics

Authors: Vladimir Hovhannisyan, Chen Yuan Dong

Abstract:

Recently nanoparticles (NPs) have been introduced in biomedicine as effective agents for cancer-targeted drug delivery and noninvasive tissue imaging. The most important requirements to these agents are their non-toxicity, biocompatibility and stability. In view of these criteria, the zeolite (ZL) nanoparticles (NPs) may be considered as perfect candidates for biomedical applications. ZLs are crystalline aluminosilicates consisting of oxygen-sharing SiO4 and AlO4 tetrahedral groups united by common vertices in three-dimensional framework and containing pores with diameters from 0.3 to 1.2 nm. Generally, the behavior and physical properties of ZLs are studied by SEM, X-ray spectroscopy, and AFM, whereas optical spectroscopic and microscopic approaches are not effective enough, because of strong scattering in common ZL bulk materials and powders. The light scattering can be reduced by using of ZL NPs. ZL NPs have large external surface area, high dispersibility in both aqueous and organic solutions, high photo- and thermal stability, and exceptional ability to adsorb various molecules and atoms in their nanopores. In this report, using multiphoton microscopy and nonlinear spectroscopy, we investigate nonlinear optical properties of clinoptilolite type of ZL micro- and nanoparticles with average diameters of 2200 nm and 240 nm, correspondingly. Multiphoton imaging is achieved using a laser scanning microscope system (LSM 510 META, Zeiss, Germany) coupled to a femtosecond titanium:sapphire laser (repetition rate- 80 MHz, pulse duration-120 fs, radiation wavelength- 720-820 nm) (Tsunami, Spectra-Physics, CA). Two Zeiss, Plan-Neofluar objectives (air immersion 20×∕NA 0.5 and water immersion 40×∕NA 1.2) are used for imaging. For the detection of the nonlinear response, we use two detection channels with 380-400 nm and 435-700 nm spectral bandwidths. We demonstrate that ZL micro- and nanoparticles can produce nonlinear optical response under the near-infrared femtosecond laser excitation. The interaction of hypericine, chlorin e6 and other dyes with ZL NPs and their photodynamic activity is investigated. Particularly, multiphoton imaging shows that individual ZL NPs particles adsorb Zn-tetraporphyrin molecules, but do not adsorb fluorescein molecules. In addition, nonlinear spectral properties of ZL NPs in native biotissues are studied. Nonlinear microscopy and spectroscopy may open new perspectives in the research and application of ZL NP in biomedicine, and the results may help to introduce novel approaches into the clinical environment.

Keywords: multiphoton microscopy, nanoparticles, nonlinear optics, zeolite

Procedia PDF Downloads 395
871 Prediction of Fluid Induced Deformation using Cavity Expansion Theory

Authors: Jithin S. Kumar, Ramesh Kannan Kandasami

Abstract:

Geomaterials are generally porous in nature due to the presence of discrete particles and interconnected voids. The porosity present in these geomaterials play a critical role in many engineering applications such as CO2 sequestration, well bore strengthening, enhanced oil and hydrocarbon recovery, hydraulic fracturing, and subsurface waste storage. These applications involves solid-fluid interactions, which govern the changes in the porosity which in turn affect the permeability and stiffness of the medium. Injecting fluid into the geomaterials results in permeation which exhibits small or negligible deformation of the soil skeleton followed by cavity expansion/ fingering/ fracturing (different forms of instabilities) due to the large deformation especially when the flow rate is greater than the ability of the medium to permeate the fluid. The complexity of this problem increases as the geomaterial behaves like a solid and fluid under certain conditions. Thus it is important to understand this multiphysics problem where in addition to the permeation, the elastic-plastic deformation of the soil skeleton plays a vital role during fluid injection. The phenomenon of permeation and cavity expansion in porous medium has been studied independently through extensive experimental and analytical/ numerical models. The analytical models generally use Darcy's/ diffusion equations to capture the fluid flow during permeation while elastic-plastic (Mohr-Coulomb and Modified Cam-Clay) models were used to predict the solid deformations. Hitherto, the research generally focused on modelling cavity expansion without considering the effect of injected fluid coming into the medium. Very few studies have considered the effect of injected fluid on the deformation of soil skeleton. However, the porosity changes during the fluid injection and coupled elastic-plastic deformation are not clearly understood. In this study, the phenomenon of permeation and instabilities such as cavity and finger/ fracture formation will be quantified extensively by performing experiments using a novel experimental setup in addition to utilizing image processing techniques. This experimental study will describe the fluid flow and soil deformation characteristics under different boundary conditions. Further, a well refined coupled semi-analytical model will be developed to capture the physics involved in quantifying the deformation behaviour of geomaterial during fluid injection.

Keywords: solid-fluid interaction, permeation, poroelasticity, plasticity, continuum model

Procedia PDF Downloads 54
870 Predicting Aggregation Propensity from Low-Temperature Conformational Fluctuations

Authors: Hamza Javar Magnier, Robin Curtis

Abstract:

There have been rapid advances in the upstream processing of protein therapeutics, which has shifted the bottleneck to downstream purification and formulation. Finding liquid formulations with shelf lives of up to two years is increasingly difficult for some of the newer therapeutics, which have been engineered for activity, but their formulations are often viscous, can phase separate, and have a high propensity for irreversible aggregation1. We explore means to develop improved predictive ability from a better understanding of how protein-protein interactions on formulation conditions (pH, ionic strength, buffer type, presence of excipients) and how these impact upon the initial steps in protein self-association and aggregation. In this work, we study the initial steps in the aggregation pathways using a minimal protein model based on square-well potentials and discontinuous molecular dynamics. The effect of model parameters, including range of interaction, stiffness, chain length, and chain sequence, implies that protein models fold according to various pathways. By reducing the range of interactions, the folding- and collapse- transition come together, and follow a single-step folding pathway from the denatured to the native state2. After parameterizing the model interaction-parameters, we developed an understanding of low-temperature conformational properties and fluctuations, and the correlation to the folding transition of proteins in isolation. The model fluctuations increase with temperature. We observe a low-temperature point, below which large fluctuations are frozen out. This implies that fluctuations at low-temperature can be correlated to the folding transition at the melting temperature. Because proteins “breath” at low temperatures, defining a native-state as a single structure with conserved contacts and a fixed three-dimensional structure is misleading. Rather, we introduce a new definition of a native-state ensemble based on our understanding of the core conservation, which takes into account the native fluctuations at low temperatures. This approach permits the study of a large range of length and time scales needed to link the molecular interactions to the macroscopically observed behaviour. In addition, these models studied are parameterized by fitting to experimentally observed protein-protein interactions characterized in terms of osmotic second virial coefficients.

Keywords: protein folding, native-ensemble, conformational fluctuation, aggregation

Procedia PDF Downloads 340
869 Multiperson Drone Control with Seamless Pilot Switching Using Onboard Camera and Openpose Real-Time Keypoint Detection

Authors: Evan Lowhorn, Rocio Alba-Flores

Abstract:

Traditional classification Convolutional Neural Networks (CNN) attempt to classify an image in its entirety. This becomes problematic when trying to perform classification with a drone’s camera in real-time due to unpredictable backgrounds. Object detectors with bounding boxes can be used to isolate individuals and other items, but the original backgrounds remain within these boxes. These basic detectors have been regularly used to determine what type of object an item is, such as “person” or “dog.” Recent advancement in computer vision, particularly with human imaging, is keypoint detection. Human keypoint detection goes beyond bounding boxes to fully isolate humans and plot points, or Regions of Interest (ROI), on their bodies within an image. ROIs can include shoulders, elbows, knees, heads, etc. These points can then be related to each other and used in deep learning methods such as pose estimation. For drone control based on human motions, poses, or signals using the onboard camera, it is important to have a simple method for pilot identification among multiple individuals while also giving the pilot fine control options for the drone. To achieve this, the OpenPose keypoint detection network was used with body and hand keypoint detection enabled. OpenPose supports the ability to combine multiple keypoint detection methods in real-time with a single network. Body keypoint detection allows simple poses to act as the pilot identifier. The hand keypoint detection with ROIs for each finger can then offer a greater variety of signal options for the pilot once identified. For this work, the individual must raise their non-control arm to be identified as the operator and send commands with the hand on their other arm. The drone ignores all other individuals in the onboard camera feed until the current operator lowers their non-control arm. When another individual wish to operate the drone, they simply raise their arm once the current operator relinquishes control, and then they can begin controlling the drone with their other hand. This is all performed mid-flight with no landing or script editing required. When using a desktop with a discrete NVIDIA GPU, the drone’s 2.4 GHz Wi-Fi connection combined with OpenPose restrictions to only body and hand allows this control method to perform as intended while maintaining the responsiveness required for practical use.

Keywords: computer vision, drone control, keypoint detection, openpose

Procedia PDF Downloads 163