Search results for: regional innovation systems
207 Single Stage Holistic Interventions: The Impact on Well-Being
Authors: L. Matthewman, J. Nowlan
Abstract:
Background: Holistic or Integrative Psychology emphasizes the interdependence of physiological, spiritual and psychological dynamics. Studying “wholeness and well-being” from a systems perspective combines innovative psychological science interventions with Eastern orientated healing wisdoms and therapies. The literature surrounding holistic/integrative psychology focuses on multi-stage interventions in attempts to enhance the mind-body experiences of well-being for participants. This study proposes a new single stage model as an intervention for UG/PG students, time-constrained workplace employees and managers/leaders for improved well-being and life enhancement. The main research objective was to investigate participants’ experiences of holistic and mindfulness interventions for impact on emotional well-being. The main research question asked was if single stage holistic interventions could impact on psychological well-being. This is of consequence because many people report that a reason for not taking part in mind-body or wellness programmes is that they believe that they do not have sufficient time to engage in such pursuits. Experimental Approach: The study employed a mixed methods pre-test/post-test research design. Data was analyzed using descriptive statistics and interpretative phenomenological analysis. Purposive sampling methods were employed. An adapted mindfulness measurement questionnaire (MAAS) was administered to 20 volunteer final year UG student participants prior to the single stage intervention and following the intervention. A further post-test longitudinal follow-up took place one week later. Intervention: The single stage model intervention consisted of a half hour session of mindfulness, yoga stretches and head and neck massage in the following sequence: Mindful awareness of the breath, yoga stretches 1, mindfulness of the body, head and neck massage, mindfulness of sounds, yoga stretches 2 and finished with pure awareness mindfulness. Results: The findings on the pre-test indicated key themes concerning: “being largely unaware of feelings”, “overwhelmed with final year exams”, “juggling other priorities” , “not feeling in control”, “stress” and “negative emotional display episodes”. Themes indicated on the post-test included: ‘more aware of self’, ‘in more control’, ‘immediately more alive’ and ‘just happier’ compared to the pre-test. Themes from post-test 2 indicated similar findings to post-test 1 in terms of themes. but on a lesser scale when scored for intensity. Interestingly, the majority of participants reported that they would now seek other similar interventions in the future and would be likely to engage with a multi-stage intervention type on a longer-term basis. Overall, participants reported increased psychological well-being after the single stage intervention. Conclusion: A single stage one-off intervention model can be effective to help towards the wellbeing of final year UG students. There is little indication to suggest that this would not be generalizable to others in different areas of life and business. However this study must be taken with caution due to low participant numbers. Implications: Single stage one-off interventions can be used to enhance peoples’ lives who might not otherwise sign up for a longer multi-stage intervention. In addition, single stage interventions can be utilized to help participants progress onto longer multiple stage interventions. Finally, further research into one stage well-being interventions is encouraged.Keywords: holistic/integrative psychology, mindfulness, well-being, yoga
Procedia PDF Downloads 353206 Asparagus racemosus Willd for Enhanced Medicinal Properties
Authors: Ashok Kumar, Parveen Parveen
Abstract:
India is bestowed with an extremely high population of plant species with medicinal value and even has two biodiversity hotspots. Indian systems of medicine including Ayurveda, Siddha and Unani have historically been serving humankind across the world since time immemorial. About 1500 plant species have well been documented in Ayurvedic Nighantus as official medicinal plants. Additionally, several hundred species of plants are being routinely used as medicines by local people especially tribes living in and around forests. The natural resources for medicinal plants have unscientifically been over-exploited forcing rapid depletion in their genetic diversity. Moreover, renewed global interest in herbal medicines may even lead to additional depletion of medicinal plant wealth of the country, as about 95% collection of medicinal plants for pharmaceutical preparation is being carried out from natural forests. On the other hand, huge export market of medicinal and aromatic plants needs to be seriously tapped for enhancing inflow of foreign currency. Asparagus racemosus Willd., a member of family Liliaceae, is one of thirty-two plant species that have been identified as priority species for cultivation and conservation by the National Medicinal Plant Board (NMPB), Government of India. Though attention is being focused on standardization of agro-techniques and extraction methods, little has been designed on genetic improvement and selection of desired types with higher root production and saponin content, a basic ingredient of medicinal value. The saponin not only improves defense mechanisms and controls diabetes but the roots of this species promote secretion of breast milk, improved lost body weight and considered as an aphrodisiac. There is ample scope for genetic improvement of this species for enhancing productivity substantially, qualitatively and quantitatively. It is emphasized to select desired genotypes with sufficient genetic diversity for important economic traits. Hybridization between two genetically divergent genotypes could result in the synthesis of new F1 hybrids consisting of useful traits of both the parents. The evaluation of twenty seed sources of Asparagus racemosus assembled different geographical locations of India revelled high degree of variability for traits of economic importance. The maximum genotypic and phenotypic variance was observed for shoot height among shoot related traits and for root length among root related traits. The shoot height, genotypic variance, phenotypic variance, genotypic coefficient of variance, the phenotypic coefficient of variance was recorded to be 231.80, 3924.80, 61.26 and 1037.32, respectively, where those of the root length were 9.55, 16.80, 23.46 and 41.27, respectively. The maximum genetic advance and genetic gain were obtained for shoot height among shoot-related traits and root length among root-related traits. Index values were developed for all seed sources based on the four most important traits, and Panthnagar (Uttrakhand), Jodhpur (Rajasthan), Dehradun (Uttarakhand), Chandigarh (Punjab), Jammu (Jammu & Kashmir) and Solan (Himachal Pradesh) were found to be promising seed sources.Keywords: asparagus, genetic, genotypes, variance
Procedia PDF Downloads 134205 A Review of Data Visualization Best Practices: Lessons for Open Government Data Portals
Authors: Bahareh Ansari
Abstract:
Background: The Open Government Data (OGD) movement in the last decade has encouraged many government organizations around the world to make their data publicly available to advance democratic processes. But current open data platforms have not yet reached to their full potential in supporting all interested parties. To make the data useful and understandable for everyone, scholars suggested that opening the data should be supplemented by visualization. However, different visualizations of the same information can dramatically change an individual’s cognitive and emotional experience in working with the data. This study reviews the data visualization literature to create a list of the methods empirically tested to enhance users’ performance and experience in working with a visualization tool. This list can be used in evaluating the OGD visualization practices and informing the future open data initiatives. Methods: Previous reviews of visualization literature categorized the visualization outcomes into four categories including recall/memorability, insight/comprehension, engagement, and enjoyment. To identify the papers, a search for these outcomes was conducted in the abstract of the publications of top-tier visualization venues including IEEE Transactions for Visualization and Computer Graphics, Computer Graphics, and proceedings of the CHI Conference on Human Factors in Computing Systems. The search results are complemented with a search in the references of the identified articles, and a search for 'open data visualization,' and 'visualization evaluation' keywords in the IEEE explore and ACM digital libraries. Articles are included if they provide empirical evidence through conducting controlled user experiments, or provide a review of these empirical studies. The qualitative synthesis of the studies focuses on identification and classifying the methods, and the conditions under which they are examined to positively affect the visualization outcomes. Findings: The keyword search yields 760 studies, of which 30 are included after the title/abstract review. The classification of the included articles shows five distinct methods: interactive design, aesthetic (artistic) style, storytelling, decorative elements that do not provide extra information including text, image, and embellishment on the graphs), and animation. Studies on decorative elements show consistency on the positive effects of these elements on user engagement and recall but are less consistent in their examination of the user performance. This inconsistency could be attributable to the particular data type or specific design method used in each study. The interactive design studies are consistent in their findings of the positive effect on the outcomes. Storytelling studies show some inconsistencies regarding the design effect on user engagement, enjoyment, recall, and performance, which could be indicative of the specific conditions required for the use of this method. Last two methods, aesthetics and animation, have been less frequent in the included articles, and provide consistent positive results on some of the outcomes. Implications for e-government: Review of the visualization best-practice methods show that each of these methods is beneficial under specific conditions. By using these methods in a potentially beneficial condition, OGD practices can promote a wide range of individuals to involve and work with the government data and ultimately engage in government policy-making procedures.Keywords: best practices, data visualization, literature review, open government data
Procedia PDF Downloads 105204 Enhancing Seismic Resilience in Colombia's Informal Housing: A Low-cost Retrofit Strategy with Buckling-restrained Braces to Protect Vulnerable Communities in Earthquake-prone Regions
Authors: Luis F. Caballero-castro, Dirsa Feliciano, Daniela Novoa, Orlando Arroyo, Jesús D. Villalba-morales
Abstract:
Colombia faces a critical challenge in seismic resilience due to the prevalence of informal housing, which constitutes approximately 70% of residential structures. More than 10 million Colombians (20% of the population), live in homes susceptible to collapse in the event of an earthquake. This, combined with the fact that 83% of the population is in intermediate and high seismic hazard areas, has brought serious consequences to the country. These consequences became evident during the 1999 Armenia earthquake, which affected nearly 100,000 properties and represented economic losses equivalent to 1.88% of that year's Gross Domestic Product (GDP). Despite previous efforts to reinforce informal housing through methods like externally reinforced masonry walls, alternatives related to seismic protection systems (SPDs), such as Buckling-Restrained Braces (BRB), have not yet been explored in the country. BRBs are reinforcement elements capable of withstanding both compression and tension, making them effective in enhancing the lateral stiffness of structures. In this study, the use of low-cost and easily installable BRBs for the retrofit of informal housing in Colombia was evaluated, considering the economic limitations of the communities. For this purpose, a case study was selected involving an informally constructed dwelling in the country, from which field information on its structural characteristics and construction materials was collected. Based on the gathered information, nonlinear models with and without BRBs were created, and their seismic performance was analyzed and compared through incremental static (pushover) and nonlinear dynamic analyses. In the first analysis, the capacity curve was identified, showcasing the sequence of failure events occurring from initial yielding to structural collapse. In the second case, the model underwent nonlinear dynamic analyses using a set of seismic records consistent with the country's seismic hazard. Based on the results, fragility curves were calculated to evaluate the probability of failure of the informal housings before and after the intervention with BRBs, providing essential information about their effectiveness in reducing seismic vulnerability. The results indicate that low-cost BRBs can significantly increase the capacity of informal housing to withstand earthquakes. The dynamic analysis revealed that retrofit structures experienced lower displacements and deformations, enhancing the safety of residents and the seismic performance of informally constructed houses. In other words, the use of low-cost BRBs in the retrofit of informal housing in Colombia is a promising strategy for improving structural safety in seismic-prone areas. This study emphasizes the importance of seeking affordable and practical solutions to address seismic risk in vulnerable communities in earthquake-prone regions in Colombia and serves as a model for addressing similar challenges of informal housing worldwide.Keywords: buckling-restrained braces, fragility curves, informal housing, incremental dynamic analysis, seismic retrofit
Procedia PDF Downloads 96203 Subcontractor Development Practices and Processes: A Conceptual Model for LEED Projects
Authors: Andrea N. Ofori-Boadu
Abstract:
The purpose is to develop a conceptual model of subcontractor development practices and processes that strengthen the integration of subcontractors into construction supply chain systems for improved subcontractor performance on Leadership in Energy and Environmental Design (LEED) certified building projects. The construction management of a LEED project has an important objective of meeting sustainability certification requirements. This is in addition to the typical project management objectives of cost, time, quality, and safety for traditional projects; and, therefore increases the complexity of LEED projects. Considering that construction management organizations rely heavily on subcontractors, poor performance on complex projects such as LEED projects has been largely attributed to the unsatisfactory preparation of subcontractors. Furthermore, the extensive use of unique and non-repetitive short term contracts limits the full integration of subcontractors into construction supply chains and hinders long-term cooperation and benefits that could enhance performance on construction projects. Improved subcontractor development practices are needed to better prepare and manage subcontractors, so that complex objectives can be met or exceeded. While supplier development and supply chain theories and practices for the manufacturing sector have been extensively investigated to address similar challenges, investigations in the construction sector are not that obvious. Consequently, the objective of this research is to investigate effective subcontractor development practices and processes to guide construction management organizations in their development of a strong network of high performing subcontractors. Drawing from foundational supply chain and supplier development theories in the manufacturing sector, a mixed interpretivist and empirical methodology is utilized to assess the body of knowledge within literature for conceptual model development. A self-reporting survey with five-point Likert scale items and open-ended questions is administered to 30 construction professionals to estimate their perceptions of the effectiveness of 37 practices, classified into five subcontractor development categories. Data analysis includes descriptive statistics, weighted means, and t-tests that guide the effectiveness ranking of practices and categories. The results inform the proposed three-phased LEED subcontractor development program model which focuses on preparation, development and implementation, and monitoring. Highly ranked LEED subcontractor pre-qualification, commitment, incentives, evaluation, and feedback practices are perceived as more effective, when compared to practices requiring more direct involvement and linkages between subcontractors and construction management organizations. This is attributed to unfamiliarity, conflicting interests, lack of trust, and resource sharing challenges. With strategic modifications, the recommended practices can be extended to other non-LEED complex projects. Additional research is needed to guide the development of subcontractor development programs that strengthen direct involvement between construction management organizations and their network of high performing subcontractors. Insights from this present research strengthen theoretical foundations to support future research towards more integrated construction supply chains. In the long-term, this would lead to increased performance, profits and client satisfaction.Keywords: construction management, general contractor, supply chain, sustainable construction
Procedia PDF Downloads 110202 Accessing Motional Quotient for All Round Development
Authors: Zongping Wang, Chengjun Cui, Jiacun Wang
Abstract:
The concept of intelligence has been widely used to access an individual's cognitive abilities to learn, form concepts, understand, apply logic, and reason. According to the multiple intelligence theory, there are eight distinguished types of intelligence. One of them is the bodily-kinaesthetic intelligence that links to the capacity of an individual controlling his body and working with objects. Motor intelligence, on the other hand, reflects the capacity to understand, perceive and solve functional problems by motor behavior. Both bodily-kinaesthetic intelligence and motor intelligence refer directly or indirectly to bodily capacity. Inspired by these two intelligence concepts, this paper introduces motional intelligence (MI). MI is two-fold. (1) Body strength, which is the capacity of various organ functions manifested by muscle activity under the control of the central nervous system during physical exercises. It can be measured by the magnitude of muscle contraction force, the frequency of repeating a movement, the time to finish a movement of body position, the duration to maintain muscles in a working status, etc. Body strength reflects the objective of MI. (2) Level of psychiatric willingness to physical events. It is a subjective thing and determined by an individual’s self-consciousness to physical events and resistance to fatigue. As such, we call it subjective MI. Subjective MI can be improved through education and proper social events. The improvement of subjective MI can lead to that of objective MI. A quantitative score of an individual’s MI is motional quotient (MQ). MQ is affected by several factors, including genetics, physical training, diet and lifestyle, family and social environment, and personal awareness of the importance of physical exercise. Genes determine one’s body strength potential. Physical training, in general, makes people stronger, faster and swifter. Diet and lifestyle have a direct impact on health. Family and social environment largely affect one’s passion for physical activities, so does personal awareness of the importance of physical exercise. The key to the success of the MQ study is developing an acceptable and efficient system that can be used to assess MQ objectively and quantitatively. We should apply different accessing systems to different groups of people according to their ages and genders. Field test, laboratory test and questionnaire are among essential components of MQ assessment. A scientific interpretation of MQ score is part of an MQ assessment system as it will help an individual to improve his MQ. IQ (intelligence quotient) and EQ (emotional quotient) and their test have been studied intensively. We argue that IQ and EQ study alone is not sufficient for an individual’s all round development. The significance of MQ study is that it offsets IQ and EQ study. MQ reflects an individual’s mental level as well as bodily level of intelligence in physical activities. It is well-known that the American Springfield College seal includes the Luther Gulick triangle with the words “spirit,” “mind,” and “body” written within it. MQ, together with IQ and EQ, echoes this education philosophy. Since its inception in 2012, the MQ research has spread rapidly in China. By now, six prestigious universities in China have established research centers on MQ and its assessment.Keywords: motional Intelligence, motional quotient, multiple intelligence, motor intelligence, all round development
Procedia PDF Downloads 162201 Post-Exercise Recovery Tracking Based on Electrocardiography-Derived Features
Authors: Pavel Bulai, Taras Pitlik, Tatsiana Kulahava, Timofei Lipski
Abstract:
The method of Electrocardiography (ECG) interpretation for post-exercise recovery tracking was developed. Metabolic indices (aerobic and anaerobic) were designed using ECG-derived features. This study reports the associations between aerobic and anaerobic indices and classical parameters of the person’s physiological state, including blood biochemistry, glycogen concentration and VO2max changes. During the study 9 participants, healthy, physically active medium trained men and women, which trained 2-4 times per week for at least 9 weeks, fulfilled (i) ECG monitoring using Apple Watch Series 4 (AWS4); (ii) blood biochemical analysis; (iii) maximal oxygen consumption (VO2max) test, (iv) bioimpedance analysis (BIA). ECG signals from a single-lead wrist-wearable device were processed with detection of QRS-complex. Aerobic index (AI) was derived as the normalized slope of QR segment. Anaerobic index (ANI) was derived as the normalized slope of SJ segment. Biochemical parameters, glycogen content and VO2max were evaluated eight times within 3-60 hours after training. ECGs were recorded 5 times per day, plus before and after training, cycloergometry and BIA. The negative correlation between AI and blood markers of the muscles functional status including creatine phosphokinase (r=-0.238, p < 0.008), aspartate aminotransferase (r=-0.249, p < 0.004) and uric acid (r = -0.293, p<0.004) were observed. ANI was also correlated with creatine phosphokinase (r= -0.265, p < 0.003), aspartate aminotransferase (r = -0.292, p < 0.001), lactate dehydrogenase (LDH) (r = -0.190, p < 0.050). So, when the level of muscular enzymes increases during post-exercise fatigue, AI and ANI decrease. During recovery, the level of metabolites is restored, and metabolic indices rising is registered. It can be concluded that AI and ANI adequately reflect the physiology of the muscles during recovery. One of the markers of an athlete’s physiological state is the ratio between testosterone and cortisol (TCR). TCR provides a relative indication of anabolic-catabolic balance and is considered to be more sensitive to training stress than measuring testosterone and cortisol separately. AI shows a strong negative correlation with TCR (r=-0.437, p < 0.001) and correctly represents post-exercise physiology. In order to reveal the relation between the ECG-derived metabolic indices and the state of the cardiorespiratory system, direct measurements of VO2max were carried out at various time points after training sessions. The negative correlation between AI and VO2max (r = -0.342, p < 0.001) was obtained. These data testifying VO2max rising during fatigue are controversial. However, some studies have revealed increased stroke volume after training, that agrees with findings. It is important to note that post-exercise increase in VO2max does not mean an athlete’s readiness for the next training session, because the recovery of the cardiovascular system occurs over a substantially longer period. Negative correlations registered for ANI with glycogen (r = -0.303, p < 0.001), albumin (r = -0.205, p < 0.021) and creatinine (r = -0.268, p < 0.002) reflect the dehydration status of participants after training. Correlations between designed metabolic indices and physiological parameters revealed in this study can be considered as the sufficient evidence to use these indices for assessing the state of person’s aerobic and anaerobic metabolic systems after training during fatigue, recovery and supercompensation.Keywords: aerobic index, anaerobic index, electrocardiography, supercompensation
Procedia PDF Downloads 115200 The Implantable MEMS Blood Pressure Sensor Model With Wireless Powering And Data Transmission
Authors: Vitaliy Petrov, Natalia Shusharina, Vitaliy Kasymov, Maksim Patrushev, Evgeny Bogdanov
Abstract:
The leading worldwide death reasons are ischemic heart disease and other cardiovascular illnesses. Generally, the common symptom is high blood pressure. Long-time blood pressure control is very important for the prophylaxis, correct diagnosis and timely therapy. Non-invasive methods which are based on Korotkoff sounds are impossible to apply often and for a long time. Implantable devices can combine longtime monitoring with high accuracy of measurements. The main purpose of this work is to create a real-time monitoring system for decreasing the death rate from cardiovascular diseases. These days implantable electronic devices began to play an important role in medicine. Usually implantable devices consist of a transmitter, powering which could be wireless with a special made battery and measurement circuit. Common problems in making implantable devices are short lifetime of the battery, big size and biocompatibility. In these work, blood pressure measure will be the focus because it’s one of the main symptoms of cardiovascular diseases. Our device will consist of three parts: the implantable pressure sensor, external transmitter and automated workstation in a hospital. The Implantable part of pressure sensors could be based on piezoresistive or capacitive technologies. Both sensors have some advantages and some limitations. The Developed circuit is based on a small capacitive sensor which is made of the technology of microelectromechanical systems (MEMS). The Capacitive sensor can provide high sensitivity, low power consumption and minimum hysteresis compared to the piezoresistive sensor. For this device, it was selected the oscillator-based circuit where frequency depends from the capacitance of sensor hence from capacitance one can calculate pressure. The external device (transmitter) used for wireless charging and signal transmission. Some implant devices for these applications are passive, the external device sends radio wave signal on internal LC circuit device. The external device gets reflected the signal from the implant and from a change of frequency is possible to calculate changing of capacitance and then blood pressure. However, this method has some disadvantages, such as the patient position dependence and static using. Developed implantable device doesn’t have these disadvantages and sends blood pressure data to the external part in real-time. The external device continuously sends information about blood pressure to hospital cloud service for analysis by a physician. Doctor’s automated workstation at the hospital also acts as a dashboard, which displays actual medical data of patients (which require attention) and stores it in cloud service. Usually, critical heart conditions occur few hours before heart attack but the device is able to send an alarm signal to the hospital for an early action of medical service. The system was tested with wireless charging and data transmission. These results can be used for ASIC design for MEMS pressure sensor.Keywords: MEMS sensor, RF power, wireless data, oscillator-based circuit
Procedia PDF Downloads 589199 Approximate-Based Estimation of Single Event Upset Effect on Statistic Random-Access Memory-Based Field-Programmable Gate Arrays
Authors: Mahsa Mousavi, Hamid Reza Pourshaghaghi, Mohammad Tahghighi, Henk Corporaal
Abstract:
Recently, Statistic Random-Access Memory-based (SRAM-based) Field-Programmable Gate Arrays (FPGAs) are widely used in aeronautics and space systems where high dependability is demanded and considered as a mandatory requirement. Since design’s circuit is stored in configuration memory in SRAM-based FPGAs; they are very sensitive to Single Event Upsets (SEUs). In addition, the adverse effects of SEUs on the electronics used in space are much higher than in the Earth. Thus, developing fault tolerant techniques play crucial roles for the use of SRAM-based FPGAs in space. However, fault tolerance techniques introduce additional penalties in system parameters, e.g., area, power, performance and design time. In this paper, an accurate estimation of configuration memory vulnerability to SEUs is proposed for approximate-tolerant applications. This vulnerability estimation is highly required for compromising between the overhead introduced by fault tolerance techniques and system robustness. In this paper, we study applications in which the exact final output value is not necessarily always a concern meaning that some of the SEU-induced changes in output values are negligible. We therefore define and propose Approximate-based Configuration Memory Vulnerability Factor (ACMVF) estimation to avoid overestimating configuration memory vulnerability to SEUs. In this paper, we assess the vulnerability of configuration memory by injecting SEUs in configuration memory bits and comparing the output values of a given circuit in presence of SEUs with expected correct output. In spite of conventional vulnerability factor calculation methods, which accounts any deviations from the expected value as failures, in our proposed method a threshold margin is considered depending on user-case applications. Given the proposed threshold margin in our model, a failure occurs only when the difference between the erroneous output value and the expected output value is more than this margin. The ACMVF is subsequently calculated by acquiring the ratio of failures with respect to the total number of SEU injections. In our paper, a test-bench for emulating SEUs and calculating ACMVF is implemented on Zynq-7000 FPGA platform. This system makes use of the Single Event Mitigation (SEM) IP core to inject SEUs into configuration memory bits of the target design implemented in Zynq-7000 FPGA. Experimental results for 32-bit adder show that, when 1% to 10% deviation from correct output is considered, the counted failures number is reduced 41% to 59% compared with the failures number counted by conventional vulnerability factor calculation. It means that estimation accuracy of the configuration memory vulnerability to SEUs is improved up to 58% in the case that 10% deviation is acceptable in output results. Note that less than 10% deviation in addition result is reasonably tolerable for many applications in approximate computing domain such as Convolutional Neural Network (CNN).Keywords: fault tolerance, FPGA, single event upset, approximate computing
Procedia PDF Downloads 198198 A Community Solution to Address Extensive Nitrate Contamination in the Lower Yakima Valley Aquifer
Authors: Melanie Redding
Abstract:
Historic widespread nitrate contamination of the Lower Yakima Valley aquifer in Washington State initiated a community-based effort to reduce nitrate concentrations to below-drinking water standards. This group commissioned studies on characterizing local nitrogen sources, deep soil assessments, drinking water, and assessing nitrate concentrations at the water table. Nitrate is the most prevalent groundwater contaminant with common sources from animal and human waste, fertilizers, plants and precipitation. It is challenging to address groundwater contamination when common sources, such as agriculture, on-site sewage systems, and animal production, are widespread. Remediation is not possible, so mitigation is essential. The Lower Yakima Valley is located over 175,000 acres, with a population of 56,000 residents. Approximately 25% of the population do not have access to safe, clean drinking water, and 20% of the population is at or below the poverty level. Agriculture is the primary economic land-use activity. Irrigated agriculture and livestock production make up the largest percentage of acreage and nitrogen load. Commodities include apples, grapes, hops, dairy, silage corn, triticale, alfalfa and cherries. These commodities are important to the economic viability of the residents of the Lower Yakima Valley, as well as Washington State. Mitigation of nitrate in groundwater is challenging. The goal is to ensure everyone has safe drinking water. There are no easy remedies due to the extensive and pervasiveness of the contamination. Monitoring at the water table indicates that 45% of the 30 spatially distributed monitoring wells exceeded the drinking water standard. This indicates that there are multiple sources that are impacting water quality. Washington State has several areas which have extensive groundwater nitrate contamination. The groundwater in these areas continues to degrade over time. However, the Lower Yakima Valley is being successful in addressing this health issue because of the following reasons: the community is engaged and committed; there is one common goal; there has been extensive public education and outreach to citizens; and generating credible data using sound scientific methods. Work in this area is continuing as an ambient groundwater monitoring network is established to assess the condition of the aquifer over time. Nitrate samples are being collected from 170 wells, spatially distributed across the aquifer. This research entails quarterly sampling for two years to characterize seasonal variability and then continue annually afterward. This assessment will provide the data to statistically determine trends in nitrate concentrations across the aquifer, over time. Thirty-three of these wells are monitoring wells that are screened across the aquifer. The water quality from these wells are indicative of activities at the land surface. Additional work is being conducted to identify land use management practices that are effective in limiting nitrate migration through the soil column. Tracking nitrate in the soil column every season is an important component of bridging land-use practices with the fate and transport of nitrate through the subsurface. Patience, tenacity, and the ability to think outside the box are essential for dealing with widespread nitrate contamination of groundwater.Keywords: community, groundwater, monitoring, nitrate
Procedia PDF Downloads 177197 Serum Concentration of the CCL7 Chemokine in Diabetic Pregnant Women during Pregnancy until the Postpartum Period
Authors: Fernanda Piculo, Giovana Vesentini, Gabriela Marini, Debora Cristina Damasceno, Angelica Mercia Pascon Barbosa, Marilza Vieira Cunha Rudge
Abstract:
Introduction: Women with previous gestational diabetes mellitus (GDM) were significantly more likely to have urinary incontinence (UI) and pelvic floor muscle dysfunction compared to non-diabetic women two years after a cesarean section. Additional results demonstrated that induced diabetes causes detrimental effects on pregnant rat urethral muscle. These results indicate the need for exploration of the mechanistic role of a recovery factor in female UI. Chemokine ligand 7 (CCL7) was significantly over expressed in rat serum, urethral and vaginal tissues immediately following induction of stress UI in a rat model simulating birth trauma. CCL7 over expression has shown potency for stimulating targeted stem cell migration and provide a translational link (clinical measurement) which further provide opportunities for treatment. The aim of this study was to investigate the CCL7 levels profile in diabetic pregnant women with urinary incontinence during pregnancy over the first year postpartum. Methods: This study was conducted in the Perinatal Diabetes Research Center of the Botucatu Medical School/UNESP, and was approved by the Research Ethics Committee of the Institution (CAAE: 20639813.0.0000.5411). The diagnosis of GDM was established between 24th and 28th gestational weeks, by the 75 g-OGTT test according to ADA’s criteria. Urinary incontinence was defined according to the International Continence Society and the CCL7 levels was measured by ELISA (R&D Systems, Catalog Number DCC700). Two hundred twelve women were classified into four study groups: normoglycemic continent (NC), normoglycemic incontinent (NI), diabetic continent (DC) and diabetic incontinent (DI). They were evaluated at six-time-points: 12-18, 24-28 and 34-38 gestational weeks, 24-48 hours, 6 weeks and 6-12 months postpartum. Results: At 12-18 weeks, it was possible to consider only two groups, continent and incontinent, because at this early gestational period has not yet been the diagnosis of GDM. The group with GDM and UI (DI group) showed lower levels of CCL7 in all time points during pregnancy and postpartum, compared to normoglycemic groups (NC and NI), indicating that these women have not recovered from child birth induced UI during the 6-12 months postpartum compared to their controls, and that the progression of UI and/or lack of recovery throughout the first postpartum year can be related with lower levels of CCL7. Instead, serum CCL7 was significantly increased in the NC group. Taken together, these findings of overexpression of CCL7 in the NC group and decreased levels in the DI group, could confirm that diabetes delays the recovery from child birth induced UI, and that CCL7 could potentially be used as a serum marker of injury. Conclusion: This study demonstrates lower levels of CCL7 in the DI group during pregnancy and postpartum and suggests that the progression of UI in diabetic women and/or lack of recovery throughout the first postpartum year can be related with low levels of CCL7. This provides a translational potential where CCL7 measurement could be used as a surrogate for injury after delivery. Successful controlled CCL7 mediated stem cell homing to the lower urinary tract could one day introduce the potential for non-operative treatment or prevention of stress urinary incontinence.Keywords: CCL7, gestational diabetes, pregnancy, urinary incontinence
Procedia PDF Downloads 336196 Scoring System for the Prognosis of Sepsis Patients in Intensive Care Units
Authors: Javier E. García-Gallo, Nelson J. Fonseca-Ruiz, John F. Duitama-Munoz
Abstract:
Sepsis is a syndrome that occurs with physiological and biochemical abnormalities induced by severe infection and carries a high mortality and morbidity, therefore the severity of its condition must be interpreted quickly. After patient admission in an intensive care unit (ICU), it is necessary to synthesize the large volume of information that is collected from patients in a value that represents the severity of their condition. Traditional severity of illness scores seeks to be applicable to all patient populations, and usually assess in-hospital mortality. However, the use of machine learning techniques and the data of a population that shares a common characteristic could lead to the development of customized mortality prediction scores with better performance. This study presents the development of a score for the one-year mortality prediction of the patients that are admitted to an ICU with a sepsis diagnosis. 5650 ICU admissions extracted from the MIMICIII database were evaluated, divided into two groups: 70% to develop the score and 30% to validate it. Comorbidities, demographics and clinical information of the first 24 hours after the ICU admission were used to develop a mortality prediction score. LASSO (least absolute shrinkage and selection operator) and SGB (Stochastic Gradient Boosting) variable importance methodologies were used to select the set of variables that make up the developed score; each of this variables was dichotomized and a cut-off point that divides the population into two groups with different mean mortalities was found; if the patient is in the group that presents a higher mortality a one is assigned to the particular variable, otherwise a zero is assigned. These binary variables are used in a logistic regression (LR) model, and its coefficients were rounded to the nearest integer. The resulting integers are the point values that make up the score when multiplied with each binary variables and summed. The one-year mortality probability was estimated using the score as the only variable in a LR model. Predictive power of the score, was evaluated using the 1695 admissions of the validation subset obtaining an area under the receiver operating characteristic curve of 0.7528, which outperforms the results obtained with Sequential Organ Failure Assessment (SOFA), Oxford Acute Severity of Illness Score (OASIS) and Simplified Acute Physiology Score II (SAPSII) scores on the same validation subset. Observed and predicted mortality rates within estimated probabilities deciles were compared graphically and found to be similar, indicating that the risk estimate obtained with the score is close to the observed mortality, it is also observed that the number of events (deaths) is indeed increasing as the outcome go from the decile with the lowest probabilities to the decile with the highest probabilities. Sepsis is a syndrome that carries a high mortality, 43.3% for the patients included in this study; therefore, tools that help clinicians to quickly and accurately predict a worse prognosis are needed. This work demonstrates the importance of customization of mortality prediction scores since the developed score provides better performance than traditional scoring systems.Keywords: intensive care, logistic regression model, mortality prediction, sepsis, severity of illness, stochastic gradient boosting
Procedia PDF Downloads 222195 Adaptive Programming for Indigenous Early Learning: The Early Years Model
Authors: Rachel Buchanan, Rebecca LaRiviere
Abstract:
Context: The ongoing effects of colonialism continue to be experienced through paternalistic policies and funding processes that cause disjuncture between and across Indigenous early childhood programming on-reserve and in urban and Northern settings in Canada. While various educational organizations and social service providers have risen to address these challenges in the short, medium and long term, there continues to be a lack in nation-wide cohesive, culturally grounded, and meaningful early learning programming for Indigenous children in Canada. Indigenous-centered early learning programs tend to face one of two scaling dilemmas: their program goals are too prescriptive to enable the program to be meaningfully replicated in different cultural/ community settings, or their program goals are too broad to be meaningfully adapted to the unique cultural and contextual needs and desires of Indigenous communities (the “franchise approach”). There are over 600 First Nations communities in Canada representing more than 50 Nations and languages. Consequently, Indigenous early learning programming cannot be applied with a universal or “one size fits all” approach. Sustainable and comprehensive programming must be responsive to each community context, building upon existing strengths and assets to avoid program duplication and irrelevance. Thesis: Community-driven and culturally adapted early childhood programming is critical but cannot be achieved on a large scale within traditional program models that are constrained by prescriptive overarching program goals. Principles, rather than goals, are an effective way to navigate and evaluate complex and dynamic systems. Principles guide an intervention to be adaptable, flexible and scalable. The Martin Family Initiative (MFI) ’s Early Years program engages a principles-based approach to programming. As will be discussed in this paper, this approach enables the program to catalyze existing community-based strengths and organizational assets toward bridging gaps across and disjuncture between Indigenous early learning programs, as well as to scale programming in sustainable, context-responsive and dynamic ways. This paper argues that using a principles-driven and adaptive scaling approach, the Early Years model establishes important learnings for culturally adapted Indigenous early learning programming in Canada. Methodology: The Early Years has leveraged this approach to develop an array of programming with partner organizations and communities across the country. The Early Years began as a singular pilot project in one First Nation. In just three years, it has expanded to five different regions and community organizations. In each context, the program supports the partner organization through different means and to different ends, the extent to which is determined in partnership with each community-based organization: in some cases, this means supporting the organization to build home visiting programming from the ground-up; in others, it means offering organization-specific culturally adapted early learning resources to support the programming that already exists in communities. Principles underpin but do not define the practices of the program in each of these relationships. This paper will explore numerous examples of principles-based adaptability with the context of the Early Years, concluding that the program model offers theadaptability and dynamism necessary to respond to unique and ever-evolving community contexts and needs of Indigenous children today.Keywords: culturally adapted programming, indigenous early learning, principles-based approach, program scaling
Procedia PDF Downloads 186194 Modelling of Reactive Methodologies in Auto-Scaling Time-Sensitive Services With a MAPE-K Architecture
Authors: Óscar Muñoz Garrigós, José Manuel Bernabeu Aubán
Abstract:
Time-sensitive services are the base of the cloud services industry. Keeping low service saturation is essential for controlling response time. All auto-scalable services make use of reactive auto-scaling. However, reactive auto-scaling has few in-depth studies. This presentation shows a model for reactive auto-scaling methodologies with a MAPE-k architecture. Queuing theory can compute different properties of static services but lacks some parameters related to the transition between models. Our model uses queuing theory parameters to relate the transition between models. It associates MAPE-k related times, the sampling frequency, the cooldown period, the number of requests that an instance can handle per unit of time, the number of incoming requests at a time instant, and a function that describes the acceleration in the service's ability to handle more requests. This model is later used as a solution to horizontally auto-scale time-sensitive services composed of microservices, reevaluating the model’s parameters periodically to allocate resources. The solution requires limiting the acceleration of the growth in the number of incoming requests to keep a constrained response time. Business benefits determine such limits. The solution can add a dynamic number of instances and remains valid under different system sizes. The study includes performance recommendations to improve results according to the incoming load shape and business benefits. The exposed methodology is tested in a simulation. The simulator contains a load generator and a service composed of two microservices, where the frontend microservice depends on a backend microservice with a 1:1 request relation ratio. A common request takes 2.3 seconds to be computed by the service and is discarded if it takes more than 7 seconds. Both microservices contain a load balancer that assigns requests to the less loaded instance and preemptively discards requests if they are not finished in time to prevent resource saturation. When load decreases, instances with lower load are kept in the backlog where no more requests are assigned. If the load grows and an instance in the backlog is required, it returns to the running state, but if it finishes the computation of all requests and is no longer required, it is permanently deallocated. A few load patterns are required to represent the worst-case scenario for reactive systems: the following scenarios test response times, resource consumption and business costs. The first scenario is a burst-load scenario. All methodologies will discard requests if the rapidness of the burst is high enough. This scenario focuses on the number of discarded requests and the variance of the response time. The second scenario contains sudden load drops followed by bursts to observe how the methodology behaves when releasing resources that are lately required. The third scenario contains diverse growth accelerations in the number of incoming requests to observe how approaches that add a different number of instances can handle the load with less business cost. The exposed methodology is compared against a multiple threshold CPU methodology allocating/deallocating 10 or 20 instances, outperforming the competitor in all studied metrics.Keywords: reactive auto-scaling, auto-scaling, microservices, cloud computing
Procedia PDF Downloads 93193 Avoidance of Brittle Fracture in Bridge Bearings: Brittle Fracture Tests and Initial Crack Size
Authors: Natalie Hoyer
Abstract:
Bridges in both roadway and railway systems depend on bearings to ensure extended service life and functionality. These bearings enable proper load distribution from the superstructure to the substructure while permitting controlled movement of the superstructure. The design of bridge bearings, according to Eurocode DIN EN 1337 and the relevant sections of DIN EN 1993, increasingly requires the use of thick plates, especially for long-span bridges. However, these plate thicknesses exceed the limits specified in the national appendix of DIN EN 1993-2. Furthermore, compliance with DIN EN 1993-1-10 regulations regarding material toughness and through-thickness properties necessitates further modifications. Consequently, these standards cannot be directly applied to the selection of bearing materials without supplementary guidance and design rules. In this context, a recommendation was developed in 2011 to regulate the selection of appropriate steel grades for bearing components. Prior to the initiation of the research project underlying this contribution, this recommendation had only been available as a technical bulletin. Since July 2023, it has been integrated into guideline 804 of the German railway. However, recent findings indicate that certain bridge-bearing components are exposed to high fatigue loads, which necessitate consideration in structural design, material selection, and calculations. Therefore, the German Centre for Rail Traffic Research called a research project with the objective of defining a proposal to expand the current standards in order to implement a sufficient choice of steel material for bridge bearings to avoid brittle fracture, even for thick plates and components subjected to specific fatigue loads. The results obtained from theoretical considerations, such as finite element simulations and analytical calculations, are validated through large-scale component tests. Additionally, experimental observations are used to calibrate the calculation models and modify the input parameters of the design concept. Within the large-scale component tests, a brittle failure is artificially induced in a bearing component. For this purpose, an artificially generated initial defect is introduced at the previously defined hotspot into the specimen using spark erosion. Then, a dynamic load is applied until the crack initiation process occurs to achieve realistic conditions in the form of a sharp notch similar to a fatigue crack. This initiation process continues until the crack length reaches a predetermined size. Afterward, the actual test begins, which requires cooling the specimen with liquid nitrogen until a temperature is reached where brittle fracture failure is expected. In the next step, the component is subjected to a quasi-static tensile test until failure occurs in the form of a brittle failure. The proposed paper will present the latest research findings, including the results of the conducted component tests and the derived definition of the initial crack size in bridge bearings.Keywords: bridge bearings, brittle fracture, fatigue, initial crack size, large-scale tests
Procedia PDF Downloads 44192 The Role of Time-Dependent Treatment of Exogenous Salicylic Acid on Endogenous Phytohormone Levels under Salinity Stress
Authors: Hülya Torun, Ondřej Novák, Jaromír Mikulík, Miroslav Strnad, Faik A. Ayaz
Abstract:
World climate is changing. Millions of people in the world still face chronic undernourishment for conducting a healthy life and the world’s population is growing steadily. To meet this growing demand, agriculture and food systems must adapt to the adverse effects of climate change and become more resilient, productive and sustainable. From this perspective, to determine tolerant cultivars for undesirable environmental conditions will be necessary food production for sustainable development. Among abiotic stresses, soil salinity is one of the most detrimental global fact restricting plant sources. Development of salt-tolerant lines is required in order to increase the crop productivity and quality in salt-treated lands. Therefore, the objective of this study was to investigate the morphological and physiological responses of barley cultivars accessions to salinity stress by NaCl. For this purpose, it was aimed to determine the crosstalk between some endogenous phytohormones and exogenous salicylic acid (SA) in two different vegetative parts (leaves and roots) of barley (Hordeum vulgare L.; Poaceae; 2n=14; Ince-04) which is detected salt-tolerant. The effects of SA on growth parameters, leaf relative water content (RWC), endogenous phytohormones; including indole-3-acetic acid (IAA), cytokinins (CKs), abscisic acid (ABA), jasmonic acid (JA) and ethylene were investigated in barley cultivars under salinity stress. SA was applied to 17-day-old seedlings of barley in two different ways including before (pre-treated for 24 h) and simultaneously with NaCl stress treatment. NaCl (0, 150, 300 mM) exposure in the hydrophonic system was associated with a rapid decrease in growth parameters and RWC, which is an indicator of plant water status, resulted in a strong up-regulation of ABA as a stress indicator. Roots were more dramatically affected than leaves. Water conservation in 150 mM NaCl treated-barley plants did not change, but decreased in 300 mM NaCl treated plants. Pre- and simultaneously treatment of SA did not significantly alter growth parameters and RWC. ABA, JA and ethylene are known to be related with stress. In the present work, ethylene also increased, similarly to ABA, but not with the same intensity. While ABA and ethylene increased by the increment of salt concentrations, JA levels rapidly decreased especially in roots. Both pre- and simultaneously SA applications alleviated salt-induced decreases in 300 mM NaCl resulted in the increment of ABA levels. CKs and IAA are related to cell growth and development. At high salinity (300 mM NaCl), CKs (cZ+cZR) contents increased in both vegetative organs while IAA levels stayed at the same level with control groups. However, IAA increased and cZ+cZR rapidly decreased in leaves of barley plants with SA treatments before salt applications (in pre- SA treated groups). Simultaneously application of SA decreased CKs levels in both leaves and roots of the cultivar. Due to increasing concentrations of NaCl in association with decreasing ABA, JA and ethylene content and increments in CKs and IAA were recorded with SA treatments. As results of the study, in view of all the phytohormones that we tested, exogenous SA induced greater tolerance to salinity particularly when applied before salinity stress.Keywords: Barley, Hordeum vulgare, phytohormones, salicylic acid, salinity
Procedia PDF Downloads 228191 DSF Elements in High-Rise Timber Buildings
Authors: Miroslav Premrov, Andrej Štrukelj, Erika Kozem Šilih
Abstract:
The utilization of prefabricated timber-wall elements with double glazing, called as double-skin façade element (DSF), represents an innovative structural approach in the context of new high-rise timber construction, simultaneously combining sustainable solutions with improved energy efficiency and living quality. In addition to the minimum energy needs of buildings, the design of modern buildings is also increasingly focused on the optimal indoor comfort, in particular on sufficient natural light indoors. An optimally energy-designed building with an optimal layout of glazed areas around the building envelope represents a great potential in modern timber construction. Usually, all these transparent façade elements, because of energy benefits, are primary asymmetrical oriented and if they are considered as non-resisting against a horizontal load impact, a strong torsion effects in the building can appear. The problem of structural stability against a strong horizontal load impact of such modern timber buildings especially increase in a case of high-rise structures where additional bracing elements have to be used. In such a case, special diagonal bracing systems or other bracing solutions with common timber wall elements have to be incorporated into the structure of the building to satisfy all prescribed resisting requirements given by the standards. However, all such structural solutions are usually not environmentally friendly and also not contribute to an improved living comfort, or they are not accepted by the architects at all. Consequently, it is a special need to develop innovative load-bearing timber-glass wall elements which are in the same time environmentally friendly, can increase internal comfort in the building, but are also load-bearing. The new developed load-bearing DSF elements can be a good answer on all these requirements. Timber-glass façade elements DSF wall elements consist of two transparent layers, thermal-insulated three-layered glass pane on the internal side and an additional single-layered glass pane on the external side of the wall. The both panes are separated by an air channel which can be of any dimensions and can have a significant influence on the thermal insulation or acoustic response of such a wall element. Most already published studies on DSF elements primarily deal only with energy and LCA solutions and do not address any structural problems. In previous studies according to experimental analysis and mathematical modeling it was already presented a possible benefit of such load-bearing DSF elements, especially comparing with previously developed load-bearing single-skin timber wall elements, but they were not applicate yet in any high-rise timber structure. Therefore, in the presented study specially selected 10-storey prefabricated timber building constructed in a cross-laminated timber (CLT) structural wall system is analyzed using the developed DSF elements in a sense to increase a structural lateral stability of the whole building. The results evidently highlight the importance the load-bearing DSF elements, as their incorporation can have a significant impact on the overall behavior of the structure through their influence on the stiffness properties. Taking these considerations into account is crucial to ensure compliance with seismic design codes and to improve the structural resilience of high-rise timber buildings.Keywords: glass, high-rise buildings, numerical analysis, timber
Procedia PDF Downloads 46190 A High-Throughput Enzyme Screening Method Using Broadband Coherent Anti-stokes Raman Spectroscopy
Authors: Ruolan Zhang, Ryo Imai, Naoko Senda, Tomoyuki Sakai
Abstract:
Enzymes have attracted increasing attentions in industrial manufacturing for their applicability in catalyzing complex chemical reactions under mild conditions. Directed evolution has become a powerful approach to optimize enzymes and exploit their full potentials under the circumstance of insufficient structure-function knowledge. With the incorporation of cell-free synthetic biotechnology, rapid enzyme synthesis can be realized because no cloning procedure such as transfection is needed. Its open environment also enables direct enzyme measurement. These properties of cell-free biotechnology lead to excellent throughput of enzymes generation. However, the capabilities of current screening methods have limitations. Fluorescence-based assay needs applicable fluorescent label, and the reliability of acquired enzymatic activity is influenced by fluorescent label’s binding affinity and photostability. To acquire the natural activity of an enzyme, another method is to combine pre-screening step and high-performance liquid chromatography (HPLC) measurement. But its throughput is limited by necessary time investment. Hundreds of variants are selected from libraries, and their enzymatic activities are then identified one by one by HPLC. The turn-around-time is 30 minutes for one sample by HPLC, which limits the acquirable enzyme improvement within reasonable time. To achieve the real high-throughput enzyme screening, i.e., obtain reliable enzyme improvement within reasonable time, a widely applicable high-throughput measurement of enzymatic reactions is highly demanded. Here, a high-throughput screening method using broadband coherent anti-Stokes Raman spectroscopy (CARS) was proposed. CARS is one of coherent Raman spectroscopy, which can identify label-free chemical components specifically from their inherent molecular vibration. These characteristic vibrational signals are generated from different vibrational modes of chemical bonds. With the broadband CARS, chemicals in one sample can be identified from their signals in one broadband CARS spectrum. Moreover, it can magnify the signal levels to several orders of magnitude greater than spontaneous Raman systems, and therefore has the potential to evaluate chemical's concentration rapidly. As a demonstration of screening with CARS, alcohol dehydrogenase, which converts ethanol and nicotinamide adenine dinucleotide oxidized form (NAD+) to acetaldehyde and nicotinamide adenine dinucleotide reduced form (NADH), was used. The signal of NADH at 1660 cm⁻¹, which is generated from nicotinamide in NADH, was utilized to measure the concentration of it. The evaluation time for CARS signal of NADH was determined to be as short as 0.33 seconds while having a system sensitivity of 2.5 mM. The time course of alcohol dehydrogenase reaction was successfully measured from increasing signal intensity of NADH. This measurement result of CARS was consistent with the result of a conventional method, UV-Vis. CARS is expected to have application in high-throughput enzyme screening and realize more reliable enzyme improvement within reasonable time.Keywords: Coherent Anti-Stokes Raman Spectroscopy, CARS, directed evolution, enzyme screening, Raman spectroscopy
Procedia PDF Downloads 141189 Magnetic Solid-Phase Separation of Uranium from Aqueous Solution Using High Capacity Diethylenetriamine Tethered Magnetic Adsorbents
Authors: Amesh P, Suneesh A S, Venkatesan K A
Abstract:
The magnetic solid-phase extraction is a relatively new method among the other solid-phase extraction techniques for the separating of metal ions from aqueous solutions, such as mine water and groundwater, contaminated wastes, etc. However, the bare magnetic particles (Fe3O4) exhibit poor selectivity due to the absence of target-specific functional groups for sequestering the metal ions. The selectivity of these magnetic particles can be remarkably improved by covalently tethering the task-specific ligands on magnetic surfaces. The magnetic particles offer a number of advantages such as quick phase separation aided by the external magnetic field. As a result, the solid adsorbent can be prepared with the particle size ranging from a few micrometers to the nanometer, which again offers the advantages such as enhanced kinetics of extraction, higher extraction capacity, etc. Conventionally, the magnetite (Fe3O4) particles were prepared by the hydrolysis and co-precipitation of ferrous and ferric salts in aqueous ammonia solution. Since the covalent linking of task-specific functionalities on Fe3O4 was difficult, and it is also susceptible to redox reaction in the presence of acid or alkali, it is necessary to modify the surface of Fe3O4 by silica coating. This silica coating is usually carried out by hydrolysis and condensation of tetraethyl orthosilicate over the surface of magnetite to yield a thin layer of silica-coated magnetite particles. Since the silica-coated magnetite particles amenable for further surface modification, it can be reacted with task-specific functional groups to obtain the functionalized magnetic particles. The surface area exhibited by such magnetic particles usually falls in the range of 50 to 150 m2.g-1, which offer advantage such as quick phase separation, as compared to the other solid-phase extraction systems. In addition, the magnetic (Fe3O4) particles covalently linked on mesoporous silica matrix (MCM-41) and task-specific ligands offer further advantages in terms of extraction kinetics, high stability, longer reusable cycles, and metal extraction capacity, due to the large surface area, ample porosity and enhanced number of functional groups per unit area on these adsorbents. In view of this, the present paper deals with the synthesis of uranium specific diethylenetriamine ligand (DETA) ligand anchored on silica-coated magnetite (Fe-DETA) as well as on magnetic mesoporous silica (MCM-Fe-DETA) and studies on the extraction of uranium from aqueous solution spiked with uranium to mimic the mine water or groundwater contaminated with uranium. The synthesized solid-phase adsorbents were characterized by FT-IR, Raman, TG-DTA, XRD, and SEM. The extraction behavior of uranium on the solid-phase was studied under several conditions like the effect of pH, initial concentration of uranium, rate of extraction and its variation with pH and initial concentration of uranium, effect of interference ions like CO32-, Na+, Fe+2, Ni+2, and Cr+3, etc. The maximum extraction capacity of 233 mg.g-1 was obtained for Fe-DETA, and a huge capacity of 1047 mg.g-1 was obtained for MCM-Fe-DETA. The mechanism of extraction, speciation of uranium, extraction studies, reusability, and the other results obtained in the present study suggests Fe-DETA and MCM-Fe-DETA are the potential candidates for the extraction of uranium from mine water, and groundwater.Keywords: diethylenetriamine, magnetic mesoporous silica, magnetic solid-phase extraction, uranium extraction, wastewater treatment
Procedia PDF Downloads 168188 The Impact of an Improved Strategic Partnership Programme on Organisational Performance and Growth of Firms in the Internet Protocol Television and Hybrid Fibre-Coaxial Broadband Industry
Authors: Collen T. Masilo, Brane Semolic, Pieter Steyn
Abstract:
The Internet Protocol Television (IPTV) and Hybrid Fibre-Coaxial (HFC) Broadband industrial sector landscape are rapidly changing and organisations within the industry need to stay competitive by exploring new business models so that they can be able to offer new services and products to customers. The business challenge in this industrial sector is meeting or exceeding high customer expectations across multiple content delivery modes. The increasing challenges in the IPTV and HFC broadband industrial sector encourage service providers to form strategic partnerships with key suppliers, marketing partners, advertisers, and technology partners. The need to form enterprise collaborative networks poses a challenge for any organisation in this sector, in selecting the right strategic partners who will ensure that the organisation’s services and products are marketed in new markets. Partners who will ensure that customers are efficiently supported by meeting and exceeding their expectations. Lastly, selecting cooperation partners who will represent the organisation in a positive manner, and contribute to improving the performance of the organisation. Companies in the IPTV and HFC broadband industrial sector tend to form informal partnerships with suppliers, vendors, system integrators and technology partners. Generally, partnerships are formed without thorough analysis of the real reason a company is forming collaborations, without proper evaluations of prospective partners using specific selection criteria, and with ineffective performance monitoring of partners to ensure that a firm gains real long term benefits from its partners and gains competitive advantage. Similar tendencies are illustrated in the research case study and are based on Skyline Communications, a global leader in end-to-end, multi-vendor network management and operational support systems (OSS) solutions. The organisation’s flagship product is the DataMiner network management platform used by many operators across multiple industries and can be referred to as a smart system that intelligently manages complex technology ecosystems for its customers in the IPTV and HFC broadband industry. The approach of the research is to develop the most efficient business model that can be deployed to improve a strategic partnership programme in order to significantly improve the performance and growth of organisations participating in a collaborative network in the IPTV and HFC broadband industrial sector. This involves proposing and implementing a new strategic partnership model and its main features within the industry which should bring about significant benefits for all involved companies to achieve value add and an optimal growth strategy. The proposed business model has been developed based on the research of existing relationships, value chains and business requirements in this industrial sector and validated in 'Skyline Communications'. The outputs of the business model have been demonstrated and evaluated in the research business case study the IPTV and HFC broadband service provider 'Skyline Communications'.Keywords: growth, partnership, selection criteria, value chain
Procedia PDF Downloads 133187 Assessment of Ecosystem Readiness for Adoption of Circularity: A Multi-Case Study Analysis of Textile Supply Chain in Pakistan
Authors: Azhar Naila, Steuer Benjamin
Abstract:
Over-exploitation of resources and the burden on natural systems have provoked worldwide concerns about the potential resource as well as supply risks in the future. It has been estimated that the consumption of materials and resources will double by 2060, substantially mounting the amount of waste and emissions produced by individuals, organizations, and businesses, which necessitates sustainable technological innovations to address the problem. Therefore, there is a need to design products and services purposefully for material resource efficiency. This directs us toward the conceptualization and implementation of the ‘Circular Economy (CE),’ which has gained considerable attention among policymakers, researchers, and businesses in the past decade. A large amount of literature focuses on the concept of CE. However, contextual empirical research on the need to embrace CE in an emerging economy like Pakistan is still scarce, where the traditional economic model of take-make-dispose is quite common. Textile exports account for approximately 61% of Pakistan's total exports, and the industry provides employment for about 40% of the country's total industrial workforce. The industry provides job opportunities to above 10 million farmers, with cotton as the main crop of Pakistan. Consumers, companies, as well as the government have explored very limited CE potential in the country. This gap has motivated us to carry out the present study. The study is based on a mixed method approach, for which key informant interviews have been conducted to get insight into the present situation of the ecosystem readiness for the adoption of CE in 20 textile manufacturing industries. The subject study has been conducted on the following areas i) the level of understanding of the CE concept among key stakeholders in the textile manufacturing industry ii) Companies are pushing boundaries to invest in circularity-based initiatives, exploring the depths of risk-taking iii) the current national policy framework support the adoption of CE. Qualitative assessment has been undertaken using MAXQDA to analyze the data received after the key informant interviews. The data has been transcribed and coded for further analysis. The results show that most of the key stakeholders have a clear understanding of the concept, whereas few consider it to be only relevant to the end-of-life treatment of waste generated from the industry. Non-governmental organizations have been observed to be key players in creating awareness among the manufacturing industries. Maximum companies have shown their consent to invest in initiatives related to the adoption of CE. Whereas a few consider themselves far behind the race due to a lack of financial resources and support from responsible institutions. Mostly, the industries have an ambitious vision for integrating CE into the company’s policy but seem not to be ready to take any significant steps to nurture a culture for experimentation. However, the government is not playing any vital role in the transition towards CE; rather, they have been busy with the state’s uncertain political situation. Presently, Pakistan does not have any policy framework that supports the transition towards CE. Acknowledging the present landscape a well-informed CE transition is immediately required.Keywords: circular economy, textile supply chain, textile manufacturing industries, resource efficiency, ecosystem readiness, multi-case study analysis
Procedia PDF Downloads 52186 Psychological Distress during the COVID-19 Pandemic in Nursing Students: A Mixed-Methods Study
Authors: Mayantoinette F. Watson
Abstract:
During such an unprecedented time of the largest public health crisis, the COVID-19 pandemic, nursing students are of the utmost concern regarding their psychological and physical well-being. Questions are emerging and circulating about what will happen to the nursing students and the long-term effects of the pandemic, especially now that hospitals are being overwhelmed with a significant need for nursing staff. Expectations, demands, change, and the fear of the unknown during this unprecedented time can only contribute to the many stressors that accompany nursing students through laborious clinical and didactic courses in nursing programs. The risk of psychological distress is at a maximum, and its effects can negatively impact not only nursing students but also nursing education and academia. The high exposures to interpersonal, economic, and academic demands contribute to the major health concerns, which include a potential risk for psychological distress. Achievement of educational success among nursing students is directly affected by the high exposure to anxiety and depression from experiences within the program. Working relationships and achieving academic success is imperative to positive student outcomes within the nursing program. The purpose of this study is to identify and establish influences and associations within multilevel factors, including the effects of the COVID-19 pandemic on psychological distress in nursing students. Neuman’s Systems Model Theory was used to determine nursing students’ responses to internal and external stressors. The research in this study utilized a mixed-methods, convergent study design. The study population included undergraduate nursing students from Southeastern U.S. The research surveyed a convenience sample of undergraduate nursing students. The quantitative survey was completed by 202 participants, and 11 participants participated in the qualitative follow-up interview surveys. Participants completed the Kessler Psychological Distress Scale (K6), the Perceived Stress Scale (PSS4), and the Dundee Readiness Educational Environment Scale (DREEM12) to measure psychological distress, perceived stress, and perceived educational environment. Participants also answered open-ended questions regarding their experience during the COVID-19 pandemic. Statistical tests, including bivariate analyses, multiple linear regression analyses, and binary logistics regression analyses were performed in effort to identify and highlight the effects of independent variables on the dependent variable, psychological distress. Coding and qualitative content analysis were performed to identify overarching themes within participants’ interviews. Quantitative data were sufficient in identifying correlations between psychological distress and multilevel factors of coping, marital status, COVID-19 stress, perceived stress, educational environment, and social support in nursing students. Qualitative data were sufficient in identifying common themes of students’ perceptions during COVID-19 and included online learning, workload, finances, experience, breaks, time, unknown, support, encouragement, unchanged, communication, and transmission. The findings are significant, specifically regarding contributing factors to nursing students’ psychological distress, which will help to improve learning in the academic environment.Keywords: nursing education, nursing students, pandemic, psychological distress
Procedia PDF Downloads 86185 Political Communication in Twitter Interactions between Government, News Media and Citizens in Mexico
Authors: Jorge Cortés, Alejandra Martínez, Carlos Pérez, Anaid Simón
Abstract:
The presence of government, news media, and general citizenry in social media allows considering interactions between them as a form of political communication (i.e. the public exchange of contradictory discourses about politics). Twitter’s asymmetrical following model (users can follow, mention or reply to other users that do not follow them) could foster alternative democratic practices and have an impact on Mexican political culture, which has been marked by a lack of direct communication channels between these actors. The research aim is to assess Twitter’s role in political communication practices through the analysis of interaction dynamics between government, news media, and citizens by extracting and visualizing data from Twitter’s API to observe general behavior patterns. The hypothesis is that regardless the fact that Twitter’s features enable direct and horizontal interactions between actors, users repeat traditional dynamics of interaction, without taking full advantage of the possibilities of this medium. Through an interdisciplinary team including Communication Strategies, Information Design, and Interaction Systems, the activity on Twitter generated by the controversy over the presence of Uber in Mexico City was analysed; an issue of public interest, involving aspects such as public opinion, economic interests and a legal dimension. This research includes techniques from social network analysis (SNA), a methodological approach focused on the comprehension of the relationships between actors through the visual representation and measurement of network characteristics. The analysis of the Uber event comprised data extraction, data categorization, corpus construction, corpus visualization and analysis. On the recovery stage TAGS, a Google Sheet template, was used to extract tweets that included the hashtags #UberSeQueda and #UberSeVa, posts containing the string Uber and tweets directed to @uber_mx. Using scripts written in Python, the data was filtered, discarding tweets with no interaction (replies, retweets or mentions) and locations outside of México. Considerations regarding bots and the omission of anecdotal posts were also taken into account. The utility of graphs to observe interactions of political communication in general was confirmed by the analysis of visualizations generated with programs such as Gephi and NodeXL. However, some aspects require improvements to obtain more useful visual representations for this type of research. For example, link¬crossings complicates following the direction of an interaction forcing users to manipulate the graph to see it clearly. It was concluded that some practices prevalent in political communication in Mexico are replicated in Twitter. Media actors tend to group together instead of interact with others. The political system tends to tweet as an advertising strategy rather than to generate dialogue. However, some actors were identified as bridges establishing communication between the three spheres, generating a more democratic exercise and taking advantage of Twitter’s possibilities. Although interactions in Twitter could become an alternative to political communication, this potential depends on the intentions of the participants and to what extent they are aiming for collaborative and direct communications. Further research is needed to get a deeper understanding on the political behavior of Twitter users and the possibilities of SNA for its analysis.Keywords: interaction, political communication, social network analysis, Twitter
Procedia PDF Downloads 221184 The Importance of School Culture in Supporting Student Mental Health Following the COVID-19 Pandemic: Insights from a Qualitative Study
Authors: Rhiannon Barker, Gregory Hartwell, Matt Egan, Karen Lock
Abstract:
Background: Evidence suggests that mental health (MH) issues in children and young people (CYP) in the UK are on the rise. Of particular concern is data that indicates that the pandemic, together with the impact of school closures, have accentuated already pronounced inequalities; children from families on low incomes or from black and minority ethnic groups are reportedly more likely to have been adversely impacted. This study aimed to help identify specific support which may facilitate the building of a positive school climate and protect student mental health, particularly in the wake of school closures following the pandemic. It has important implications for integrated working between schools and statutory health services. Methods: The research comprised of three parts; scoping, case studies, and a stakeholder workshop to explore and consolidate results. The scoping phase included a literature review alongside interviews with a range of stakeholders from government, academia, and the third sector. Case studies were then conducted in two London state schools. Results: Our research identified how student MH was being impacted by a range of factors located at different system levels, both internal to the school and in the wider community. School climate, relating both to a shared system of beliefs and values, as well as broader factors including style of leadership, teaching, discipline, safety, and relationships -all played a role in the experience of school life and, consequently, the MH of both students and staff. Participants highlighted the importance of a whole school approach and ensuring that support for student MH was not separated from academic achievement, as well as the importance of identifying and applying universal measuring systems to establish levels of MH need. Our findings suggest that a school’s climate is influenced by the style and strength of its leadership, while this school climate - together with mechanisms put in place to respond to MH needs (both statutory and non-statutory) - plays a key role in supporting student MH. Implications: Schools in England have a responsibility to decide on the nature of MH support provided for their students, and there is no requirement for them to report centrally on the form this provision takes. The reality on the ground, as our study suggests, is that MH provision varies significantly between schools, particularly in relation to ‘lower’ levels of need which are not covered by statutory requirements. A valid concern may be that in the huge raft of possible options schools have to support CYP wellbeing, too much is left to chance. Work to support schools in rebuilding their cultures post-lockdowns must include the means to identify and promote appropriate tools and techniques to facilitate regular measurement of student MH. This will help establish both the scale of the problem and monitor the effectiveness of the response. A strong vision from a school’s leadership team that emphasises the importance of student wellbeing, running alongside (but not overshadowed by) academic attainment, should help shape a school climate to promote beneficial MH outcomes. The sector should also be provided with support to improve the consistency and efficacy of MH provision in schools across the country.Keywords: mental health, schools, young people, whole-school culture
Procedia PDF Downloads 63183 A Hardware-in-the-loop Simulation for the Development of Advanced Control System Design for a Spinal Joint Wear Simulator
Authors: Kaushikk Iyer, Richard M Hall, David Keeling
Abstract:
Hardware-in-the-loop (HIL) simulation is an advanced technique for developing and testing complex real-time control systems. This paper presents the benefits of HIL simulation and how it can be implemented and used effectively to develop, test, and validate advanced control algorithms used in a spinal joint Wear simulator for the Tribological testing of spinal disc prostheses. spinal wear simulator is technologically the most advanced machine currently employed For the in-vitro testing of newly developed spinal Discimplants. However, the existing control techniques, such as a simple position control Does not allow the simulator to test non-sinusoidal waveforms. Thus, there is a need for better and advanced control methods that can be developed and tested Rigorouslybut safely before deploying it into the real simulator. A benchtop HILsetupis was created for experimentation, controller verification, and validation purposes, allowing different control strategies to be tested rapidly in a safe environment. The HIL simulation aspect in this setup attempts to replicate similar spinal motion and loading conditions. The spinal joint wear simulator containsa four-Barlinkpowered by electromechanical actuators. LabVIEW software is used to design a kinematic model of the spinal wear Simulator to Validatehow each link contributes towards the final motion of the implant under test. As a result, the implant articulates with an angular motion specified in the international standards, ISO-18192-1, that define fixed, simplified, and sinusoid motion and load profiles for wear testing of cervical disc implants. Using a PID controller, a velocity-based position control algorithm was developed to interface with the benchtop setup that performs HIL simulation. In addition to PID, a fuzzy logic controller (FLC) was also developed that acts as a supervisory controller. FLC provides intelligence to the PID controller by By automatically tuning the controller for profiles that vary in amplitude, shape, and frequency. This combination of the fuzzy-PID controller is novel to the wear testing application for spinal simulators and demonstrated superior performance against PIDwhen tested for a spectrum of frequency. Kaushikk Iyer is a Ph.D. Student at the University of Leeds and an employee at Key Engineering Solutions, Leeds, United Kingdom, (e-mail: [email protected], phone: +44 740 541 5502). Richard M Hall is with the University of Leeds, the United Kingdom as a professor in the Mechanical Engineering Department (e-mail: [email protected]). David Keeling is the managing director of Key Engineering Solutions, Leeds, United Kingdom (e-mail: [email protected]). Results obtained are successfully validated against the load and motion tolerances specified by the ISO18192-1 standard and fall within limits, that is, ±0.5° at the maxima and minima of the motion and ±2 % of the complete cycle for phasing. The simulation results prove the efficacy of the test setup using HIL simulation to verify and validate the accuracy and robustness of the prospective controller before its deployment into the spinal wear simulator. This method of testing controllers enables a wide range of possibilities to test advanced control algorithms that can potentially test even profiles of patients performing various dailyliving activities.Keywords: Fuzzy-PID controller, hardware-in-the-loop (HIL), real-time simulation, spinal wear simulator
Procedia PDF Downloads 171182 A Framework for Automated Nuclear Waste Classification
Authors: Seonaid Hume, Gordon Dobie, Graeme West
Abstract:
Detecting and localizing radioactive sources is a necessity for safe and secure decommissioning of nuclear facilities. An important aspect for the management of the sort-and-segregation process is establishing the spatial distributions and quantities of the waste radionuclides, their type, corresponding activity, and ultimately classification for disposal. The data received from surveys directly informs decommissioning plans, on-site incident management strategies, the approach needed for a new cell, as well as protecting the workforce and the public. Manual classification of nuclear waste from a nuclear cell is time-consuming, expensive, and requires significant expertise to make the classification judgment call. Also, in-cell decommissioning is still in its relative infancy, and few techniques are well-developed. As with any repetitive and routine tasks, there is the opportunity to improve the task of classifying nuclear waste using autonomous systems. Hence, this paper proposes a new framework for the automatic classification of nuclear waste. This framework consists of five main stages; 3D spatial mapping and object detection, object classification, radiological mapping, source localisation based on gathered evidence and finally, waste classification. The first stage of the framework, 3D visual mapping, involves object detection from point cloud data. A review of related applications in other industries is provided, and recommendations for approaches for waste classification are made. Object detection focusses initially on cylindrical objects since pipework is significant in nuclear cells and indeed any industrial site. The approach can be extended to other commonly occurring primitives such as spheres and cubes. This is in preparation of stage two, characterizing the point cloud data and estimating the dimensions, material, degradation, and mass of the objects detected in order to feature match them to an inventory of possible items found in that nuclear cell. Many items in nuclear cells are one-offs, have limited or poor drawings available, or have been modified since installation, and have complex interiors, which often and inadvertently pose difficulties when accessing certain zones and identifying waste remotely. Hence, this may require expert input to feature match objects. The third stage, radiological mapping, is similar in order to facilitate the characterization of the nuclear cell in terms of radiation fields, including the type of radiation, activity, and location within the nuclear cell. The fourth stage of the framework takes the visual map for stage 1, the object characterization from stage 2, and radiation map from stage 3 and fuses them together, providing a more detailed scene of the nuclear cell by identifying the location of radioactive materials in three dimensions. The last stage involves combining the evidence from the fused data sets to reveal the classification of the waste in Bq/kg, thus enabling better decision making and monitoring for in-cell decommissioning. The presentation of the framework is supported by representative case study data drawn from an application in decommissioning from a UK nuclear facility. This framework utilises recent advancements of the detection and mapping capabilities of complex radiation fields in three dimensions to make the process of classifying nuclear waste faster, more reliable, cost-effective and safer.Keywords: nuclear decommissioning, radiation detection, object detection, waste classification
Procedia PDF Downloads 200181 Functional Outcome of Speech, Voice and Swallowing Following Excision of Glomus Jugulare Tumor
Authors: B. S. Premalatha, Kausalya Sahani
Abstract:
Background: Glomus jugulare tumors arise within the jugular foramen and are commonly seen in females particularly on the left side. Surgical excision of the tumor may cause lower cranial nerve deficits. Cranial nerve involvement produces hoarseness of voice, slurred speech, and dysphagia along with other physical symptoms, thereby affecting the quality of life of individuals. Though oncological clearance is mainly emphasized on while treating these individuals, little importance is given to their communication, voice and swallowing problems, which play a crucial part in daily functioning. Objective: To examine the functions of voice, speech and swallowing outcomes of the subjects, following excision of glomus jugulare tumor. Methods: Two female subjects aged 56 and 62 years had come with a complaint of change in voice, inability to swallow and reduced clarity of speech following surgery for left glomus jugulare tumor were participants of the study. Their surgical information revealed multiple cranial nerve palsies involving the left facial, left superior and recurrent branches of the vagus nerve, left pharyngeal, left soft palate, left hypoglossal and vestibular nerves. Functional outcomes of voice, speech and swallowing were evaluated by perceptual and objective assessment procedures. Assessment included the examination of oral structures and functions, dysarthria by Frenchey dysarthria assessment, cranial nerve functions and swallowing functions. MDVP and Dr. Speech software were used to evaluate acoustic parameters of voice and quality of voice respectively. Results: The study revealed that both the subjects, subsequent to excision of glomus jugulare tumor, showed a varied picture of affected oral structure and functions, articulation, voice and swallowing functions. The cranial nerve assessment showed impairment of the vagus, hypoglossal, facial and glossopharyngeal nerves. Voice examination indicated vocal cord paralysis associated with breathy quality of voice, weak voluntary cough, reduced pitch and loudness range, and poor respiratory support. Perturbation parameters as jitter, shimmer were affected along with s/z ratio indicative of voice fold pathology. Reduced MPD(Maximum Phonation Duration) of vowels indicated that disturbed coordination between respiratory and laryngeal systems. Hypernasality was found to be a prominent feature which reduced speech intelligibility. Imprecise articulation was seen in both the subjects as the hypoglossal nerve was affected following surgery. Injury to vagus, hypoglossal, gloss pharyngeal and facial nerves disturbed the function of swallowing. All the phases of swallow were affected. Aspiration was observed before and during the swallow, confirming the oropharyngeal dysphagia. All the subsystems were affected as per Frenchey Dysarthria Assessment signifying the diagnosis of flaccid dysarthria. Conclusion: There is an observable communication and swallowing difficulty seen following excision of glomus jugulare tumor. Even with complete resection, extensive rehabilitation may be necessary due to significant lower cranial nerve dysfunction. The finding of the present study stresses the need for involvement of as speech and swallowing therapist for pre-operative counseling and assessment of functional outcomes.Keywords: functional outcome, glomus jugulare tumor excision, multiple cranial nerve impairment, speech and swallowing
Procedia PDF Downloads 252180 Transition Metal Bis(Dicarbollide) Complexes in Design of Molecular Switches
Authors: Igor B. Sivaev
Abstract:
Design of molecular machines is an extraordinary growing and very important area of research that it was recognized by awarding Sauvage, Stoddart and Feringa the Nobel Prize in Chemistry in 2016 'for the design and synthesis of molecular machines'. Based on the type of motion being performed, molecular machines can be divided into two main types: molecular motors and molecular switches. Molecular switches are molecules or supramolecular complexes having bistability, i.e., the ability to exist in two or more stable forms, among which may be reversible transitions under external influence (heating, lighting, changing the medium acidity, the action of chemicals, exposure to magnetic or electric field). Molecular switches are the main structural element of any molecular electronics devices. Therefore, the design and the study of molecules and supramolecular systems capable of performing mechanical movement is an important and urgent problem of modern chemistry. There is growing interest in molecular switches and other devices of molecular electronics based on transition metal complexes; therefore choice of suitable stable organometallic unit is of great importance. An example of such unit is bis(dicarbollide) complexes of transition metals [3,3’-M(1,2-C₂B₉H₁₁)₂]ⁿ⁻. The control on the ligand rotation in such complexes can be reached by introducing substituents which could provide stabilization of certain rotamers due to specific interactions between the ligands, on the one hand, and which can participate as Lewis bases in complex formation with external metals resulting in a change in the rotation angle of the ligands, on the other hand. A series of isomeric methyl sulfide derivatives of cobalt bis(dicarbollide) complexes containing methyl sulfide substituents at boron atoms in different positions of the pentagonal face of the dicarbollide ligands [8,8’-(MeS)₂-3,3’-Co(1,2-C₂B₉H₁₀)₂]⁻, rac-[4,4’-(MeS)₂-3,3’-Co(1,2-C₂B₉H₁₀)₂]⁻ and meso-[4,7’-(MeS)₂-3,3’-Co(1,2-C₂B₉H₁₀)₂]⁻ were synthesized by the reaction of CoCl₂ with the corresponding methyl sulfide carborane derivatives [10-MeS-7,8-C₂B₉H₁₁)₂]⁻ and [10-MeS-7,8-C₂B₉H₁₁)₂]⁻. In the case of asymmetrically substituted cobalt bis(dicarbollide) complexes the corresponding rac- and meso-isomers were successfully separated by column chromatography as the tetrabutylammonium salts. The compounds obtained were studied by the methods of ¹H, ¹³C, and ¹¹B NMR spectroscopy, single crystal X-ray diffraction, cyclic voltammetry, controlled potential coulometry and quantum chemical calculations. It was found that in the solid state, the transoid- and gauche-conformations of the 8,8’- and 4,4’-isomers are stabilized by four intramolecular CH···S(Me)B hydrogen bonds each one (2.683-2.712 Å and 2.709-2.752 Å, respectively), whereas gauche-conformation of the 4,7’-isomer is stabilized by two intramolecular CH···S hydrogen bonds (2.699-2.711 Å). The existence of the intramolecular CH·S(Me)B hydrogen bonding in solutions was supported by the 1H NMR spectroscopy. These data are in a good agreement with results of the quantum chemical calculations. The corresponding iron and nickel complexes were synthesized as well. The reaction of the methyl sulfide derivatives of cobalt bis(dicarbollide) with various labile transition metal complexes results in rupture of intramolecular hydrogen bonds and complexation of the methyl sulfide groups with external metal. This results in stabilization of other rotational conformation of cobalt bis(dicarbollide) and can be used in design of molecular switches. This work was supported by the Russian Science Foundation (16-13-10331).Keywords: molecular switches, NMR spectroscopy, single crystal X-ray diffraction, transition metal bis(dicarbollide) complexes, quantum chemical calculations
Procedia PDF Downloads 172179 International Collaboration: Developing the Practice of Social Work Curriculum through Study Abroad and Participatory Research
Authors: Megan Lindsey
Abstract:
Background: Globalization presents international social work with both opportunities and challenges. Thus, the design of this international experience aligns with the three charges of the Commission on Global Social Work Education. An international collaborative effort between an American and Scottish University Social Work Program was based on an established University agreement. The presentation provides an overview of an international study abroad among American and Scottish Social Work students. Further, presenters will discuss the opportunities of international collaboration and the challenges of the project. First, we will discuss the process of a successful international collaboration. This discussion will include the planning, collaboration, execution of the experience, along with its application to the international field of social work. Second, we will discuss the development and implementation of participatory action research in which the student engage to enhance their learning experience. A collaborative qualitative research project was undertaken with three goals. First, students gained experience in Scottish social services, including agency visits and presentations. Second, a collaboration between American and Scottish MSW Students allowed the exchange of ideas and knowledge about services and social work education. Third, students collaborated on a qualitative research method to reflect on their social work education and the formation of their professional identity. Methods/Methodology: American and Scottish students engaged in participatory action research by using Photovoice methods while studying together in Scotland. The collaboration between faculty researchers framed a series of research questions. Both universities obtained IRB approval and trained students in Photovoice methods. The student teams used the research question and Photovoice method to discover images that represented their professional identity formation. Two Photovoice goals grounded the study's research question. First, the methods enabled the individual students to record and reflect on their professional strengths and concerns. Second, student teams promoted critical dialogue and knowledge about personal and professional issues through large and small group discussions of photographs. Results: The international participatory approach generated the ability for students to contextualize their common social work education and practice experiences. Team discussions between representatives of each country resulted in understanding professional identity formation and the processes of social work education that contribute to that identity. Students presented the photograph narration of their knowledge and understanding of international social work education and practice. Researchers then collaborated on finding common themes. The results found commonalities in the quality and depth of social work education. The themes found differences regarding how professional identity is formed. Students found great differences between their and American accreditation and certification. Conclusions: Faculty researchers’ collaboration themes sought to categorize the students’ experiences of their professional identity. While the social work education systems are similar, there are vast differences. The Scottish themes noted structures within American social work not found in the United Kingdom. The American researchers noted that Scotland, as does the United Kingdom, relies on programs, agencies, and the individual social worker to provide structure to identity formation. Other themes will be presented.Keywords: higher education curriculum, international collaboration, social sciences, action research
Procedia PDF Downloads 126178 Identification of Failures Occurring on a System on Chip Exposed to a Neutron Beam for Safety Applications
Authors: S. Thomet, S. De-Paoli, F. Ghaffari, J. M. Daveau, P. Roche, O. Romain
Abstract:
In this paper, we present a hardware module dedicated to understanding the fail reason of a System on Chip (SoC) exposed to a particle beam. Impact of Single-Event Effects (SEE) on processor-based SoCs is a concern that has increased in the past decade, particularly for terrestrial applications with automotive safety increasing requirements, as well as consumer and industrial domains. The SEE created by the impact of a particle on an SoC may have consequences that can end to instability or crashes. Specific hardening techniques for hardware and software have been developed to make such systems more reliable. SoC is then qualified using cosmic ray Accelerated Soft-Error Rate (ASER) to ensure the Soft-Error Rate (SER) remains in mission profiles. Understanding where errors are occurring is another challenge because of the complexity of operations performed in an SoC. Common techniques to monitor an SoC running under a beam are based on non-intrusive debug, consisting of recording the program counter and doing some consistency checking on the fly. To detect and understand SEE, we have developed a module embedded within the SoC that provide support for recording probes, hardware watchpoints, and a memory mapped register bank dedicated to software usage. To identify CPU failure modes and the most important resources to probe, we have carried out a fault injection campaign on the RTL model of the SoC. Probes are placed on generic CPU registers and bus accesses. They highlight the propagation of errors and allow identifying the failure modes. Typical resulting errors are bit-flips in resources creating bad addresses, illegal instructions, longer than expected loops, or incorrect bus accesses. Although our module is processor agnostic, it has been interfaced to a RISC-V by probing some of the processor registers. Probes are then recorded in a ring buffer. Associated hardware watchpoints are allowing to do some control, such as start or stop event recording or halt the processor. Finally, the module is also providing a bank of registers where the firmware running on the SoC can log information. Typical usage is for operating system context switch recording. The module is connected to a dedicated debug bus and is interfaced to a remote controller via a debugger link. Thus, a remote controller can interact with the monitoring module without any intrusiveness on the SoC. Moreover, in case of CPU unresponsiveness, or system-bus stall, the recorded information can still be recovered, providing the fail reason. A preliminary version of the module has been integrated into a test chip currently being manufactured at ST in 28-nm FDSOI technology. The module has been triplicated to provide reliable information on the SoC behavior. As the primary application domain is automotive and safety, the efficiency of the module will be evaluated by exposing the test chip under a fast-neutron beam by the end of the year. In the meantime, it will be tested with alpha particles and electromagnetic fault injection (EMFI). We will report in the paper on fault-injection results as well as irradiation results.Keywords: fault injection, SoC fail reason, SoC soft error rate, terrestrial application
Procedia PDF Downloads 229