Search results for: supply chain design and performance
359 Assessment of Efficiency of Underwater Undulatory Swimming Strategies Using a Two-Dimensional CFD Method
Authors: Dorian Audot, Isobel Margaret Thompson, Dominic Hudson, Joseph Banks, Martin Warner
Abstract:
In competitive swimming, after dives and turns, athletes perform underwater undulatory swimming (UUS), copying marine mammals’ method of locomotion. The body, performing this wave-like motion, accelerates the fluid downstream in its vicinity, generating propulsion with minimal resistance. Through this technique, swimmers can maintain greater speeds than surface swimming and take advantage of the overspeed granted by the dive (or push-off). Almost all previous work has considered UUS when performed at maximum effort. Critical parameters to maximize UUS speed are frequently discussed; however, this does not apply to most races. In only 3 out of the 16 individual competitive swimming events are athletes likely to attempt to perform UUS with the greatest speed, without thinking of the cost of locomotion. In the other cases, athletes will want to control the speed of their underwater swimming, attempting to maximise speed whilst considering energy expenditure appropriate to the duration of the event. Hence, there is a need to understand how swimmers adapt their underwater strategies to optimize the speed within the allocated energetic cost. This paper develops a consistent methodology that enables different sets of UUS kinematics to be investigated. These may have different propulsive efficiencies and force generation mechanisms (e.g.: force distribution along with the body and force magnitude). The developed methodology, therefore, needs to: (i) provide an understanding of the UUS propulsive mechanisms at different speeds, (ii) investigate the key performance parameters when UUS is not performed solely for maximizing speed; (iii) consistently determine the propulsive efficiency of a UUS technique. The methodology is separated into two distinct parts: kinematic data acquisition and computational fluid dynamics (CFD) analysis. For the kinematic acquisition, the position of several joints along the body and their sequencing were either obtained by video digitization or by underwater motion capture (Qualisys system). During data acquisition, the swimmers were asked to perform UUS at a constant depth in a prone position (facing the bottom of the pool) at different speeds: maximum effort, 100m pace, 200m pace and 400m pace. The kinematic data were input to a CFD algorithm employing a two-dimensional Large Eddy Simulation (LES). The algorithm adopted was specifically developed in order to perform quick unsteady simulations of deforming bodies and is therefore suitable for swimmers performing UUS. Despite its approximations, the algorithm is applied such that simulations are performed with the inflow velocity updated at every time step. It also enables calculations of the resistive forces (total and applied to each segment) and the power input of the modeled swimmer. Validation of the methodology is achieved by comparing the data obtained from the computations with the original data (e.g.: sustained swimming speed). This method is applied to the different kinematic datasets and provides data on swimmers’ natural responses to pacing instructions. The results show how kinematics affect force generation mechanisms and hence how the propulsive efficiency of UUS varies for different race strategies.Keywords: CFD, efficiency, human swimming, hydrodynamics, underwater undulatory swimming
Procedia PDF Downloads 219358 Knowledge and Attitude Towards Strabismus Among Adult Residents in Woreta Town, Northwest Ethiopia: A Community-Based Study
Authors: Henok Biruk Alemayehu, Kalkidan Berhane Tsegaye, Fozia Seid Ali, Nebiyat Feleke Adimassu, Getasew Alemu Mersha
Abstract:
Background: Strabismus is a visual disorder where the eyes are misaligned and point in different directions. Untreated strabismus can lead to amblyopia, loss of binocular vision, and social stigma due to its appearance. Since it is assumed that knowledge is pertinent for early screening and prevention of strabismus, the main objective of this study was to assess knowledge and attitudes toward strabismus in Woreta town, Northwest Ethiopia. Providing data in this area is important for planning health policies. Methods: A community-based cross-sectional study was done in Woreta town from April–May 2020. The sample size was determined using a single population proportion formula by taking a 50% proportion of good knowledge, 95% confidence level, 5% margin of errors, and 10% non- response rate. Accordingly, the final computed sample size was 424. All four kebeles were included in the study. There were 42,595 people in total, with 39,684 adults and 9229 house holds. A sample fraction ’’k’’ was obtained by dividing the number of the household by the calculated sample size of 424. Systematic random sampling with proportional allocation was used to select the participating households with a sampling fraction (K) of 21 i.e. each household was approached in every 21 households included in the study. One individual was selected ran- domly from each household with more than one adult, using the lottery method to obtain a final sample size. The data was collected through a face-to-face interview with a pretested and semi-structured questionnaire which was translated from English to Amharic and back to English to maintain its consistency. Data were entered using epi-data version 3.1, then processed and analyzed via SPSS version- 20. Descriptive and analytical statistics were employed to summarize the data. A p-value of less than 0.05 was used to declare statistical significance. Result: A total of 401 individuals aged over 18 years participated, with a response rate of 94.5%. Of those who responded, 56.6% were males. Of all the participants, 36.9% were illiterate. The proportion of people with poor knowledge of strabismus was 45.1%. It was shown that 53.9% of the respondents had a favorable attitude. Older age, higher educational level, having a history of eye examination, and a having a family history of strabismus were significantly associated with good knowledge of strabismus. A higher educational level, older age, and hearing about strabismus were significantly associated with a favorable attitude toward strabismus. Conclusion and recommendation: The proportion of good knowledge and favorable attitude towards strabismus were lower than previously reported in Gondar City, Northwest Ethiopia. There is a need to provide health education and promotion campaigns on strabismus to the community: what strabismus is, its’ possible treatments and the need to bring children to the eye care center for early diagnosis and treatment. it advocate for prospective research endeavors to employ qualitative study design.Additionally, it suggest the exploration of studies that investigate causal-effect relationship.Keywords: strabismus, knowledge, attitude, Woreta
Procedia PDF Downloads 62357 Small Town Big Urban Issues the Case of Kiryat Ono, Israel
Authors: Ruth Shapira
Abstract:
Introduction: The rapid urbanization of the last century confronts planners, regulatory bodies, developers and most of all – the public with seemingly unsolved conflicts regarding values, capital, and wellbeing of the built and un-built urban space. This is reflected in the quality of the urban form and life which has known no significant progress in the last 2-3 decades despite the on-growing urban population. It is the objective of this paper to analyze some of these fundamental issues through the case study of a relatively small town in the center of Israel (Kiryat-Ono, 100,000 inhabitants), unfold the deep structure of qualities versus disruptors, present some cure that we have developed to bridge over and humbly suggest a practice that may be generic for similar cases. Basic Methodologies: The OBJECT, the town of Kiryat Ono, shall be experimented upon in a series of four action processes: De-composition, Re-composition, the Centering process and, finally, Controlled Structural Disintegration. Each stage will be based on facts, analysis of previous multidisciplinary interventions on various layers – and the inevitable reaction of the OBJECT, leading to the conclusion based on innovative theoretical and practical methods that we have developed and that we believe are proper for the open ended network, setting the rules for the contemporary urban society to cluster by. The Study: Kiryat Ono, was founded 70 years ago as an agricultural settlement and rapidly turned into an urban entity. In spite the massive intensification, the original DNA of the old small town was still deeply embedded, mostly in the quality of the public space and in the sense of clustered communities. In the past 20 years, the recent demand for housing has been addressed to on the national level with recent master plans and urban regeneration policies mostly encouraging individual economic initiatives. Unfortunately, due to the obsolete existing planning platform the present urban renewal is characterized by pressure of developers, a dramatic change in building scale and widespread disintegration of the existing urban and social tissue. Our office was commissioned to conceptualize two master plans for the two contradictory processes of Kiryat Ono’s future: intensification and conservation. Following a comprehensive investigation into the deep structures and qualities of the existing town, we developed a new vocabulary of conservation terms thus redefying the sense of PLACE. The main challenge was to create master plans that should offer a regulatory basis to the accelerated and sporadic development providing for the public good and preserving the characteristics of the PLACE consisting of a tool box of design guidelines that will have the ability to reorganize space along the time axis in a coherent way. In Conclusion: The system of rules that we have developed can generate endless possible patterns making sure that at each implementation fragment an event is created, and a better place is revealed. It takes time and perseverance but it seems to be the way to provide a healthy framework for the accelerated urbanization of our chaotic present.Keywords: housing, architecture, urban qualities, urban regeneration, conservation, intensification
Procedia PDF Downloads 361356 Identification of Odorant Receptors through the Antennal Transcriptome of the Grapevine Pest, Lobesia botrana (Lepidoptera: Tortricidae)
Authors: Ricardo Godoy, Herbert Venthur, Hector Jimenez, Andres Quiroz, Ana Mutis
Abstract:
In agriculture, grape production has great economic importance at global level, considering that in 2013 it reached 7.4 million hectares (ha) covered by plantations of this fruit worldwide. Chile is the number one exporter in the world with 800,000 tons. However, these values have been threatened by the attack of the grapevine moth, Lobesia botrana (Denis & Schiffermuller) (Lepidoptera: Tortricidae), since its detection in 2008. Nowadays, the use of semiochemicals, in particular the major component of the sex pheromone, (E,Z)-7.9-dodecadienil acetate, are part of mating disruption methods to control L. botrana. How insect pests can recognize these molecules, is being part of huge efforts to deorphanize their olfactory mechanism at molecular level. Thus, an interesting group of proteins has been identified in the antennae of insects, where odorant-binding proteins (OBPs) are known by transporting molecules to odorant receptors (ORs) and a co-receptor (ORCO) causing a behavioral change in the insect. Other proteins such as chemosensory proteins (CSPs), ionotropic receptors (IRs), odorant degrading enzymes (ODEs) and sensory neuron membrane proteins (SNMPs) seem to be involved, but few studies have been performed so far. The above has led to an increasing interest in insect communication at a molecular level, which has contributed to both a better understanding of the olfaction process and the design of new pest management strategies. To date, it has been reported that the ORs can detect one or a small group of odorants in a specific way. Therefore, the objective of this study is the identification of genes that encode these ORs using the antennal transcriptome of L. botrana. Total RNA was extracted for females and males of L. botrana, and the antennal transcriptome sequenced by Next Generation Sequencing service using an Illumina HiSeq2500 platform with 50 million reads per sample. Unigenes were assembled using Trinity v2.4.0 package and transcript abundance was obtained using edgeR. Genes were identified using BLASTN and BLASTX locally installed in a Unix system and based on our own Tortricidae database. Those Unigenes related to ORs were characterized using ORFfinder and protein Blastp server. Finally, a phylogenetic analysis was performed with the candidate amino acid sequences for LbotORs including amino acid sequences of other moths ORs, such as Bombyx mori, Cydia pomonella, among others. Our findings suggest 61 genes encoding ORs and one gene encoding an ORCO in both sexes, where the greatest difference was found in the OR6 because of the transcript abundance according to the value of FPKM in females and males was 1.48 versus 324.00. In addition, according to phylogenetic analysis OR6 is closely related to OR1 in Cydia pomonella and OR6, OR7 in Epiphyas postvittana, which have been described as pheromonal receptors (PRs). These results represent the first evidence of ORs present in the antennae of L. botrana and a suitable starting point for further functional studies with selected ORs, such as OR6, which is potentially related to pheromonal recognition.Keywords: antennal transcriptome, lobesia botrana, odorant receptors (ORs), phylogenetic analysis
Procedia PDF Downloads 200355 Design of an Automated Deep Learning Recurrent Neural Networks System Integrated with IoT for Anomaly Detection in Residential Electric Vehicle Charging in Smart Cities
Authors: Wanchalerm Patanacharoenwong, Panaya Sudta, Prachya Bumrungkun
Abstract:
The paper focuses on the development of a system that combines Internet of Things (IoT) technologies and deep learning algorithms for anomaly detection in residential Electric Vehicle (EV) charging in smart cities. With the increasing number of EVs, ensuring efficient and reliable charging systems has become crucial. The aim of this research is to develop an integrated IoT and deep learning system for detecting anomalies in residential EV charging and enhancing EV load profiling and event detection in smart cities. This approach utilizes IoT devices equipped with infrared cameras to collect thermal images and household EV charging profiles from the database of Thailand utility, subsequently transmitting this data to a cloud database for comprehensive analysis. The methodology includes the use of advanced deep learning techniques such as Recurrent Neural Networks (RNN) and Long Short-Term Memory (LSTM) algorithms. IoT devices equipped with infrared cameras are used to collect thermal images and EV charging profiles. The data is transmitted to a cloud database for comprehensive analysis. The researchers also utilize feature-based Gaussian mixture models for EV load profiling and event detection. Moreover, the research findings demonstrate the effectiveness of the developed system in detecting anomalies and critical profiles in EV charging behavior. The system provides timely alarms to users regarding potential issues and categorizes the severity of detected problems based on a health index for each charging device. The system also outperforms existing models in event detection accuracy. This research contributes to the field by showcasing the potential of integrating IoT and deep learning techniques in managing residential EV charging in smart cities. The system ensures operational safety and efficiency while also promoting sustainable energy management. The data is collected using IoT devices equipped with infrared cameras and is stored in a cloud database for analysis. The collected data is then analyzed using RNN, LSTM, and feature-based Gaussian mixture models. The approach includes both EV load profiling and event detection, utilizing a feature-based Gaussian mixture model. This comprehensive method aids in identifying unique power consumption patterns among EV owners and outperforms existing models in event detection accuracy. In summary, the research concludes that integrating IoT and deep learning techniques can effectively detect anomalies in residential EV charging and enhance EV load profiling and event detection accuracy. The developed system ensures operational safety and efficiency, contributing to sustainable energy management in smart cities.Keywords: cloud computing framework, recurrent neural networks, long short-term memory, Iot, EV charging, smart grids
Procedia PDF Downloads 64354 Participatory Monitoring Strategy to Address Stakeholder Engagement Impact in Co-creation of NBS Related Project: The OPERANDUM Case
Authors: Teresa Carlone, Matteo Mannocchi
Abstract:
In the last decade, a growing number of International Organizations are pushing toward green solutions for adaptation to climate change. This is particularly true in the field of Disaster Risk Reduction (DRR) and land planning, where Nature-Based Solutions (NBS) had been sponsored through funding programs and planning tools. Stakeholder engagement and co-creation of NBS is growing as a practice and research field in environmental projects, fostering the consolidation of a multidisciplinary socio-ecological approach in addressing hydro-meteorological risk. Even thou research and financial interests are constantly spread, the NBS mainstreaming process is still at an early stage as innovative concepts and practices make it difficult to be fully accepted and adopted by a multitude of different actors to produce wide scale societal change. The monitoring and impact evaluation of stakeholders’ participation in these processes represent a crucial aspect and should be seen as a continuous and integral element of the co-creation approach. However, setting up a fit for purpose-monitoring strategy for different contexts is not an easy task, and multiple challenges emerge. In this scenario, the Horizon 2020 OPERANDUM project, designed to address the major hydro-meteorological risks that negatively affect European rural and natural territories through the co-design, co-deployment, and assessment of Nature-based Solution, represents a valid case study to test a monitoring strategy from which set a broader, general and scalable monitoring framework. Applying a participative monitoring methodology, based on selected indicators list that combines quantitative and qualitative data developed within the activity of the project, the paper proposes an experimental in-depth analysis of the stakeholder engagement impact in the co-creation process of NBS. The main focus will be to spot and analyze which factors increase knowledge, social acceptance, and mainstreaming of NBS, promoting also a base-experience guideline to could be integrated with the stakeholder engagement strategy in current and future similar strongly collaborative approach-based environmental projects, such as OPERANDUM. Measurement will be carried out through survey submitted at a different timescale to the same sample (stakeholder: policy makers, business, researchers, interest groups). Changes will be recorded and analyzed through focus groups in order to highlight causal explanation and to assess the proposed list of indicators to steer the conduction of similar activities in other projects and/or contexts. The idea of the paper is to contribute to the construction of a more structured and shared corpus of indicators that can support the evaluation of the activities of involvement and participation of various levels of stakeholders in the co-production, planning, and implementation of NBS to address climate change challenges.Keywords: co-creation and collaborative planning, monitoring, nature-based solution, participation & inclusion, stakeholder engagement
Procedia PDF Downloads 114353 Innovation Outputs from Higher Education Institutions: A Case Study of the University of Waterloo, Canada
Authors: Wendy De Gomez
Abstract:
The University of Waterloo is situated in central Canada in the Province of Ontario- one hour from the metropolitan city of Toronto. For over 30 years, it has held Canada’s top spot as the most innovative university; and has been consistently ranked in the top 25 computer science and top 50 engineering schools in the world. Waterloo benefits from the federal government’s over 100 domestic innovation policies which have assisted in the country’s 15th place global ranking in the World Intellectual Property Organization’s (WIPO) 2022 Global Innovation Index. Yet undoubtedly, the University of Waterloo’s unique characteristics are what propels its innovative creativeness forward. This paper will provide a contextual definition of innovation in higher education and then demonstrate the five operational attributes that contribute to the University of Waterloo’s innovative reputation. The methodology is based on statistical analyses obtained from ranking bodies such as the QS World University Rankings, a secondary literature review related to higher education innovation in Canada, and case studies that exhibit the operationalization of the attributes outlined below. The first attribute is geography. Specifically, the paper investigates the network structure effect of the Toronto-Waterloo high-tech corridor and the resultant industrial relationships built there. The second attribute is University Policy 73-Intellectal Property Rights. This creator-owned policy grants all ownership to the creator/inventor regardless of the use of the University of Waterloo property or funding. Essentially, through the incentivization of IP ownership by all researchers, further commercialization and entrepreneurship are formed. Third, this IP policy works hand in hand with world-renowned business incubators such as the Accelerator Centre in the dedicated research and technology park and velocity, a 14-year-old facility that equips and guides founders to build and scale companies. Communitech, a 25-year-old provincially backed facility in the region, also works closely with the University of Waterloo to build strong teams, access capital, and commercialize products. Fourth, Waterloo’s co-operative education program contributes 31% of all co-op participants to the Canadian economy. Home to the world’s largest co-operative education program, data shows that over 7,000 from around the world recruit Waterloo students for short- and long-term placements- directly contributing to the student’s ability to learn and optimize essential employment skills when they graduate. Finally, the students themselves at Waterloo are exceptional. The entrance average ranges from the low 80s to the mid-90s depending on the program. In computer, electrical, mechanical, mechatronics, and systems design engineering, to have a 66% chance of acceptance, the applicant’s average must be 95% or above. Singularly, none of these five attributes could lead to the university’s outstanding track record of innovative creativity, but when bundled up into a 1000 acre- 100 building main campus with 6 academic faculties, 40,000+ students, and over 1300 world-class faculty, the recipe for success becomes quite evident.Keywords: IP policy, higher education, economy, innovation
Procedia PDF Downloads 70352 Effects of Live Webcast-Assisted Teaching on Physical Assessment Technique Learning of Young Nursing Majors
Authors: Huey-Yeu Yan, Ching-Ying Lee, Hung-Ru Lin
Abstract:
Background: Physical assessment is a vital clinical nursing competence. The gap between conventional teaching method and the way e-generation students’ preferred could be bridged owing to the support of Internet technology, i.e. interacting with online media to manage learning works. Nursing instructors in the wake of new learning pattern of the e-generation students are challenged to actively adjust and make teaching contents and methods more versatile. Objective: The objective of this research is to explore the effects on teaching and learning with live webcast-assisted on a specific topic, Physical Assessment technique, on a designated group of young nursing majors. It’s hoped that, with a way of nursing instructing, more versatile learning resources may be provided to facilitate self-directed learning. Design: This research adopts a cross-sectional descriptive survey. The instructor demonstrated physical assessment techniques and operation procedures via live webcast broadcasted online to all students. It increased both the off-time interaction between teacher and students concerning teaching materials. Methods: A convenient sampling was used to recruit a total of 52 nursing-majors at a certain university. The nursing majors took two-hour classes of Physical Assessment per week for 18 weeks (36 hrs. in total). The instruction covered four units with live webcasting and then conducted an online anonymous survey of learning outcomes by questionnaire. The research instrument was the online questionnaire, covering three major domains—online media used, learning outcome evaluation and evaluation result. The data analysis was conducted via IBM SPSS Statistics Version 2.0. The descriptive statistics was undertaken to describe the analysis of basic data and learning outcomes. Statistical methods such as descriptive statistics, t-test, ANOVA, and Pearson’s correlation were employed in verification. Results: Results indicated the following five major findings. (1) learning motivation, about four fifth of the participants agreed the online instruction resources are very helpful in improving learning motivation and raising the learning interest. (2) learning needs, about four fifth of participants agreed it was helpful to plan self-directed practice after the instruction, and meet their needs of repetitive learning and/or practice at their leisure time. (3) learning effectiveness, about two third agreed it was helpful to reduce pre-exam anxiety, and improve their test scores. (4) course objects, about three fourth agreed that it was helpful to achieve the goal of ‘executing the complete Physical Assessment procedures with proper skills’. (5) finally, learning reflection, about all of participants agreed this experience of online instructing, learning, and practicing is beneficial to them, they recommend instructor to share with other nursing majors, and they will recommend it to fellow students too. Conclusions: Live webcasting is a low-cost, convenient, efficient and interactive resource to facilitate nursing majors’ motivation of learning, need of self-directed learning and practice, outcome of learning. When live webcasting is integrated into nursing teaching, it provides an opportunity of self-directed learning to promote learning effectiveness, as such to fulfill the teaching objective.Keywords: innovative teaching, learning effectiveness, live webcasting, physical assessment technique
Procedia PDF Downloads 132351 Balloon Analogue Risk Task (BART) Performance Indicators Help Predict Outcomes of Matched Savings Program
Authors: Carlos M. Parra, Matthew Sutherland, Ranjita Poudel
Abstract:
Reduced mental-bandwidth related to low socioeconomic status (low-SES) might lead to impulsivity and risk-taking behavior, which poses as a major hurdle towards asset building (savings) behavior. Understanding the relationship between risk-related personality metrics as well as laboratory risk behavior and real-life savings behavior can help facilitate the development of effective asset building programs, which are vital for mitigating financial vulnerability and income inequality. As such, this study explored the relationship between personality metrics, laboratory behavior in a risky decision-making task and real-life asset building (savings) behaviors among individuals with low-SES from Miami, Florida (FL). Study participants (12 male, 15 female) included racially and ethnically diverse adults (mean age 41.22 ± 12.65 years), with incomplete higher education (18% had High School Diploma, 30% Associates, and 52% Some College), and low annual income (mean $13,872 ± $8020.43). Participants completed eight self-report surveys and played a widely used risky decision-making paradigm called the Balloon Analogue Risk Task (BART). Specifically, participants played three runs of BART (20 trials in each run; total 60 trials). In addition, asset building behavior data was collected for 24 participants who opened and used savings accounts and completed a 6-month savings program that involved monthly matches, and a final reward for completing the savings program without any interim withdrawals. Each participant’s total savings at the end of this program was the main asset building indicator considered. In addition, a new effective use of average pump bet (EUAPB) indicator was developed to characterize each participant’s ability to place winning bets. This indicator takes the ratio of each participant’s total BART earnings to average pump bet (APB) in all 60 trials. Our findings indicated that EUAPB explained more than a third of the variation in total savings among participants. Moreover, participants who managed to obtain BART earnings of at least 30 cents out of their APB, also tended to exhibit better asset building (savings) behavior. In particular, using this criterion to separate participants into high and low EUAPB groups, the nine participants with high EUAPB (mean BART earnings of 35.64 cents per APB) ended up with higher mean total savings ($255.11), while the 15 participants with low EUAPB (mean BART earnings of 22.50 cents per APB) obtained lower mean total savings ($40.01). All mean differences are statistically significant (2-tailed p .0001) indicating that the relation between higher EUAPB and higher total savings is robust. Overall, these findings can help refine asset building interventions implemented by policy makers and practitioners interested in reducing financial vulnerability among low-SES population. Specifically, by helping identify individuals who are likely to readily take advantage of savings opportunities (such as matched savings programs) and avoiding the stipulation of unnecessary and expensive financial coaching programs to these individuals. This study was funded by J.P. Morgan Chase (JPMC) and carried out by scientists from Florida International University (FIU) in partnership with Catalyst Miami.Keywords: balloon analogue risk task (BART), matched savings programs, asset building capability, low-SES participants
Procedia PDF Downloads 145350 Examining Influence of The Ultrasonic Power and Frequency on Microbubbles Dynamics Using Real-Time Visualization of Synchrotron X-Ray Imaging: Application to Membrane Fouling Control
Authors: Masoume Ehsani, Ning Zhu, Huu Doan, Ali Lohi, Amira Abdelrasoul
Abstract:
Membrane fouling poses severe challenges in membrane-based wastewater treatment applications. Ultrasound (US) has been considered an effective fouling remediation technique in filtration processes. Bubble cavitation in the liquid medium results from the alternating rarefaction and compression cycles during the US irradiation at sufficiently high acoustic pressure. Cavitation microbubbles generated under US irradiation can cause eddy current and turbulent flow within the medium by either oscillating or discharging energy to the system through microbubble explosion. Turbulent flow regime and shear forces created close to the membrane surface cause disturbing the cake layer and dislodging the foulants, which in turn improve the cleaning efficiency and filtration performance. Therefore, the number, size, velocity, and oscillation pattern of the microbubbles created in the liquid medium play a crucial role in foulant detachment and permeate flux recovery. The goal of the current study is to gain in depth understanding of the influence of the US power intensity and frequency on the microbubble dynamics and its characteristics generated under US irradiation. In comparison with other imaging techniques, the synchrotron in-line Phase Contrast Imaging technique at the Canadian Light Source (CLS) allows in-situ observation and real-time visualization of microbubble dynamics. At CLS biomedical imaging and therapy (BMIT) polychromatic beamline, the effective parameters were optimized to enhance the contrast gas/liquid interface for the accuracy of the qualitative and quantitative analysis of bubble cavitation within the system. With the high flux of photons and the high-speed camera, a typical high projection speed was achieved; and each projection of microbubbles in water was captured in 0.5 ms. ImageJ software was used for post-processing the raw images for the detailed quantitative analyses of microbubbles. The imaging has been performed under the US power intensity levels of 50 W, 60 W, and 100 W, in addition to the US frequency levels of 20 kHz, 28 kHz, and 40 kHz. For the duration of 2 seconds of imaging, the effect of the US power and frequency on the average number, size, and fraction of the area occupied by bubbles were analyzed. Microbubbles’ dynamics in terms of their velocity in water was also investigated. For the US power increase of 50 W to 100 W, the average bubble number and the average bubble diameter were increased from 746 to 880 and from 36.7 µm to 48.4 µm, respectively. In terms of the influence of US frequency, a fewer number of bubbles were created at 20 kHz (average of 176 bubbles rather than 808 bubbles at 40 kHz), while the average bubble size was significantly larger than that of 40 kHz (almost seven times). The majority of bubbles were captured close to the membrane surface in the filtration unit. According to the study observations, membrane cleaning efficiency is expected to be improved at higher US power and lower US frequency due to the higher energy release to the system by increasing the number of bubbles or growing their size during oscillation (optimum condition is expected to be at 20 kHz and 100 W).Keywords: bubble dynamics, cavitational bubbles, membrane fouling, ultrasonic cleaning
Procedia PDF Downloads 149349 Investigating the Thermal Comfort Properties of Mohair Fabrics
Authors: Adine Gericke, Jiri Militky, Mohanapriya Venkataraman
Abstract:
Mohair, obtained from the Angora goat, is a luxury fiber and recognized as one of the best quality natural fibers. Expansion of the use of mohair into technical and functional textile products necessitates the need for a better understanding of how the use of mohair in fabrics will impact on its thermo-physiological comfort related properties. Despite its popularity, very little information is available on the quantification of the thermal and moisture management properties of mohair fabrics. This study investigated the effect of fibrous matter composition and fabric structural parameters on conductive and convective heat transfers to attain more information on the thermal comfort properties of mohair fabrics. Dry heat transfer through textiles may involve conduction through the fibrous phase, radiation through fabric interstices and convection of air within the structure. Factors that play a major role in heat transfer by conduction are fabric areal density (g/m2) and derived quantities such as cover factor and porosity. Convective heat transfer through fabrics is found in environmental conditions where there is wind-flow or the object is moving (e.g. running or walking). The thermal comfort properties of mohair fibers were objectively evaluated firstly in comparison with other textile fibers and secondly in a variety of fabric structures. Two sample sets were developed for this purpose, with fibre content, yarn structure and fabric design as main variables. SEM and microscopic images were obtained to closely examine the physical structures of the fibers and fabrics. Thermal comfort properties such as thermal resistance and thermal conductivity, as well as fabric thickness, were measured on the well-known Alambeta test instrument. Clothing insulation (clo) was calculated from the above. The thermal properties of fabrics under heat convection was evaluated using a laboratory model device developed at the Technical University of Liberec (referred to as the TP2-instrument). The effects of the different variables on fabric thermal comfort properties were analyzed statistically using TIBCO Statistica Software. The results showed that fabric structural properties, specifically sample thickness, played a significant role in determining the thermal comfort properties of the fabrics tested. It was found that regarding thermal resistance related to conductive heat flow, the effect of fiber type was not always statistically significant, probably as a result of the amount of trapped air within the fabric structure. The very low thermal conductivity of air, compared to that of the fibers, had a significant influence on the total conductivity and thermal resistance of the samples. This was confirmed by the high correlation of these factors with sample thickness. Regarding convective heat flow, the most important factor influencing the ability of the fabric to allow dry heat to move through the structure, was again fabric thickness. However, it would be wrong to totally disregard the effect of fiber composition on the thermal resistance of textile fabrics. In this study, the samples containing mohair or mohair/wool were consistently thicker than the others even though weaving parameters were kept constant. This can be ascribed to the physical properties of the mohair fibers that renders it exceptionally well towards trapping air among fibers (in a yarn) as well as among yarns (inside a fabric structure). The thicker structures trap more air to provide higher thermal insulation, but also prevent the free flow of air that allow thermal convection.Keywords: mohair fabrics, convective heat transfer, thermal comfort properties, thermal resistance
Procedia PDF Downloads 144348 Hydrogeomatic System for the Economic Evaluation of Damage by Flooding in Mexico
Authors: Alondra Balbuena Medina, Carlos Diaz Delgado, Aleida Yadira Vilchis Fránces
Abstract:
In Mexico, each year news is disseminated about the ravages of floods, such as the total loss of housing, damage to the fields; the increase of the costs of the food, derived from the losses of the harvests, coupled with health problems such as skin infection, etc. In addition to social problems such as delinquency, damage in education institutions and the population in general. The flooding is a consequence of heavy rains, tropical storms and or hurricanes that generate excess water in drainage systems that exceed its capacity. In urban areas, heavy rains can be one of the main factors in causing flooding, in addition to excessive precipitation, dam breakage, and human activities, for example, excessive garbage in the strainers. In agricultural areas, these can hardly achieve large areas of cultivation. It should be mentioned that for both areas, one of the significant impacts of floods is that they can permanently affect the livelihoods of many families, cause damage, for example in their workplaces such as farmlands, commercial or industry areas and where services are provided. In recent years, Information and Communication Technologies (ICT) have had an accelerated development, being reflected in the growth and the exponential evolution of the innovation giving; as a result, the daily generation of new technologies, updates, and applications. Innovation in the development of Information Technology applications has impacted on all areas of human activity. They influence all the orders of life of individuals, reconfiguring the way of perceiving and analyzing the world such as, for instance, interrelating with people as individuals and as a society, in the economic, political, social, cultural, educational, environmental, etc. Therefore the present work describes the creation of a system of calculation of flood costs for housing areas, retail establishments and agricultural areas from the Mexican Republic, based on the use and application of geotechnical tools being able to be useful for the benefit of the sectors of public, education and private. To generate analysis of hydrometereologic affections and with the obtained results to realize the Geoinformatics tool was constructed from two different points of view: the geoinformatic (design and development of GIS software) and the methodology of flood damage validation in order to integrate a tool that provides the user the monetary estimate of the effects caused by the floods. With information from the period 2000-2014, the functionality of the application was corroborated. For the years 2000 to 2009 only the analysis of the agricultural and housing areas was carried out, incorporating for the commercial establishment's information of the period 2010 - 2014. The method proposed for the resolution of this research project is a fundamental contribution to society, in addition to the tool itself. Therefore, it can be summarized that the problems that are in the physical-geographical environment, conceiving them from the point of view of the spatial analysis, allow to offer different alternatives of solution and also to open up slopes towards academia and research.Keywords: floods, technological innovation, monetary estimation, spatial analysis
Procedia PDF Downloads 224347 Modelling of Reactive Methodologies in Auto-Scaling Time-Sensitive Services With a MAPE-K Architecture
Authors: Óscar Muñoz Garrigós, José Manuel Bernabeu Aubán
Abstract:
Time-sensitive services are the base of the cloud services industry. Keeping low service saturation is essential for controlling response time. All auto-scalable services make use of reactive auto-scaling. However, reactive auto-scaling has few in-depth studies. This presentation shows a model for reactive auto-scaling methodologies with a MAPE-k architecture. Queuing theory can compute different properties of static services but lacks some parameters related to the transition between models. Our model uses queuing theory parameters to relate the transition between models. It associates MAPE-k related times, the sampling frequency, the cooldown period, the number of requests that an instance can handle per unit of time, the number of incoming requests at a time instant, and a function that describes the acceleration in the service's ability to handle more requests. This model is later used as a solution to horizontally auto-scale time-sensitive services composed of microservices, reevaluating the model’s parameters periodically to allocate resources. The solution requires limiting the acceleration of the growth in the number of incoming requests to keep a constrained response time. Business benefits determine such limits. The solution can add a dynamic number of instances and remains valid under different system sizes. The study includes performance recommendations to improve results according to the incoming load shape and business benefits. The exposed methodology is tested in a simulation. The simulator contains a load generator and a service composed of two microservices, where the frontend microservice depends on a backend microservice with a 1:1 request relation ratio. A common request takes 2.3 seconds to be computed by the service and is discarded if it takes more than 7 seconds. Both microservices contain a load balancer that assigns requests to the less loaded instance and preemptively discards requests if they are not finished in time to prevent resource saturation. When load decreases, instances with lower load are kept in the backlog where no more requests are assigned. If the load grows and an instance in the backlog is required, it returns to the running state, but if it finishes the computation of all requests and is no longer required, it is permanently deallocated. A few load patterns are required to represent the worst-case scenario for reactive systems: the following scenarios test response times, resource consumption and business costs. The first scenario is a burst-load scenario. All methodologies will discard requests if the rapidness of the burst is high enough. This scenario focuses on the number of discarded requests and the variance of the response time. The second scenario contains sudden load drops followed by bursts to observe how the methodology behaves when releasing resources that are lately required. The third scenario contains diverse growth accelerations in the number of incoming requests to observe how approaches that add a different number of instances can handle the load with less business cost. The exposed methodology is compared against a multiple threshold CPU methodology allocating/deallocating 10 or 20 instances, outperforming the competitor in all studied metrics.Keywords: reactive auto-scaling, auto-scaling, microservices, cloud computing
Procedia PDF Downloads 93346 Evaluation of Alternative Approaches for Additional Damping in Dynamic Calculations of Railway Bridges under High-Speed Traffic
Authors: Lara Bettinelli, Bernhard Glatz, Josef Fink
Abstract:
Planning engineers and researchers use various calculation models with different levels of complexity, calculation efficiency and accuracy in dynamic calculations of railway bridges under high-speed traffic. When choosing a vehicle model to depict the dynamic loading on the bridge structure caused by passing high-speed trains, different goals are pursued: On the one hand, the selected vehicle models should allow the calculation of a bridge’s vibrations as realistic as possible. On the other hand, the computational efficiency and manageability of the models should be preferably high to enable a wide range of applications. The commonly adopted and straightforward vehicle model is the moving load model (MLM), which simplifies the train to a sequence of static axle loads moving at a constant speed over the structure. However, the MLM can significantly overestimate the structure vibrations, especially when resonance events occur. More complex vehicle models, which depict the train as a system of oscillating and coupled masses, can reproduce the interaction dynamics between the vehicle and the bridge superstructure to some extent and enable the calculation of more realistic bridge accelerations. At the same time, such multi-body models require significantly greater processing capacities and precise knowledge of various vehicle properties. The European standards allow for applying the so-called additional damping method when simple load models, such as the MLM, are used in dynamic calculations. An additional damping factor depending on the bridge span, which should take into account the vibration-reducing benefits of the vehicle-bridge interaction, is assigned to the supporting structure in the calculations. However, numerous studies show that when the current standard specifications are applied, the calculation results for the bridge accelerations are in many cases still too high compared to the measured bridge accelerations, while in other cases, they are not on the safe side. A proposal to calculate the additional damping based on extensive dynamic calculations for a parametric field of simply supported bridges with a ballasted track was developed to address this issue. In this contribution, several different approaches to determine the additional damping of the supporting structure considering the vehicle-bridge interaction when using the MLM are compared with one another. Besides the standard specifications, this includes the approach mentioned above and two additional recently published alternative formulations derived from analytical approaches. For a bridge catalogue of 65 existing bridges in Austria in steel, concrete or composite construction, calculations are carried out with the MLM for two different high-speed trains and the different approaches for additional damping. The results are compared with the calculation results obtained by applying a more sophisticated multi-body model of the trains used. The evaluation and comparison of the results allow assessing the benefits of different calculation concepts for the additional damping regarding their accuracy and possible applications. The evaluation shows that by applying one of the recently published redesigned additional damping methods, the calculation results can reflect the influence of the vehicle-bridge interaction on the design-relevant structural accelerations considerably more reliable than by using normative specifications.Keywords: Additional Damping Method, Bridge Dynamics, High-Speed Railway Traffic, Vehicle-Bridge-Interaction
Procedia PDF Downloads 161345 Modeling Discrimination against Gay People: Predictors of Homophobic Behavior against Gay Men among High School Students in Switzerland
Authors: Patrick Weber, Daniel Gredig
Abstract:
Background and Purpose: Research has well documented the impact of discrimination and micro-aggressions on the wellbeing of gay men and, especially, adolescents. For the prevention of homophobic behavior against gay adolescents, however, the focus has to shift on those who discriminate: For the design and tailoring of prevention and intervention, it is important to understand the factors responsible for homophobic behavior such as, for example, verbal abuse. Against this background, the present study aimed to assess homophobic – in terms of verbally abusive – behavior against gay people among high school students. Furthermore, it aimed to establish the predictors of the reported behavior by testing an explanatory model. This model posits that homophobic behavior is determined by negative attitudes and knowledge. These variables are supposed to be predicted by the acceptance of traditional gender roles, religiosity, orientation toward social dominance, contact with gay men, and by the perceived expectations of parents, friends and teachers. These social-cognitive variables in turn are assumed to be determined by students’ gender, age, immigration background, formal school level, and the discussion of gay issues in class. Method: From August to October 2016, we visited 58 high school classes in 22 public schools in a county in Switzerland, and asked the 8th and 9th year students on three formal school levels to participate in survey about gender and gay issues. For data collection, we used an anonymous self-administered questionnaire filled in during class. Data were analyzed using descriptive statistics and structural equation modelling (Generalized Least Square Estimates method). The sample included 897 students, 334 in the 8th and 563 in the 9th year, aged 12–17, 51.2% being female, 48.8% male, 50.3% with immigration background. Results: A proportion of 85.4% participants reported having made homophobic statements in the 12 month before survey, 4.7% often and very often. Analysis showed that respondents’ homophobic behavior was predicted directly by negative attitudes (β=0.20), as well as by the acceptance of traditional gender roles (β=0.06), religiosity (β=–0.07), contact with gay people (β=0.10), expectations of parents (β=–0.14) and friends (β=–0.19), gender (β=–0.22) and having a South-East-European or Western- and Middle-Asian immigration background (β=0.09). These variables were predicted, in turn, by gender, age, immigration background, formal school level, and discussion of gay issues in class (GFI=0.995, AGFI=0.979, SRMR=0.0169, CMIN/df=1.199, p>0.213, adj. R2 =0.384). Conclusion: Findings evidence a high prevalence of homophobic behavior in the responding high school students. The tested explanatory model explained 38.4% of the assessed homophobic behavior. However, data did not found full support of the model. Knowledge did not turn out to be a predictor of behavior. Except for the perceived expectation of teachers and orientation toward social dominance, the social-cognitive variables were not fully mediated by attitudes. Equally, gender and immigration background predicted homophobic behavior directly. These findings demonstrate the importance of prevention and provide also leverage points for interventions against anti-gay bias in adolescents – also in social work settings as, for example, in school social work, open youth work or foster care.Keywords: discrimination, high school students, gay men, predictors, Switzerland
Procedia PDF Downloads 329344 Catalytic Dehydrogenation of Formic Acid into H2/CO2 Gas: A Novel Approach
Authors: Ayman Hijazi, Witold Kwapinski, J. J. Leahy
Abstract:
Finding a sustainable alternative energy to fossil fuel is an urgent need as various environmental challenges in the world arise. Therefore, formic acid (FA) decomposition has been an attractive field that lies at the center of biomass platform, comprising a potential pool of hydrogen energy that stands as a new energy vector. Liquid FA features considerable volumetric energy density of 6.4 MJ/L and a specific energy density of 5.3 MJ/Kg that qualifies it in the prime seat as an energy source for transportation infrastructure. Additionally, the increasing research interest in FA decomposition is driven by the need of in-situ H2 production, which plays a key role in the hydrogenation reactions of biomass into higher value components. It is reported elsewhere in literature that catalytic decomposition of FA is usually performed in poorly designed setup using simple glassware under magnetic stirring, thus demanding further energy investment to retain the used catalyst. it work suggests an approach that integrates designing a novel catalyst featuring magnetic property with a robust setup that minimizes experimental & measurement discrepancies. One of the most prominent active species for dehydrogenation/hydrogenation of biomass compounds is palladium. Accordingly, we investigate the potential of engrafting palladium metal onto functionalized magnetic nanoparticles as a heterogeneous catalyst to favor the production of CO-free H2 gas from FA. Using ordinary magnet to collect the spent catalyst renders core-shell magnetic nanoparticles as the backbone of the process. Catalytic experiments were performed in a jacketed batch reactor equipped with an overhead stirrer under inert medium. Through a novel approach, FA is charged into the reactor via high-pressure positive displacement pump at steady state conditions. The produced gas (H2+CO2) was measured by connecting the gas outlet to a measuring system based on the amount of the displaced water. The novelty of this work lies in designing a very responsive catalyst, pumping consistent amount of FA into a sealed reactor running at steady state mild temperatures, and continuous gas measurement, along with collecting the used catalyst without the need for centrifugation. Catalyst characterization using TEM, XRD, SEM, and CHN elemental analyzer provided us with details of catalyst preparation and facilitated new venues to alter the nanostructure of the catalyst framework. Consequently, the introduction of amine groups has led to appreciable improvements in terms of dispersion of the doped metals and eventually attaining nearly complete conversion (100%) of FA after 7 hours. The relative importance of the process parameters such as temperature (35-85°C), stirring speed (150-450rpm), catalyst loading (50-200mgr.), and Pd doping ratio (0.75-1.80wt.%) on gas yield was assessed by a Taguchi design-of-experiment based model. Experimental results showed that operating at lower temperature range (35-50°C) yielded more gas while the catalyst loading and Pd doping wt.% were found to be the most significant factors with a P-values 0.026 & 0.031, respectively.Keywords: formic acid decomposition, green catalysis, hydrogen, mesoporous silica, process optimization, nanoparticles
Procedia PDF Downloads 52343 Measuring Digital Literacy in the Chilean Workforce
Authors: Carolina Busco, Daniela Osses
Abstract:
The development of digital literacy has become a fundamental element that allows for citizen inclusion, access to quality jobs, and a labor market capable of responding to the digital economy. There are no methodological instruments available in Chile to measure the workforce’s digital literacy and improve national policies on this matter. Thus, the objective of this research is to develop a survey to measure digital literacy in a sample of 200 Chilean workers. Dimensions considered in the instrument are sociodemographics, access to infrastructure, digital education, digital skills, and the ability to use e-government services. To achieve the research objective of developing a digital literacy model of indicators and a research instrument for this purpose, along with an exploratory analysis of data using factor analysis, we used an empirical, quantitative-qualitative, exploratory, non-probabilistic, and cross-sectional research design. The research instrument is a survey created to measure variables that make up the conceptual map prepared from the bibliographic review. Before applying the survey, a pilot test was implemented, resulting in several adjustments to the phrasing of some items. A validation test was also applied using six experts, including their observations on the final instrument. The survey contained 49 items that were further divided into three sets of questions: sociodemographic data; a Likert scale of four values ranked according to the level of agreement; iii) multiple choice questions complementing the dimensions. Data collection occurred between January and March 2022. For the factor analysis, we used the answers to 12 items with the Likert scale. KMO showed a value of 0.626, indicating a medium level of correlation, whereas Bartlett’s test yielded a significance value of less than 0.05 and a Cronbach’s Alpha of 0.618. Taking all factor selection criteria into account, we decided to include and analyze four factors that together explain 53.48% of the accumulated variance. We identified the following factors: i) access to infrastructure and opportunities to develop digital skills at the workplace or educational establishment (15.57%), ii) ability to solve everyday problems using digital tools (14.89%), iii) online tools used to stay connected with others (11.94%), and iv) residential Internet access and speed (11%). Quantitative results were discussed within six focus groups using heterogenic selection criteria related to the most relevant variables identified in the statistical analysis: upper-class school students; middle-class university students; Ph.D. professors; low-income working women, elderly individuals, and a group of rural workers. The digital divide and its social and economic correlations are evident in the results of this research. In Chile, the items that explain the acquisition of digital tools focus on access to infrastructure, which ultimately puts the first filter on the development of digital skills. Therefore, as expressed in the literature review, the advance of these skills is radically different when sociodemographic variables are considered. This increases socioeconomic distances and exclusion criteria, putting those who do not have these skills at a disadvantage and forcing them to seek the assistance of others.Keywords: digital literacy, digital society, workforce digitalization, digital skills
Procedia PDF Downloads 67342 Genomic and Proteomic Variability in Glycine Max Genotypes in Response to Salt Stress
Authors: Faheema Khan
Abstract:
To investigate the ability of sensitive and tolerant genotype of Glycine max to adapt to a saline environment in a field, we examined the growth performance, water relation and activities of antioxidant enzymes in relation to photosynthetic rate, chlorophyll a fluorescence, photosynthetic pigment concentration, protein and proline in plants exposed to salt stress. Ten soybean genotypes (Pusa-20, Pusa-40, Pusa-37, Pusa-16, Pusa-24, Pusa-22, BRAGG, PK-416, PK-1042, and DS-9712) were selected and grown hydroponically. After 3 days of proper germination, the seedlings were transferred to Hoagland’s solution (Hoagland and Arnon 1950). The growth chamber was maintained at a photosynthetic photon flux density of 430 μmol m−2 s−1, 14 h of light, 10 h of dark and a relative humidity of 60%. The nutrient solution was bubbled with sterile air and changed on alternate days. Ten-day-old seedlings were given seven levels of salt in the form of NaCl viz., T1 = 0 mM NaCl, T2=25 mM NaCl, T3=50 mM NaCl, T4=75 mM NaCl, T5=100 mM NaCl, T6=125 mM NaCl, T7=150 mM NaCl. The investigation showed that genotype Pusa-24, PK-416 and Pusa-20 appeared to be the most salt-sensitive. genotypes as inferred from their significantly reduced length, fresh weight and dry weight in response to the NaCl exposure. Pusa-37 appeared to be the most tolerant genotype since no significant effect of NaCl treatment on growth was found. We observed a greater decline in the photosynthetic variables like photosynthetic rate, chlorophyll fluorescence and chlorophyll content, in salt-sensitive (Pusa-24) genotype than in salt-tolerant Pusa-37 under high salinity. Numerous primers were verified on ten soybean genotypes obtained from Operon technologies among which 30 RAPD primers shown high polymorphism and genetic variation. The Jaccard’s similarity coefficient values for each pairwise comparison between cultivars were calculated and similarity coefficient matrix was constructed. The closer varieties in the cluster behaved similar in their response to salinity tolerance. Intra-clustering within the two clusters precisely grouped the 10 genotypes in sub-cluster as expected from their physiological findings.Salt tolerant genotype Pusa-37, was further analysed by 2-Dimensional gel electrophoresis to analyse the differential expression of proteins at high salt stress. In the Present study, 173 protein spots were identified. Of these, 40 proteins responsive to salinity were either up- or down-regulated in Pusa-37. Proteomic analysis in salt-tolerant genotype (Pusa-37) led to the detection of proteins involved in a variety of biological processes, such as protein synthesis (12 %), redox regulation (19 %), primary and secondary metabolism (25 %), or disease- and defence-related processes (32 %). In conclusion, the soybean plants in our study responded to salt stress by changing their protein expression pattern. The photosynthetic, biochemical and molecular study showed that there is variability in salt tolerance behaviour in soybean genotypes. Pusa-24 is the salt-sensitive and Pusa-37 is the salt-tolerant genotype. Moreover this study gives new insights into the salt-stress response in soybean and demonstrates the power of genomic and proteomic approach in plant biology studies which finally could help us in identifying the possible regulatory switches (gene/s) controlling the salt tolerant genotype of the crop plants and their possible role in defence mechanism.Keywords: glycine max, salt stress, RAPD, genomic and proteomic variability
Procedia PDF Downloads 423341 Worldwide GIS Based Earthquake Information System/Alarming System for Microzonation/Liquefaction and It’s Application for Infrastructure Development
Authors: Rajinder Kumar Gupta, Rajni Kant Agrawal, Jaganniwas
Abstract:
One of the most frightening phenomena of nature is the occurrence of earthquake as it has terrible and disastrous effects. Many earthquakes occur every day worldwide. There is need to have knowledge regarding the trends in earthquake occurrence worldwide. The recoding and interpretation of data obtained from the establishment of the worldwide system of seismological stations made this possible. From the analysis of recorded earthquake data, the earthquake parameters and source parameters can be computed and the earthquake catalogues can be prepared. These catalogues provide information on origin, time, epicenter locations (in term of latitude and longitudes) focal depths, magnitude and other related details of the recorded earthquakes. Theses catalogues are used for seismic hazard estimation. Manual interpretation and analysis of these data is tedious and time consuming. A geographical information system is a computer based system designed to store, analyzes and display geographic information. The implementation of integrated GIS technology provides an approach which permits rapid evaluation of complex inventor database under a variety of earthquake scenario and allows the user to interactively view results almost immediately. GIS technology provides a powerful tool for displaying outputs and permit to users to see graphical distribution of impacts of different earthquake scenarios and assumptions. An endeavor has been made in present study to compile the earthquake data for the whole world in visual Basic on ARC GIS Plate form so that it can be used easily for further analysis to be carried out by earthquake engineers. The basic data on time of occurrence, location and size of earthquake has been compiled for further querying based on various parameters. A preliminary analysis tool is also provided in the user interface to interpret the earthquake recurrence in region. The user interface also includes the seismic hazard information already worked out under GHSAP program. The seismic hazard in terms of probability of exceedance in definite return periods is provided for the world. The seismic zones of the Indian region are included in the user interface from IS 1893-2002 code on earthquake resistant design of buildings. The City wise satellite images has been inserted in Map and based on actual data the following information could be extracted in real time: • Analysis of soil parameters and its effect • Microzonation information • Seismic hazard and strong ground motion • Soil liquefaction and its effect in surrounding area • Impacts of liquefaction on buildings and infrastructure • Occurrence of earthquake in future and effect on existing soil • Propagation of earth vibration due of occurrence of Earthquake GIS based earthquake information system has been prepared for whole world in Visual Basic on ARC GIS Plate form and further extended micro level based on actual soil parameters. Individual tools has been developed for liquefaction, earthquake frequency etc. All information could be used for development of infrastructure i.e. multi story structure, Irrigation Dam & Its components, Hydro-power etc in real time for present and future.Keywords: GIS based earthquake information system, microzonation, analysis and real time information about liquefaction, infrastructure development
Procedia PDF Downloads 316340 A High-Throughput Enzyme Screening Method Using Broadband Coherent Anti-stokes Raman Spectroscopy
Authors: Ruolan Zhang, Ryo Imai, Naoko Senda, Tomoyuki Sakai
Abstract:
Enzymes have attracted increasing attentions in industrial manufacturing for their applicability in catalyzing complex chemical reactions under mild conditions. Directed evolution has become a powerful approach to optimize enzymes and exploit their full potentials under the circumstance of insufficient structure-function knowledge. With the incorporation of cell-free synthetic biotechnology, rapid enzyme synthesis can be realized because no cloning procedure such as transfection is needed. Its open environment also enables direct enzyme measurement. These properties of cell-free biotechnology lead to excellent throughput of enzymes generation. However, the capabilities of current screening methods have limitations. Fluorescence-based assay needs applicable fluorescent label, and the reliability of acquired enzymatic activity is influenced by fluorescent label’s binding affinity and photostability. To acquire the natural activity of an enzyme, another method is to combine pre-screening step and high-performance liquid chromatography (HPLC) measurement. But its throughput is limited by necessary time investment. Hundreds of variants are selected from libraries, and their enzymatic activities are then identified one by one by HPLC. The turn-around-time is 30 minutes for one sample by HPLC, which limits the acquirable enzyme improvement within reasonable time. To achieve the real high-throughput enzyme screening, i.e., obtain reliable enzyme improvement within reasonable time, a widely applicable high-throughput measurement of enzymatic reactions is highly demanded. Here, a high-throughput screening method using broadband coherent anti-Stokes Raman spectroscopy (CARS) was proposed. CARS is one of coherent Raman spectroscopy, which can identify label-free chemical components specifically from their inherent molecular vibration. These characteristic vibrational signals are generated from different vibrational modes of chemical bonds. With the broadband CARS, chemicals in one sample can be identified from their signals in one broadband CARS spectrum. Moreover, it can magnify the signal levels to several orders of magnitude greater than spontaneous Raman systems, and therefore has the potential to evaluate chemical's concentration rapidly. As a demonstration of screening with CARS, alcohol dehydrogenase, which converts ethanol and nicotinamide adenine dinucleotide oxidized form (NAD+) to acetaldehyde and nicotinamide adenine dinucleotide reduced form (NADH), was used. The signal of NADH at 1660 cm⁻¹, which is generated from nicotinamide in NADH, was utilized to measure the concentration of it. The evaluation time for CARS signal of NADH was determined to be as short as 0.33 seconds while having a system sensitivity of 2.5 mM. The time course of alcohol dehydrogenase reaction was successfully measured from increasing signal intensity of NADH. This measurement result of CARS was consistent with the result of a conventional method, UV-Vis. CARS is expected to have application in high-throughput enzyme screening and realize more reliable enzyme improvement within reasonable time.Keywords: Coherent Anti-Stokes Raman Spectroscopy, CARS, directed evolution, enzyme screening, Raman spectroscopy
Procedia PDF Downloads 141339 Development of Building Information Modeling in Property Industry: Beginning with Building Information Modeling Construction
Authors: B. Godefroy, D. Beladjine, K. Beddiar
Abstract:
In France, construction BIM actors commonly evoke the BIM gains for exploitation by integrating of the life cycle of a building. The standardization of level 7 of development would achieve this stage of the digital model. The householders include local public authorities, social landlords, public institutions (health and education), enterprises, facilities management companies. They have a dual role: owner and manager of their housing complex. In a context of financial constraint, the BIM of exploitation aims to control costs, make long-term investment choices, renew the portfolio and enable environmental standards to be met. It assumes a knowledge of the existing buildings, marked by its size and complexity. The information sought must be synthetic and structured, it concerns, in general, a real estate complex. We conducted a study with professionals about their concerns and ways to use it to see how householders could benefit from this development. To obtain results, we had in mind the recurring interrogation of the project management, on the needs of the operators, we tested the following stages: 1) Inculcate a minimal culture of BIM with multidisciplinary teams of the operator then by business, 2) Learn by BIM tools, the adaptation of their trade in operations, 3) Understand the place and creation of a graphic and technical database management system, determine the components of its library so their needs, 4) Identify the cross-functional interventions of its managers by business (operations, technical, information system, purchasing and legal aspects), 5) Set an internal protocol and define the BIM impact in their digital strategy. In addition, continuity of management by the integration of construction models in the operation phase raises the question of interoperability in the control of the production of IFC files in the operator’s proprietary format and the export and import processes, a solution rivaled by the traditional method of vectorization of paper plans. Companies that digitize housing complex and those in FM produce a file IFC, directly, according to their needs without recourse to the model of construction, they produce models business for the exploitation. They standardize components, equipment that are useful for coding. We observed the consequences resulting from the use of the BIM in the property industry and, made the following observations: a) The value of data prevail over the graphics, 3D is little used b) The owner must, through his organization, promote the feedback of technical management information during the design phase c) The operator's reflection on outsourcing concerns the acquisition of its information system and these services, observing the risks and costs related to their internal or external developments. This study allows us to highlight: i) The need for an internal organization of operators prior to a response to the construction management ii) The evolution towards automated methods for creating models dedicated to the exploitation, a specialization would be required iii) A review of the communication of the project management, management continuity not articulating around his building model, it must take into account the environment of the operator and reflect on its scope of action.Keywords: information system, interoperability, models for exploitation, property industry
Procedia PDF Downloads 144338 The Incident of Concussion across Popular American Youth Sports: A Retrospective Review
Authors: Rami Hashish, Manon Limousis-Gayda, Caitlin H. McCleery
Abstract:
Introduction: A leading cause of emergency room visits among youth (in the United States), is sports-related traumatic brain injuries. Mild traumatic brain injuries (mTBIs), also called concussions, are caused by linear and/or angular acceleration experienced at the head and represent an increasing societal burden. Due to the developing nature of the brain in youth, there is a great risk for long-term neuropsychological deficiencies following a concussion. Accordingly, the purpose of this paper is to investigate incidence rates of concussion across gender for the five most common youth sports in the United States. These include basketball, track and field, soccer, baseball (boys), softball (girls), football (boys), and volleyball (girls). Methods: A PubMed search was performed for four search themes combined. The first theme identified the outcomes (concussion, brain injuries, mild traumatic brain injury, etc.). The second theme identified the sport (American football, soccer, basketball, softball, volleyball, track, and field, etc.). The third theme identified the population (adolescence, children, youth, boys, girls). The last theme identified the study design (prevalence, frequency, incidence, prospective). Ultimately, 473 studies were surveyed, with 15 fulfilling the criteria: prospective study presenting original data and incidence of concussion in the relevant youth sport. The following data were extracted from the selected studies: population age, total study population, total athletic exposures (AE) and incidence rate per 1000 athletic exposures (IR/1000). Two One-Way ANOVA and a Tukey’s post hoc test were conducted using SPSS. Results: From the 15 selected studies, statistical analysis revealed the incidence of concussion per 1000 AEs across the considered sports ranged from 0.014 (girl’s track and field) to 0.780 (boy’s football). Average IR/1000 across all sports was 0.483 and 0.268 for boys and girls, respectively; this difference in IR was found to be statistically significant (p=0.013). Tukey’s post hoc test showed that football had significantly higher IR/1000 than boys’ basketball (p=0.022), soccer (p=0.033) and track and field (p=0.026). No statistical difference was found for concussion incidence between girls’ sports. Removal of football was found to lower the IR/1000 for boys without a statistical difference (p=0.101) compared to girls. Discussion: Football was the only sport showing a statistically significant difference in concussion incidence rate relative to other sports (within gender). Males were overall more likely to be concussed than females when football was included (1.8x), whereas concussion was more likely for females when football was excluded. While the significantly higher rate of concussion in football is not surprising because of the nature and rules of the sport, it is concerning that research has shown higher incidence of concussion in practices than games. Interestingly, findings indicate that girls’ sports are more concussive overall when football is removed. This appears to counter the common notion that boys’ sports are more physically taxing and dangerous. Future research should focus on understanding the concussive mechanisms of injury in each sport to enable effective rule changes.Keywords: gender, football, soccer, traumatic brain injury
Procedia PDF Downloads 141337 The Reality of Gender Equality in Universities Libraries: A Case of Pakistan
Authors: Qurat Ul Ain Saleem, Kanwal Ameen
Abstract:
The library and information science discipline is universally known as a feminist profession. It is considered a suitable field for females in Pakistan like a few other fields such as teaching and healthcare. It is also reflected through the uneven enrollment at graduate levels in library schools across the country as there are more females as compared to males. However, that uneven ratio does not really translate in the profession after passing out. There are more males in the professional as compared to females, as well as males can be seen on managerial and administrative posts majorly. A few females who joined the perception remain underrated and are hardly seen at managerial or administrative positions in the academic libraries. Therefore, this study was designed to highlight the perceptions of those females who have joined the profession to identify the issues related to equality faced by them as a professional. A qualitative research design based on a semi-structured interview was selected as an appropriate method to achieve the objectives of this study. Female librarians working in the higher education commission’s recognized public and private sector universities of Punjab, Pakistan, were selected as the population for this study. Female librarians shared that inequalities and discrimination based on face value, experience, communication, and relationship with the manager are common at their workplaces. They added that managers prefer male professionals to deal with delegation or presentations though we both can do that. Female professionals from the private sector believed that library managers make final hiring and selection decisions based on job duties and gender. However, the one with strong references will be preferred for the job. Also, private-sector employees suffered more prejudice due to the non-availability of proper patterns of promotions and increments. The government personnel said there is always a proper board/procedure for hiring and promotions; therefore, it is difficult for them to identify any inequality. Participants were dissatisfied with their managers for not allowing them to attend training and conferences. The majority of participants from the private sector said they wouldn't speak up to prejudice because they are afraid of losing their jobs and their voice is lost in a male-dominated society where males hold numerous authoritative positions and females are considered less competent. Nonetheless, the discrimination and inequalities affected the work motivation and enthusiasm of employees. Therefore, organizations should not discriminate against the staff in terms of facilities and benefits. The sample may not represent the true picture of gender equality in university libraries of Pakistan due to less number of participants and limited geographical boundaries. It is also assumed that some females may refrain from disclosing factual information or some may exaggerate the facts as a large number of participants requested to become part of the study. Equal opportunities should be offered to female library professionals to uplift and involve them to mitigate the perception of gender dominance. The organizations or immediate authorities should allow their staff to participate in training opportunities to learn modern practices to better serve the community.Keywords: equality-workplace, libraries as workplace, female professionals, librarians-Pakistan
Procedia PDF Downloads 113336 Fully Instrumented Small-Scale Fire Resistance Benches for Aeronautical Composites Assessment
Authors: Fabienne Samyn, Pauline Tranchard, Sophie Duquesne, Emilie Goncalves, Bruno Estebe, Serge Boubigot
Abstract:
Stringent fire safety regulations are enforced in the aeronautical industry due to the consequences that potential fire event on an aircraft might imply. This is so much true that the fire issue is considered right from the design of the aircraft structure. Due to the incorporation of an increasing amount of polymer matrix composites in replacement of more conventional materials like metals, the nature of the fire risks is changing. The choice of materials used is consequently of prime importance as well as the evaluation of its resistance to fire. The fire testing is mostly done using the so-called certification tests according to standards such as the ISO2685:1998(E). The latter describes a protocol to evaluate the fire resistance of structures located in fire zone (ability to withstand fire for 5min). The test consists in exposing an at least 300x300mm² sample to an 1100°C propane flame with a calibrated heat flux of 116kW/m². This type of test is time-consuming, expensive and gives access to limited information in terms of fire behavior of the materials (pass or fail test). Consequently, it can barely be used for material development purposes. In this context, the laboratory UMET in collaboration with industrial partners has developed a horizontal and a vertical small-scale instrumented fire benches for the characterization of the fire behavior of composites. The benches using smaller samples (no more than 150x150mm²) enables to cut downs costs and hence to increase sampling throughput. However, the main added value of our benches is the instrumentation used to collect useful information to understand the behavior of the materials. Indeed, measurements of the sample backside temperature are performed using IR camera in both configurations. In addition, for the vertical set up, a complete characterization of the degradation process, can be achieved via mass loss measurements and quantification of the gasses released during the tests. These benches have been used to characterize and study the fire behavior of aeronautical carbon/epoxy composites. The horizontal set up has been used in particular to study the performances and durability of protective intumescent coating on 2mm thick 2D laminates. The efficiency of this approach has been validated, and the optimized coating thickness has been determined as well as the performances after aging. Reductions of the performances after aging were attributed to the migration of some of the coating additives. The vertical set up has enabled to investigate the degradation process of composites under fire. An isotropic and a unidirectional 4mm thick laminates have been characterized using the bench and post-fire analyses. The mass loss measurements and the gas phase analyses of both composites do not present significant differences unlike the temperature profiles in the thickness of the samples. The differences have been attributed to differences of thermal conductivity as well as delamination that is much more pronounced for the isotropic composite (observed on the IR-images). This has been confirmed by X-ray microtomography. The developed benches have proven to be valuable tools to develop fire safe composites.Keywords: aeronautical carbon/epoxy composite, durability, intumescent coating, small-scale ‘ISO 2685 like’ fire resistance test, X-ray microtomography
Procedia PDF Downloads 271335 Improved Morphology in Sequential Deposition of the Inverted Type Planar Heterojunction Solar Cells Using Cheap Additive (DI-H₂O)
Authors: Asmat Nawaz, Ceylan Zafer, Ali K. Erdinc, Kaiying Wang, M. Nadeem Akram
Abstract:
Hybrid halide Perovskites with the general formula ABX₃, where X = Cl, Br or I, are considered as an ideal candidates for the preparation of photovoltaic devices. The most commonly and successfully used hybrid halide perovskite for photovoltaic applications is CH₃NH₃PbI₃ and its analogue prepared from lead chloride, commonly symbolized as CH₃NH₃PbI₃_ₓClₓ. Some researcher groups are using lead free (Sn replaces Pb) and mixed halide perovskites for the fabrication of the devices. Both mesoporous and planar structures have been developed. By Comparing mesoporous structure in which the perovskite materials infiltrate into mesoporous metal oxide scaffold, the planar architecture is much simpler and easy for device fabrication. In a typical perovskite solar cell, a perovskite absorber layer is sandwiched between the hole and electron transport. Upon the irradiation, carriers are created in the absorber layer that can travel through hole and electron transport layers and the interface in between. We fabricated inverted planar heterojunction structure ITO/PEDOT/ Perovskite/PCBM/Al, based solar cell via two-step spin coating method. This is also called Sequential deposition method. A small amount of cheap additive H₂O was added into PbI₂/DMF to make a homogeneous solution. We prepared four different solution such as (W/O H₂O, 1% H₂O, 2% H₂O, 3% H₂O). After preparing, the whole night stirring at 60℃ is essential for the homogenous precursor solutions. We observed that the solution with 1% H₂O was much more homogenous at room temperature as compared to others. The solution with 3% H₂O was precipitated at once at room temperature. The four different films of PbI₂ were formed on PEDOT substrates by spin coating and after that immediately (before drying the PbI₂) the substrates were immersed in the methyl ammonium iodide solution (prepared in isopropanol) for the completion of the desired perovskite film. After getting desired films, rinse the substrates with isopropanol to remove the excess amount of methyl ammonium iodide and finally dried it on hot plate only for 1-2 minutes. In this study, we added H₂O in the PbI₂/DMF precursor solution. The concept of additive is widely used in the bulk- heterojunction solar cells to manipulate the surface morphology, leading to the enhancement of the photovoltaic performance. There are two most important parameters for the selection of additives. (a) Higher boiling point w.r.t host material (b) good interaction with the precursor materials. We observed that the morphology of the films was improved and we achieved a denser, uniform with less cavities and almost full surface coverage films but only using precursor solution having 1% H₂O. Therefore, we fabricated the complete perovskite solar cell by sequential deposition technique with precursor solution having 1% H₂O. We concluded that with the addition of additives in the precursor solutions one can easily be manipulate the morphology of the perovskite film. In the sequential deposition method, thickness of perovskite film is in µm and the charge diffusion length of PbI₂ is in nm. Therefore, by controlling the thickness using other deposition methods for the fabrication of solar cells, we can achieve the better efficiency.Keywords: methylammonium lead iodide, perovskite solar cell, precursor composition, sequential deposition
Procedia PDF Downloads 246334 Comparative Study on Effectiveness and Safety of Oral Antidiabetic Medications in Patients with Type 2 Diabetes Mellitus in a Tertiary Care Hospital of Bangalore, South India - A Prospective Cohort Study
Authors: Shobha Rani R. Hiremath, Keerthana R., Madhuvan H. S.
Abstract:
1.BACKGROUND: Type 2 Diabetes is a chronic health condition where the body cannot effectively use the insulin it produces, leading to elevated blood sugar levels It is often associated with lifestyle factors and insulin resistance. Globally, diabetes is on the rise, with urban areas like Bangalore seeing a surge due to lifestyle changes, stress, and dietary habits. To manage diabetes effectively, over 50 medications are available, each serving to regulate blood sugar through different mechanisms. This reflects the complex and individualized nature of diabetes treatment. Given the increase in medications for Type 2 diabetes mellitus, it is important to evaluate their effectiveness and safety so that clinicians can make informed choices while treating their patients. 2.OBJECTIVES: To evaluate the effectiveness of various anti-diabetic medications used in the hospital in Type 2 diabetes patients by monitoring their HbA1c levels. To assess the safety of these medications by monitoring Adverse drug reactions if any. 3. METHODOLOGY: Design : Prospective Cohort study, Study period: 8 months, Sample Size: 100 patients. Inclusion Criteria: Patients above 18 years of both genders who were diagnosed with T2DM and who were prescribed oral hypoglycaemic agents. Exclusion Criteria: Diabetic patients with hepatic/renal failure, patients prescribed with insulin /not prescribed with OHAs and patients who were terminally ill, pregnant and lactating patients. Source of Data: Prescriptions, lab reports, ECG reports. Data collection forms were used to collect data which consisted of patient demographic details, drugs prescribed, laboratory investigations such as HbA1C, FBS, PPBS , ECG and any ADRs experienced. Data was collected at baseline, 3 months, and 6 months. It was statistically analyzed using SPSS (version 26) software. 4. RESULTS: Greater number of patients (46%) were in the age group of greater than 61 years. 43 patients had no co-morbidities whereas 51 patients had Hypertension as the co morbidity. Basically 5 Drug combinations were prescribed as follows. Combination 1: Tablet Metformin HCL + Glimepiride (500, 2 mg) : 1-0-1, Combination 2: Tablet Sitagliptin + Metformin HCL (50, 500 mg) : 1-0-1, Combination 3: Tablet Vildagliptin + Metformin HCL (50, 500 mg): 1-0-1, Combination 4: Tablet Voglibose+ Glimepiride+ Metformin HCL (0.2 ,2 ,500mg): : 1-0-1, Combination 5: Tablet Voglibose+ Glimepiride+ Metformin HCL (0.2, 2 ,500mg): : 1-0-1 and T. Sitagliptin +Metformin HCL (50, 500 mg): 0-1-0. Combination 5 (Voglibose, Glimepiride, Metformin, Sitagliptin) was most effective in reducing HbA1c levels, showing a significant decrease (-0.00682, p = 0.001), followed by Combinations 4 and 3. However, Combination 5 also had the highest incidence of gastrointestinal side effects (72.7%) and ECG abnormalities (27.3%). Combination 1 (Metformin + Glimepiride) had the highest occurrence of hypoglycemia due to Glimepiride's insulin-stimulating effects. Weight loss was most notable in Combination 5, affecting 36.36% of patients. 5.CONCLUSION: The enhanced effectiveness of Combinations 3, 4, and 5 suggests that a multi-drug approach that incorporates different mechanisms of action is more effective in managing HbA1c levels in patients with diabetes. Adverse effect profiles should be considered for personalized treatment strategies.Keywords: type 2 diabetes, safety, oral anti diabetic medications, effectiveness
Procedia PDF Downloads 10333 Effects of the Exit from Budget Support on Good Governance: Findings from Four Sub-Saharan Countries
Authors: Magdalena Orth, Gunnar Gotz
Abstract:
Background: Domestic accountability, budget transparency and public financial management (PFM) are considered vital components of good governance in developing countries. The aid modality budget support (BS) promotes these governance functions in developing countries. BS engages in political decision-making and provides financial and technical support to poverty reduction strategies of the partner countries. Nevertheless, many donors have withdrawn their support from this modality due to cases of corruption, fraud or human rights violations. This exit from BS is leaving a finance and governance vacuum in the countries. The evaluation team analyzed the consequences of terminating the use of this modality and found particularly negative effects for good governance outcomes. Methodology: The evaluation uses a qualitative (theory-based) approach consisting of a comparative case study design, which is complemented by a process-tracing approach. For the case studies, the team conducted over 100 semi-structured interviews in Malawi, Uganda, Rwanda and Zambia and used four country-specific, tailor-made budget analysis. In combination with a previous DEval evaluation synthesis on the effects of BS, the team was able to create a before-and-after comparison that yields causal effects. Main Findings: In all four countries domestic accountability and budget transparency declined if other forms of pressure are not replacing BS´s mutual accountability mechanisms. In Malawi a fraud scandal created pressure from the society and from donors so that accountability was improved. In the other countries, these pressure mechanisms were absent so that domestic accountability declined. BS enables donors to actively participate in political processes of the partner country as donors transfer funds into the treasury of the partner country and conduct a high-level political dialogue. The results confirm that the exit from BS created a governance vacuum that, if not compensated through external/internal pressure, leads to a deterioration of good governance. For example, in the case of highly aid dependent Malawi did the possibility of a relaunch of BS provide sufficient incentives to push for governance reforms. Overall the results show that the three good governance areas are negatively affected by the exit from BS. This stands in contrast to positive effects found before the exit. The team concludes that the relationship is causal, because the before-and-after comparison coherently shows that the presence of BS correlates with positive effects and the absence with negative effects. Conclusion: These findings strongly suggest that BS is an effective modality to promote governance and its abolishment is likely to cause governance disruptions. Donors and partner governments should find ways to re-engage in closely coordinated policy-based aid modalities. In addition, a coordinated and carefully managed exit-strategy should be in place before an exit from similar modalities is considered. Particularly a continued framework of mutual accountability and a high-level political dialogue should be aspired to maintain pressure and oversight that is required to achieve good governance.Keywords: budget support, domestic accountability, public financial management and budget transparency, Sub-Sahara Africa
Procedia PDF Downloads 151332 Formulation and Optimization of Self Nanoemulsifying Drug Delivery System of Rutin for Enhancement of Oral Bioavailability Using QbD Approach
Authors: Shrestha Sharma, Jasjeet K. Sahni, Javed Ali, Sanjula Baboota
Abstract:
Introduction: Rutin is a naturally occurring strong antioxidant molecule belonging to bioflavonoid category. Due to its free radical scavenging properties, it has been found to be beneficial in the treatment of various diseases including inflammation, cancer, diabetes, allergy, cardiovascular disorders and various types of microbial infections. Despite its beneficial effects, it suffers from the problem of low aqueous solubility which is responsible for low oral bioavailability. The aim of our study was to optimize and characterize self-nanoemulsifying drug delivery system (SNEDDS) of rutin using Box-Behnken design (BBD) combined with a desirability function. Further various antioxidant, pharmacokinetic and pharmacodynamic studies were performed for the optimized rutin SNEDDS formulation. Methodologies: Selection of oil, surfactant and co-surfactant was done on the basis of solubility/miscibility studies. Sefsol+ Vitamin E, Solutol HS 15 and Transcutol P were selected as oil phase, surfactant and co-surfactant respectively. Optimization of SNEDDS formulations was done by a three-factor, three-level (33)BBD. The independent factors were Sefsol+ Vitamin E, Solutol HS15, and Transcutol P. The dependent variables were globule size, self emulsification time (SEF), % transmittance and cumulative percentage drug released. Various response surface graphs and contour plots were constructed to understand the effect of different factor, their levels and combinations on the responses. The optimized Rutin SNEDDS formulation was characterized for various parameters such as globule size, zeta potential, viscosity, refractive index , % Transmittance and in vitro drug release. Ex vivo permeation studies and pharmacokinetic studies were performed for optimized formulation. Antioxidant activity was determined by DPPH and reducing power assays. Anti-inflammatory activity was determined by using carrageenan induced rat paw oedema method. Permeation of rutin across small intestine was assessed using confocal laser scanning microscopy (CLSM). Major findings:The optimized SNEDDS formulation consisting of Sefsol+ Vitamin E - Solutol HS15 -Transcutol HP at proportions of 25:35:17.5 (w/w) was prepared and a comparison of the predicted values and experimental values were found to be in close agreement. The globule size and PDI of optimized SNEDDS formulation was found to be 16.08 ± 0.02 nm and 0.124±0.01 respectively. Significant (p˂0.05) increase in percentage drug release was achieved in the case of optimized SNEDDS formulation (98.8 %) as compared to rutin suspension. Furthermore, pharmacokinetic study showed a 2.3-fold increase in relative oral bioavailability compared with that of the suspension. Antioxidant assay results indicated better efficacy of the developed formulation than the pure drug and it was found to be comparable with ascorbic acid. The results of anti-inflammatory studies showed 72.93 % inhibition for the SNEDDS formulation which was significantly higher than the drug suspension 46.56%. The results of CLSM indicated that the absorption of SNEDDS formulation was considerably higher than that from rutin suspension. Conclusion: Rutin SNEDDS have been successfully prepared and they can serve as an effective tool in enhancing oral bioavailability and efficacy of Rutin.Keywords: rutin, oral bioavilability, pharamacokinetics, pharmacodynamics
Procedia PDF Downloads 500331 CD97 and Its Role in Glioblastoma Stem Cell Self-Renewal
Authors: Niklas Ravn-Boess, Nainita Bhowmick, Takamitsu Hattori, Shohei Koide, Christopher Park, Dimitris Placantonakis
Abstract:
Background: Glioblastoma (GBM) is the most common and deadly primary brain malignancy in adults. Tumor propagation, brain invasion, and resistance to therapy critically depend on GBM stem-like cells (GSCs); however, the mechanisms that regulate GSC self-renewal are incompletely understood. Given the aggressiveness and poor prognosis of GBM, it is imperative to find biomarkers that could also translate into novel drug targets. Along these lines, we have identified a cell surface antigen, CD97 (ADGRE5), an adhesion G protein-coupled receptor (GPCR), that is expressed on GBM cells but is absent from non-neoplastic brain tissue. CD97 has been shown to promote invasiveness, angiogenesis, and migration in several human cancers, but its frequency of expression and functional role in regulating GBM growth and survival, and its potential as a therapeutic target has not been investigated. Design: We assessed CD97 mRNA and protein expression in patient derived GBM samples and cell lines using publicly available RNA-sequencing datasets and flow cytometry, respectively. To assess CD97 function, we generated shRNA lentiviral constructs that target a sequence in the CD97 extracellular domain (ECD). A scrambled shRNA (scr) with no predicted targets in the genome was used as a control. We evaluated CD97 shRNA lentivirally transduced GBM cells for Ki67, Annexin V, and DAPI. We also tested CD97 KD cells for their ability to self-renew using clonogenic tumorsphere formation assays. Further, we utilized synthetic Abs (sAbs) generated against the ECD of CD97 to test for potential antitumor effects using patient-derived GBM cell lines. Results: CD97 mRNA expression was expressed at high levels in all GBM samples available in the TCGA cohort. We found high levels of surface CD97 protein expression in 6/6 patient-derived GBM cell cultures, but not human neural stem cells. Flow cytometry confirmed downregulation of CD97 in CD97 shRNA lentivirally transduced cells. CD97 KD induced a significant reduction in cell growth in 3 independent GBM cell lines representing mesenchymal and proneural subtypes, which was accompanied by reduced (~20%) Ki67 staining and increased (~30%) apoptosis. Incubation of GBM cells with sAbs (20 ug/ ml) against the ECD of CD97 for 3 days induced GSC differentiation, as determined by the expression of GFAP and Tubulin. Using three unique GBM patient derived cultures, we found that CD97 KD attenuated the ability of GBM cells to initiate sphere formation by over 300 fold, consistent with an impairment in GSC self-renewal. Conclusion: Loss of CD97 expression in patient-derived GBM cells markedly decreases proliferation, induces cell death, and reduces tumorsphere formation. sAbs against the ECD of CD97 reduce tumorsphere formation, recapitulating the phenotype of CD97 KD, suggesting that sAbs that inhibit CD97 function exhibit anti-tumor activity. Collectively, these findings indicate that CD97 is necessary for the proliferation and survival of human GBM cells and identify CD97 as a promising therapeutically targetable vulnerability in GBM.Keywords: adhesion GPCR, CD97, GBM stem cell, glioblastoma
Procedia PDF Downloads 137330 Rural-To-Urban Migrants' Experiences with Primary Care in Four Types of Medical Institutions in Guangzhou, China
Authors: Jiazhi Zeng, Leiyu Shi, Xia Zou, Wen Chen, Li Ling
Abstract:
Background: China is facing the unprecedented challenge of rapidly increasing rural-to-urban migration. Due to the household registration system, migrants are in a vulnerable state when they attempt to access to primary care services. A strong primary care system can reduce health inequities and mitigate socioeconomic disparities in healthcare utilization. Literature indicated that migrants were more reliant on the primary care system than local residents. Although the Chinese government has attached great importance to creating an efficient health system, primary care services are still underutilized. The referral system between primary care institutions and hospitals has not yet been completely established in China. The general populations often go directly to hospitals instead of primary care institutions for their primary care. Primary care institutions generally consist of community health centers (CHCs) and community health stations (CHSs) in urban areas, and township health centers (THCs) and rural health stations (THSs) in rural areas. In addition, primary care services are also provided by the outpatient department of municipal hospitals and tertiary hospitals. A better understanding of migrants’ experiences with primary care in the above-mentioned medical institutions is critical for improving the performance of primary care institutions and providing indications of the attributes that require further attention. The purpose of this pioneering study is to explore rural-to-urban migrants’ experiences in primary care, compare their primary care experiences in four types of medical institutions in Guangzhou, China, and suggest implications for targeted interventions to improve primary care for the migrants. Methods: This was a cross-sectional study conducted with 736 rural-to-urban migrants in Guangzhou, China, in 2014. A multistage sampling method was employed. A validated Chinese version of Primary Care Assessment Tool - Adult Short Version (PCAT-AS) was used to collect information on migrants’ primary care experiences. The PCAT-AS consists of 10 domains. Analysis of covariance was conducted for comparison on PCAT domain scores and total scores among migrants accessing four types of medical institutions. Multiple linear regression models were used to explore factors associated with PCAT total scores. Results: After controlling for socio-demographic characteristics, migrant characteristics, health status and health insurance status, migrants accessing primary care in tertiary hospitals had the highest PCAT total scores when compared with those accessing primary care THCs/ RHSs (25.49 vs. 24.18, P=0.007) and CHCs/ CHSs(25.49 vs. 24.24, P=0.006). There was no statistical significant difference for PCAT total scores between migrants accessing primary care in CHCs/CHSs and those in municipal hospitals (24.24 vs. 25.02, P=0.436). Factors positively associated with higher PCAT total scores also included insurance covering parts of healthcare payment (P < 0.001). Conclusions: This study highlights the need for improvement in primary care provided by primary care institutions for rural-to-urban migrants. Migrants receiving primary care from THCs, RHSs, CHSs and CHSs reported worse primary care experiences than those receiving primary care from tertiary hospitals. Relevant policies related to medical insurance should be implemented for providing affordable healthcare services for migrants accessing primary care. Further research exploring the specific reasons for poorer PCAT scores of primary care institutions users will be needed.Keywords: China, PCAT, primary care, rural-to-urban migrants
Procedia PDF Downloads 356