Search results for: tomato yield prediction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4593

Search results for: tomato yield prediction

123 Development and Validation of a Quantitative Measure of Engagement in the Analysing Aspect of Dialogical Inquiry

Authors: Marcus Goh Tian Xi, Alicia Chua Si Wen, Eunice Gan Ghee Wu, Helen Bound, Lee Liang Ying, Albert Lee

Abstract:

The Map of Dialogical Inquiry provides a conceptual look at the underlying nature of future-oriented skills. According to the Map, learning is learner-oriented, with conversational time shifted from teachers to learners, who play a strong role in deciding what and how they learn. For example, in courses operating on the principles of Dialogical Inquiry, learners were able to leave the classroom with a deeper understanding of the topic, broader exposure to differing perspectives, and stronger critical thinking capabilities, compared to traditional approaches to teaching. Despite its contributions to learning, the Map is grounded in a qualitative approach both in its development and its application for providing feedback to learners and educators. Studies hinge on openended responses by Map users, which can be time consuming and resource intensive. The present research is motivated by this gap in practicality by aiming to develop and validate a quantitative measure of the Map. In addition, a quantifiable measure may also strengthen applicability by making learning experiences trackable and comparable. The Map outlines eight learning aspects that learners should holistically engage. This research focuses on the Analysing aspect of learning. According to the Map, Analysing has four key components: liking or engaging in logic, using interpretative lenses, seeking patterns, and critiquing and deconstructing. Existing scales of constructs (e.g., critical thinking, rationality) related to these components were identified so that the current scale could adapt items from. Specifically, items were phrased beginning with an “I”, followed by an action phrase, to fulfil the purpose of assessing learners' engagement with Analysing either in general or in classroom contexts. Paralleling standard scale development procedure, the 26-item Analysing scale was administered to 330 participants alongside existing scales with varying levels of association to Analysing, to establish construct validity. Subsequently, the scale was refined and its dimensionality, reliability, and validity were determined. Confirmatory factor analysis (CFA) revealed if scale items loaded onto the four factors corresponding to the components of Analysing. To refine the scale, items were systematically removed via an iterative procedure, according to their factor loadings and results of likelihood ratio tests at each step. Eight items were removed this way. The Analysing scale is better conceptualised as unidimensional, rather than comprising the four components identified by the Map, for three reasons: 1) the covariance matrix of the model specified for the CFA was not positive definite, 2) correlations among the four factors were high, and 3) exploratory factor analyses did not yield an easily interpretable factor structure of Analysing. Regarding validity, since the Analysing scale had higher correlations with conceptually similar scales than conceptually distinct scales, with minor exceptions, construct validity was largely established. Overall, satisfactory reliability and validity of the scale suggest that the current procedure can result in a valid and easy-touse measure for each aspect of the Map.

Keywords: analytical thinking, dialogical inquiry, education, lifelong learning, pedagogy, scale development

Procedia PDF Downloads 74
122 Three-Stage Least Squared Models of a Station-Level Subway Ridership: Incorporating an Analysis on Integrated Transit Network Topology Measures

Authors: Jungyeol Hong, Dongjoo Park

Abstract:

The urban transit system is a critical part of a solution to the economic, energy, and environmental challenges. Furthermore, it ultimately contributes the improvement of people’s quality of lives. For taking these kinds of advantages, the city of Seoul has tried to construct an integrated transit system including both subway and buses. The effort led to the fact that approximately 6.9 million citizens use the integrated transit system every day for their trips. Diagnosing the current transit network is a significant task to provide more convenient and pleasant transit environment. Therefore, the critical objective of this study is to establish a methodological framework for the analysis of an integrated bus-subway network and to examine the relationship between subway ridership and parameters such as network topology measures, bus demand, and a variety of commercial business facilities. Regarding a statistical approach to estimate subway ridership at a station level, many previous studies relied on Ordinary Least Square regression, but there was lack of studies considering the endogeneity issues which might show in the subway ridership prediction model. This study focused on both discovering the impacts of integrated transit network topology measures and endogenous effect of bus demand on subway ridership. It could ultimately contribute to developing more accurate subway ridership estimation accounting for its statistical bias. The spatial scope of the study covers Seoul city in South Korea, and it includes 243 subway stations and 10,120 bus stops with the temporal scope set during twenty-four hours with one-hour interval time panels each. The subway and bus ridership information in detail was collected from the Seoul Smart Card data in 2015 and 2016. First, integrated subway-bus network topology measures which have characteristics regarding connectivity, centrality, transitivity, and reciprocity were estimated based on the complex network theory. The results of integrated transit network topology analysis were compared to subway-only network topology. Also, the non-recursive approach which is Three-Stage Least Square was applied to develop the daily subway ridership model as capturing the endogeneity between bus and subway demands. Independent variables included roadway geometry, commercial business characteristics, social-economic characteristics, safety index, transit facility attributes, and dummies for seasons and time zone. Consequently, it was found that network topology measures were significant size effect. Especially, centrality measures showed that the elasticity was a change of 4.88% for closeness centrality, 24.48% for betweenness centrality while the elasticity of bus ridership was 8.85%. Moreover, it was proved that bus demand and subway ridership were endogenous in a non-recursive manner as showing that predicted bus ridership and predicted subway ridership is statistically significant in OLS regression models. Therefore, it shows that three-stage least square model appears to be a plausible model for efficient subway ridership estimation. It is expected that the proposed approach provides a reliable guideline that can be used as part of the spectrum of tools for evaluating a city-wide integrated transit network.

Keywords: integrated transit system, network topology measures, three-stage least squared, endogeneity, subway ridership

Procedia PDF Downloads 157
121 Statistical Models and Time Series Forecasting on Crime Data in Nepal

Authors: Dila Ram Bhandari

Abstract:

Throughout the 20th century, new governments were created where identities such as ethnic, religious, linguistic, caste, communal, tribal, and others played a part in the development of constitutions and the legal system of victim and criminal justice. Acute issues with extremism, poverty, environmental degradation, cybercrimes, human rights violations, crime against, and victimization of both individuals and groups have recently plagued South Asian nations. Everyday massive number of crimes are steadfast, these frequent crimes have made the lives of common citizens restless. Crimes are one of the major threats to society and also for civilization. Crime is a bone of contention that can create a societal disturbance. The old-style crime solving practices are unable to live up to the requirement of existing crime situations. Crime analysis is one of the most important activities of the majority of intelligent and law enforcement organizations all over the world. The South Asia region lacks such a regional coordination mechanism, unlike central Asia of Asia Pacific regions, to facilitate criminal intelligence sharing and operational coordination related to organized crime, including illicit drug trafficking and money laundering. There have been numerous conversations in recent years about using data mining technology to combat crime and terrorism. The Data Detective program from Sentient as a software company, uses data mining techniques to support the police (Sentient, 2017). The goals of this internship are to test out several predictive model solutions and choose the most effective and promising one. First, extensive literature reviews on data mining, crime analysis, and crime data mining were conducted. Sentient offered a 7-year archive of crime statistics that were daily aggregated to produce a univariate dataset. Moreover, a daily incidence type aggregation was performed to produce a multivariate dataset. Each solution's forecast period lasted seven days. Statistical models and neural network models were the two main groups into which the experiments were split. For the crime data, neural networks fared better than statistical models. This study gives a general review of the applied statistics and neural network models. A detailed image of each model's performance on the available data and generalizability is provided by a comparative analysis of all the models on a comparable dataset. Obviously, the studies demonstrated that, in comparison to other models, Gated Recurrent Units (GRU) produced greater prediction. The crime records of 2005-2019 which was collected from Nepal Police headquarter and analysed by R programming. In conclusion, gated recurrent unit implementation could give benefit to police in predicting crime. Hence, time series analysis using GRU could be a prospective additional feature in Data Detective.

Keywords: time series analysis, forecasting, ARIMA, machine learning

Procedia PDF Downloads 146
120 South African Multiple Deprivation-Concentration Index Quantiles Differentiated by Components of Success and Impediment to Tuberculosis Control Programme Using Mathematical Modelling in Rural O. R. Tambo District Health Facilities

Authors: Ntandazo Dlatu, Benjamin Longo-Mbenza, Andre Renzaho, Ruffin Appalata, Yolande Yvonne Valeria Matoumona Mavoungou, Mbenza Ben Longo, Kenneth Ekoru, Blaise Makoso, Gedeon Longo Longo

Abstract:

Background: The gap between complexities related to the integration of Tuberculosis /HIV control and evidence-based knowledge motivated the initiation of the study. Therefore, the objective of this study was to explore correlations between national TB management guidelines, multiple deprivation indexes, quantiles, components and levels of Tuberculosis control programme using mathematical modeling in rural O.R. Tambo District Health Facilities, South Africa. Methods: The study design used mixed secondary data analysis and cross-sectional analysis between 2009 and 2013 across O.R Tambo District, Eastern Cape, South Africa using univariate/ bivariate analysis, linear multiple regression models, and multivariate discriminant analysis. Health inequalities indicators and component of an impediment to the tuberculosis control programme were evaluated. Results: In total, 62 400 records for TB notification were analyzed for the period 2009-2013. There was a significant but negative between Financial Year Expenditure (r= -0.894; P= 0.041) Seropositive HIV status(r= -0.979; P= 0.004), Population Density (r = -0.881; P= 0.048) and the number of TB defaulter in all TB cases. It was shown unsuccessful control of TB management program through correlations between numbers of new PTB smear positive, TB defaulter new smear-positive, TB failure all TB, Pulmonary Tuberculosis case finding index and deprivation-concentration-dispersion index. It was shown successful TB program control through significant and negative associations between declining numbers of death in co-infection of HIV and TB, TB deaths all TB and SMIAD gradient/ deprivation-concentration-dispersion index. The multivariate linear model was summarized by unadjusted r of 96%, adjusted R2 of 95 %, Standard Error of estimate of 0.110, R2 changed of 0.959 and significance for variance change for P=0.004 to explain the prediction of TB defaulter in all TB with equation y= 8.558-0.979 x number of HIV seropositive. After adjusting for confounding factors (PTB case finding the index, TB defaulter new smear-positive, TB death in all TB, TB defaulter all TB, and TB failure in all TB). The HIV and TB death, as well as new PTB smear positive, were identified as the most important, significant, and independent indicator to discriminate most deprived deprivation index far from other deprivation quintiles 2-5 using discriminant analysis. Conclusion: Elimination of poverty such as overcrowding, lack of sanitation and environment of highest burden of HIV might end the TB threat in O.R Tambo District, Eastern Cape, South Africa. Furthermore, ongoing adequate budget comprehensive, holistic and collaborative initiative towards Sustainable Developmental Goals (SDGs) is necessary for complete elimination of TB in poor O.R Tambo District.

Keywords: tuberculosis, HIV/AIDS, success, failure, control program, health inequalities, South Africa

Procedia PDF Downloads 141
119 Cross-cultural Training in International Cooperation Efforts

Authors: Shawn Baker-Garcia, Janna O. Schaeffer

Abstract:

As the global and national communities and governments strive to address ongoing and evolving threats to humanity and pervasive or emerging “shared” global priorities on environmental, economic, political, and security, it is more urgent than ever before to understand each other, communicate effectively with one another, identify models of cooperation that yield improved, mutually reinforcing outcomes across and within cultures. It is within the backdrop of this reality that the presentation examines whether cultural training as we have approached it in recent decades is sufficiently meeting our current needs and what changes may be applied to foster better and more productive and sustainable intercultural interactions. Domestic and global relations face multiple challenges to peaceable cooperation. The last two years, in particular, have been defined by a travel-restricted COVID-19 pandemic yielding increased intercultural interactions over virtual platforms, polarized politics dividing nations and regions, and the commensurate rise in weaponized social and traditional media communication. These societal and cultural fissures are noticeably challenging our collective and individual abilities to constructively interact both at home and abroad. It is within this pressure cooker environment that the authors believe it is time to reexamine existing and broadly accepted inter- and cross- cultural training approaches and concepts to determine their level of effectiveness in setting conditions for optimal human understanding and relationships both in the national and international context. In order to better understand the amount and the type of intercultural training practitioners professionally engaging in international partnership building have received throughout their careers and its perceived effectiveness, a survey was designed and distributed to US and international professionals presently engaged in the fields of diplomacy, military, academia, and international business. The survey questions were deigned to address the two primary research questions investigators posed in this exploratory study. Research questions aimed to examine practitioners’ view of the role and effectiveness of current and traditional cultural training and education as a means to fostering improved communication, interactions, understanding, and cooperation among inter, cross, or multi-cultural communities or efforts.Responses were then collected and analyzed for themes present in the participants’ reflections. In their responses, the practitioners identified the areas of improvement and desired outcomes in regards to intercultural training and awareness raising curricular approaches. They also raised issues directly and indirectly pertaining to the role of foreign language proficiency in intercultural interactions and a need for a solid grasp on cultural and regional issues (regional expertise) to facilitate such an interaction. Respondents indicated knowledge, skills, abilities, and capabilities that the participants were not trained on but learned through ad hoc personal and professional intercultural interactions, which they found most valuable and wished they had acquired prior to the intercultural experience.

Keywords: cultural training, improved communication, intercultural competence, international cooperation

Procedia PDF Downloads 111
118 Towards Visual Personality Questionnaires Based on Deep Learning and Social Media

Authors: Pau Rodriguez, Jordi Gonzalez, Josep M. Gonfaus, Xavier Roca

Abstract:

Image sharing in social networks has increased exponentially in the past years. Officially, there are 600 million Instagrammers uploading around 100 million photos and videos per day. Consequently, there is a need for developing new tools to understand the content expressed in shared images, which will greatly benefit social media communication and will enable broad and promising applications in education, advertisement, entertainment, and also psychology. Following these trends, our work aims to take advantage of the existing relationship between text and personality, already demonstrated by multiple researchers, so that we can prove that there exists a relationship between images and personality as well. To achieve this goal, we consider that images posted on social networks are typically conditioned on specific words, or hashtags, therefore any relationship between text and personality can also be observed with those posted images. Our proposal makes use of the most recent image understanding models based on neural networks to process the vast amount of data generated by social users to determine those images most correlated with personality traits. The final aim is to train a weakly-supervised image-based model for personality assessment that can be used even when textual data is not available, which is an increasing trend. The procedure is described next: we explore the images directly publicly shared by users based on those accompanying texts or hashtags most strongly related to personality traits as described by the OCEAN model. These images will be used for personality prediction since they have the potential to convey more complex ideas, concepts, and emotions. As a result, the use of images in personality questionnaires will provide a deeper understanding of respondents than through words alone. In other words, from the images posted with specific tags, we train a deep learning model based on neural networks, that learns to extract a personality representation from a picture and use it to automatically find the personality that best explains such a picture. Subsequently, a deep neural network model is learned from thousands of images associated with hashtags correlated to OCEAN traits. We then analyze the network activations to identify those pictures that maximally activate the neurons: the most characteristic visual features per personality trait will thus emerge since the filters of the convolutional layers of the neural model are learned to be optimally activated depending on each personality trait. For example, among the pictures that maximally activate the high Openness trait, we can see pictures of books, the moon, and the sky. For high Conscientiousness, most of the images are photographs of food, especially healthy food. The high Extraversion output is mostly activated by pictures of a lot of people. In high Agreeableness images, we mostly see flower pictures. Lastly, in the Neuroticism trait, we observe that the high score is maximally activated by animal pets like cats or dogs. In summary, despite the huge intra-class and inter-class variabilities of the images associated to each OCEAN traits, we found that there are consistencies between visual patterns of those images whose hashtags are most correlated to each trait.

Keywords: emotions and effects of mood, social impact theory in social psychology, social influence, social structure and social networks

Procedia PDF Downloads 174
117 Management of Myofascial Temporomandibular Disorder in Secondary Care: A Quality Improvement Project

Authors: Rishana Bilimoria, Selina Tang, Sajni Shah, Marianne Henien, Christopher Sproat

Abstract:

Temporomandibular disorders (TMD) may affect up to a third of the general population, and there is evidence demonstrating the majority of Myofascial TMD cases improve after education and conservative measures. In 2015 our department implemented a modified care pathway for myofascial TMD patients in an attempt to improve the patient journey. This involved the use of an interactive group therapy approach to deliver education, reinforce conservative measures and promote self-management. Patient reported experience measures from the new group clinic revealed 71% patient satisfaction. This service is efficient in improving aspects of health status while reducing health-care costs and redistributing clinical time. Since its’ establishment, 52 hours of clinical time, resources and funding have been redirected effectively. This Quality Improvement Project was initiated because it was felt that this new service was being underutilised by our surgical teams. The ‘Plan-Do-Study-Act cycle’ (PDSA) framework was employed to analyse utilisation of the service: The ‘plan’ stage involved outlining our aims: to raise awareness amongst clinicians of the unified care pathway and to increase referral to this clinic. The ‘do’ stage involved collecting data from a sample of 96 patients over 4 month period to ascertain the proportion of Myofascial TMD patients who were correctly referred to the designated clinic. ‘Suitable’ patients who weren’t referred were identified. The ‘Study’ phase involved analysis of results, which revealed that 77% of suitable patients weren’t referred to the designated clinic. They were reviewed on other clinics, which are often overbooked, or managed by junior staff members. This correlated with our original prediction. Barriers to referral included: lack of awareness of the clinic, individual consultant treatment preferences and patient, reluctance to be referred to a ‘group’ clinic. The ‘Act’ stage involved presenting our findings to the team at a clinical governance meeting. This included demonstration of the clinical effectiveness of the care-pathway and explaining the referral route and criteria. In light of the evaluation results, it was decided to keep the group clinic and maximize utilisation. The second cycle of data collection following these changes revealed that of 66 Myofascial TMD patients over a 4 month period, only 9% of suitable patients were not seen via the designated pathway; therefore this QIP was successful in meeting the set objectives. Overall, employing the PDSA cycle in this QIP resulted in appropriate utilisation of the modified care pathway for patients with myofascial TMD in Guy’s Oral Surgery Department. In turn, this leads to high patient satisfaction with the service and effectively redirected 52 hours of clinical time. It permitted adoption of a collaborative working style with oral surgery colleagues to investigate problems, identify solutions, and collectively raise standards of clinical care to ensure we adopt a unified care pathway in secondary care management of Myofascial TMD patients.

Keywords: myofascial, quality Improvement, PDSA, TMD

Procedia PDF Downloads 124
116 Isolation and Transplantation of Hepatocytes in an Experimental Model

Authors: Inas Raafat, Azza El Bassiouny, Waldemar L. Olszewsky, Nagui E. Mikhail, Mona Nossier, Nora E. I. El-Bassiouni, Mona Zoheiry, Houda Abou Taleb, Noha Abd El-Aal, Ali Baioumy, Shimaa Attia

Abstract:

Background: Orthotopic liver transplantation is an established treatment for patients with severe acute and end-stage chronic liver disease. The shortage of donor organs continues to be the rate-limiting factor for liver transplantation throughout the world. Hepatocyte transplantation is a promising treatment for several liver diseases and can, also, be used as a "bridge" to liver transplantation in cases of liver failure. Aim of the work: This study was designed to develop a highly efficient protocol for isolation and transplantation of hepatocytes in experimental Lewis rat model to provide satisfactory guidelines for future application on humans.Materials and Methods: Hepatocytes were isolated from the liver by double perfusion technique and bone marrow cells were isolated by centrifugation of shafts of tibia and femur of donor Lewis rats. Recipient rats were subjected to sub-lethal dose of irradiation 2 days before transplantation. In a laparotomy operation the spleen was injected by freshly isolated hepatocytes and bone marrow cells were injected intravenously. The animals were sacrificed 45 day latter and splenic sections were prepared and stained with H & E, PAS AFP and Prox1. Results: The data obtained from this study showed that the double perfusion technique is successful in separation of hepatocytes regarding cell number and viability. Also the method used for bone marrow cells separation gave excellent results regarding cell number and viability. Intrasplenic engraftment of hepatocytes and live tissue formation within the splenic tissue were found in 70% of cases. Hematoxylin and eosin stained splenic sections from 7 rats showed sheets and clusters of cells among the splenic tissues. Periodic Acid Schiff stained splenic sections from 7 rats showed clusters of hepatocytes with intensely stained pink cytoplasmic granules denoting the presence of glycogen. Splenic sections from 7 rats stained with anti-α-fetoprotein antibody showed brownish cytoplasmic staining of the hepatocytes denoting positive expression of AFP. Splenic sections from 7 rats stained with anti-Prox1 showed brownish nuclear staining of the hepatocytes denoting positive expression of Prox1 gene on these cells. Also, positive expression of Prox1 gene was detected on lymphocytes aggregations in the spleens. Conclusions: Isolation of liver cells by double perfusion technique using collagenase buffer is a reliable method that has a very satisfactory yield regarding cell number and viability. The intrasplenic route of transplantation of the freshly isolated liver cells in an immunocompromised model was found to give good results regarding cell engraftment and tissue formation. Further studies are needed to assess function of engrafted hepatocytes by measuring prothrombin time, serum albumin and bilirubin levels.

Keywords: Lewis rats, hepatocytes, BMCs, transplantation, AFP, Prox1

Procedia PDF Downloads 296
115 Predicting Suicidal Behavior by an Accurate Monitoring of RNA Editing Biomarkers in Blood Samples

Authors: Berengere Vire, Nicolas Salvetat, Yoann Lannay, Guillaume Marcellin, Siem Van Der Laan, Franck Molina, Dinah Weissmann

Abstract:

Predicting suicidal behaviors is one of the most complex challenges of daily psychiatric practices. Today, suicide risk prediction using biological tools is not validated and is only based on subjective clinical reports of the at-risk individual. Therefore, there is a great need to identify biomarkers that would allow early identification of individuals at risk of suicide. Alterations of adenosine-to-inosine (A-to-I) RNA editing of neurotransmitter receptors and other proteins have been shown to be involved in etiology of different psychiatric disorders and linked to suicidal behavior. RNA editing is a co- or post-transcriptional process leading to a site-specific alteration in RNA sequences. It plays an important role in the epi transcriptomic regulation of RNA metabolism. On postmortem human brain tissue (prefrontal cortex) of depressed suicide victims, Alcediag found specific alterations of RNA editing activity on the mRNA coding for the serotonin 2C receptor (5-HT2cR). Additionally, an increase in expression levels of ADARs, the RNA editing enzymes, and modifications of RNA editing profiles of prime targets, such as phosphodiesterase 8A (PDE8A) mRNA, have also been observed. Interestingly, the PDE8A gene is located on chromosome 15q25.3, a genomic region that has recurrently been associated with the early-onset major depressive disorder (MDD). In the current study, we examined whether modifications in RNA editing profile of prime targets allow identifying disease-relevant blood biomarkers and evaluating suicide risk in patients. To address this question, we performed a clinical study to identify an RNA editing signature in blood of depressed patients with and without the history of suicide attempts. Patient’s samples were drawn in PAXgene tubes and analyzed on Alcediag’s proprietary RNA editing platform using next generation sequencing technology. In addition, gene expression analysis by quantitative PCR was performed. We generated a multivariate algorithm comprising various selected biomarkers to detect patients with a high risk to attempt suicide. We evaluated the diagnostic performance using the relative proportion of PDE8A mRNA editing at different sites and/or isoforms as well as the expression of PDE8A and the ADARs. The significance of these biomarkers for suicidality was evaluated using the area under the receiver-operating characteristic curve (AUC). The generated algorithm comprising the biomarkers was found to have strong diagnostic performances with high specificity and sensitivity. In conclusion, we developed tools to measure disease-specific biomarkers in blood samples of patients for identifying individuals at the greatest risk for future suicide attempts. This technology not only fosters patient management but is also suitable to predict the risk of drug-induced psychiatric side effects such as iatrogenic increase of suicidal ideas/behaviors.

Keywords: blood biomarker, next-generation-sequencing, RNA editing, suicide

Procedia PDF Downloads 237
114 Estimating Poverty Levels from Satellite Imagery: A Comparison of Human Readers and an Artificial Intelligence Model

Authors: Ola Hall, Ibrahim Wahab, Thorsteinn Rognvaldsson, Mattias Ohlsson

Abstract:

The subfield of poverty and welfare estimation that applies machine learning tools and methods on satellite imagery is a nascent but rapidly growing one. This is in part driven by the sustainable development goal, whose overarching principle is that no region is left behind. Among other things, this requires that welfare levels can be accurately and rapidly estimated at different spatial scales and resolutions. Conventional tools of household surveys and interviews do not suffice in this regard. While they are useful for gaining a longitudinal understanding of the welfare levels of populations, they do not offer adequate spatial coverage for the accuracy that is needed, nor are their implementation sufficiently swift to gain an accurate insight into people and places. It is this void that satellite imagery fills. Previously, this was near-impossible to implement due to the sheer volume of data that needed processing. Recent advances in machine learning, especially the deep learning subtype, such as deep neural networks, have made this a rapidly growing area of scholarship. Despite their unprecedented levels of performance, such models lack transparency and explainability and thus have seen limited downstream applications as humans generally are apprehensive of techniques that are not inherently interpretable and trustworthy. While several studies have demonstrated the superhuman performance of AI models, none has directly compared the performance of such models and human readers in the domain of poverty studies. In the present study, we directly compare the performance of human readers and a DL model using different resolutions of satellite imagery to estimate the welfare levels of demographic and health survey clusters in Tanzania, using the wealth quintile ratings from the same survey as the ground truth data. The cluster-level imagery covers all 608 cluster locations, of which 428 were classified as rural. The imagery for the human readers was sourced from the Google Maps Platform at an ultra-high resolution of 0.6m per pixel at zoom level 18, while that of the machine learning model was sourced from the comparatively lower resolution Sentinel-2 10m per pixel data for the same cluster locations. Rank correlation coefficients of between 0.31 and 0.32 achieved by the human readers were much lower when compared to those attained by the machine learning model – 0.69-0.79. This superhuman performance by the model is even more significant given that it was trained on the relatively lower 10-meter resolution satellite data while the human readers estimated welfare levels from the higher 0.6m spatial resolution data from which key markers of poverty and slums – roofing and road quality – are discernible. It is important to note, however, that the human readers did not receive any training before ratings, and had this been done, their performance might have improved. The stellar performance of the model also comes with the inevitable shortfall relating to limited transparency and explainability. The findings have significant implications for attaining the objective of the current frontier of deep learning models in this domain of scholarship – eXplainable Artificial Intelligence through a collaborative rather than a comparative framework.

Keywords: poverty prediction, satellite imagery, human readers, machine learning, Tanzania

Procedia PDF Downloads 78
113 Unravelling Glyphosates Disruptive Effects on the Photochemical Efficiency of Amaranthus cruentus

Authors: Jacques M. Berner, Lehlogonolo Maloma

Abstract:

Context: Glyphosate, a widely used herbicide, has raised concerns about its impact on various crops. Amaranthus cruentus, an important grain crop species, is particularly susceptible to glyphosate. Understanding the specific disruptions caused by glyphosate on the photosynthetic process in Amaranthus cruentus is crucial for assessing its effects on crop productivity and ecological sustainability. Research Aim: This study aimed to investigate the dose-dependent impact of glyphosate on the photochemical efficiency of Amaranthus cruentus using the OJIP transient analysis. The goal was to assess the specific disruptions caused by glyphosate on key parameters of photosystem II. Methodology: The experiment was conducted in a controlled greenhouse environment. Amaranthus cruentus plants were exposed to different concentrations of glyphosate, including half, recommended, and double the recommended application rates. The photochemical efficiency of the plants was evaluated using non-invasive chlorophyll a fluorescence measurements and subsequent analysis of OJIP transients. Measurements were taken on 1-hour dark-adapted leaves using a Hansatech Handy PEA+ chlorophyll fluorimeter. Findings: The study's results demonstrated a significant reduction in the photochemical efficiency of Amaranthus cruentus following glyphosate treatment. The OJIP transients showed distinct alterations in the glyphosate-treated plants compared to the control group. These changes included a decrease in maximal fluorescence (FP) and a delay in the rise of the fluorescence signal, indicating impairment in the energy conversion process within the photosystem II. Glyphosate exposure also led to a substantial decrease in the maximum quantum yield efficiency of photosystem II (FV/FM) and the total performance index (PItotal), which reflects the overall photochemical efficiency of photosystem II. These reductions in photochemical efficiency were observed even at half the recommended dose of glyphosate. Theoretical Importance: The study provides valuable insights into the specific disruptions caused by glyphosate on the photochemical efficiency of Amaranthus cruentus. Data Collection and Analysis Procedures: Data collection involved non-invasive chlorophyll a fluorescence measurements using a chlorophyll fluorimeter on dark-adapted leaves. The OJIP transients were then analyzed to assess specific disruptions in key parameters of photosystem II. Statistical analysis was conducted to determine the significance of the differences observed between glyphosate-treated plants and the control group. Question Addressed: The study aimed to address the question of how glyphosate exposure affects the photochemical efficiency of Amaranthus cruentus, specifically examining disruptions in the photosynthetic electron transport chain and overall photochemical efficiency. Conclusion: The study demonstrates that glyphosate severely impairs the photochemical efficiency of Amaranthus cruentus, as indicated by the alterations in OJIP transients. Even at half the recommended dose, glyphosate caused significant reductions in photochemical efficiency. These findings highlight the detrimental effects of glyphosate on crop productivity and emphasize the need for further research to evaluate its long-term consequences and ecological implications in agriculture. The authors gratefully acknowledge the support from North-West University for making this research possible.

Keywords: glyphosate, amaranthus cruentus, ojip transient analysis, pitotal, photochemical efficiency, chlorophyll fluorescence, weeds

Procedia PDF Downloads 70
112 Catalytic Dehydrogenation of Formic Acid into H2/CO2 Gas: A Novel Approach

Authors: Ayman Hijazi, Witold Kwapinski, J. J. Leahy

Abstract:

Finding a sustainable alternative energy to fossil fuel is an urgent need as various environmental challenges in the world arise. Therefore, formic acid (FA) decomposition has been an attractive field that lies at the center of biomass platform, comprising a potential pool of hydrogen energy that stands as a new energy vector. Liquid FA features considerable volumetric energy density of 6.4 MJ/L and a specific energy density of 5.3 MJ/Kg that qualifies it in the prime seat as an energy source for transportation infrastructure. Additionally, the increasing research interest in FA decomposition is driven by the need of in-situ H2 production, which plays a key role in the hydrogenation reactions of biomass into higher value components. It is reported elsewhere in literature that catalytic decomposition of FA is usually performed in poorly designed setup using simple glassware under magnetic stirring, thus demanding further energy investment to retain the used catalyst. it work suggests an approach that integrates designing a novel catalyst featuring magnetic property with a robust setup that minimizes experimental & measurement discrepancies. One of the most prominent active species for dehydrogenation/hydrogenation of biomass compounds is palladium. Accordingly, we investigate the potential of engrafting palladium metal onto functionalized magnetic nanoparticles as a heterogeneous catalyst to favor the production of CO-free H2 gas from FA. Using ordinary magnet to collect the spent catalyst renders core-shell magnetic nanoparticles as the backbone of the process. Catalytic experiments were performed in a jacketed batch reactor equipped with an overhead stirrer under inert medium. Through a novel approach, FA is charged into the reactor via high-pressure positive displacement pump at steady state conditions. The produced gas (H2+CO2) was measured by connecting the gas outlet to a measuring system based on the amount of the displaced water. The novelty of this work lies in designing a very responsive catalyst, pumping consistent amount of FA into a sealed reactor running at steady state mild temperatures, and continuous gas measurement, along with collecting the used catalyst without the need for centrifugation. Catalyst characterization using TEM, XRD, SEM, and CHN elemental analyzer provided us with details of catalyst preparation and facilitated new venues to alter the nanostructure of the catalyst framework. Consequently, the introduction of amine groups has led to appreciable improvements in terms of dispersion of the doped metals and eventually attaining nearly complete conversion (100%) of FA after 7 hours. The relative importance of the process parameters such as temperature (35-85°C), stirring speed (150-450rpm), catalyst loading (50-200mgr.), and Pd doping ratio (0.75-1.80wt.%) on gas yield was assessed by a Taguchi design-of-experiment based model. Experimental results showed that operating at lower temperature range (35-50°C) yielded more gas while the catalyst loading and Pd doping wt.% were found to be the most significant factors with a P-values 0.026 & 0.031, respectively.

Keywords: formic acid decomposition, green catalysis, hydrogen, mesoporous silica, process optimization, nanoparticles

Procedia PDF Downloads 29
111 Automation of Finite Element Simulations for the Design Space Exploration and Optimization of Type IV Pressure Vessel

Authors: Weili Jiang, Simon Cadavid Lopera, Klaus Drechsler

Abstract:

Fuel cell vehicle has become the most competitive solution for the transportation sector in the hydrogen economy. Type IV pressure vessel is currently the most popular and widely developed technology for the on-board storage, based on their high reliability and relatively low cost. Due to the stringent requirement on mechanical performance, the pressure vessel is subject to great amount of composite material, a major cost driver for the hydrogen tanks. Evidently, the optimization of composite layup design shows great potential in reducing the overall material usage, yet requires comprehensive understanding on underlying mechanisms as well as the influence of different design parameters on mechanical performance. Given the type of materials and manufacturing processes by which the type IV pressure vessels are manufactured, the design and optimization are a nuanced subject. The manifold of stacking sequence and fiber orientation variation possibilities have an out-standing effect on vessel strength due to the anisotropic property of carbon fiber composites, which make the design space high dimensional. Each variation of design parameters requires computational resources. Using finite element analysis to evaluate different designs is the most common method, however, the model-ing, setup and simulation process can be very time consuming and result in high computational cost. For this reason, it is necessary to build a reliable automation scheme to set up and analyze the di-verse composite layups. In this research, the simulation process of different tank designs regarding various parameters is conducted and automatized in a commercial finite element analysis framework Abaqus. Worth mentioning, the modeling of the composite overwrap is automatically generated using an Abaqus-Python scripting interface. The prediction of the winding angle of each layer and corresponding thickness variation on dome region is the most crucial step of the modeling, which is calculated and implemented using analytical methods. Subsequently, these different composites layups are simulated as axisymmetric models to facilitate the computational complexity and reduce the calculation time. Finally, the results are evaluated and compared regarding the ultimate tank strength. By automatically modeling, evaluating and comparing various composites layups, this system is applicable for the optimization of the tanks structures. As mentioned above, the mechanical property of the pressure vessel is highly dependent on composites layup, which requires big amount of simulations. Consequently, to automatize the simulation process gains a rapid way to compare the various designs and provide an indication of the optimum one. Moreover, this automation process can also be operated for creating a data bank of layups and corresponding mechanical properties with few preliminary configuration steps for the further case analysis. Subsequently, using e.g. machine learning to gather the optimum by the data pool directly without the simulation process.

Keywords: type IV pressure vessels, carbon composites, finite element analy-sis, automation of simulation process

Procedia PDF Downloads 106
110 Balancing Biodiversity and Agriculture: A Broad-Scale Analysis of the Land Sparing/Land Sharing Trade-Off for South African Birds

Authors: Chevonne Reynolds, Res Altwegg, Andrew Balmford, Claire N. Spottiswoode

Abstract:

Modern agriculture has revolutionised the planet’s capacity to support humans, yet has simultaneously had a greater negative impact on biodiversity than any other human activity. Balancing the demand for food with the conservation of biodiversity is one of the most pressing issues of our time. Biodiversity-friendly farming (‘land sharing’), or alternatively, separation of conservation and production activities (‘land sparing’), are proposed as two strategies for mediating the trade-off between agriculture and biodiversity. However, there is much debate regarding the efficacy of each strategy, as this trade-off has typically been addressed by short term studies at fine spatial scales. These studies ignore processes that are relevant to biodiversity at larger scales, such as meta-population dynamics and landscape connectivity. Therefore, to better understand species response to agricultural land-use and provide evidence to underpin the planning of better production landscapes, we need to determine the merits of each strategy at larger scales. In South Africa, a remarkable citizen science project - the South African Bird Atlas Project 2 (SABAP2) – collates an extensive dataset describing the occurrence of birds at a 5-min by 5-min grid cell resolution. We use these data, along with fine-resolution data on agricultural land-use, to determine which strategy optimises the agriculture-biodiversity trade-off in a southern African context, and at a spatial scale never considered before. To empirically test this trade-off, we model bird species population density, derived for each 5-min grid cell by Royle-Nicols single-species occupancy modelling, against both the amount and configuration of different types of agricultural production in the same 5-min grid cell. In using both production amount and configuration, we can show not only how species population densities react to changes in yield, but also describe the production landscape patterns most conducive to conservation. Furthermore, the extent of both the SABAP2 and land-cover datasets allows us to test this trade-off across multiple regions to determine if bird populations respond in a consistent way and whether results can be extrapolated to other landscapes. We tested the land sparing/sharing trade-off for 281 bird species across three different biomes in South Africa. Overall, a higher proportion of species are classified as losers, and would benefit from land sparing. However, this proportion of loser-sparers is not consistent and varies across biomes and the different types of agricultural production. This is most likely because of differences in the intensity of agricultural land-use and the interactions between the differing types of natural vegetation and agriculture. Interestingly, we observe a higher number of species that benefit from agriculture than anticipated, suggesting that agriculture is a legitimate resource for certain bird species. Our results support those seen at smaller scales and across vastly different agricultural systems, that land sparing benefits the most species. However, our analysis suggests that land sparing needs to be implemented at spatial scales much larger than previously considered. Species persistence in agricultural landscapes will require the conservation of large tracts of land, and is an important consideration in developing countries, which are undergoing rapid agricultural development.

Keywords: agriculture, birds, land sharing, land sparing

Procedia PDF Downloads 190
109 Testicular Differential MicroRNA Expression Derived Occupational Risk Factor Assessment in Idiopathic Non-obstructive Azoospermia Cases

Authors: Nisha Sharma, Mili Kaur, Ashutosh Halder, Seema Kaushal, Manoj Kumar, Manish Jain

Abstract:

Purpose: To investigate microRNAs (miRNA) as an epigenomic etiological factor in idiopathic non-obstructive azoospermia (NOA). In order to achieve the same, an association was seen between occupational exposure to radiation, thermal, and chemical factors and idiopathic cases of non-obstructive azoospermia, and later, testicular differential miRNA expression profiling was done in exposure group NOA cases. Method: It is a prospective study in which 200 apparent idiopathic male factor infertility cases, who have been advised to undergo testicular fine needle aspiration (FNA) evaluation, are recruited. A detailed occupational history was taken to understand the possible type of exposure due to the nature and duration of work. A total of 26 patients were excluded upon XY-FISH and Yq microdeletion tests due to the presence of genetic causes of infertility, 6 hypospermatogeneis (HS), six Sertoli cell-only syndrome (SCOS), and six normospermatogeneis patients testicular FNA samples were used for RNA isolation followed by small RNA sequencing and nCounter miRNA expression analysis. Differential miRNA expression profile of HS and SCOS patients was done. A web-based tool, miRNet, was used to predict the interacting compounds or chemicals using the shortlisted miRNAs with high fold change. The major limitation encountered in this study was the insufficient quantity of testicular FNA sample used for total RNA isolation, which resulted in a low yield and RNA integrity number (RIN) value. Therefore, the number of RNA samples admissible for differential miRNA expression analysis was very small in comparison to the total number of patients recruited. Results: Differential expression analysis revealed 69 down-regulated and 40 up-regulated miRNAs in HS and 66 down-regulated and 33 up-regulated miRNAs in SCOS in comparison to normospermatogenesis controls. The miRNA interaction analysis using the miRNet tool showed that the differential expression profiles of HS and SCOS patients were associated with arsenic trioxide, bisphenol-A, calcium sulphate, lithium, and cadmium. These compounds are reproductive toxins and might be responsible for miRNA-mediated epigenetic deregulation leading to NOA. The association between occupational risk factor exposure and the non-exposure group of NOA patients was not statistically significant, with ꭓ2 (3, N= 178) = 6.70, p= 0.082. The association between individual exposure groups (radiation, thermal, and chemical) and various sub-types of NOA is also not significant, with ꭓ2 (9, N= 178) = 15.06, p= 0.089. Functional analysis of HS and SCOS patients' miRNA profiles revealed some important miR-family members in terms of male fertility. The miR-181 family plays a role in the differentiation of spermatogonia and spermatocytes, as well as the transcriptional regulation of haploid germ cells. The miR-34 family is expressed in spermatocytes and round spermatids and is involved in the regulation of SSCs differentiation. Conclusion: The reproductive toxins might adopt the miRNA-mediated mechanism of disease development in idiopathic cases of NOA. Chemical compound induced; miRNA-mediated epigenetic deregulation can give a future perspective on the etiopathogenesis of the disease.

Keywords: microRNA, non-obstructive azoospermia (NOA), occupational exposure, hypospermatogenesis (HS), Sertoli cell only syndrome (SCOS)

Procedia PDF Downloads 66
108 Surviral: An Agent-Based Simulation Framework for Sars-Cov-2 Outcome Prediction

Authors: Sabrina Neururer, Marco Schweitzer, Werner Hackl, Bernhard Tilg, Patrick Raudaschl, Andreas Huber, Bernhard Pfeifer

Abstract:

History and the current outbreak of Covid-19 have shown the deadly potential of infectious diseases. However, infectious diseases also have a serious impact on areas other than health and healthcare, such as the economy or social life. These areas are strongly codependent. Therefore, disease control measures, such as social distancing, quarantines, curfews, or lockdowns, have to be adopted in a very considerate manner. Infectious disease modeling can support policy and decision-makers with adequate information regarding the dynamics of the pandemic and therefore assist in planning and enforcing appropriate measures that will prevent the healthcare system from collapsing. In this work, an agent-based simulation package named “survival” for simulating infectious diseases is presented. A special focus is put on SARS-Cov-2. The presented simulation package was used in Austria to model the SARS-Cov-2 outbreak from the beginning of 2020. Agent-based modeling is a relatively recent modeling approach. Since our world is getting more and more complex, the complexity of the underlying systems is also increasing. The development of tools and frameworks and increasing computational power advance the application of agent-based models. For parametrizing the presented model, different data sources, such as known infections, wastewater virus load, blood donor antibodies, circulating virus variants and the used capacity for hospitalization, as well as the availability of medical materials like ventilators, were integrated with a database system and used. The simulation result of the model was used for predicting the dynamics and the possible outcomes and was used by the health authorities to decide on the measures to be taken in order to control the pandemic situation. The survival package was implemented in the programming language Java and the analytics were performed with R Studio. During the first run in March 2020, the simulation showed that without measures other than individual personal behavior and appropriate medication, the death toll would have been about 27 million people worldwide within the first year. The model predicted the hospitalization rates (standard and intensive care) for Tyrol and South Tyrol with an accuracy of about 1.5% average error. They were calculated to provide 10-days forecasts. The state government and the hospitals were provided with the 10-days models to support their decision-making. This ensured that standard care was maintained for as long as possible without restrictions. Furthermore, various measures were estimated and thereafter enforced. Among other things, communities were quarantined based on the calculations while, in accordance with the calculations, the curfews for the entire population were reduced. With this framework, which is used in the national crisis team of the Austrian province of Tyrol, a very accurate model could be created on the federal state level as well as on the district and municipal level, which was able to provide decision-makers with a solid information basis. This framework can be transferred to various infectious diseases and thus can be used as a basis for future monitoring.

Keywords: modelling, simulation, agent-based, SARS-Cov-2, COVID-19

Procedia PDF Downloads 158
107 Implementation of Synthesis and Quality Control Procedures of ¹⁸F-Fluoromisonidazole Radiopharmaceutical

Authors: Natalia C. E. S. Nascimento, Mercia L. Oliveira, Fernando R. A. Lima, Leonardo T. C. do Nascimento, Marina B. Silveira, Brigida G. A. Schirmer, Andrea V. Ferreira, Carlos Malamut, Juliana B. da Silva

Abstract:

Tissue hypoxia is a common characteristic of solid tumors leading to decreased sensitivity to radiotherapy and chemotherapy. In the clinical context, tumor hypoxia assessment employing the positron emission tomography (PET) tracer ¹⁸F-fluoromisonidazole ([¹⁸F]FMISO) is helpful for physicians for planning and therapy adjusting. The aim of this work was to implement the synthesis of 18F-FMISO in a TRACERlab® MXFDG module and also to establish the quality control procedure. [¹⁸F]FMISO was synthesized at Centro de Desenvolvimento da Tecnologia Nuclear (CDTN/CNEN/Brazil) using an automated synthesizer (TRACERlab® MXFDG, GE) adapted for the production of [¹⁸F]FMISO. The FMISO chemical standard was purchased from ABX. 18O- enriched water was acquired from Center of Molecular Research. Reagent kits containing eluent solution, acetonitrile, ethanol, 2.0 M HCl solution, buffer solution, water for injections and [¹⁸F]FMISO precursor (dissolved in 2 ml acetonitrile) were purchased from ABX. The [¹⁸F]FMISO samples were purified by Solid Phase Extraction method. The quality requirements of [¹⁸F]FMISO are established in the European Pharmacopeia. According to that reference, quality control of [¹⁸F]FMISO should include appearance, pH, radionuclidic identity and purity, radiochemical identity and purity, chemical purity, residual solvents, bacterial endotoxins, and sterility. The duration of the synthesis process was 53 min, with radiochemical yield of (37.00 ± 0.01) % and the specific activity was more than 70 GBq/µmol. The syntheses were reproducible and showed satisfactory results. In relation to the quality control analysis, the samples were clear and colorless at pH 6.0. The spectrum emission, measured by using a High-Purity Germanium Detector (HPGe), presented a single peak at 511 keV and the half-life, determined by the decay method in an activimeter, was (111.0 ± 0.5) min, indicating no presence of radioactive contaminants, besides the desirable radionuclide (¹⁸F). The samples showed concentration of tetrabutylammonium (TBA) < 50μg/mL, assessed by visual comparison to TBA standard applied in the same thin layer chromatographic plate. Radiochemical purity was determined by high performance liquid chromatography (HPLC) and the results were 100%. Regarding the residual solvents tested, ethanol and acetonitrile presented concentration lower than 10% and 0.04%, respectively. Healthy female mice were injected via lateral tail vein with [¹⁸F]FMISO, microPET imaging studies (15 min) were performed after 2 h post injection (p.i), and the biodistribution was analyzed in five-time points (30, 60, 90, 120 and 180 min) after injection. Subsequently, organs/tissues were assayed for radioactivity with a gamma counter. All parameters of quality control test were in agreement to quality criteria confirming that [¹⁸F]FMISO was suitable for use in non-clinical and clinical trials, following the legal requirements for the production of new radiopharmaceuticals in Brazil.

Keywords: automatic radiosynthesis, hypoxic tumors, pharmacopeia, positron emitters, quality requirements

Procedia PDF Downloads 176
106 Clinical Presentation and Immune Response to Intramammary Infection of Holstein-Friesian Heifers with Isolates from Two Staphylococcus aureus Lineages

Authors: Dagmara A. Niedziela, Mark P. Murphy, Orla M. Keane, Finola C. Leonard

Abstract:

Staphylococcus aureus is the most frequent cause of clinical and subclinical bovine mastitis in Ireland. Mastitis caused by S. aureus is often chronic and tends to recur after antibiotic treatment. This may be due to several virulence factors, including attributes that enable the bacterium to internalize into bovine mammary epithelial cells, where it may evade antibiotic treatment, or evade the host immune response. Four bovine-adapted lineages (CC71, CC97, CC151 and ST136) were identified among a collection of Irish S. aureus mastitis isolates. Genotypic variation of mastitis-causing strains may contribute to different presentations of the disease, including differences in milk somatic cell count (SCC), the main method of mastitis detection. The objective of this study was to investigate the influence of bacterial strain and lineage on host immune response, by employing cell culture methods in vitro as well as an in vivo infection model. Twelve bovine adapted S. aureus strains were examined for internalization into bovine mammary epithelial cells (bMEC) and their ability to induce an immune response from bMEC (using qPCR and ELISA). In vitro studies found differences in a variety of virulence traits between the lineages. Strains from lineages CC97 and CC71 internalized more efficiently into bovine mammary epithelial cells (bMEC) than CC151 and ST136. CC97 strains also induced immune genes in bMEC more strongly than strains from the other 3 lineages. One strain each of CC151 and CC97 that differed in their ability to cause an immune response in bMEC were selected on the basis of the above in vitro experiments. Fourteen first-lactation Holstein-Friesian cows were purchased from 2 farms on the basis of low SCC (less than 50 000 cells/ml) and infection free status. Seven cows were infected with 1.73 x 102 c.f.u. of the CC97 strain (Group 1) and another seven with 5.83 x 102 c.f.u. of the CC151 strain (Group 2). The contralateral quarter of each cow was inoculated with PBS (vehicle). Clinical signs of infection (temperature, milk and udder appearance, milk yield) were monitored for 30 days. Blood and milk samples were taken to determine bacterial counts in milk, SCC, white blood cell populations and cytokines. Differences in disease presentation in vivo between groups were observed, with two animals from Group 2 developing clinical mastitis and requiring antibiotic treatment, while one animal from Group 1 did not develop an infection for the duration of the study. Fever (temperature > 39.5⁰C) was observed in 3 animals from Group 2 and in none from Group 1. Significant differences in SCC and bacterial load between groups were observed in the initial stages of infection (week 1). Data is also being collected on cytokines and chemokines secreted during the course of infection. The results of this study suggest that a strain from lineage CC151 may cause more severe clinical mastitis, while a strain from lineage CC97 may cause mild, subclinical mastitis. Diversity between strains of S. aureus may therefore influence the clinical presentation of mastitis, which in turn may influence disease detection and treatment needs.

Keywords: Bovine mastitis, host immune response, host-pathogen interactions, Staphylococcus aureus

Procedia PDF Downloads 140
105 An Argument for Agile, Lean, and Hybrid Project Management in Museum Conservation Practice: A Qualitative Evaluation of the Morris Collection Conservation Project at the Sainsbury Centre for Visual Arts

Authors: Maria Ledinskaya

Abstract:

This paper is part case study and part literature review. It seeks to introduce Agile, Lean, and Hybrid project management concepts from business, software development, and manufacturing fields to museum conservation by looking at their practical application on a recent conservation project at the Sainsbury Centre for Visual Arts. The author outlines the advantages of leaner and more agile conservation practices in today’s faster, less certain, and more budget-conscious museum climate where traditional project structures are no longer as relevant or effective. The Morris Collection Conservation Project was carried out in 2019-2021 in Norwich, UK, and concerned the remedial conservation of around 150 Abstract Constructivist artworks bequeathed to the Sainsbury Centre by private collectors Michael and Joyce Morris. It was a medium-sized conservation project of moderate complexity, planned and delivered in an environment with multiple known unknowns – unresearched collection, unknown conditions and materials, unconfirmed budget. The project was later impacted by the COVID-19 pandemic, introducing indeterminate lockdowns, budget cuts, staff changes, and the need to accommodate social distancing and remote communications. The author, then a staff conservator at the Sainsbury Centre who acted as project manager on the Morris Project, presents an incremental, iterative, and value-based approach to managing a conservation project in an uncertain environment. The paper examines the project from the point of view of Traditional, Agile, Lean, and Hybrid project management. The author argues that most academic writing on project management in conservation has focussed on a Traditional plan-driven approach – also known as Waterfall project management – which has significant drawbacks in today’s museum environment due to its over-reliance on prediction-based planning and its low tolerance to change. In the last 20 years, alternative Agile, Lean and Hybrid approaches to project management have been widely adopted in software development, manufacturing, and other industries, although their recognition in the museum sector has been slow. Using examples from the Morris Project, the author introduces key principles and tools of Agile, Lean, and Hybrid project management and presents a series of arguments on the effectiveness of these alternative methodologies in museum conservation, including the ethical and practical challenges to their implementation. These project management approaches are discussed in the context of consequentialist, relativist, and utilitarian developments in contemporary conservation ethics. Although not intentionally planned as such, the Morris Project had a number of Agile and Lean features which were instrumental to its successful delivery. These key features are identified as distributed decision-making, a co-located cross-disciplinary team, servant leadership, focus on value-added work, flexible planning done in shorter sprint cycles, light documentation, and emphasis on reducing procedural, financial, and logistical waste. Overall, the author’s findings point in favour of a hybrid model, which combines traditional and alternative project processes and tools to suit the specific needs of the project.

Keywords: agile project management, conservation, hybrid project management, lean project management, waterfall project management

Procedia PDF Downloads 56
104 Catalytic Decomposition of Formic Acid into H₂/CO₂ Gas: A Distinct Approach

Authors: Ayman Hijazi, Witold Kwapinski, J. J. Leahy

Abstract:

Finding a sustainable alternative energy to fossil fuel is an urgent need as various environmental challenges in the world arise. Therefore, formic acid (FA) decomposition has been an attractive field that lies at the center of the biomass platform, comprising a potential pool of hydrogen energy that stands as a distinct energy vector. Liquid FA features considerable volumetric energy density of 6.4 MJ/L and a specific energy density of 5.3 MJ/Kg that qualifies it in the prime seat as an energy source for transportation infrastructure. Additionally, the increasing research interest in FA decomposition is driven by the need for in-situ H₂ production, which plays a key role in the hydrogenation reactions of biomass into higher-value components. It is reported elsewhere in the literature that catalytic decomposition of FA is usually performed in poorly designed setups using simple glassware under magnetic stirring, thus demanding further energy investment to retain the used catalyst. Our work suggests an approach that integrates designing a distinct catalyst featuring magnetic properties with a robust setup that minimizes experimental & measurement discrepancies. One of the most prominent active species for the dehydrogenation/hydrogenation of biomass compounds is palladium. Accordingly, we investigate the potential of engrafting palladium metal onto functionalized magnetic nanoparticles as a heterogeneous catalyst to favor the production of CO-free H₂ gas from FA. Using an ordinary magnet to collect the spent catalyst renders core-shell magnetic nanoparticles as the backbone of the process. Catalytic experiments were performed in a jacketed batch reactor equipped with an overhead stirrer under an inert medium. Through a distinct approach, FA is charged into the reactor via a high-pressure positive displacement pump at steady-state conditions. The produced gas (H₂+CO₂) was measured by connecting the gas outlet to a measuring system based on the amount of the displaced water. The uniqueness of this work lies in designing a very responsive catalyst, pumping a consistent amount of FA into a sealed reactor running at steady-state mild temperatures, and continuous gas measurement, along with collecting the used catalyst without the need for centrifugation. Catalyst characterization using TEM, XRD, SEM, and CHN elemental analyzer provided us with details of catalyst preparation and facilitated new venues to alter the nanostructure of the catalyst framework. Consequently, the introduction of amine groups has led to appreciable improvements in terms of dispersion of the doped metals and eventually attaining nearly complete conversion (100%) of FA after 7 hours. The relative importance of the process parameters such as temperature (35-85°C), stirring speed (150-450rpm), catalyst loading (50-200mgr.), and Pd doping ratio (0.75-1.80wt.%) on gas yield was assessed by a Taguchi design-of-experiment based model. Experimental results showed that operating at a lower temperature range (35-50°C) yielded more gas, while the catalyst loading and Pd doping wt.% were found to be the most significant factors with P-values 0.026 & 0.031, respectively.

Keywords: formic acid decomposition, green catalysis, hydrogen, mesoporous silica, process optimization, nanoparticles

Procedia PDF Downloads 31
103 Bacteriophages for Sustainable Wastewater Treatment: Application in Black Water Decontamination with an Emphasis to DRDO Biotoilet

Authors: Sonika Sharma, Mohan G. Vairale, Sibnarayan Datta, Soumya Chatterjee, Dharmendra Dubey, Rajesh Prasad, Raghvendra Budhauliya, Bidisha Das, Vijay Veer

Abstract:

Bacteriophages are viruses that parasitize specific bacteria and multiply in metabolising host bacteria. Bacteriophages hunt for a single or a subset of bacterial species, making them potential antibacterial agents. Utilizing the ability of phages to control bacterial populations has several applications from medical to the fields of agriculture, aquaculture and the food industry. However, harnessing phage based techniques in wastewater treatments to improve quality of effluent and sludge release into the environment is a potential area for R&D application. Phage mediated bactericidal effect in any wastewater treatment process has many controlling factors that lead to treatment performance. In laboratory conditions, titer of bacteriophages (coliphages) isolated from effluent water of a specially designed anaerobic digester of human night soil (DRDO Biotoilet) was successfully increased with a modified protocol of the classical double layer agar technique. Enrichment of the same was carried out and efficacy of the phage enriched medium was evaluated at different conditions (specific media, temperature, storage conditions). Growth optimization study was carried out on different media like soybean casein digest medium (Tryptone soya medium), Luria-Bertani medium, phage deca broth medium and MNA medium (Modified nutrient medium). Further, temperature-phage yield relationship was also observed at three different temperatures 27˚C, 37˚C and 44˚C at laboratory condition. Results showed the higher activity of coliphage 27˚C and at 37˚C. Further, addition of divalent ions (10mM MgCl2, 5mM CaCl2) and 5% glycerol resulted in a significant increase in phage titer. Besides this, effect of antibiotics addition like ampicillin and kanamycin at different concentration on plaque formation was analysed and reported that ampicillin at a concentration of 1mg/ml ampicillin stimulates phage infection and results in more number of plaques. Experiments to test viability of phage showed that it can remain active for 6 months at 4˚C in fresh tryptone soya broth supplemented with fresh culture of coliforms (early log phase). The application of bacteriophages (especially coliphages) for treatment of effluent of human faecal matter contaminated effluent water is unique. This environment-friendly treatment system not only reduces the pathogenic coliforms, but also decreases the competition between nuisance bacteria and functionally important microbial populations. Therefore, the phage based cocktail to treat fecal pathogenic bacteria present in black water has many implication in wastewater treatment processes including ‘DRDO Biotoilet’, which is an ecofriendly appropriate and affordable human faecal matter treatment technology for different climates and situations.

Keywords: wastewater, microbes, virus, biotoilet, phage viability

Procedia PDF Downloads 418
102 Carlos Guillermo 'Cubena' Wilson's Literary Texts as Platforms for Social Commentary and Critique of Panamanian Society

Authors: Laverne Seales

Abstract:

When most people think of Panama, they immediately think of the Canal; however, the construction and the people who made it possible are often omitted and seldom acknowledged. The reality is that the construction of this waterway was achieved through forced migration and discriminatory practices toward people of African descent, specifically black people from the Caribbean. From the colonial period to the opening and subsequent operation of the Panama Canal by the United States, this paper goes through the rich layers of Panamanian history to examine the life of Afro-Caribbeans and their descendants in Panama. It also considers the role of the United States in Panama; it explores how the United States in Panama forged a racially complex country that made the integration of Afro-Caribbeans and their descendants difficult. After laying a historical foundation, the exploration of Afro-Caribbean people and Panamanians of Afro-Caribbean descent are analyzed through Afro-Panamanian writer Carlos Guillermo ‘Cubena’ Wilson's novels, short stories, and poetry. This study focuses on how Cubena addresses racism, discrimination, inequality, and social justice issues towards Afro-Caribbeans and their descendants who traveled to Panama to construct the Canal. Content analysis methodology can yield several significant contributions, and analyzing Carlos Guillermo Wilson's literature under this framework allows us to consider social commentary and critique of Panamanian society. It identifies the social issues and concerns of Afro-Caribbeans and people of Afro-Caribbean descent, such as inequality, corruption, racism, political oppression, and cultural identity. Analysis methodology allows us to explore how Cubena's literature engages with questions of cultural identity and belonging in Panamanian society. By examining themes related to race, ethnicity, language, and heritage, this research uncovers the complexities of Panamanian cultural identity, allowing us to interrogate power dynamics and social hierarchies in Panamanian society. Analyzing the portrayal of different social groups, institutions, and power structures helps uncover how power is wielded, contested, and resisted; Cubena's fictional world allows us to see how it functions in Panama. Content analysis methodology also provides for critiquing political systems and governance in Panama. By examining the representation and presentation of political figures, institutions, and events in Cubena's literature, we uncover his commentary on corruption, authoritarianism, governance, and the role of the United States in Panama. Content analysis highlights how Wilson's literature amplifies the voices and experiences of marginalized individuals and communities in Panamanian society. By centering the narratives of Afro-Panamanians and other marginalized groups, this researcher uncovers Cubena's commitment to social justice and inclusion in his writing and helps the reader engage with historical narratives and collective memory in Panama. Overall, analyzing Carlos Guillermo ‘Cubena’ Wilson's literature as a platform for social commentary and critique of Panamanian society using content analysis methodology provides valuable insights into the cultural, social, and political dimensions of Afro-Panamanians during and after the construction of the Panama Canal.

Keywords: Afro-Caribbean, Panama Canal, race, Afro-Panamanian, identity, history

Procedia PDF Downloads 18
101 Development of a Psychometric Testing Instrument Using Algorithms and Combinatorics to Yield Coupled Parameters and Multiple Geometric Arrays in Large Information Grids

Authors: Laith F. Gulli, Nicole M. Mallory

Abstract:

The undertaking to develop a psychometric instrument is monumental. Understanding the relationship between variables and events is important in structural and exploratory design of psychometric instruments. Considering this, we describe a method used to group, pair and combine multiple Philosophical Assumption statements that assisted in development of a 13 item psychometric screening instrument. We abbreviated our Philosophical Assumptions (PA)s and added parameters, which were then condensed and mathematically modeled in a specific process. This model produced clusters of combinatorics which was utilized in design and development for 1) information retrieval and categorization 2) item development and 3) estimation of interactions among variables and likelihood of events. The psychometric screening instrument measured Knowledge, Assessment (education) and Beliefs (KAB) of New Addictions Research (NAR), which we called KABNAR. We obtained an overall internal consistency for the seven Likert belief items as measured by Cronbach’s α of .81 in the final study of 40 Clinicians, calculated by SPSS 14.0.1 for Windows. We constructed the instrument to begin with demographic items (degree/addictions certifications) for identification of target populations that practiced within Outpatient Substance Abuse Counseling (OSAC) settings. We then devised education items, beliefs items (seven items) and a modifiable “barrier from learning” item that consisted of six “choose any” choices. We also conceptualized a close relationship between identifying various degrees and certifications held by Outpatient Substance Abuse Therapists (OSAT) (the demographics domain) and all aspects of their education related to EB-NAR (past and present education and desired future training). We placed a descriptive (PA)1tx in both demographic and education domains to trace relationships of therapist education within these two domains. The two perceptions domains B1/b1 and B2/b2 represented different but interrelated perceptions from the therapist perspective. The belief items measured therapist perceptions concerning EB-NAR and therapist perceptions using EB-NAR during the beginning of outpatient addictions counseling. The (PA)s were written in simple words and descriptively accurate and concise. We then devised a list of parameters and appropriately matched them to each PA and devised descriptive parametric (PA)s in a domain categorized information grid. Descriptive parametric (PA)s were reduced to simple mathematical symbols. This made it easy to utilize parametric (PA)s into algorithms, combinatorics and clusters to develop larger information grids. By using matching combinatorics we took paired demographic and education domains with a subscript of 1 and matched them to the column with each B domain with subscript 1. Our algorithmic matching formed larger information grids with organized clusters in columns and rows. We repeated the process using different demographic, education and belief domains and devised multiple information grids with different parametric clusters and geometric arrays. We found benefit combining clusters by different geometric arrays, which enabled us to trace parametric variables and concepts. We were able to understand potential differences between dependent and independent variables and trace relationships of maximum likelihoods.

Keywords: psychometric, parametric, domains, grids, therapists

Procedia PDF Downloads 256
100 Feasibility of an Extreme Wind Risk Assessment Software for Industrial Applications

Authors: Francesco Pandolfi, Georgios Baltzopoulos, Iunio Iervolino

Abstract:

The impact of extreme winds on industrial assets and the built environment is gaining increasing attention from stakeholders, including the corporate insurance industry. This has led to a progressively more in-depth study of building vulnerability and fragility to wind. Wind vulnerability models are used in probabilistic risk assessment to relate a loss metric to an intensity measure of the natural event, usually a gust or a mean wind speed. In fact, vulnerability models can be integrated with the wind hazard, which consists of associating a probability to each intensity level in a time interval (e.g., by means of return periods) to provide an assessment of future losses due to extreme wind. This has also given impulse to the world- and regional-scale wind hazard studies.Another approach often adopted for the probabilistic description of building vulnerability to the wind is the use of fragility functions, which provide the conditional probability that selected building components will exceed certain damage states, given wind intensity. In fact, in wind engineering literature, it is more common to find structural system- or component-level fragility functions rather than wind vulnerability models for an entire building. Loss assessment based on component fragilities requires some logical combination rules that define the building’s damage state given the damage state of each component and the availability of a consequence model that provides the losses associated with each damage state. When risk calculations are based on numerical simulation of a structure’s behavior during extreme wind scenarios, the interaction of component fragilities is intertwined with the computational procedure. However, simulation-based approaches are usually computationally demanding and case-specific. In this context, the present work introduces the ExtReMe wind risk assESsment prototype Software, ERMESS, which is being developed at the University of Naples Federico II. ERMESS is a wind risk assessment tool for insurance applications to industrial facilities, collecting a wide assortment of available wind vulnerability models and fragility functions to facilitate their incorporation into risk calculations based on in-built or user-defined wind hazard data. This software implements an alternative method for building-specific risk assessment based on existing component-level fragility functions and on a number of simplifying assumptions for their interactions. The applicability of this alternative procedure is explored by means of an illustrative proof-of-concept example, which considers four main building components, namely: the roof covering, roof structure, envelope wall and envelope openings. The application shows that, despite the simplifying assumptions, the procedure can yield risk evaluations that are comparable to those obtained via more rigorous building-level simulation-based methods, at least in the considered example. The advantage of this approach is shown to lie in the fact that a database of building component fragility curves can be put to use for the development of new wind vulnerability models to cover building typologies not yet adequately covered by existing works and whose rigorous development is usually beyond the budget of portfolio-related industrial applications.

Keywords: component wind fragility, probabilistic risk assessment, vulnerability model, wind-induced losses

Procedia PDF Downloads 170
99 Experimental-Numerical Inverse Approaches in the Characterization and Damage Detection of Soft Viscoelastic Layers from Vibration Test Data

Authors: Alaa Fezai, Anuj Sharma, Wolfgang Mueller-Hirsch, André Zimmermann

Abstract:

Viscoelastic materials have been widely used in the automotive industry over the last few decades with different functionalities. Besides their main application as a simple and efficient surface damping treatment, they may ensure optimal operating conditions for on-board electronics as thermal interface or sealing layers. The dynamic behavior of viscoelastic materials is generally dependent on many environmental factors, the most important being temperature and strain rate or frequency. Prior to the reliability analysis of systems including viscoelastic layers, it is, therefore, crucial to accurately predict the dynamic and lifetime behavior of these materials. This includes the identification of the dynamic material parameters under critical temperature and frequency conditions along with a precise damage localization and identification methodology. The goal of this work is twofold. The first part aims at applying an inverse viscoelastic material-characterization approach for a wide frequency range and under different temperature conditions. For this sake, dynamic measurements are carried on a single lap joint specimen using an electrodynamic shaker and an environmental chamber. The specimen consists of aluminum beams assembled to adapter plates through a viscoelastic adhesive layer. The experimental setup is reproduced in finite element (FE) simulations, and frequency response functions (FRF) are calculated. The parameters of both the generalized Maxwell model and the fractional derivatives model are identified through an optimization algorithm minimizing the difference between the simulated and the measured FRFs. The second goal of the current work is to guarantee an on-line detection of the damage, i.e., delamination in the viscoelastic bonding of the described specimen during frequency monitored end-of-life testing. For this purpose, an inverse technique, which determines the damage location and size based on the modal frequency shift and on the change of the mode shapes, is presented. This includes a preliminary FE model-based study correlating the delamination location and size to the change in the modal parameters and a subsequent experimental validation achieved through dynamic measurements of specimen with different, pre-generated crack scenarios and comparing it to the virgin specimen. The main advantage of the inverse characterization approach presented in the first part resides in the ability of adequately identifying the material damping and stiffness behavior of soft viscoelastic materials over a wide frequency range and under critical temperature conditions. Classic forward characterization techniques such as dynamic mechanical analysis are usually linked to limitations under critical temperature and frequency conditions due to the material behavior of soft viscoelastic materials. Furthermore, the inverse damage detection described in the second part guarantees an accurate prediction of not only the damage size but also its location using a simple test setup and outlines; therefore, the significance of inverse numerical-experimental approaches in predicting the dynamic behavior of soft bonding layers applied in automotive electronics.

Keywords: damage detection, dynamic characterization, inverse approaches, vibration testing, viscoelastic layers

Procedia PDF Downloads 191
98 Web-Based Decision Support Systems and Intelligent Decision-Making: A Systematic Analysis

Authors: Serhat Tüzün, Tufan Demirel

Abstract:

Decision Support Systems (DSS) have been investigated by researchers and technologists for more than 35 years. This paper analyses the developments in the architecture and software of these systems, provides a systematic analysis for different Web-based DSS approaches and Intelligent Decision-making Technologies (IDT), with the suggestion for future studies. Decision Support Systems literature begins with building model-oriented DSS in the late 1960s, theory developments in the 1970s, and the implementation of financial planning systems and Group DSS in the early and mid-80s. Then it documents the origins of Executive Information Systems, online analytic processing (OLAP) and Business Intelligence. The implementation of Web-based DSS occurred in the mid-1990s. With the beginning of the new millennia, intelligence is the main focus on DSS studies. Web-based technologies are having a major impact on design, development and implementation processes for all types of DSS. Web technologies are being utilized for the development of DSS tools by leading developers of decision support technologies. Major companies are encouraging its customers to port their DSS applications, such as data mining, customer relationship management (CRM) and OLAP systems, to a web-based environment. Similarly, real-time data fed from manufacturing plants are now helping floor managers make decisions regarding production adjustment to ensure that high-quality products are produced and delivered. Web-based DSS are being employed by organizations as decision aids for employees as well as customers. A common usage of Web-based DSS has been to assist customers configure product and service according to their needs. These systems allow individual customers to design their own products by choosing from a menu of attributes, components, prices and delivery options. The Intelligent Decision-making Technologies (IDT) domain is a fast growing area of research that integrates various aspects of computer science and information systems. This includes intelligent systems, intelligent technology, intelligent agents, artificial intelligence, fuzzy logic, neural networks, machine learning, knowledge discovery, computational intelligence, data science, big data analytics, inference engines, recommender systems or engines, and a variety of related disciplines. Innovative applications that emerge using IDT often have a significant impact on decision-making processes in government, industry, business, and academia in general. This is particularly pronounced in finance, accounting, healthcare, computer networks, real-time safety monitoring and crisis response systems. Similarly, IDT is commonly used in military decision-making systems, security, marketing, stock market prediction, and robotics. Even though lots of research studies have been conducted on Decision Support Systems, a systematic analysis on the subject is still missing. Because of this necessity, this paper has been prepared to search recent articles about the DSS. The literature has been deeply reviewed and by classifying previous studies according to their preferences, taxonomy for DSS has been prepared. With the aid of the taxonomic review and the recent developments over the subject, this study aims to analyze the future trends in decision support systems.

Keywords: decision support systems, intelligent decision-making, systematic analysis, taxonomic review

Procedia PDF Downloads 258
97 An Engineer-Oriented Life Cycle Assessment Tool for Building Carbon Footprint: The Building Carbon Footprint Evaluation System in Taiwan

Authors: Hsien-Te Lin

Abstract:

The purpose of this paper is to introduce the BCFES (building carbon footprint evaluation system), which is a LCA (life cycle assessment) tool developed by the Low Carbon Building Alliance (LCBA) in Taiwan. A qualified BCFES for the building industry should fulfill the function of evaluating carbon footprint throughout all stages in the life cycle of building projects, including the production, transportation and manufacturing of materials, construction, daily energy usage, renovation and demolition. However, many existing BCFESs are too complicated and not very designer-friendly, creating obstacles in the implementation of carbon reduction policies. One of the greatest obstacle is the misapplication of the carbon footprint inventory standards of PAS2050 or ISO14067, which are designed for mass-produced goods rather than building projects. When these product-oriented rules are applied to building projects, one must compute a tremendous amount of data for raw materials and the transportation of construction equipment throughout the construction period based on purchasing lists and construction logs. This verification method is very cumbersome by nature and unhelpful to the promotion of low carbon design. With a view to provide an engineer-oriented BCFE with pre-diagnosis functions, a component input/output (I/O) database system and a scenario simulation method for building energy are proposed herein. Most existing BCFESs base their calculations on a product-oriented carbon database for raw materials like cement, steel, glass, and wood. However, data on raw materials is meaningless for the purpose of encouraging carbon reduction design without a feedback mechanism, because an engineering project is not designed based on raw materials but rather on building components, such as flooring, walls, roofs, ceilings, roads or cabinets. The LCBA Database has been composited from existing carbon footprint databases for raw materials and architectural graphic standards. Project designers can now use the LCBA Database to conduct low carbon design in a much more simple and efficient way. Daily energy usage throughout a building's life cycle, including air conditioning, lighting, and electric equipment, is very difficult for the building designer to predict. A good BCFES should provide a simplified and designer-friendly method to overcome this obstacle in predicting energy consumption. In this paper, the author has developed a simplified tool, the dynamic Energy Use Intensity (EUI) method, to accurately predict energy usage with simple multiplications and additions using EUI data and the designed efficiency levels for the building envelope, AC, lighting and electrical equipment. Remarkably simple to use, it can help designers pre-diagnose hotspots in building carbon footprint and further enhance low carbon designs. The BCFES-LCBA offers the advantages of an engineer-friendly component I/O database, simplified energy prediction methods, pre-diagnosis of carbon hotspots and sensitivity to good low carbon designs, making it an increasingly popular carbon management tool in Taiwan. To date, about thirty projects have been awarded BCFES-LCBA certification and the assessment has become mandatory in some cities.

Keywords: building carbon footprint, life cycle assessment, energy use intensity, building energy

Procedia PDF Downloads 125
96 Fungal Cellulase/Xylanase Complex and Their Industrial Applications

Authors: L. Kutateldze, T. Urushadze, R. Khvedelidze, N. Zakariashvili, I. Khokhashvili, T. Sadunishvili

Abstract:

Microbial cellulase/xylanase have shown their potential application in various industries including pulp and paper, textile, laundry, biofuel production, food and feed industry, brewing, and agriculture. Extremophilic micromycetes and their enzymes that are resistant to critical values of temperature and pH, and retaining enzyme activity for a long time are of great industrial interest. Among strains of microscopic fungi from the collection of S. Durmishidze Institute of Biochemistry and Biotechnology, strains isolated from different ecological niches of Southern Caucasus-active producers of cellulase/xylanase have been selected by means of screening under deep cultivation conditions. Extremophilic micromycetes and their enzymes that are resistant to critical values of temperature and pH, and retaining enzyme activity for a long time are of great industrial interest. Among strains of microscopic fungi from the collection of S. Durmishidze Institute of Biochemistry and Biotechnology, strains isolated from different ecological niches of Southern Caucasus-active producers of cellulase/xylanase have been selected by means of screening under deep cultivation conditions. Representatives of the genera Aspergillus, Penicillium and Trichoderma are outstanding by relatively high activities of these enzymes. Among the producers were revealed thermophilic strains, representatives of the genus Aspergillus-Aspergillus terreus, Aspergillus versicolor, Aspergillus wentii, also strains of Sporotrichum pulverulentum and Chaetomium thermophile. As a result of optimization of cultivation media and conditions, activities of enzymes produced by the strains have been increased by 4 -189 %. Two strains, active producers of cellulase/xylanase – Penicillium canescence E2 (mesophile) and Aspergillus versicolor Z17 (thermophile) were chosen for further studies. Cellulase/xylanase enzyme preparations from two different genera of microscopic fungi Penicillium canescence E2 and Aspergillus versicolor Z 17 were obtained with activities 220 U/g /1200 U/g and 125 U/g /940 U/g, correspondingly. Main technical characteristics were as follows: the highest enzyme activities were obtained for mesophilic strain Penicillium canescence E2 at 45-500C, while almost the same enzyme activities were fixed for the thermophilic strain Aspergillus versicolor Z 17 at temperature 60-65°C, exceeding the temperature optimum of the mesophile by 150C. Optimum pH of action of the studied cellulase/xylanases from mesophileic and thermophilic strains were similar and equaled to 4.5-5.0 It has been shown that cellulase/xylanase technical preparations from selected strains of Penicillium canescence E2 and Aspergillus versicolor Z17 hydrolyzed cellulose of untreated wheat straw to reducible sugars by 46-52%, and to glucose by 22-27%. However the thermophilic enzyme preparations from the thermophilic A.versicolor strains conducted the process at 600C higher by 100C as compared to mesophlic analogue. Rate of hydrolyses of the pretreated substrate by the same enzyme preparations to reducible sugars and glucose conducted at optimum for their action 60 and 500C was 52-61% and 29-33%, correspondingly. Thus, maximum yield of glucose and reducible sugars form untreated and pretreated wheat straw was achieved at higher temperature (600C) by enzyme preparations from thermophilic strain, which gives advantage for their industrial application.

Keywords: cellulase/xylanase, cellulose hydrolysis, microscopic fungi, thermophilic strain

Procedia PDF Downloads 266
95 The Applications of Zero Water Discharge (ZWD) Systems for Environmental Management

Authors: Walter W. Loo

Abstract:

China declared the “zero discharge rules which leave no toxics into our living environment and deliver blue sky, green land and clean water to many generations to come”. The achievement of ZWD will provide conservation of water, soil and energy and provide drastic increase in Gross Domestic Products (GDP). Our society’s engine needs a major tune up; it is sputtering. ZWD is achieved in world’s space stations – no toxic air emission and the water is totally recycled and solid wastes all come back to earth. This is all done with solar power. These are all achieved under extreme temperature, pressure and zero gravity in space. ZWD can be achieved on earth under much less fluctuations in temperature, pressure and normal gravity environment. ZWD systems are not expensive and will have multiple beneficial returns on investment which are both financially and environmentally acceptable. The paper will include successful case histories since the mid-1970s. ZWD discharge can be applied to the following types of projects: nuclear and coal fire power plants with a closed loop system that will eliminate thermal water discharge; residential communities with wastewater treatment sump and recycle the water use as a secondary water supply; waste water treatment Plants with complete water recycling including water distillation to produce distilled water by very economical 24-hours solar power plant. Landfill remediation is based on neutralization of landfilled gas odor and preventing anaerobic leachate formation. It is an aerobic condition which will render landfill gas emission explosion proof. Desert development is the development of recovering soil moisture from soil and completing a closed loop water cycle by solar energy within and underneath an enclosed greenhouse. Salt-alkali land development can be achieved by solar distillation of salty shallow water into distilled water. The distilled water can be used for soil washing and irrigation and complete a closed loop water cycle with energy and water conservation. Heavy metals remediation can be achieved by precipitation of dissolved toxic metals below the plant or vegetation root zone by solar electricity without pumping and treating. Soil and groundwater remediation - abandoned refineries, chemical and pesticide factories can be remediated by in-situ electrobiochemical and bioventing treatment method without pumping or excavation. Toxic organic chemicals are oxidized into carbon dioxide and heavy metals precipitated below plant and vegetation root zone. New water sources: low temperature distilled water can be recycled for repeated use within a greenhouse environment by solar distillation; nano bubble water can be made from the distilled water with nano bubbles of oxygen, nitrogen and carbon dioxide from air (fertilizer water) and also eliminate the use of pesticides because the nano oxygen will break the insect growth chain in the larvae state. Three dimensional high yield greenhouses can be constructed by complete water recycling using the vadose zone soil as a filter with no farming wastewater discharge.

Keywords: greenhouses, no discharge, remediation of soil and water, wastewater

Procedia PDF Downloads 333
94 Leveraging the HDAC Inhibitory Pharmacophore to Construct Deoxyvasicinone Based Tractable Anti-Lung Cancer Agent and pH-Responsive Nanocarrier

Authors: Ram Sharma, Esha Chatterjee, Santosh Kumar Guru, Kunal Nepali

Abstract:

A tractable anti-lung cancer agent was identified via the installation of a Ring C expanded synthetic analogue of the alkaloid vasicinone [7,8,9,10-tetrahydroazepino[2,1-b] quinazolin-12(6H)-one (TAZQ)] as a surface recognition part in the HDAC inhibitory three-component model. Noteworthy to mention that the candidature of TAZQ was deemed suitable for accommodation in HDAC inhibitory pharmacophore as per the results of the fragment recruitment process conducted by our laboratory. TAZQ was pinpointed through the fragment screening program as a synthetically flexible fragment endowed with some moderate cell growth inhibitory activity against the lung cancer cell lines, and it was anticipated that the use of the aforementioned fragment to generate hydroxamic acid functionality (zinc-binding motif) bearing HDAC inhibitors would boost the antitumor efficacy of TAZQ. Consistent with our aim of applying epigenetic targets to the treatment of lung cancer, a strikingly potent anti-lung cancer scaffold (compound 6) was pinpointed through a series of in-vitro experiments. Notably, the compounds manifested a magnificent activity profile against KRAS and EGFR mutant lung cancer cell lines (IC50 = 0.80 - 0.96 µM), and the effects were found to be mediated through preferential HDAC6 inhibition (IC50 = 12.9 nM). In addition to HDAC6 inhibition, the compounds also elicited HDAC1 and HDAC3 inhibitory activity with an IC50 value of 49.9 nM and 68.5 nM, respectively. The HDAC inhibitory ability of compound 6 was also confirmed from the results of the western blot experiment that revealed its potential to decrease the expression levels of HDAC isoforms (HDAC1, HDAC3, and HDAC6). Noteworthy to mention that complete downregulation of the HDAC6 isoform was exerted by compound 6 at 0.5 and 1 µM. Moreover, in another western blot experiment, treatment with hydroxamic acid 6 led to upregulation of H3 acK9 and α-Tubulin acK40 levels, ascertaining its inhibitory activity toward both the class I HDACs and Class II B HDACs. The results of other assays were also encouraging as treatment with compound 6 led to the suppression of the colony formation ability of A549 cells, induction of apoptosis, and increase in autophagic flux. In silico studies led us to rationalize the results of the experimental assay, and some key interactions of compound 6 with the amino acid residues of HDAC isoforms were identified. In light of the impressive activity spectrum of compound 6, a pH-responsive nanocarrier (hyaluronic acid-compound 6 nanoparticles) was prepared. The dialysis bag approach was used for the assessment of the nanoparticles under both normal and acidic circumstances, and the pH-sensitive nature of hyaluronic acid-compound 6 nanoparticles was confirmed. Delightfully, the nanoformulation was devoid of cytotoxicity against the L929 mouse fibroblast cells (normal settings) and exhibited selective cytotoxicity towards the A549 lung cancer cell lines. In a nutshell, compound 6 appears to be a promising adduct, and a detailed investigation of this compound might yield a therapeutic for the treatment of lung cancer.

Keywords: HDAC inhibitors, lung cancer, scaffold, hyaluronic acid, nanoparticles

Procedia PDF Downloads 79