Search results for: application lifecycle management
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16223

Search results for: application lifecycle management

3503 Comparison between Conventional Bacterial and Algal-Bacterial Aerobic Granular Sludge Systems in the Treatment of Saline Wastewater

Authors: Philip Semaha, Zhongfang Lei, Ziwen Zhao, Sen Liu, Zhenya Zhang, Kazuya Shimizu

Abstract:

The increasing generation of saline wastewater through various industrial activities is becoming a global concern for activated sludge (AS) based biological treatment which is widely applied in wastewater treatment plants (WWTPs). As for the AS process, an increase in wastewater salinity has negative impact on its overall performance. The advent of conventional aerobic granular sludge (AGS) or bacterial AGS biotechnology has gained much attention because of its superior performance. The development of algal-bacterial AGS could enhance better nutrients removal, potentially reduce aeration cost through symbiotic algae-bacterial activity, and thus, can also reduce overall treatment cost. Nonetheless, the potential of salt stress to decrease biomass growth, microbial activity and nutrient removal exist. Up to the present, little information is available on saline wastewater treatment by algal-bacterial AGS. To the authors’ best knowledge, a comparison of the two AGS systems has not been done to evaluate nutrients removal capacity in the context of salinity increase. This study sought to figure out the impact of salinity on the algal-bacterial AGS system in comparison to bacterial AGS one, contributing to the application of AGS technology in the real world of saline wastewater treatment. In this study, the salt concentrations tested were 0 g/L, 1 g/L, 5 g/L, 10 g/L and 15 g/L of NaCl with 24-hr artificial illuminance of approximately 97.2 µmol m¯²s¯¹, and mature bacterial and algal-bacterial AGS were used for the operation of two identical sequencing batch reactors (SBRs) with a working volume of 0.9 L each, respectively. The results showed that salinity increase caused no apparent change in the color of bacterial AGS; while for algal-bacterial AGS, its color was progressively changed from green to dark green. A consequent increase in granule diameter and fluffiness was observed in the bacterial AGS reactor with the increase of salinity in comparison to a decrease in algal-bacterial AGS diameter. However, nitrite accumulation peaked from 1.0 mg/L and 0.4 mg/L at 1 g/L NaCl in the bacterial and algal-bacterial AGS systems, respectively to 9.8 mg/L in both systems when NaCl concentration varied from 5 g/L to 15 g/L. Almost no ammonia nitrogen was detected in the effluent except at 10 g/L NaCl concentration, where it averaged 4.2 mg/L and 2.4 mg/L, respectively, in the bacterial and algal-bacterial AGS systems. Nutrients removal in the algal-bacterial system was relatively higher than the bacterial AGS in terms of nitrogen and phosphorus removals. Nonetheless, the nutrient removal rate was almost 50% or lower. Results show that algal-bacterial AGS is more adaptable to salinity increase and could be more suitable for saline wastewater treatment. Optimization of operation conditions for algal-bacterial AGS system would be important to ensure its stably high efficiency in practice.

Keywords: algal-bacterial aerobic granular sludge, bacterial aerobic granular sludge, Nutrients removal, saline wastewater, sequencing batch reactor

Procedia PDF Downloads 124
3502 Assessment of the Impacts of Climate Change on Watershed Runoff Using Soil and Water Assessment Tool Model in Southeast Nigeria

Authors: Samuel Emeka Anarah, Kingsley Nnaemeka Ogbu, Obasi Arinze

Abstract:

Quantifying the hydrological response due to changes in climate change is imperative for proper management of water resources within a watershed. The impact of climate change on the hydrology of the Upper Ebony River (UER) watershed, South East Nigeria, was studied using the Soil and Water Assessment Tool (SWAT) hydrological model. A climatological time series analysis from 1985 - 2014 using non-parametric test showed significant negative trends in precipitation and relative humidity trend while minimum and maximum temperature, solar radiation and wind speed showed significant positive trends. Future hypothetical land-use change scenarios (Scenarios 1, 2, 3 and 4) representing urbanization and conversion of forest to agricultural land were combined with future downscaled climate model (CSIRO-Mk3-6-0) and simulated in SWAT model. Relative to the Baseline scenario (2005 - 2014), the results showed a decrease in streamflow by 10.29%, 26.20%, 11.80% and 26.72% for Scenarios 1, 2, 3, and 4 respectively. Model results suggest development of adaptation strategies to cope with the predicted hydrological conditions under future climate change in the watershed.

Keywords: climate change, hydrology, runoff, SWAT model

Procedia PDF Downloads 118
3501 Prevalence of Dactylogyrus spp. and Butriocephalus spp. in Common Carp (Cyprinus carpio) from Mazandaran Province, Iran

Authors: Mohamad Bolandmrtabeh, Nafiseh Hasani, Saeid Abdi darakeh, Amir Heydari, Maryam Asghari

Abstract:

Dactylogyrus spp. and Butriocephalus spp. are among the most common parasites infecting farmed fish in Iran and worldwide. These parasites infect fish of all ages and are responsible for many lesions and mortalities in fingerlings and occasionally in grow-out and brood fish. Understanding the prevalence and intensity of these parasites appears necessary for their effective management and control. In this study, 60 common carp (Cyprinus carpio) were collected from 6 stations in Mazandaran Province, Iran, during summer 2023 and examined parasitologically. Common carp is a freshwater fish of the family Cyprinidae. The fish were transferred live to the laboratory after capture and were examined for parasites after anaesthetization and biometrics. The identified parasitic species included a Dactylogyrus sp. and a Butriocephalus sp. A total of 11 Dactylogyrus spp. were recovered from the gills of 60 fish examined. Additionally, 6 Butriocephalus spp. were collected from the intestine of infected fish. Based on taxonomic keys, the parasites were identified to genus level. The overall prevalence of infection was 23.3%, with approximately one quarter of all fish infected. A significant association was found between infections in different sampling stations. These results highlight the need for further research and planning for treatment and control in order to mitigate the impact of these parasitic diseases.

Keywords: dactylogyrus, butriocephalus, common carp, cyprinus carpio

Procedia PDF Downloads 48
3500 Evaluation of Pheromone and Tree Trap Efficiency in Orthotomicus erosus (Col: Curculionidae: Scolytinae) Monitoring in Pine Forests of Iran

Authors: Sudabe Amini, Jamasb Nozari, Somaye Rahimi

Abstract:

Bark beetles are one of the most destructive groups of pests in the forest and green space. Mediterranean pine Engraver Orthotomicus erosus (Wollston) is the dominant species in the pine forests of Iran. Pine forests are considered a crucial region in the world and need high protection. Although there is no effective control method, mass trapping is the most common method to suppress the bark beetle population. Due to this, from 2018-to 2020, a survey was conducted on bark beetles mass trapping by using two kinds of traps, including pheromone and tree trap. These traps were evaluated in 10 different sites of pine forests. The statistical results proved that significant differences between the pheromone trap and tree trap were observed. It confirmed that the pheromone trap attracted more beetles than the tree trap. The results of this study suggest that the most effective and applicable method in bark beetle’s management of pines forest is using a pheromone trap that suppresses and maintains bark beetle’s population at an economic level, although tree traps attract bark beetles too. In the future, using tree-pheromone traps, which would synergist attraction of more bark beetles, is recommended.

Keywords: bark beetle, pines forest, Orthotomicus erosus, pheromone trap, tree trap

Procedia PDF Downloads 135
3499 Feeding Ecology and Habitat Preference of Red Panda in Dhorpatan Hunting Reserve, Nepal

Authors: Saroj Panthi

Abstract:

The red panda (Ailurus fulgens fulgens) is distributed throughout the Himalayas and is found in both protected and unprotected areas of Nepal. Loss and fragmentation of habitat threaten red panda population throughout its range, and as a consequence, it is listed as vulnerable on the IUCN Red List of Threatened Species. Despite this pressing situation, data on the ecology of the red panda in western Nepal are lacking. Our aim in the current study was to determine the distribution, associated habitats, and summer diet of the red panda in Dhorpatan Hunting Reserve (DHR), Nepal. Evidence of red pandas was found in all 7 blocks of the reserve, spanning an area of 345.8 km2, between elevations of 2800 m and 4000 m and predominantly (> 75%) in forests comprising plant communities dominated by Abies spectabilis, Acer caesium, Tsuga domusa, and Betula utilis, with ground cover of Arundinaria spp. The dominant plant found in scat of the red panda was Arundinaria spp. (81.7%), with Acer spp., B. utilis, and lichen also frequently present. Livestock grazing and human activities were significantly higher in habitats where signs of pandas were recorded than in areas where they were absent. This habitat overlap between the red panda and livestock potentially poses a major threat to the panda’s survival in the DHR, a fact that should be taken into account in devising management strategies for this threatened species.

Keywords: red panda, Dhorpatan hunting reserve, diet, habitat preference

Procedia PDF Downloads 326
3498 Cognitive Linguistic Features Underlying Spelling Development in a Second Language: A Case Study of L2 Spellers in South Africa

Authors: A. Van Staden, A. Tolmie, E. Vorster

Abstract:

Research confirms the multifaceted nature of spelling development and underscores the importance of both cognitive and linguistic skills that affect sound spelling development such as working and long-term memory, phonological and orthographic awareness, mental orthographic images, semantic knowledge and morphological awareness. This has clear implications for many South African English second language spellers (L2) who attempt to become proficient spellers. Since English has an opaque orthography, with irregular spelling patterns and insufficient sound/grapheme correspondences, L2 spellers can neither rely, nor draw on the phonological awareness skills of their first language (for example Sesotho and many other African languages), to assist them to spell the majority of English words. Epistemologically, this research is informed by social constructivism. In addition the researchers also hypothesized that the principles of the Overlapping Waves Theory was an appropriate lens through which to investigate whether L2 spellers could significantly improve their spelling skills via the implementation of an alternative route to spelling development, namely the orthographic route, and more specifically via the application of visual imagery. Post-test results confirmed the results of previous research that argues for the interactive nature of different cognitive and linguistic systems such as working memory and its subsystems and long-term memory, as learners were systematically guided to store visual orthographic images of words in their long-term lexicons. Moreover, the results have shown that L2 spellers in the experimental group (n = 9) significantly outperformed L2 spellers (n = 9) in the control group whose intervention involved phonological awareness (and coding) including the teaching of spelling rules. Consequently, L2 learners in the experimental group significantly improved in all the post-test measures included in this investigation, namely the four sub-tests of short-term memory; as well as two spelling measures (i.e. diagnostic and standardized measures). Against this background, the findings of this study look promising and have shown that, within a social-constructivist learning environment, learners can be systematically guided to apply higher-order thinking processes such as visual imagery to successfully store and retrieve mental images of spelling words from their output lexicons. Moreover, results from the present study could play an important role in directing research into this under-researched aspect of L2 literacy development within the South African education context.

Keywords: English second language spellers, phonological and orthographic coding, social constructivism, visual imagery as spelling strategy

Procedia PDF Downloads 328
3497 Quality of Life of Patients on Oral Antiplatelet Therapy in Outpatient Cardiac Department Dr. Hasan Sadikin Central General Hospital Bandung

Authors: Andhiani Sharfina Arnellya, Mochammad Indra Permana, Dika Pramita Destiani, Ellin Febrina

Abstract:

Health Research Data, Ministry of Health of Indonesia in 2007, showed coronary heart disease (CHD) or coronary artery disease (CAD) was the third leading cause of death in Indonesia after hypertension and stroke with 7.2% incidence rate. Antiplatelet is one of the important therapy in management of patients with CHD. In addition to therapeutic effect on patients, quality of life is one aspect of another assessment to see the success of antiplatelet therapy. The purpose of this study was to determine the quality of life of patients on oral antiplatelet therapy in outpatient cardiac department Dr. Hasan Sadikin central general hospital, Bandung, Indonesia. This research is a cross sectional by collecting data through quality of life questionnaire of patients which performed prospectively as primary data and secondary data from medical record of patients. The results of this study showed that 54.3% of patients had a good quality of life, 45% had a moderate quality of life, and 0.7% had a poor quality of life. There are no significant differences in quality of life-based on age, gender, diagnosis, and duration of drug use.

Keywords: antiplatelet, quality of life, coronary artery disease, coronary heart disease

Procedia PDF Downloads 297
3496 Efficacy of Preimplantation Genetic Screening in Women with a Spontaneous Abortion History with Eukaryotic or Aneuploidy Abortus

Authors: Jayeon Kim, Eunjung Yu, Taeki Yoon

Abstract:

Most spontaneous miscarriage is believed to be a consequence of embryo aneuploidies. Transferring eukaryotic embryos selected by PGS is expected to decrease the miscarriage rate. Current PGS indications include advanced maternal age, recurrent pregnancy loss, repeated implantation failure. Recently, use of PGS for healthy women without above indications for the purpose of improving in vitro fertilization (IVF) outcomes is on the rise. However, it is still controversy about the beneficial effect of PGS in this population, especially, in women with a history of no more than 2 miscarriages or miscarriage of eukaryotic abortus. This study aimed to investigate if karyotyping result of abortus is a good indicator of preimplantation genetic screening (PGS) in subsequent IVF cycle in women with a history of spontaneous abortion. A single-center retrospective cohort study was performed. Women who had spontaneous abortion(s) (less than 3) and dilatation and evacuation, and subsequent IVF from January 2016 to November 2016 were included. Their medical information was extracted from the charts. Clinical pregnancy was defined as presence of a gestational sac with fetal heart beat detected on ultrasound in week 7. Statistical analysis was performed using SPSS software. Total 234 women were included. 121 out of 234 (51.7%) underwent karyotyping of the abortus, and 113 did not have the abortus karyotyped. Embryo biopsy was performed on 3 or 5 days after oocyte retrieval, followed by embryo transfer (ET) on a fresh or frozen cycle. The biopsied materials were subjected to microarray comparative genomic hybridization. Clinical pregnancy rate per ET was compared between PGS and non-PGS group in each study group. Patients were grouped by two criteria: karyotype of the abortus from previous miscarriage (unknown fetal karyotype (n=89, Group 1), eukaryotic abortus (n=36, Group 2) or aneuploidy abortus (n=67, Group 3)), and pursuing PGS in subsequent IVF cycle (pursuing PGS (PGS group, n=105) or not pursuing PGS (non-PGS group, n=87)). The PGS group was significantly older and had higher number of retrieved oocytes and prior miscarriages compared to non-PGS group. There were no differences in BMI and AMH level between those two groups. In PGS group, the mean number of transferable embryos (eukaryotic embryo) was 1.3 ± 0.7, 1.5 ± 0.5 and 1.4 ± 0.5, respectively (p = 0.049). In 42 cases, ET was cancelled because all embryos biopsied turned out to be abnormal. In all three groups (group 1, 2, and 3), clinical pregnancy rates were not statistically different between PGS and non-PGS group (Group 1: 48.8% vs. 52.2% (p=0.858), Group 2: 70% vs. 73.1% (p=0.730), Group 3: 42.3% vs. 46.7% (p=0.640), in PGS and non-PGS group, respectively). In both groups who had miscarriage with eukaryotic and aneuploidy abortus, the clinical pregnancy rate between IVF cycles with and without PGS was not different. When we compare miscarriage and ongoing pregnancy rate, there were no significant differences between PGS and non-PGS group in all three groups. Our results show that the routine application of PGS in women who had less than 3 miscarriages would not be beneficial, even in cases that previous miscarriage had been caused by fetal aneuploidy.

Keywords: preimplantation genetic diagnosis, miscarriage, kpryotyping, in vitro fertilization

Procedia PDF Downloads 155
3495 Development of Social Competence in the Preparation and Continuing Training of Adult Educators

Authors: Genute Gedviliene, Vidmantas Tutlys

Abstract:

The aim of this paper is to reveal the deployment and development of the social competence in the higher education programmes of adult education and in the continuing training and competence development of the andragogues. There will be compared how the issues of cooperation and communication in the learning and teaching processes are treated in the study programmes and in the courses of continuing training of andragogues. Theoretical and empirical research methods were combined for research analysis. For the analysis the following methods were applied: 1) Literature and document analysis helped to highlight the communication and cooperation as fundamental phenomena of the social competence, it’s important for the adult education in the context of digitalization and globalization. There were also analyzed the research studies on the development of social competence in the field of andragogy, as well as on the place and weight of the social competence in the overall competence profile of the andragogue. 2) The empirical study is based on questionnaire survey method. The population of survey consists of 240 students of bachelor and master degree studies of andragogy in Lithuania and of 320 representatives of the different bodies and institutions involved in the continuing training and professional development of the adult educators in Lithuania. The themes of survey questionnaire were defined on the basis of findings of the literature review and included the following: 1) opinions of the respondents on the role and place of a social competence in the work of andragogue; 2) opinions of the respondents on the role and place of the development of social competence in the curricula of higher education studies and continuing training courses; 3) judgements on the implications of the higher education studies and courses of continuing training for the development of social competence and it’s deployment in the work of andragogue. Data analysis disclosed a wide range of ways and modalities of the deployment and development of social competence in the preparation and continuing training of the adult educators. Social competence is important for the students and adult education providers not only as the auxiliary capability for the communication and transfer of information, but also as the outcome of collective learning leading to the development of new capabilities applied by the learners in the learning process, their professional field of adult education and their social life. Equally so, social competence is necessary for the effective adult education activities not only as an auxiliary capacity applied in the teaching process, but also as a potential for improvement, development and sustainability of the didactic competence and know-how in this field. The students of the higher education programmes in the field of adult education treat social competence as important generic capacity important for the work of adult educator, whereas adult education providers discern the concrete issues of application of social competence in the different processes of adult education, starting from curriculum design and ending with assessment of learning outcomes.

Keywords: adult education, andragogues, social competence, curriculum

Procedia PDF Downloads 122
3494 Sceletium Tortuosum: A review on its Phytochemistry, Pharmacokinetics, Biological and Clinical Activities

Authors: Tomi Lois Olatunji, Frances Siebert, Ademola Emmanuel Adetunji, Brian Harvey, Johane Gericke, Josias Hamman, Frank Van Der Kooy

Abstract:

Ethnopharmacological relevance: Sceletium tortuosum (L.) N.E.Br, the most sought after and widely researched species in the genus Sceletium is a succulent forb endemic to South Africa. Traditionally, this medicinal plant is mainly masticated or smoked and used for the relief of toothache, abdominal pain, and as a mood-elevator, analgesic, hypnotic, anxiolytic, thirst and hunger suppressant, and for its intoxicating/euphoric effects. Sceletium tortuosum is currently of widespread scientific interest due to its clinical potential in treating anxiety and depression, relieving stress in healthy individuals, and enhancing cognitive functions. These pharmacological actions are attributed to its phytochemical constituents referred to as mesembrine-type alkaloids. Aim of the review: The aim of this review was to comprehensively summarize and critically evaluate recent research advances on the phytochemistry, pharmacokinetics, biological and clinical activities of the medicinal plant S. tortuosum. Additionally, current ongoing research and future perspectives are also discussed. Methods: All relevant scientific articles, books, MSc and Ph.D. dissertations on botany, behavioral pharmacology, traditional uses, and phytochemistry of S. tortuosum were retrieved from different databases (including Science Direct, PubMed, Google Scholar, Scopus and Web of Science). For pharmacokinetics and pharmacological effects of S. tortuosum, the focus fell on relevant publications published between 2009 and 2021. Results: Twenty-five alkaloids belonging to four structural classes viz: mesembrine, Sceletium A4, joubertiamine, and tortuosamine, have been identified from S. tortuosum, of which the mesembrine class is predominant. The crude extracts and commercially available standardized extracts of S. tortuosum have displayed a wide spectrum of biological activities (e.g. antimalarial, anti-oxidant, immunomodulatory, anti-HIV, neuroprotection, enhancement of cognitive function) in in vitro or in vivo studies. This plant has not yet been studied in a clinical population, but has potential for enhancing cognitive function, and managing anxiety and depression. Conclusion: As an important South African medicinal plant, S. tortuosum has garnered many research advances on its phytochemistry and biological activities over the last decade. These scientific studies have shown that S. tortuosum has various bioactivities. The findings have further established the link between the phytochemistry and pharmacological application, and support the traditional use of S. tortuosum in the indigenous medicine of South Africa.

Keywords: Aizoaceae, Mesembrine, Serotonin, Sceletium tortuosum, Zembrin®, psychoactive, antidepressant

Procedia PDF Downloads 186
3493 Interference Management in Long Term Evolution-Advanced System

Authors: Selma Sbit, Mohamed Bechir Dadi, Belgacem Chibani Rhaimi

Abstract:

Incorporating Home eNodeB (HeNB) in cellular networks, e.g. Long Term Evolution Advanced (LTE-A), is beneficial for extending coverage and enhancing capacity at low price especially within the non-line-of sight (NLOS) environments such as homes. HeNB or femtocell is a small low powered base station which provides radio coverage to the mobile users in an indoor environment. This deployment results in a heterogeneous network where the available spectrum becomes shared between two layers. Therefore, a problem of Inter Cell Interference (ICI) appears. This issue is the main challenge in LTE-A. To deal with this challenge, various techniques based on frequency, time and power control are proposed. This paper deals with the impact of carrier aggregation and higher order MIMO (Multiple Input Multiple Output) schemes on the LTE-Advanced performance. Simulation results show the advantages of these schemes on the system capacity (4.109 b/s/Hz when bandwidth B=100 MHz and when applying MIMO 8x8 for SINR=30 dB), maximum theoretical peak data rate (more than 4 Gbps for B=100 MHz and when MIMO 8x8 is used) and spectral efficiency (15 b/s/Hz and 30b/s/Hz when MIMO 4x4 and MIMO 8x8 are applying respectively for SINR=30 dB).

Keywords: capacity, carrier aggregation, LTE-Advanced, MIMO (Multiple Input Multiple Output), peak data rate, spectral efficiency

Procedia PDF Downloads 229
3492 The Key to the Success of Hotel Loyalty Program Enrolments: An Assessment of the Best Practices within the Three Hospitality Tiers

Authors: Nicole Anne Bass Monsanto

Abstract:

The objective of this study were to identify the numerous effective ways of capturing new members to participate in the brand hotel loyalty programs through the following objectives: (1) To differentiate the appropriate approach on selling the loyalty program to the different consumer market within their tier; (2) to evaluate the different behavioural degree of guests from different aspects of class and preference; (3) to compare the 3 hospitality tiers and the best tactic of approach according to their target market; (4) to assess the best practices suitable to the category of hotel in their specific tier. This study is a mixed methods research which involves both qualitative and quantitative. The qualitative research consisted of twelve management level respondents from 6 different hotels. These hotels were subdivided into 3 hotel tiers -- Luxury Hotels Upper Upscale Hotels and Mid-Scale Hotels—and the research studied two hotels of each hotel tier. Meanwhile, the quantitative research was conducted to further confirm and enhance the results by the means of guest post-stay surveys. Research findings of the qualitative research were as follows: (1) There are no differentiated best practices in capturing new enrolments based on the hotel tier. (2) There is only one key tool that determines the hotel loyalty program enrolment success. Meanwhile, the research findings of the quantitative research lead to the conclusion that most guests do not participate in the loyalty program was because they were not invited from the start.

Keywords: best practices, loyalty program, hospitality tier, enrollment success

Procedia PDF Downloads 258
3491 The Two Question Challenge: Embedding the Serious Illness Conversation in Acute Care Workflows

Authors: D. M. Lewis, L. Frisby, U. Stead

Abstract:

Objective: Many patients are receiving invasive treatments in acute care or are dying in hospital without having had comprehensive goals of care conversations. Some of these treatments may not align with the patient’s wishes, may be futile, and may cause unnecessary suffering. While many staff may recognize the benefits of engaging patients and families in Serious Illness Conversations (a goal of care framework developed by Ariadne Labs in Boston), few staff feel confident and/or competent in having these conversations in acute care. Another barrier to having these conversations may be due to a lack of incorporation in the current workflow. An educational exercise, titled the Two Question Challenge, was initiated on four medical units across two Vancouver Coastal Health (VCH) hospitals in attempt to engage the entire interdisciplinary team in asking patients and families questions around goals of care and to improve the documentation of these expressed wishes and preferences. Methods: Four acute care units across two separate hospitals participated in the Two Question Challenge. On each unit, over the course of two eight-hour shifts, all members of the interdisciplinary team were asked to select at least two questions from a selection of nine goals of care questions. They were asked to pose these questions of a patient or family member throughout their shift and then asked to document their conversations in a centralized Advance Care Planning/Goals of Care discussion record in the patient’s chart. A visual representation of conversation outcomes was created to demonstrate to staff and patients the breadth of conversations that took place throughout the challenge. Staff and patients were interviewed about their experiences throughout the challenge. Two palliative approach leads remained present on the units throughout the challenge to support, guide, or role model these conversations. Results: Across four acute care medical units, 47 interdisciplinary staff participated in the Two Question Challenge, including nursing, allied health, and a physician. A total of 88 questions were asked of patients, or their families around goals of care and 50 newly documented goals of care conversations were charted. Two code statuses were changed as a result of the conversations. Patients voiced an appreciation for these conversations and staff were able to successfully incorporate these questions into their daily care. Conclusion: The Two Question Challenge proved to be an effective way of having teams explore the goals of care of patients and families in an acute care setting. Staff felt that they gained confidence and competence. Both staff and patients found these conversations to be meaningful and impactful and felt they were notably different from their usual interactions. Documentation of these conversations in a centralized location that is easily accessible to all care providers increased significantly. Application of the Two Question Challenge in non-medical units or other care settings, such as long-term care facilities or community health units, should be explored in the future.

Keywords: advance care planning, goals of care, interdisciplinary, palliative approach, serious illness conversations

Procedia PDF Downloads 86
3490 Use of Machine Learning in Data Quality Assessment

Authors: Bruno Pinto Vieira, Marco Antonio Calijorne Soares, Armando Sérgio de Aguiar Filho

Abstract:

Nowadays, a massive amount of information has been produced by different data sources, including mobile devices and transactional systems. In this scenario, concerns arise on how to maintain or establish data quality, which is now treated as a product to be defined, measured, analyzed, and improved to meet consumers' needs, which is the one who uses these data in decision making and companies strategies. Information that reaches low levels of quality can lead to issues that can consume time and money, such as missed business opportunities, inadequate decisions, and bad risk management actions. The step of selecting, identifying, evaluating, and selecting data sources with significant quality according to the need has become a costly task for users since the sources do not provide information about their quality. Traditional data quality control methods are based on user experience or business rules limiting performance and slowing down the process with less than desirable accuracy. Using advanced machine learning algorithms, it is possible to take advantage of computational resources to overcome challenges and add value to companies and users. In this study, machine learning is applied to data quality analysis on different datasets, seeking to compare the performance of the techniques according to the dimensions of quality assessment. As a result, we could create a ranking of approaches used, besides a system that is able to carry out automatically, data quality assessment.

Keywords: machine learning, data quality, quality dimension, quality assessment

Procedia PDF Downloads 122
3489 Event Data Representation Based on Time Stamp for Pedestrian Detection

Authors: Yuta Nakano, Kozo Kajiwara, Atsushi Hori, Takeshi Fujita

Abstract:

In association with the wave of electric vehicles (EV), low energy consumption systems have become more and more important. One of the key technologies to realize low energy consumption is a dynamic vision sensor (DVS), or we can call it an event sensor, neuromorphic vision sensor and so on. This sensor has several features, such as high temporal resolution, which can achieve 1 Mframe/s, and a high dynamic range (120 DB). However, the point that can contribute to low energy consumption the most is its sparsity; to be more specific, this sensor only captures the pixels that have intensity change. In other words, there is no signal in the area that does not have any intensity change. That is to say, this sensor is more energy efficient than conventional sensors such as RGB cameras because we can remove redundant data. On the other side of the advantages, it is difficult to handle the data because the data format is completely different from RGB image; for example, acquired signals are asynchronous and sparse, and each signal is composed of x-y coordinate, polarity (two values: +1 or -1) and time stamp, it does not include intensity such as RGB values. Therefore, as we cannot use existing algorithms straightforwardly, we have to design a new processing algorithm to cope with DVS data. In order to solve difficulties caused by data format differences, most of the prior arts make a frame data and feed it to deep learning such as Convolutional Neural Networks (CNN) for object detection and recognition purposes. However, even though we can feed the data, it is still difficult to achieve good performance due to a lack of intensity information. Although polarity is often used as intensity instead of RGB pixel value, it is apparent that polarity information is not rich enough. Considering this context, we proposed to use the timestamp information as a data representation that is fed to deep learning. Concretely, at first, we also make frame data divided by a certain time period, then give intensity value in response to the timestamp in each frame; for example, a high value is given on a recent signal. We expected that this data representation could capture the features, especially of moving objects, because timestamp represents the movement direction and speed. By using this proposal method, we made our own dataset by DVS fixed on a parked car to develop an application for a surveillance system that can detect persons around the car. We think DVS is one of the ideal sensors for surveillance purposes because this sensor can run for a long time with low energy consumption in a NOT dynamic situation. For comparison purposes, we reproduced state of the art method as a benchmark, which makes frames the same as us and feeds polarity information to CNN. Then, we measured the object detection performances of the benchmark and ours on the same dataset. As a result, our method achieved a maximum of 7 points greater than the benchmark in the F1 score.

Keywords: event camera, dynamic vision sensor, deep learning, data representation, object recognition, low energy consumption

Procedia PDF Downloads 71
3488 Microseismicity of the Tehran Region Based on Three Seismic Networks

Authors: Jamileh Vasheghani Farahani

Abstract:

The main purpose of this research is to show the current active faults and active tectonic of the area by three seismic networks in Tehran region: 1-Tehran Disaster Mitigation and Management Organization (TDMMO), 2-Broadband Iranian National Seismic Network Center (BIN), 3-Iranian Seismological Center (IRSC). In this study, we analyzed microearthquakes happened in Tehran city and its surroundings using the Tehran networks from 1996 to 2015. We found some active faults and trends in the region. There is a 200-year history of historical earthquakes in Tehran. Historical and instrumental seismicity show that the east of Tehran is more active than the west. The Mosha fault in the North of Tehran is one of the active faults of the central Alborz. Moreover, other major faults in the region are Kahrizak, Eyvanakey, Parchin and North Tehran faults. An important seismicity region is an intersection of the Mosha and North Tehran fault systems (Kalan village in Lavasan). This region shows a cluster of microearthquakes. According to the historical and microseismic events analyzed in this research, there is a seismic gap in SE of Tehran. The empirical relationship is used to assess the Mmax based on the rupture length. There is a probability of occurrence of a strong motion of 7.0 to 7.5 magnitudes in the region (based on the assessed capability of the major faults such as Parchin and Eyvanekey faults and historical earthquakes).

Keywords: Iran, major faults, microseismicity, Tehran

Procedia PDF Downloads 343
3487 Higher Education Institution Students’ Perception on Educational Technology

Authors: Kuek Teik Sheng, Leaw Zee Guan, Lim Wah Kien, Ting Tin Tin

Abstract:

Educational technology such as YouTube and Kahoot have arisen as an alternative to effective learning among higher education institutions. There are many researches done in carrying out experiments to test different educational technologies and received positive feedback from students. Yet, similar study is hardly found in Malaysia especially study that includes the latest educational technologies. As a developing country, it is crucial to ensure that these emerging technologies are assisting students in learning process before it is widely adopted in institutions. This paper conducted a study to explore the perception of higher education institution students on the current educational technologies in Malaysia which include online educational games, online videos/course, social media, presentation tools and resource management tool. Some of these technologies have not been looked into its potential in effective learning process. An online survey using questionnaire is conducted among a target of 300 university/college. In the survey, the result shows that majority of the target students in Malaysia agree that the current educational technologies help them in learning, understanding and manage their studies. It is necessary to discover students’ perceptions on the educational technologies in order to provide guidelines for the educators/institutions in selecting appropriate technology to conduct the lecture/tutorial efficiently and effectively.

Keywords: education, educational technology, Facebook, PowerPoint, YouTube

Procedia PDF Downloads 219
3486 Budget Discipline and National Prosperity: The Nigerian Experience

Authors: Ben-Caleb Egbide, Iyoha Francis, Egharevba Mathew, Oduntan Emmanuel

Abstract:

The prosperity of any nation is determined not just by the availability of resources, but also by the discipline exercised in the management of those resources. This paper examines the functional association between adherence to budgetary estimates or budget discipline (BDISC) and national prosperity proxied by Real Gross Domestic Product (RGDP) and Relative Poverty Index (RPI)/Human Development Index (HDI). Adopting a longitudinal retrospective research strategy, time series data relating to both the endogenous and exogenous variables were extracted from official government publications for 36 years’ (1980-2015 in the case of RGDP and RPI), and for 26 years (1990-2015 in the case of HDI). Ordinary Least Square (OLS), as well as cointegration regressions, were employed to gauge both the short term and long term impact of BDISC on RPI/HDI and RGDP. The results indicated that BDISC is directly related with RGDP but indirectly related with RPI. The implication is that while adherence to budgetary estimate can enhance economic growth, it has the capacity to slow down the rate of poverty in the long run. The paper, therefore, recommend stricter adherence to budgets as a way out of economic under performance in Nigeria and engender the process of promoting human development and national prosperity.

Keywords: budget discipline, human development index, national prosperity, Nigeria

Procedia PDF Downloads 208
3485 Change of Substrate in Solid State Fermentation Can Produce Proteases and Phytases with Extremely Distinct Biochemical Characteristics and Promising Applications for Animal Nutrition

Authors: Paula K. Novelli, Margarida M. Barros, Luciana F. Flueri

Abstract:

Utilization of agricultural by-products, wheat ban and soybean bran, as substrate for solid state fermentation (SSF) was studied, aiming the achievement of different enzymes from Aspergillus sp. with distinct biological characteristics and its application and improvement on animal nutrition. Aspergillus niger and Aspergillus oryzea were studied as they showed very high yield of phytase and protease production, respectively. Phytase activity was measure using p-nitrophenilphosphate as substrate and a standard curve of p-nitrophenol, as the enzymatic activity unit was the quantity of enzyme necessary to release one μmol of p-nitrophenol. Protease activity was measure using azocasein as substrate. Activity for phytase and protease substantially increased when the different biochemical characteristics were considered in the study. Optimum pH and stability of the phytase produced by A. niger with wheat bran as substrate was between 4.0 - 5.0 and optimum temperature of activity was 37oC. Phytase fermented in soybean bran showed constant values at all pHs studied, for optimal and stability, but low production. Phytase with both substrates showed stable activity for temperatures higher than 80oC. Protease from A. niger showed very distinct behavior of optimum pH, acid for wheat bran and basic for soybean bran, respectively and optimal values of temperature and stability at 50oC. Phytase produced by A. oryzae in wheat bran had optimum pH and temperature of 9 and 37oC, respectively, but it was very unstable. On the other hand, proteases were stable at high temperatures, all pH’s studied and showed very high yield when fermented in wheat bran, however when it was fermented in soybean bran the production was very low. Subsequently the upscale production of phytase from A. niger and proteases from A. oryzae were applied as an enzyme additive in fish fed for digestibility studies. Phytases and proteases were produced with stable enzyme activity of 7,000 U.g-1 and 2,500 U.g-1, respectively. When those enzymes were applied in a plant protein based fish diet for digestibility studies, they increased protein, mineral, energy and lipids availability, showing that these new enzymes can improve animal production and performance. In conclusion, the substrate, as well as, the microorganism species can affect the biochemical character of the enzyme produced. Moreover, the production of these enzymes by SSF can be up to 90% cheaper than commercial ones produced with the same fungi species but submerged fermentation. Add to that these cheap enzymes can be easily applied as animal diet additives to improve production and performance.

Keywords: agricultural by-products, animal nutrition, enzymes production, solid state fermentation

Procedia PDF Downloads 302
3484 Trimma: Trimming Metadata Storage and Latency for Hybrid Memory Systems

Authors: Yiwei Li, Boyu Tian, Mingyu Gao

Abstract:

Hybrid main memory systems combine both performance and capacity advantages from heterogeneous memory technologies. With larger capacities, higher associativities, and finer granularities, hybrid memory systems currently exhibit significant metadata storage and lookup overheads for flexibly remapping data blocks between the two memory tiers. To alleviate the inefficiencies of existing designs, we propose Trimma, the combination of a multi-level metadata structure and an efficient metadata cache design. Trimma uses a multilevel metadata table to only track truly necessary address remap entries. The saved memory space is effectively utilized as extra DRAM cache capacity to improve performance. Trimma also uses separate formats to store the entries with non-identity and identity mappings. This improves the overall remap cache hit rate, further boosting the performance. Trimma is transparent to software and compatible with various types of hybrid memory systems. When evaluated on a representative DDR4 + NVM hybrid memory system, Trimma achieves up to 2.4× and on average 58.1% speedup benefits, compared with a state-of-the-art design that only leverages the unallocated fast memory space for caching. Trimma addresses metadata management overheads and targets future scalable large-scale hybrid memory architectures.

Keywords: memory system, data cache, hybrid memory, non-volatile memory

Procedia PDF Downloads 36
3483 Customized Temperature Sensors for Sustainable Home Appliances

Authors: Merve Yünlü, Nihat Kandemir, Aylin Ersoy

Abstract:

Temperature sensors are used in home appliances not only to monitor the basic functions of the machine but also to minimize energy consumption and ensure safe operation. In parallel with the development of smart home applications and IoT algorithms, these sensors produce important data such as the frequency of use of the machine, user preferences, and the compilation of critical data in terms of diagnostic processes for fault detection throughout an appliance's operational lifespan. Commercially available thin-film resistive temperature sensors have a well-established manufacturing procedure that allows them to operate over a wide temperature range. However, these sensors are over-designed for white goods applications. The operating temperature range of these sensors is between -70°C and 850°C, while the temperature range requirement in home appliance applications is between 23°C and 500°C. To ensure the operation of commercial sensors in this wide temperature range, usually, a platinum coating of approximately 1-micron thickness is applied to the wafer. However, the use of platinum in coating and the high coating thickness extends the sensor production process time and therefore increases sensor costs. In this study, an attempt was made to develop a low-cost temperature sensor design and production method that meets the technical requirements of white goods applications. For this purpose, a custom design was made, and design parameters (length, width, trim points, and thin film deposition thickness) were optimized by using statistical methods to achieve the desired resistivity value. To develop thin film resistive temperature sensors, one side polished sapphire wafer was used. To enhance adhesion and insulation 100 nm silicon dioxide was coated by inductively coupled plasma chemical vapor deposition technique. The lithography process was performed by a direct laser writer. The lift-off process was performed after the e-beam evaporation of 10 nm titanium and 280 nm platinum layers. Standard four-point probe sheet resistance measurements were done at room temperature. The annealing process was performed. Resistivity measurements were done with a probe station before and after annealing at 600°C by using a rapid thermal processing machine. Temperature dependence between 25-300 °C was also tested. As a result of this study, a temperature sensor has been developed that has a lower coating thickness than commercial sensors but can produce reliable data in the white goods application temperature range. A relatively simplified but optimized production method has also been developed to produce this sensor.

Keywords: thin film resistive sensor, temperature sensor, household appliance, sustainability, energy efficiency

Procedia PDF Downloads 50
3482 Comparison of Compression Properties of Stretchable Knitted Fabrics and Bi-Stretch Woven Fabrics for Compression Garments

Authors: Muhammad Maqsood, Yasir Nawab, Syed Talha Ali Hamdani

Abstract:

Stretchable fabrics have diverse applications ranging from casual apparel to performance sportswear and compression therapy. Compression therapy is the universally accepted treatment for the management of hypertrophic scarring after severe burns. Mostly stretchable knitted fabrics are used in compression therapy but in the recent past, some studies have also been found on bi-stretch woven fabrics being used as compression garments as they also have been found quite effective in the treatment of oedema. Therefore, the objective of the present study is to compare the compression properties of stretchable knitted and bi-stretch woven fabrics for compression garments. For this purpose four woven structures and four knitted structures were produced having the same areal density and their compression, comfort and mechanical properties were compared before and after 5, 10 and 15 washes. Four knitted structures used were single jersey, single locaste, plain pique and the honeycomb, whereas four woven structures produced were 1/1 plain, 2/1 twill, 3/1 twill and 4/1 twill. The compression properties of the produced samples were tested by using kikuhime pressure sensor and it was found that bi-stretch woven fabrics possessed better compression properties before and after washes and retain their durability after repeated use, whereas knitted stretchable fabrics lost their compression ability after repeated use and the required sub garment pressure of the knitted structures after 15 washes was almost half to that of woven bi-stretch fabrics.

Keywords: compression garments, knitted structures, medical textiles, woven bi-stretch

Procedia PDF Downloads 379
3481 Extrudable Foamed Concrete: General Benefits in Prefabrication and Comparison in Terms of Fresh Properties and Compressive Strength with Classic Foamed Concrete

Authors: D. Falliano, G. Ricciardi, E. Gugliandolo

Abstract:

Foamed concrete belongs to the category of lightweight concrete. It is characterized by a density which is generally ranging from 200 to 2000 kg/m³ and typically comprises cement, water, preformed foam, fine sand and eventually fine particles such as fly ash or silica fume. The foam component mixed with the cement paste give rise to the development of a system of air-voids in the cementitious matrix. The peculiar characteristics of foamed concrete elements are summarized in the following aspects: 1) lightness which allows reducing the dimensions of the resisting frame structure and is advantageous in the scope of refurbishment or seismic retrofitting in seismically vulnerable areas; 2) thermal insulating properties, especially in the case of low densities; 3) the good resistance against fire as compared to ordinary concrete; 4) the improved workability; 5) cost-effectiveness due to the usage of rather simple constituting elements that are easily available locally. Classic foamed concrete cannot be extruded, as the dimensional stability is not permitted in the green state and this severely limits the possibility of industrializing them through a simple and cost-effective process, characterized by flexibility and high production capacity. In fact, viscosity enhancing agents (VEA) used to extrude traditional concrete, in the case of foamed concrete cause the collapsing of air bubbles, so that it is impossible to extrude a lightweight product. These requirements have suggested the study of a particular additive that modifies the rheology of foamed concrete fresh paste by increasing cohesion and viscosity and, at the same time, stabilizes the bubbles into the cementitious matrix, in order to allow the dimensional stability in the green state and, consequently, the extrusion of a lightweight product. There are plans to submit the additive’s formulation to patent. In addition to the general benefits of using the extrusion process, extrudable foamed concrete allow other limits to be exceeded: elimination of formworks, expanded application spectrum, due to the possibility of extrusion in a range varying between 200 and 2000 kg/m³, which allows the prefabrication of both structural and non-structural constructive elements. Besides, this contribution aims to present the significant differences regarding extrudable and classic foamed concrete fresh properties in terms of slump. Plastic air content, plastic density, hardened density and compressive strength have been also evaluated. The outcomes show that there are no substantial differences between extrudable and classic foamed concrete compression resistances.

Keywords: compressive strength, extrusion, foamed concrete, fresh properties, plastic air content, slump.

Procedia PDF Downloads 151
3480 An Inquiry of the Impact of Flood Risk on Housing Market with Enhanced Geographically Weighted Regression

Authors: Lin-Han Chiang Hsieh, Hsiao-Yi Lin

Abstract:

This study aims to determine the impact of the disclosure of flood potential map on housing prices. The disclosure is supposed to mitigate the market failure by reducing information asymmetry. On the other hand, opponents argue that the official disclosure of simulated results will only create unnecessary disturbances on the housing market. This study identifies the impact of the disclosure of the flood potential map by comparing the hedonic price of flood potential before and after the disclosure. The flood potential map used in this study is published by Taipei municipal government in 2015, which is a result of a comprehensive simulation based on geographical, hydrological, and meteorological factors. The residential property sales data of 2013 to 2016 is used in this study, which is collected from the actual sales price registration system by the Department of Land Administration (DLA). The result shows that the impact of flood potential on residential real estate market is statistically significant both before and after the disclosure. But the trend is clearer after the disclosure, suggesting that the disclosure does have an impact on the market. Also, the result shows that the impact of flood potential differs by the severity and frequency of precipitation. The negative impact for a relatively mild, high frequency flood potential is stronger than that for a heavy, low possibility flood potential. The result indicates that home buyers are of more concern to the frequency, than the intensity of flood. Another contribution of this study is in the methodological perspective. The classic hedonic price analysis with OLS regression suffers from two spatial problems: the endogeneity problem caused by omitted spatial-related variables, and the heterogeneity concern to the presumption that regression coefficients are spatially constant. These two problems are seldom considered in a single model. This study tries to deal with the endogeneity and heterogeneity problem together by combining the spatial fixed-effect model and geographically weighted regression (GWR). A series of literature indicates that the hedonic price of certain environmental assets varies spatially by applying GWR. Since the endogeneity problem is usually not considered in typical GWR models, it is arguable that the omitted spatial-related variables might bias the result of GWR models. By combing the spatial fixed-effect model and GWR, this study concludes that the effect of flood potential map is highly sensitive by location, even after controlling for the spatial autocorrelation at the same time. The main policy application of this result is that it is improper to determine the potential benefit of flood prevention policy by simply multiplying the hedonic price of flood risk by the number of houses. The effect of flood prevention might vary dramatically by location.

Keywords: flood potential, hedonic price analysis, endogeneity, heterogeneity, geographically-weighted regression

Procedia PDF Downloads 268
3479 An A-Star Approach for the Quickest Path Problem with Time Windows

Authors: Christofas Stergianos, Jason Atkin, Herve Morvan

Abstract:

As air traffic increases, more airports are interested in utilizing optimization methods. Many processes happen in parallel at an airport, and complex models are needed in order to have a reliable solution that can be implemented for ground movement operations. The ground movement for aircraft in an airport, allocating a path to each aircraft to follow in order to reach their destination (e.g. runway or gate), is one process that could be optimized. The Quickest Path Problem with Time Windows (QPPTW) algorithm has been developed to provide a conflict-free routing of vehicles and has been applied to routing aircraft around an airport. It was subsequently modified to increase the accuracy for airport applications. These modifications take into consideration specific characteristics of the problem, such as: the pushback process, which considers the extra time that is needed for pushing back an aircraft and turning its engines on; stand holding where any waiting should be allocated to the stand; and runway sequencing, where the sequence of the aircraft that take off is optimized and has to be respected. QPPTW involves searching for the quickest path by expanding the search in all directions, similarly to Dijkstra’s algorithm. Finding a way to direct the expansion can potentially assist the search and achieve a better performance. We have further modified the QPPTW algorithm to use a heuristic approach in order to guide the search. This new algorithm is based on the A-star search method but estimates the remaining time (instead of distance) in order to assess how far the target is. It is important to consider the remaining time that it is needed to reach the target, so that delays that are caused by other aircraft can be part of the optimization method. All of the other characteristics are still considered and time windows are still used in order to route multiple aircraft rather than a single aircraft. In this way the quickest path is found for each aircraft while taking into account the movements of the previously routed aircraft. After running experiments using a week of real aircraft data from Zurich Airport, the new algorithm (A-star QPPTW) was found to route aircraft much more quickly, being especially fast in routing the departing aircraft where pushback delays are significant. On average A-star QPPTW could route a full day (755 to 837 aircraft movements) 56% faster than the original algorithm. In total the routing of a full week of aircraft took only 12 seconds with the new algorithm, 15 seconds faster than the original algorithm. For real time application, the algorithm needs to be very fast, and this speed increase will allow us to add additional features and complexity, allowing further integration with other processes in airports and leading to more optimized and environmentally friendly airports.

Keywords: a-star search, airport operations, ground movement optimization, routing and scheduling

Procedia PDF Downloads 206
3478 Experiences and Aspirations of Hearing Impaired Learners in Inclusive Classrooms

Authors: Raymon P. Española

Abstract:

Hearing impaired students are admitted to regular high schools in the context of inclusive education. In this setting, several academic difficulties and social struggles are disregarded by many educators. The study aimed to describe the aspirations and lived experiences in mainstream classrooms of hearing impaired students. In the research process, the participants were interviewed using sign language. Thematic analysis of interview responses was done, supplemented by interviews with teachers and classroom observations. The study revealed four patterns of experiences: academic difficulties, coping mechanisms, identification with hearing peers, and impression management. This means that these learners were struggling in inclusive classrooms, where identification with and modeling the positive qualities of hearing peers were done to cope with academic difficulties and alter negative impressions about them. By implication, these learners tended to socially immerse themselves rather than resort to isolation. Along with this tendency was the aspiration for achievement as they were eager to finish post-secondary technical-vocational education. This means aspiring for continuing social immersion into the mainstream. All these findings provide insights to K-12 educators to increase the use of collaborative techniques and experiential learning strategies, as well as to adequately address the special educational needs of these students.

Keywords: descriptive, experiences and aspirations of hearing impaired learners, inclusive classrooms, Surigao City Philippines

Procedia PDF Downloads 386
3477 The Application of Raman Spectroscopy in Olive Oil Analysis

Authors: Silvia Portarena, Chiara Anselmi, Chiara Baldacchini, Enrico Brugnoli

Abstract:

Extra virgin olive oil (EVOO) is a complex matrix mainly composed by fatty acid and other minor compounds, among which carotenoids are well known for their antioxidative function that is a key mechanism of protection against cancer, cardiovascular diseases, and macular degeneration in humans. EVOO composition in terms of such constituents is generally the result of a complex combination of genetic, agronomical and environmental factors. To selectively improve the quality of EVOOs, the role of each factor on its biochemical composition need to be investigated. By selecting fruits from four different cultivars similarly grown and harvested, it was demonstrated that Raman spectroscopy, combined with chemometric analysis, is able to discriminate the different cultivars, also as a function of the harvest date, based on the relative content and composition of fatty acid and carotenoids. In particular, a correct classification up to 94.4% of samples, according to the cultivar and the maturation stage, was obtained. Moreover, by using gas chromatography and high-performance liquid chromatography as reference techniques, the Raman spectral features further allowed to build models, based on partial least squares regression, that were able to predict the relative amount of the main fatty acids and the main carotenoids in EVOO, with high coefficients of determination. Besides genetic factors, climatic parameters, such as light exposition, distance from the sea, temperature, and amount of precipitations could have a strong influence on EVOO composition of both major and minor compounds. This suggests that the Raman spectra could act as a specific fingerprint for the geographical discrimination and authentication of EVOO. To understand the influence of environment on EVOO Raman spectra, samples from seven regions along the Italian coasts were selected and analyzed. In particular, it was used a dual approach combining Raman spectroscopy and isotope ratio mass spectrometry (IRMS) with principal component and linear discriminant analysis. A correct classification of 82% EVOO based on their regional geographical origin was obtained. Raman spectra were obtained by Super Labram spectrometer equipped with an Argon laser (514.5 nm wavelenght). Analyses of stable isotope content ratio were performed using an isotope ratio mass spectrometer connected to an elemental analyzer and to a pyrolysis system. These studies demonstrate that RR spectroscopy is a valuable and useful technique for the analysis of EVOO. In combination with statistical analysis, it makes possible the assessment of specific samples’ content and allows for classifying oils according to their geographical and varietal origin.

Keywords: authentication, chemometrics, olive oil, raman spectroscopy

Procedia PDF Downloads 304
3476 Safety Validation of Black-Box Autonomous Systems: A Multi-Fidelity Reinforcement Learning Approach

Authors: Jared Beard, Ali Baheri

Abstract:

As autonomous systems become more prominent in society, ensuring their safe application becomes increasingly important. This is clearly demonstrated with autonomous cars traveling through a crowded city or robots traversing a warehouse with heavy equipment. Human environments can be complex, having high dimensional state and action spaces. This gives rise to two problems. One being that analytic solutions may not be possible. The other is that in simulation based approaches, searching the entirety of the problem space could be computationally intractable, ruling out formal methods. To overcome this, approximate solutions may seek to find failures or estimate their likelihood of occurrence. One such approach is adaptive stress testing (AST) which uses reinforcement learning to induce failures in the system. The premise of which is that a learned model can be used to help find new failure scenarios, making better use of simulations. In spite of these failures AST fails to find particularly sparse failures and can be inclined to find similar solutions to those found previously. To help overcome this, multi-fidelity learning can be used to alleviate this overuse of information. That is, information in lower fidelity can simulations can be used to build up samples less expensively, and more effectively cover the solution space to find a broader set of failures. Recent work in multi-fidelity learning has passed information bidirectionally using “knows what it knows” (KWIK) reinforcement learners to minimize the number of samples in high fidelity simulators (thereby reducing computation time and load). The contribution of this work, then, is development of the bidirectional multi-fidelity AST framework. Such an algorithm, uses multi-fidelity KWIK learners in an adversarial context to find failure modes. Thus far, a KWIK learner has been used to train an adversary in a grid world to prevent an agent from reaching its goal; thus demonstrating the utility of KWIK learners in an AST framework. The next step is implementation of the bidirectional multi-fidelity AST framework described. Testing will be conducted in a grid world containing an agent attempting to reach a goal position and adversary tasked with intercepting the agent as demonstrated previously. Fidelities will be modified by adjusting the size of a time-step, with higher-fidelity effectively allowing for more responsive closed loop feedback. Results will compare the single KWIK AST learner with the multi-fidelity algorithm with respect to number of samples, distinct failure modes found, and relative effect of learning after a number of trials.

Keywords: multi-fidelity reinforcement learning, multi-fidelity simulation, safety validation, falsification

Procedia PDF Downloads 130
3475 Enrichment and Flux of Heavy Metals along the Coastal Sediments of Pakistan

Authors: Asmat Siddiqui, Noor Us Saher

Abstract:

Heavy metal contamination in the marine environment is a global issue, and in past decades, this problem has intensified due to an increase in urbanization and industrialization, especially in developing countries. Marine sediments act as a preliminary indicator of heavy metal contamination in the coastal and estuarine environment, which has adverse effects on biota as well as in the marine system. The aim of the current study was to evaluate the contamination status, enrichment, and flux of heavy metals in two monitoring years from coastal sediments of Pakistan. A total of 74 sediment samples were collected from seven coastal areas of Pakistan in two monitoring years, 2001-03 (MY-I) and 2011-13 (MY-II). The geochemical properties (grain size analysis, organic contents and eight heavy metals, i.e. Fe, Zn, Cu, Cr, Ni, Co, Pb, and Cd) of all sediment samples were analyzed. A significant increase in Fe, Ni and Cr concentrations detected between the years, whereas no significant differences were exhibited in Cu, Zn, Co, Pb and Cd concentrations. The extremely high enrichment (>50) of Cu, Zn, Pb and Cd were scrutinized in both monitoring years. The annual deposition flux of heavy metals ranged from 0.63 to 66.44 and 0.78 to 68.27 tons per year in MY-I and MY-II, respectively, with the lowest flux evaluated for Cd and highest for Zn in both monitoring years. A significant increase (p <0.05) was observed in the burial flux of Cr and Ni during the last decade in coastal sediments. The use of geo-indicators is helpful to assess the contamination analysis for management and conservation of the marine environment.

Keywords: coastal contamination, enrichment factor, geo-indicator, heavy metal flux

Procedia PDF Downloads 352
3474 Sequential and Combinatorial Pre-Treatment Strategy of Lignocellulose for the Enhanced Enzymatic Hydrolysis of Spent Coffee Waste

Authors: Rajeev Ravindran, Amit K. Jaiswal

Abstract:

Waste from the food-processing industry is produced in large amount and contains high levels of lignocellulose. Due to continuous accumulation throughout the year in large quantities, it creates a major environmental problem worldwide. The chemical composition of these wastes (up to 75% of its composition is contributed by polysaccharide) makes it inexpensive raw material for the production of value-added products such as biofuel, bio-solvents, nanocrystalline cellulose and enzymes. In order to use lignocellulose as the raw material for the microbial fermentation, the substrate is subjected to enzymatic treatment, which leads to the release of reducing sugars such as glucose and xylose. However, the inherent properties of lignocellulose such as presence of lignin, pectin, acetyl groups and the presence of crystalline cellulose contribute to recalcitrance. This leads to poor sugar yields upon enzymatic hydrolysis of lignocellulose. A pre-treatment method is generally applied before enzymatic treatment of lignocellulose that essentially removes recalcitrant components in biomass through structural breakdown. Present study is carried out to find out the best pre-treatment method for the maximum liberation of reducing sugars from spent coffee waste (SPW). SPW was subjected to a range of physical, chemical and physico-chemical pre-treatment followed by a sequential, combinatorial pre-treatment strategy is also applied on to attain maximum sugar yield by combining two or more pre-treatments. All the pre-treated samples were analysed for total reducing sugar followed by identification and quantification of individual sugar by HPLC coupled with RI detector. Besides, generation of any inhibitory compounds such furfural, hydroxymethyl furfural (HMF) which can hinder microbial growth and enzyme activity is also monitored. Results showed that ultrasound treatment (31.06 mg/L) proved to be the best pre-treatment method based on total reducing content followed by dilute acid hydrolysis (10.03 mg/L) while galactose was found to be the major monosaccharide present in the pre-treated SPW. Finally, the results obtained from the study were used to design a sequential lignocellulose pre-treatment protocol to decrease the formation of enzyme inhibitors and increase sugar yield on enzymatic hydrolysis by employing cellulase-hemicellulase consortium. Sequential, combinatorial treatment was found better in terms of total reducing yield and low content of the inhibitory compounds formation, which could be due to the fact that this mode of pre-treatment combines several mild treatment methods rather than formulating a single one. It eliminates the need for a detoxification step and potential application in the valorisation of lignocellulosic food waste.

Keywords: lignocellulose, enzymatic hydrolysis, pre-treatment, ultrasound

Procedia PDF Downloads 344