Search results for: automatic mapping
352 Results of Three-Year Operation of 220kV Pilot Superconducting Fault Current Limiter in Moscow Power Grid
Authors: M. Moyzykh, I. Klichuk, L. Sabirov, D. Kolomentseva, E. Magommedov
Abstract:
Modern city electrical grids are forced to increase their density due to the increasing number of customers and requirements for reliability and resiliency. However, progress in this direction is often limited by the capabilities of existing network equipment. New energy sources or grid connections increase the level of short-circuit currents in the adjacent network, which can exceed the maximum rating of equipment–breaking capacity of circuit breakers, thermal and dynamic current withstand qualities of disconnectors, cables, and transformers. Superconducting fault current limiter (SFCL) is a modern solution designed to deal with the increasing fault current levels in power grids. The key feature of this device is its instant (less than 2 ms) limitation of the current level due to the nature of the superconductor. In 2019 Moscow utilities installed SuperOx SFCL in the city power grid to test the capabilities of this novel technology. The SFCL became the first SFCL in the Russian energy system and is currently the most powerful SFCL in the world. Modern SFCL uses second-generation high-temperature superconductor (2G HTS). Despite its name, HTS still requires low temperatures of liquid nitrogen for operation. As a result, Moscow SFCL is built with a cryogenic system to provide cooling to the superconductor. The cryogenic system consists of three cryostats that contain a superconductor part and are filled with liquid nitrogen (three phases), three cryocoolers, one water chiller, three cryopumps, and pressure builders. All these components are controlled by an automatic control system. SFCL has been continuously operating on the city grid for over three years. During that period of operation, numerous faults occurred, including cryocooler failure, chiller failure, pump failure, and others (like a cryogenic system power outage). All these faults were eliminated without an SFCL shut down due to the specially designed cryogenic system backups and quick responses of grid operator utilities and the SuperOx crew. The paper will describe in detail the results of SFCL operation and cryogenic system maintenance and what measures were taken to solve and prevent similar faults in the future.Keywords: superconductivity, current limiter, SFCL, HTS, utilities, cryogenics
Procedia PDF Downloads 83351 Little Girls and Big Stories: A Thematic Analysis of Gender Representations in Selected Asian Room to Read Storybooks
Authors: Cheeno Marlo Sayuno
Abstract:
Room to Read is an international nonprofit organization aimed at empowering young readers through literature and literacy education. In particular, the organization is focused on girls’ education in schools and bettering their social status through crafting stories and making sure that these stories are accessible to them. In 2019, Room to Read visited the Philippines and partnered with Philippine children’s literature publishers Adarna House, Lampara Books, Anvil Publishing, and OMF-Hiyas with the goal of producing contextualized stories that Filipino children can read. The result is a set of 20 storybooks developed by Filipino writers and illustrators, the author of this paper included. The project led to narratives of experiences in storybook production from conceptualization to publication, towards translations and reimagining in online repository, storytelling, and audiobook formats. During the production process, we were particularly reminded of gender representations, child’s rights, and telling stories that can empower the children in vulnerable communities, who are the beneficiaries of the project. The storybooks, along with many others produced in Asia and the world, are available online through the literacycloud.org website of Room to Read. In this study, the goal is to survey the stories produced in Asia and look at how gender is represented in the storybooks. By analyzing both the texts and the illustrations of the storybooks produced across Asian countries, themes of portrayals of young boys and girls, their characteristics and narratives, and how they are empowered in the stories are identified, with the goal of mapping how Room to Read is able to address the problem of access to literacy among young girls and ensuring them that they can do anything, the way they are portrayed in the stories. The paper hopes to determine how gender is represented in Asian storybooks produced by the international nonprofit organization Room to Read. Thematic textual analysis was used as methodology, where the storybooks are analyzed qualitatively to identify arising themes of gender representation. This study will shed light on the importance of responsible portrayal of gender in storybooks and how it can impact and empower children. The results of the study can also aid writers and illustrators in developing gender-sensitive storybooks.Keywords: room to read, asian storybooks, young girls, thematic analysis, child empowerment, literacy, education
Procedia PDF Downloads 79350 The Need for Automation in the Domestic Food Processing Sector and its Impact
Authors: Shantam Gupta
Abstract:
The objective of this study is to address the critical need for automation in the domestic food processing sector and study its impact. Food is the one of the most basic physiological needs essential for the survival of a living being. Some of them have the capacity to prepare their own food (like most plants) and henceforth are designated as primary food producers; those who depend on these primary food producers for food form the primary consumers’ class (herbivores). Some of the organisms relying on the primary food are the secondary food consumers (carnivores). There is a third class of consumers called tertiary food consumers/apex food consumers that feed on both the primary and secondary food consumers. Humans form an essential part of the apex predators and are generally at the top of the food chain. But still further disintegration of the food habits of the modern human i.e. Homo sapiens, reveals that humans depend on other individuals for preparing their own food. The old notion of eating raw/brute food is long gone and food processing has become very trenchant in lives of modern human. This has led to an increase in dependence on other individuals for ‘processing’ the food before it can be actually consumed by the modern human. This has led to a further shift of humans in the classification of food chain of consumers. The effects of the shifts shall be systematically investigated in this paper. The processing of food has a direct impact on the economy of the individual (consumer). Also most individuals depend on other processing individuals for the preparation of food. This dependency leads to establishment of a vital link of dependency in the food web which when altered can adversely affect the food web and can have dire consequences on the health of the individual. This study investigates the challenges arising out due to this dependency and the impact of food processing on the economy of the individual. A comparison of Industrial food processing and processing at domestic platforms (households and restaurants) has been made to provide an idea about the present scenario of automation in the food processing sector. A lot of time and energy is also consumed while processing food at home for consumption. The high frequency of consumption of meals (greater than 2 times a day) makes it even more laborious. Through the medium of this study a pressing need for development of an automatic cooking machine is proposed with a mission to reduce the inter-dependency & human effort of individuals required for the preparation of food (by automation of the food preparation process) and make them more self-reliant The impact of development of this product has also further been profoundly discussed. Assumption used: The individuals those who process food also consume the food that they produce. (They are also termed as ‘independent’ or ‘self-reliant’ modern human beings.)Keywords: automation, food processing, impact on economy, processing individual
Procedia PDF Downloads 470349 Spanish Language Violence Corpus: An Analysis of Offensive Language in Twitter
Authors: Beatriz Botella-Gil, Patricio Martínez-Barco, Lea Canales
Abstract:
The Internet and ICT are an integral element of and omnipresent in our daily lives. Technologies have changed the way we see the world and relate to it. The number of companies in the ICT sector is increasing every year, and there has also been an increase in the work that occurs online, from sending e-mails to the way companies promote themselves. In social life, ICT’s have gained momentum. Social networks are useful for keeping in contact with family or friends that live far away. This change in how we manage our relationships using electronic devices and social media has been experienced differently depending on the age of the person. According to currently available data, people are increasingly connected to social media and other forms of online communication. Therefore, it is no surprise that violent content has also made its way to digital media. One of the important reasons for this is the anonymity provided by social media, which causes a sense of impunity in the victim. Moreover, it is not uncommon to find derogatory comments, attacking a person’s physical appearance, hobbies, or beliefs. This is why it is necessary to develop artificial intelligence tools that allow us to keep track of violent comments that relate to violent events so that this type of violent online behavior can be deterred. The objective of our research is to create a guide for detecting and recording violent messages. Our annotation guide begins with a study on the problem of violent messages. First, we consider the characteristics that a message should contain for it to be categorized as violent. Second, the possibility of establishing different levels of aggressiveness. To download the corpus, we chose the social network Twitter for its ease of obtaining free messages. We chose two recent, highly visible violent cases that occurred in Spain. Both of them experienced a high degree of social media coverage and user comments. Our corpus has a total of 633 messages, manually tagged, according to the characteristics we considered important, such as, for example, the verbs used, the presence of exclamations or insults, and the presence of negations. We consider it necessary to create wordlists that are present in violent messages as indicators of violence, such as lists of negative verbs, insults, negative phrases. As a final step, we will use automatic learning systems to check the data obtained and the effectiveness of our guide.Keywords: human language technologies, language modelling, offensive language detection, violent online content
Procedia PDF Downloads 133348 Understanding the Coping Experience of Mothers with Childhood Trauma Histories: A Qualitative Study
Authors: Chan Yan Nok
Abstract:
The present study is a qualitative study based on the coping experiences of six Hong Kong Chinese mothers who had childhood trauma from their first-person perspective. Expanding the perspective beyond the dominant discourse of “inter-generation transmission of trauma”, this study explores the experiences and meanings of child trauma embedded in their narratives through the process of thematic analysis and narrative analysis. The interviewees painted a nuanced picture of their process of coping and trauma resolution. First, acknowledgement; second, feel safe and start to tell the story of trauma; third, feel the feelings and expression of emotions; fourth, clarifying and coping with the impacts of trauma; fifth, integration and transformation; and sixth, using their new understanding of experience to have a better life. It was seen that there was no “end” within the process of trauma resolution. Instead, this is an ongoing process with positive healing trajectory. Analysis of the stories of the mothers revealed recurrent themes around continuous self-reflective awareness in the process of trauma coping. Rather than being necessarily negative and detrimental, childhood trauma could highlight the meanings of being a mother and reveal opportunities for continuous personal growth and self-enhancement. Utilizing the sense of inadequacy as a core driver in the trauma recovery process while developing a heightened awareness of the unfinished business embedded in their “automatic pattern” of behaviors, emotions, and thoughts can help these mothers become more flexible to formulate new methods in facing future predicaments. Future social work and parent education practices should help mothers deal with unresolved trauma, make sense of their impacts of childhood trauma and discover the growth embedded in the past traumatic experience. They should be facilitated in “acknowledging the reality of the trauma”, including understanding their complicated emotions arising from the traumatic experiences and voicing their struggles. In addition, helping these mothers to be aware of short-term and long-term trauma impacts (i.e., secondary responses to the trauma) and explore their effective coping strategies in “overcoming secondary responses to the trauma” are crucial for their future positive adjustment and transformation. Through affirming their coping abilities and lessons learnt from past experiences, mothers can reduce feelings of shame and powerlessness and enhance their parental capacity.Keywords: childhood trauma, coping, mothers, self-awareness, self-reflection, trauma resolution
Procedia PDF Downloads 167347 Generative Pre-Trained Transformers (GPT-3) and Their Impact on Higher Education
Authors: Sheelagh Heugh, Michael Upton, Kriya Kalidas, Stephen Breen
Abstract:
This article aims to create awareness of the opportunities and issues the artificial intelligence (AI) tool GPT-3 (Generative Pre-trained Transformer-3) brings to higher education. Technological disruptors have featured in higher education (HE) since Konrad Klaus developed the first functional programmable automatic digital computer. The flurry of technological advances, such as personal computers, smartphones, the world wide web, search engines, and artificial intelligence (AI), have regularly caused disruption and discourse across the educational landscape around harnessing the change for the good. Accepting AI influences are inevitable; we took mixed methods through participatory action research and evaluation approach. Joining HE communities, reviewing the literature, and conducting our own research around Chat GPT-3, we reviewed our institutional approach to changing our current practices and developing policy linked to assessments and the use of Chat GPT-3. We review the impact of GPT-3, a high-powered natural language processing (NLP) system first seen in 2020 on HE. Historically HE has flexed and adapted with each technological advancement, and the latest debates for educationalists are focusing on the issues around this version of AI which creates natural human language text from prompts and other forms that can generate code and images. This paper explores how Chat GPT-3 affects the current educational landscape: we debate current views around plagiarism, research misconduct, and the credibility of assessment and determine the tool's value in developing skills for the workplace and enhancing critical analysis skills. These questions led us to review our institutional policy and explore the effects on our current assessments and the development of new assessments. Conclusions: After exploring the pros and cons of Chat GTP-3, it is evident that this form of AI cannot be un-invented. Technology needs to be harnessed for positive outcomes in higher education. We have observed that materials developed through AI and potential effects on our development of future assessments and teaching methods. Materials developed through Chat GPT-3 can still aid student learning but lead to redeveloping our institutional policy around plagiarism and academic integrity.Keywords: artificial intelligence, Chat GPT-3, intellectual property, plagiarism, research misconduct
Procedia PDF Downloads 89346 Assessment of Hypersaline Outfalls via Computational Fluid Dynamics Simulations: A Case Study of the Gold Coast Desalination Plant Offshore Multiport Brine Diffuser
Authors: Mitchell J. Baum, Badin Gibbes, Greg Collecutt
Abstract:
This study details a three-dimensional field-scale numerical investigation conducted for the Gold Coast Desalination Plant (GCDP) offshore multiport brine diffuser. Quantitative assessment of diffuser performance with regard to trajectory, dilution and mapping of seafloor concentration distributions was conducted for 100% plant operation. The quasi-steady Computational Fluid Dynamics (CFD) simulations were performed using the Reynolds averaged Navier-Stokes equations with a k-ω shear stress transport turbulence closure scheme. The study compliments a field investigation, which measured brine plume characteristics under similar conditions. CFD models used an iterative mesh in a domain with dimensions 400 m long, 200 m wide and an average depth of 24.2 m. Acoustic Doppler current profiler measurements conducted in the companion field study exhibited considerable variability over the water column. The effect of this vertical variability on simulated discharge outcomes was examined. Seafloor slope was also accommodated into the model. Ambient currents varied predominantly in the longshore direction – perpendicular to the diffuser structure. Under these conditions, the alternating port orientation of the GCDP diffuser resulted in simultaneous subjection to co-propagating and counter-propagating ambient regimes. Results from quiescent ambient simulations suggest broad agreement with empirical scaling arguments traditionally employed in design and regulatory assessments. Simulated dynamic ambient regimes showed the influence of ambient crossflow upon jet trajectory, dilution and seafloor concentration is significant. The effect of ambient flow structure and the subsequent influence on jet dynamics is discussed, along with the implications for using these different simulation approaches to inform regulatory decisions.Keywords: computational fluid dynamics, desalination, field-scale simulation, multiport brine diffuser, negatively buoyant jet
Procedia PDF Downloads 215345 Nursing Students' Experience of Using Electronic Health Record System in Clinical Placements
Authors: Nurten Tasdemir, Busra Baloglu, Zeynep Cingoz, Can Demirel, Zeki Gezer, Barıs Efe
Abstract:
Student nurses are increasingly exposed to technology in the workplace after graduation with the growing numbers of electric health records (EHRs), handheld computers, barcode scanner medication dispensing systems, and automatic capture of patient data such as vital signs. Internationally, electronic health records (EHRs) systems are being implemented and evaluated. Students will inevitably encounter EHRs in the clinical learning environment and their professional practice. Nursing students must develop competency in the use of EHR. Aim: The study aimed to examine nursing students’ experiences of learning to use electronic health records (EHR) in clinical placements. Method: This study adopted a descriptive approach. The study population consisted of second and third-year nursing students at the Zonguldak School of Health in the West Black Sea Region of Turkey; the study was conducted during the 2015–2016 academic year. The sample consisted of 315 (74.1% of 425 students) nursing students who volunteered to participate. The students, who were involved in clinical practice, were invited to participate in the study Data were collected by a questionnaire designed by the researchers based on the relevant literature. Data were analyzed descriptively using the Statistical Package for Social Sciences (SPSS) for Windows version 16.0. The data are presented as means, standard deviations, and percentages. Approval for the study was obtained from the Ethical Committee of the University (Reg. Number: 29/03/2016/112) and the director of Nursing Department. Findings: A total of 315 students enrolled in this study, for a response rate of 74.1%. The mean age of the sample was 22.24 ± 1.37 (min: 19, max: 32) years, and most participants (79.7%) were female. Most of the nursing students (82.3%) stated that they use information technologies in clinical practice. Nearly half of the students (42.5%) reported that they have not accessed to EHR system. In addition, 61.6% of the students reported that insufficient computers available in clinical placement. Of the students, 84.7% reported that they prefer to have patient information from EHR system, and 63.8% of them found more effective to preparation for the clinical reporting. Conclusion: This survey indicated that nursing students experience to learn about EHR systems in clinical placements. For more effective learning environment nursing education should prepare nursing students for EHR systems in their educational life.Keywords: electronic health record, clinical placement, nursing student, nursing education
Procedia PDF Downloads 292344 Efficient Reuse of Exome Sequencing Data for Copy Number Variation Callings
Authors: Chen Wang, Jared Evans, Yan Asmann
Abstract:
With the quick evolvement of next-generation sequencing techniques, whole-exome or exome-panel data have become a cost-effective way for detection of small exonic mutations, but there has been a growing desire to accurately detect copy number variations (CNVs) as well. In order to address this research and clinical needs, we developed a sequencing coverage pattern-based method not only for copy number detections, data integrity checks, CNV calling, and visualization reports. The developed methodologies include complete automation to increase usability, genome content-coverage bias correction, CNV segmentation, data quality reports, and publication quality images. Automatic identification and removal of poor quality outlier samples were made automatically. Multiple experimental batches were routinely detected and further reduced for a clean subset of samples before analysis. Algorithm improvements were also made to improve somatic CNV detection as well as germline CNV detection in trio family. Additionally, a set of utilities was included to facilitate users for producing CNV plots in focused genes of interest. We demonstrate the somatic CNV enhancements by accurately detecting CNVs in whole exome-wide data from the cancer genome atlas cancer samples and a lymphoma case study with paired tumor and normal samples. We also showed our efficient reuses of existing exome sequencing data, for improved germline CNV calling in a family of the trio from the phase-III study of 1000 Genome to detect CNVs with various modes of inheritance. The performance of the developed method is evaluated by comparing CNV calling results with results from other orthogonal copy number platforms. Through our case studies, reuses of exome sequencing data for calling CNVs have several noticeable functionalities, including a better quality control for exome sequencing data, improved joint analysis with single nucleotide variant calls, and novel genomic discovery of under-utilized existing whole exome and custom exome panel data.Keywords: bioinformatics, computational genetics, copy number variations, data reuse, exome sequencing, next generation sequencing
Procedia PDF Downloads 257343 Status of the European Atlas of Natural Radiation
Authors: G. Cinelli, T. Tollefsen, P. Bossew, V. Gruber, R. Braga, M. A. Hernández-Ceballos, M. De Cort
Abstract:
In 2006, the Joint Research Centre (JRC) of the European Commission started the project of the 'European Atlas of Natural Radiation'. The Atlas aims at preparing a collection of maps of Europe displaying the levels of natural radioactivity caused by different sources (indoor and outdoor radon, cosmic radiation, terrestrial radionuclides, terrestrial gamma radiation, etc). The overall goal of the project is to estimate, in geographical resolution, the annual dose that the public may receive from natural radioactivity, combining all the information from the different radiation components. The first map which has been developed is the European map of indoor radon (Rn) since in most cases Rn is the most important contribution to exposure. New versions of the map are realised when new countries join the project or when already participating countries send new data. We show the latest status of this map which currently includes 25 European countries. Second, the JRC has undertaken to map a variable which measures 'what earth delivers' in terms of Rn. The corresponding quantity is called geogenic radon potential (RP). Due to the heterogeneity of data sources across the Europe there is need to develop a harmonized quantity which at the one hand adequately measures or classifies the RP, and on the other hand is suited to accommodate the variety of input data used to estimate this target quantity. Candidates for input quantities which may serve as predictors of the RP, and for which data are available across Europe, to different extent, are Uranium (U) concentration in rocks and soils, soil gas radon and soil permeability, terrestrial gamma dose rate, geological information and indoor data from ground floor. The European Geogenic Radon Map gives the possibility to characterize areas, on European geographical scale, for radon hazard where indoor radon measurements are not available. Parallel to ongoing work on the European Indoor Radon, Geogenic Radon and Cosmic Radiation Maps, we made progress in the development of maps of terrestrial gamma radiation and U, Th and K concentrations in soil and bedrock. We show the first, preliminary map of the terrestrial gamma dose rate, estimated using the data of ambient dose equivalent rate available from the EURDEP system (about 5000 fixed monitoring stations across Europe). Also, the first maps of U, Th, and K concentrations in soil and bedrock are shown in the present work.Keywords: Europe, natural radiation, mapping, indoor radon
Procedia PDF Downloads 293342 Macroscopic Support Structure Design for the Tool-Free Support Removal of Laser Powder Bed Fusion-Manufactured Parts Made of AlSi10Mg
Authors: Tobias Schmithuesen, Johannes Henrich Schleifenbaum
Abstract:
The additive manufacturing process laser powder bed fusion offers many advantages over conventional manufacturing processes. For example, almost any complex part can be produced, such as topologically optimized lightweight parts, which would be inconceivable with conventional manufacturing processes. A major challenge posed by the LPBF process, however, is, in most cases, the need to use and remove support structures on critically inclined part surfaces (α < 45 ° regarding substrate plate). These are mainly used for dimensionally accurate mapping of part contours and to reduce distortion by absorbing process-related internal stresses. Furthermore, they serve to transfer the process heat to the substrate plate and are, therefore, indispensable for the LPBF process. A major challenge for the economical use of the LPBF process in industrial process chains is currently still the high manual effort involved in removing support structures. According to the state of the art (SoA), the parts are usually treated by simple hand tools (e.g., pliers, chisels) or by machining (e.g., milling, turning). New automatable approaches are the removal of support structures by means of wet chemical ablation and thermal deburring. According to the state of the art, the support structures are essentially adapted to the LPBF process and not to potential post-processing steps. The aim of this study is the determination of support structure designs that are adapted to the mentioned post-processing approaches. In the first step, the essential boundary conditions for complete removal by means of the respective approaches are identified. Afterward, a representative demonstrator part with various macroscopic support structure designs will be LPBF-manufactured and tested with regard to a complete powder and support removability. Finally, based on the results, potentially suitable support structure designs for the respective approaches will be derived. The investigations are carried out on the example of the aluminum alloy AlSi10Mg.Keywords: additive manufacturing, laser powder bed fusion, laser beam melting, selective laser melting, post processing, tool-free, wet chemical ablation, thermal deburring, aluminum alloy, AlSi10Mg
Procedia PDF Downloads 91341 The Impact of Artificial Intelligence on Medicine Production
Authors: Yasser Ahmed Mahmoud Ali Helal
Abstract:
The use of CAD (Computer Aided Design) technology is ubiquitous in the architecture, engineering and construction (AEC) industry. This has led to its inclusion in the curriculum of architecture schools in Nigeria as an important part of the training module. This article examines the ethical issues involved in implementing CAD (Computer Aided Design) content into the architectural education curriculum. Using existing literature, this study begins with the benefits of integrating CAD into architectural education and the responsibilities of different stakeholders in the implementation process. It also examines issues related to the negative use of information technology and the perceived negative impact of CAD use on design creativity. Using a survey method, data from the architecture department of University was collected to serve as a case study on how the issues raised were being addressed. The article draws conclusions on what ensures successful ethical implementation. Millions of people around the world suffer from hepatitis C, one of the world's deadliest diseases. Interferon (IFN) is treatment options for patients with hepatitis C, but these treatments have their side effects. Our research focused on developing an oral small molecule drug that targets hepatitis C virus (HCV) proteins and has fewer side effects. Our current study aims to develop a drug based on a small molecule antiviral drug specific for the hepatitis C virus (HCV). Drug development using laboratory experiments is not only expensive, but also time-consuming to conduct these experiments. Instead, in this in silicon study, we used computational techniques to propose a specific antiviral drug for the protein domains of found in the hepatitis C virus. This study used homology modeling and abs initio modeling to generate the 3D structure of the proteins, then identifying pockets in the proteins. Acceptable lagans for pocket drugs have been developed using the de novo drug design method. Pocket geometry is taken into account when designing ligands. Among the various lagans generated, a new specific for each of the HCV protein domains has been proposed.Keywords: drug design, anti-viral drug, in-silicon drug design, hepatitis C virus (HCV) CAD (Computer Aided Design), CAD education, education improvement, small-size contractor automatic pharmacy, PLC, control system, management system, communication
Procedia PDF Downloads 85340 Understanding the Influence of Fibre Meander on the Tensile Properties of Advanced Composite Laminates
Authors: Gaoyang Meng, Philip Harrison
Abstract:
When manufacturing composite laminates, the fibre directions within the laminate are never perfectly straight and inevitably contain some degree of stochastic in-plane waviness or ‘meandering’. In this work we aim to understand the relationship between the degree of meandering of the fibre paths, and the resulting uncertainty in the laminate’s final mechanical properties. To do this, a numerical tool is developed to automatically generate meandering fibre paths in each of the laminate's 8 plies (using Matlab) and after mapping this information into finite element simulations (using Abaqus), the statistical variability of the tensile mechanical properties of a [45°/90°/-45°/0°]s carbon/epoxy (IM7/8552) laminate is predicted. The stiffness, first ply failure strength and ultimate failure strength are obtained. Results are generated by inputting the degree of variability in the fibre paths and the laminate is then examined in all directions (from 0° to 359° in increments of 1°). The resulting predictions are output as flower (polar) plots for convenient analysis. The average fibre orientation of each ply in a given laminate is determined by the laminate layup code [45°/90°/-45°/0°]s. However, in each case, the plies contain increasingly large amounts of in-plane waviness (quantified by the standard deviation of the fibre direction in each ply across the laminate. Four different amounts of variability in the fibre direction are tested (2°, 4°, 6° and 8°). Results show that both the average tensile stiffness and the average tensile strength decrease, while the standard deviations increase, with an increasing degree of fibre meander. The variability in stiffness is found to be relatively insensitive to the rotation angle, but the variability in strength is sensitive. Specifically, the uncertainty in laminate strength is relatively low at orientations centred around multiples of 45° rotation angle, and relatively high between these rotation angles. To concisely represent all the information contained in the various polar plots, rotation-angle dependent Weibull distribution equations are fitted to the data. The resulting equations can be used to quickly estimate the size of the errors bars for the different mechanical properties, resulting from the amount of fibre directional variability contained within the laminate. A longer term goal is to use these equations to quickly introduce realistic variability at the component level.Keywords: advanced composite laminates, FE simulation, in-plane waviness, tensile properties, uncertainty quantification
Procedia PDF Downloads 89339 The Sociocultural, Economic, and Environmental Contestations of Agbogbloshie: A Critical Review
Authors: Khiddir Iddris, Martin Oteng – Ababio, Andreas Bürkert, Christoph Scherrer, Katharina Hemmler
Abstract:
Agbogbloshie, as an informal settlement and economy where the e-waste sector thrives, has become a global hub of complex urban contestations involving sociocultural, economic, and environmental dimensions due to the implication that e-waste and informal economic patterns have on livelihoods, urbanisation, development and sustainability. Multi-author collaborations have produced an ever-growing body of literature on Agbogbloshie and the informal e-waste economy. There is, however, a dearth of an assessment of Agbogbloshie as an urban informal settlement's intricate nexus of socioecological contestations. We address this gap by systematising, from literature, the context knowledge, navigating the complex terrain of Agbogbloshie's challenges, and employing a multidimensional lens to unravel the sociocultural intricacies, economic dynamics, and environmental complexities shaping its identity. A systematic critical review approach was espoused, with a pragmatic consolidation of content analysis and controversy mapping, grounded on the concept of ‘sustainable rurbanism,’ highlighted core themes and identified contrasting viewpoints. An analytical framework is presented. Five categories – geohistorical, sociocultural, economic, environmental and future trends - are proposed as an approach to systematising the literature. The review finds that the sociocultural dimension unveils a mosaic of cultural amalgamation, communal identity, and tensions impacting community cohesion. The analysis of economic intricacies reveals the prevalence of informal economies sustaining livelihoods yet entrenching economic disparities and marginalisation. Environmental scrutiny exposes the grim realities of e-waste disposal, pollution, and land use conflicts. The findings suggest that there is a high resilience within the community and the potential for sustainable trajectories. Theoretical and conceptual synergy is limited. This review provides a comprehensive exploration, offering insights and directions for future research, policy formulation, and community-driven interventions aimed at fostering sustainable transformations in Agbogbloshie and analogous urban contexts.Keywords: Agbogbloshie, economic complexities, environmental challenges, resilience, sociocultural dynamics, sustainability, urban informal settlement
Procedia PDF Downloads 73338 Portable and Parallel Accelerated Development Method for Field-Programmable Gate Array (FPGA)-Central Processing Unit (CPU)- Graphics Processing Unit (GPU) Heterogeneous Computing
Authors: Nan Hu, Chao Wang, Xi Li, Xuehai Zhou
Abstract:
The field-programmable gate array (FPGA) has been widely adopted in the high-performance computing domain. In recent years, the embedded system-on-a-chip (SoC) contains coarse granularity multi-core CPU (central processing unit) and mobile GPU (graphics processing unit) that can be used as general-purpose accelerators. The motivation is that algorithms of various parallel characteristics can be efficiently mapped to the heterogeneous architecture coupled with these three processors. The CPU and GPU offload partial computationally intensive tasks from the FPGA to reduce the resource consumption and lower the overall cost of the system. However, in present common scenarios, the applications always utilize only one type of accelerator because the development approach supporting the collaboration of the heterogeneous processors faces challenges. Therefore, a systematic approach takes advantage of write-once-run-anywhere portability, high execution performance of the modules mapped to various architectures and facilitates the exploration of design space. In this paper, A servant-execution-flow model is proposed for the abstraction of the cooperation of the heterogeneous processors, which supports task partition, communication and synchronization. At its first run, the intermediate language represented by the data flow diagram can generate the executable code of the target processor or can be converted into high-level programming languages. The instantiation parameters efficiently control the relationship between the modules and computational units, including two hierarchical processing units mapping and adjustment of data-level parallelism. An embedded system of a three-dimensional waveform oscilloscope is selected as a case study. The performance of algorithms such as contrast stretching, etc., are analyzed with implementations on various combinations of these processors. The experimental results show that the heterogeneous computing system with less than 35% resources achieves similar performance to the pure FPGA and approximate energy efficiency.Keywords: FPGA-CPU-GPU collaboration, design space exploration, heterogeneous computing, intermediate language, parameterized instantiation
Procedia PDF Downloads 118337 An Experimental Study of Scalar Implicature Processing in Chinese
Authors: Liu Si, Wang Chunmei, Liu Huangmei
Abstract:
A prominent component of the semantic versus pragmatic debate, scalar implicature (SI) has been gaining great attention ever since it was proposed by Horn. The constant debate is between the structural and pragmatic approach. The former claims that generation of SI is costless, automatic, and dependent mostly on the structural properties of sentences, whereas the latter advocates both that such generation is largely dependent upon context, and that the process is costly. Many experiments, among which Katsos’s text comprehension experiments are influential, have been designed and conducted in order to verify their views, but the results are not conclusive. Besides, most of the experiments were conducted in English language materials. Katsos conducted one off-line and three on-line text comprehension experiments, in which the previous shortcomings were addressed on a certain extent and the conclusion was in favor of the pragmatic approach. We intend to test the results of Katsos’s experiment in Chinese scalar implicature. Four experiments in both off-line and on-line conditions to examine the generation and response time of SI in Chinese "yixie" (some) and "quanbu (dou)" (all) will be conducted in order to find out whether the structural or the pragmatic approach could be sustained. The study mainly aims to answer the following questions: (1) Can SI be generated in the upper- and lower-bound contexts as Katsos confirmed when Chinese language materials are used in the experiment? (2) Can SI be first generated, then cancelled as default view claimed or can it not be generated in a neutral context when Chinese language materials are used in the experiment? (3) Is SI generation costless or costly in terms of processing resources? (4) In line with the SI generation process, what conclusion can be made about the cognitive processing model of language meaning? Is it a parallel model or a linear model? Or is it a dynamic and hierarchical model? According to previous theoretical debates and experimental conflicts, presumptions could be made that SI, in Chinese language, might be generated in the upper-bound contexts. Besides, the response time might be faster in upper-bound than that found in lower-bound context. SI generation in neutral context might be the slowest. At last, a conclusion would be made that the processing model of SI could not be verified by either absolute structural or pragmatic approaches. It is, rather, a dynamic and complex processing mechanism, in which the interaction of language forms, ad hoc context, mental context, background knowledge, speakers’ interaction, etc. are involved.Keywords: cognitive linguistics, pragmatics, scalar implicture, experimental study, Chinese language
Procedia PDF Downloads 363336 Visualising Charles Bonnet Syndrome: Digital Co-Creation of Pseudohallucinations
Authors: Victoria H. Hamilton
Abstract:
Charles Bonnet Syndrome (CBS) is when a person experiences pseudohallucinations that fill in visual information from any type of sight loss. CBS arises from an epiphenomenal process, with the physical actions of sight resulting in the mental formations of images. These pseudohallucinations—referred to as visions by the CBS community—manifest in a wide range of forms, from complex scenes to simple geometric shapes. To share these unique visual experiences, a remote co-creation website was created where CBS participants communicated their lived experiences. This created a reflexive process, and we worked to produce true representations of these interesting and little-known phenomena. Digital reconstruction of the visions is utilised as it echoes the vivid, experiential movie-like nature of what is being perceived. This paper critically analyses co-creation as a method for making digital assets. The implications of the participants' vision impairments and the application of ethical safeguards are examined in this context. Important to note, this research is of a medical syndrome for a non-medical, practice-based design. CBS research to date is primarily conducted by the ophthalmic, neurological, and psychiatric fields and approached with the primary concerns of these specialties. This research contributes a distinct approach incorporating practice-based digital design, autoethnography, and phenomenology. Autoethnography and phenomenology combine as a foundation, with the first bringing understanding and insights, balanced by the second philosophical, bigger picture, and established approach. With further refining, it is anticipated that the research may be applied to other conditions. Conditions where articulating internal experiences proves challenging and the use of digital methods could aid communication. Both the research and CBS communities will benefit from the insights regarding the relationship between cognitive perceptions and the vision process. This research combines the digital visualising of visions with interest in the link between metaphor, embodied cognition, and image. The argument for a link between CBS visions and metaphor may appear evident due to the cross-category mapping of images that is necessary for comprehension. They both are— CBS visions and metaphors—the experience of picturing images, often with lateral connections and imaginative associations.Keywords: Charles Bonnet Syndrome, digital design, visual hallucinations, visual perception
Procedia PDF Downloads 46335 Tactile Sensory Digit Feedback for Cochlear Implant Electrode Insertion
Authors: Yusuf Bulale, Mark Prince, Geoff Tansley, Peter Brett
Abstract:
Cochlear Implantation (CI) which became a routine procedure for the last decades is an electronic device that provides a sense of sound for patients who are severely and profoundly deaf. Today, cochlear implantation technology uses electrode array (EA) implanted manually into the cochlea. The optimal success of this implantation depends on the electrode technology and deep insertion techniques. However, this manual insertion procedure may cause mechanical trauma which can lead to a severe destruction of the delicate intracochlear structure. Accordingly, future improvement of the cochlear electrode implant insertion needs reduction of the excessive force application during the cochlear implantation which causes tissue damage and trauma. This study is examined tool-tissue interaction of large prototype scale digit embedded with distributive tactile sensor based upon cochlear electrode and large prototype scale cochlea phantom for simulating the human cochlear which could lead to small-scale digit requirements. The digit, distributive tactile sensors embedded with silicon-substrate was inserted into the cochlea phantom to measure any digit/phantom interaction and position of the digit in order to minimize tissue and trauma damage during the electrode cochlear insertion. The digit has provided tactile information from the digit-phantom insertion interaction such as contact status, tip penetration, obstacles, relative shape and location, contact orientation and multiple contacts. The tests demonstrated that even devices of such a relative simple design with low cost have a potential to improve cochlear implant surgery and other lumen mapping applications by providing tactile sensory feedback information and thus controlling the insertion through sensing and control of the tip of the implant during the insertion. In that approach, the surgeon could minimize the tissue damage and potential damage to the delicate structures within the cochlear caused by current manual electrode insertion of the cochlear implantation. This approach also can be applied to other minimally invasive surgery applications as well as diagnosis and path navigation procedures.Keywords: cochlear electrode insertion, distributive tactile sensory feedback information, flexible digit, minimally invasive surgery, tool/tissue interaction
Procedia PDF Downloads 398334 Investigation of the EEG Signal Parameters during Epileptic Seizure Phases in Consequence to the Application of External Healing Therapy on Subjects
Authors: Karan Sharma, Ajay Kumar
Abstract:
Epileptic seizure is a type of disease due to which electrical charge in the brain flows abruptly resulting in abnormal activity by the subject. One percent of total world population gets epileptic seizure attacks.Due to abrupt flow of charge, EEG (Electroencephalogram) waveforms change. On the display appear a lot of spikes and sharp waves in the EEG signals. Detection of epileptic seizure by using conventional methods is time-consuming. Many methods have been evolved that detect it automatically. The initial part of this paper provides the review of techniques used to detect epileptic seizure automatically. The automatic detection is based on the feature extraction and classification patterns. For better accuracy decomposition of the signal is required before feature extraction. A number of parameters are calculated by the researchers using different techniques e.g. approximate entropy, sample entropy, Fuzzy approximate entropy, intrinsic mode function, cross-correlation etc. to discriminate between a normal signal & an epileptic seizure signal.The main objective of this review paper is to present the variations in the EEG signals at both stages (i) Interictal (recording between the epileptic seizure attacks). (ii) Ictal (recording during the epileptic seizure), using most appropriate methods of analysis to provide better healthcare diagnosis. This research paper then investigates the effects of a noninvasive healing therapy on the subjects by studying the EEG signals using latest signal processing techniques. The study has been conducted with Reiki as a healing technique, beneficial for restoring balance in cases of body mind alterations associated with an epileptic seizure. Reiki is practiced around the world and is recommended for different health services as a treatment approach. Reiki is an energy medicine, specifically a biofield therapy developed in Japan in the early 20th century. It is a system involving the laying on of hands, to stimulate the body’s natural energetic system. Earlier studies have shown an apparent connection between Reiki and the autonomous nervous system. The Reiki sessions are applied by an experienced therapist. EEG signals are measured at baseline, during session and post intervention to bring about effective epileptic seizure control or its elimination altogether.Keywords: EEG signal, Reiki, time consuming, epileptic seizure
Procedia PDF Downloads 407333 GPU-Based Back-Projection of Synthetic Aperture Radar (SAR) Data onto 3D Reference Voxels
Authors: Joshua Buli, David Pietrowski, Samuel Britton
Abstract:
Processing SAR data usually requires constraints in extent in the Fourier domain as well as approximations and interpolations onto a planar surface to form an exploitable image. This results in a potential loss of data requires several interpolative techniques, and restricts visualization to two-dimensional plane imagery. The data can be interpolated into a ground plane projection, with or without terrain as a component, all to better view SAR data in an image domain comparable to what a human would view, to ease interpretation. An alternate but computationally heavy method to make use of more of the data is the basis of this research. Pre-processing of the SAR data is completed first (matched-filtering, motion compensation, etc.), the data is then range compressed, and lastly, the contribution from each pulse is determined for each specific point in space by searching the time history data for the reflectivity values for each pulse summed over the entire collection. This results in a per-3D-point reflectivity using the entire collection domain. New advances in GPU processing have finally allowed this rapid projection of acquired SAR data onto any desired reference surface (called backprojection). Mathematically, the computations are fast and easy to implement, despite limitations in SAR phase history data size and 3D-point cloud size. Backprojection processing algorithms are embarrassingly parallel since each 3D point in the scene has the same reflectivity calculation applied for all pulses, independent of all other 3D points and pulse data under consideration. Therefore, given the simplicity of the single backprojection calculation, the work can be spread across thousands of GPU threads allowing for accurate reflectivity representation of a scene. Furthermore, because reflectivity values are associated with individual three-dimensional points, a plane is no longer the sole permissible mapping base; a digital elevation model or even a cloud of points (collected from any sensor capable of measuring ground topography) can be used as a basis for the backprojection technique. This technique minimizes any interpolations and modifications of the raw data, maintaining maximum data integrity. This innovative processing will allow for SAR data to be rapidly brought into a common reference frame for immediate exploitation and data fusion with other three-dimensional data and representations.Keywords: backprojection, data fusion, exploitation, three-dimensional, visualization
Procedia PDF Downloads 86332 House Extension Strategy in High-Density Informal Settlement: A Case Study in Kampung Cikini, Jakarta, Indonesia
Authors: Meidesta Pitria, Akiko Okabe
Abstract:
In high-density informal settlement, extension area at the outside of the houses could primarily happen as a spatial modification response. House extension in high-density informal settlement is not only becoming a physical spatial modification that makes a blur zone between private and public but also supporting the growth and existence of informal economy and other daily activities in both individuals and communities. This research took a case study in an informal settlement named Kampung Cikini, a densely populated area in Central Jakarta. The aim of this study is to identify and clarify house extension as a strategy in dealing with urbanization in an informal settlement. By using the perspective and information from housewives, the analysis is based on the assumption that land ownership transformation and the activities in house extension area influence the different kinds of house extension’s spatial modification and local planning policy in relation with the implementation of house extension strategy. The data collection was done in four sites, two sites are located in outer-wide alley and another two sites are located in inner-narrow alley. In this research, data of 104 housewives in 86 houses were collected through representatives of housewives and local leader of each sites. The research was started from participatory mapping process, deep interview with local leader, and initiated collaboration with housewives community in having a celebration as communal event to cultivate together the issue. This study shows that land ownership, activities, and alley are indispensable in the decision of extension space making. The more permanency status of land ownership the more permanent and various extension could be implemented. However, in some blocks, the existence of origin house or first land owner also has a significant role in coordination and agreement in using and modifying extension space. In outer-wide alley, the existence of more various activities in front area of the houses is significantly related with the chance given by having wider alley, particularly for informal income generating activities. In the inner-narrow alley, limited space in front of the houses affects more negotiations in the community for having more shared spaces, even inside their private space.Keywords: house extension, housewives, informal settlement, kampung, high density
Procedia PDF Downloads 208331 The importance of Clinical Pharmacy and Computer Aided Drug Design
Authors: Peter Edwar Mortada Nasif
Abstract:
The use of CAD (Computer Aided Design) technology is ubiquitous in the architecture, engineering and construction (AEC) industry. This has led to its inclusion in the curriculum of architecture schools in Nigeria as an important part of the training module. This article examines the ethical issues involved in implementing CAD (Computer Aided Design) content into the architectural education curriculum. Using existing literature, this study begins with the benefits of integrating CAD into architectural education and the responsibilities of different stakeholders in the implementation process. It also examines issues related to the negative use of information technology and the perceived negative impact of CAD use on design creativity. Using a survey method, data from the architecture department of Chukwuemeka Odumegwu Ojukwu Uli University was collected to serve as a case study on how the issues raised were being addressed. The article draws conclusions on what ensures successful ethical implementation. Millions of people around the world suffer from hepatitis C, one of the world's deadliest diseases. Interferon (IFN) is treatment options for patients with hepatitis C, but these treatments have their side effects. Our research focused on developing an oral small molecule drug that targets hepatitis C virus (HCV) proteins and has fewer side effects. Our current study aims to develop a drug based on a small molecule antiviral drug specific for the hepatitis C virus (HCV). Drug development using laboratory experiments is not only expensive, but also time-consuming to conduct these experiments. Instead, in this in silicon study, we used computational techniques to propose a specific antiviral drug for the protein domains of found in the hepatitis C virus. This study used homology modeling and abs initio modeling to generate the 3D structure of the proteins, then identifying pockets in the proteins. Acceptable lagans for pocket drugs have been developed using the de novo drug design method. Pocket geometry is taken into account when designing ligands. Among the various lagans generated, a new specific for each of the HCV protein domains has been proposed.Keywords: drug design, anti-viral drug, in-silicon drug design, hepatitis C virus, computer aided design, CAD education, education improvement, small-size contractor automatic pharmacy, PLC, control system, management system, communication
Procedia PDF Downloads 26330 Using Hemicellulosic Liquor from Sugarcane Bagasse to Produce Second Generation Lactic Acid
Authors: Regiane A. Oliveira, Carlos E. Vaz Rossell, Rubens Maciel Filho
Abstract:
Lactic acid, besides a valuable chemical may be considered a platform for other chemicals. In fact, the feasibility of hemicellulosic sugars as feedstock for lactic acid production process, may represent the drop of some of the barriers for the second generation bioproducts, especially bearing in mind the 5-carbon sugars from the pre-treatment of sugarcane bagasse. Bearing this in mind, the purpose of this study was to use the hemicellulosic liquor from sugarcane bagasse as a substrate to produce lactic acid by fermentation. To release of sugars from hemicellulose it was made a pre-treatment with a diluted sulfuric acid in order to obtain a xylose's rich liquor with low concentration of inhibiting compounds for fermentation (≈ 67% of xylose, ≈ 21% of glucose, ≈ 10% of cellobiose and arabinose, and around 1% of inhibiting compounds as furfural, hydroxymethilfurfural and acetic acid). The hemicellulosic sugars associated with 20 g/L of yeast extract were used in a fermentation process with Lactobacillus plantarum to produce lactic acid. The fermentation process pH was controlled with automatic injection of Ca(OH)2 to keep pH at 6.00. The lactic acid concentration remained stable from the time when the glucose was depleted (48 hours of fermentation), with no further production. While lactic acid is produced occurs the concomitant consumption of xylose and glucose. The yield of fermentation was 0.933 g lactic acid /g sugars. Besides, it was not detected the presence of by-products, what allows considering that the microorganism uses a homolactic fermentation to produce its own energy using pentose-phosphate pathway. Through facultative heterofermentative metabolism the bacteria consume pentose, as is the case of L. plantarum, but the energy efficiency for the cell is lower than during the hexose consumption. This implies both in a slower cell growth, as in a reduction in lactic acid productivity compared with the use of hexose. Also, L. plantarum had shown to have a capacity for lactic acid production from hemicellulosic hydrolysate without detoxification, which is very attractive in terms of robustness for an industrial process. Xylose from hydrolyzed bagasse and without detoxification is consumed, although the hydrolyzed bagasse inhibitors (especially aromatic inhibitors) affect productivity and yield of lactic acid. The use of sugars and the lack of need for detoxification of the C5 liquor from sugarcane bagasse hydrolyzed is a crucial factor for the economic viability of second generation processes. Taking this information into account, the production of second generation lactic acid using sugars from hemicellulose appears to be a good alternative to the complete utilization of sugarcane plant, directing molasses and cellulosic carbohydrates to produce 2G-ethanol, and hemicellulosic carbohydrates to produce 2G-lactic acid.Keywords: fermentation, lactic acid, hemicellulosic sugars, sugarcane
Procedia PDF Downloads 374329 Linguistic Analysis of Argumentation Structures in Georgian Political Speeches
Authors: Mariam Matiashvili
Abstract:
Argumentation is an integral part of our daily communications - formal or informal. Argumentative reasoning, techniques, and language tools are used both in personal conversations and in the business environment. Verbalization of the opinions requires the use of extraordinary syntactic-pragmatic structural quantities - arguments that add credibility to the statement. The study of argumentative structures allows us to identify the linguistic features that make the text argumentative. Knowing what elements make up an argumentative text in a particular language helps the users of that language improve their skills. Also, natural language processing (NLP) has become especially relevant recently. In this context, one of the main emphases is on the computational processing of argumentative texts, which will enable the automatic recognition and analysis of large volumes of textual data. The research deals with the linguistic analysis of the argumentative structures of Georgian political speeches - particularly the linguistic structure, characteristics, and functions of the parts of the argumentative text - claims, support, and attack statements. The research aims to describe the linguistic cues that give the sentence a judgmental/controversial character and helps to identify reasoning parts of the argumentative text. The empirical data comes from the Georgian Political Corpus, particularly TV debates. Consequently, the texts are of a dialogical nature, representing a discussion between two or more people (most often between a journalist and a politician). The research uses the following approaches to identify and analyze the argumentative structures Lexical Classification & Analysis - Identify lexical items that are relevant in argumentative texts creating process - Creating the lexicon of argumentation (presents groups of words gathered from a semantic point of view); Grammatical Analysis and Classification - means grammatical analysis of the words and phrases identified based on the arguing lexicon. Argumentation Schemas - Describe and identify the Argumentation Schemes that are most likely used in Georgian Political Speeches. As a final step, we analyzed the relations between the above mentioned components. For example, If an identified argument scheme is “Argument from Analogy”, identified lexical items semantically express analogy too, and they are most likely adverbs in Georgian. As a result, we created the lexicon with the words that play a significant role in creating Georgian argumentative structures. Linguistic analysis has shown that verbs play a crucial role in creating argumentative structures.Keywords: georgian, argumentation schemas, argumentation structures, argumentation lexicon
Procedia PDF Downloads 74328 Revisiting Hospital Ward Design Basics for Sustainable Family Integration
Authors: Ibrahim Abubakar Alkali, Abubakar Sarkile Kawuwa, Ibrahim Sani Khalil
Abstract:
The concept of space and function forms the bedrock for spatial configuration in architectural design. Thus, the effectiveness and functionality of an architectural product depends their cordial relationship. This applies to all buildings especially to a hospital ward setting designed to accommodate various complex and diverse functions. Health care facilities design, especially an inpatient setting, is governed by many regulations and technical requirements. It is also affected by many less defined needs, particularly, response to culture and the need to provide for patient families’ presence and participation. The spatial configuration of the hospital ward setting in developing countries has no consideration for the patient’s families despite the significant role they play in promoting recovery. Attempts to integrate facilities for patients’ families have always been challenging, especially in developing countries like Nigeria, where accommodation for inpatients is predominantly in an open ward system. In addition, the situation is compounded by culture, which significantly dictates healthcare practices in Africa. Therefore, achieving such a hospital ward setting that is patient and family-centered requires careful assessment of family care actions and transaction spaces so as to arrive at an evidence based solution. Therefore, the aim of this study is to identify how hospital ward spaces can be reconfigured to provide for sustainable family integration. In achieving this aim, a qualitative approach using the principles of behavioral mapping was employed in male and female medical wards of the Federal Teaching Hospital (FTH) Gombe, Nigeria. The data obtained was analysed using classical and comparative content analysis. Patients’ families have been found to be a critical component of hospital ward design that cannot be undermined. Accordingly, bedsides, open yards, corridors and foyers have been identified as patient families’ transaction spaces that require design attention. Arriving at sustainable family integration can be achieved by revisiting the design requirements of the family transaction spaces based on the findings in order to avoid the rowdiness of the wards and uncoordinated sprawl.Keywords: caregiving, design basics, family integration, hospital ward, sustainability
Procedia PDF Downloads 307327 Algorithm Development of Individual Lumped Parameter Modelling for Blood Circulatory System: An Optimization Study
Authors: Bao Li, Aike Qiao, Gaoyang Li, Youjun Liu
Abstract:
Background: Lumped parameter model (LPM) is a common numerical model for hemodynamic calculation. LPM uses circuit elements to simulate the human blood circulatory system. Physiological indicators and characteristics can be acquired through the model. However, due to the different physiological indicators of each individual, parameters in LPM should be personalized in order for convincing calculated results, which can reflect the individual physiological information. This study aimed to develop an automatic and effective optimization method to personalize the parameters in LPM of the blood circulatory system, which is of great significance to the numerical simulation of individual hemodynamics. Methods: A closed-loop LPM of the human blood circulatory system that is applicable for most persons were established based on the anatomical structures and physiological parameters. The patient-specific physiological data of 5 volunteers were non-invasively collected as personalized objectives of individual LPM. In this study, the blood pressure and flow rate of heart, brain, and limbs were the main concerns. The collected systolic blood pressure, diastolic blood pressure, cardiac output, and heart rate were set as objective data, and the waveforms of carotid artery flow and ankle pressure were set as objective waveforms. Aiming at the collected data and waveforms, sensitivity analysis of each parameter in LPM was conducted to determine the sensitive parameters that have an obvious influence on the objectives. Simulated annealing was adopted to iteratively optimize the sensitive parameters, and the objective function during optimization was the root mean square error between the collected waveforms and data and simulated waveforms and data. Each parameter in LPM was optimized 500 times. Results: In this study, the sensitive parameters in LPM were optimized according to the collected data of 5 individuals. Results show a slight error between collected and simulated data. The average relative root mean square error of all optimization objectives of 5 samples were 2.21%, 3.59%, 4.75%, 4.24%, and 3.56%, respectively. Conclusions: Slight error demonstrated good effects of optimization. The individual modeling algorithm developed in this study can effectively achieve the individualization of LPM for the blood circulatory system. LPM with individual parameters can output the individual physiological indicators after optimization, which are applicable for the numerical simulation of patient-specific hemodynamics.Keywords: blood circulatory system, individual physiological indicators, lumped parameter model, optimization algorithm
Procedia PDF Downloads 138326 Development of an Integrated Methodology for Fouling Control in Membrane Bioreactors
Authors: Petros Gkotsis, Anastasios Zouboulis, Manasis Mitrakas, Dimitrios Zamboulis, E. Peleka
Abstract:
The most serious drawback in wastewater treatment using membrane bioreactors (MBRs) is membrane fouling which gradually leads to membrane permeability decrease and efficiency deterioration. This work is part of a research project that aims to develop an integrated methodology for membrane fouling control, using specific chemicals which will enhance the coagulation and flocculation of compounds responsible for fouling, hence reducing biofilm formation on the membrane surface and limiting the fouling rate acting as a pre-treatment step. For this purpose, a pilot-scale plant with fully automatic operation achieved by means of programmable logic controller (PLC) has been constructed and tested. The experimental set-up consists of four units: wastewater feed unit, bioreactor, membrane (side-stream) filtration unit and permeate collection unit. Synthetic wastewater was fed as the substrate for the activated sludge. The dissolved oxygen (DO) concentration of the aerobic tank was maintained in the range of 2-3 mg/L during the entire operation by using an aerator below the membrane module. The membranes were operated at a flux of 18 LMH while membrane relaxation steps of 1 min were performed every 10 min. Both commercial and composite coagulants are added in different concentrations in the pilot-scale plant and their effect on the overall performance of the ΜΒR system is presented. Membrane fouling was assessed in terms of TMP, membrane permeability, sludge filterability tests, total resistance and the unified modified fouling index (UMFI). Preliminary tests showed that particular attention should be paid to the addition of the coagulant solution, indicating that pipe flocculation effectively increases hydraulic retention time and leads to voluminous sludge flocs. The most serious drawback in wastewater treatment using MBRs is membrane fouling, which gradually leads to membrane permeability decrease and efficiency deterioration. This results in increased treatment cost, due to high energy consumption and the need for frequent membrane cleaning and replacement. Due to the widespread application of MBR technology over the past few years, it becomes clear that the development of a methodology to mitigate membrane fouling is of paramount importance. The present work aims to develop an integrated technique for membrane fouling control in MBR systems and, thus, contribute to sustainable wastewater treatment.Keywords: coagulation, membrane bioreactor, membrane fouling, pilot plant
Procedia PDF Downloads 310325 A Four-Step Ortho-Rectification Procedure for Geo-Referencing Video Streams from a Low-Cost UAV
Authors: B. O. Olawale, C. R. Chatwin, R. C. D. Young, P. M. Birch, F. O. Faithpraise, A. O. Olukiran
Abstract:
Ortho-rectification is the process of geometrically correcting an aerial image such that the scale is uniform. The ortho-image formed from the process is corrected for lens distortion, topographic relief, and camera tilt. This can be used to measure true distances, because it is an accurate representation of the Earth’s surface. Ortho-rectification and geo-referencing are essential to pin point the exact location of targets in video imagery acquired at the UAV platform. This can only be achieved by comparing such video imagery with an existing digital map. However, it is only when the image is ortho-rectified with the same co-ordinate system as an existing map that such a comparison is possible. The video image sequences from the UAV platform must be geo-registered, that is, each video frame must carry the necessary camera information before performing the ortho-rectification process. Each rectified image frame can then be mosaicked together to form a seamless image map covering the selected area. This can then be used for comparison with an existing map for geo-referencing. In this paper, we present a four-step ortho-rectification procedure for real-time geo-referencing of video data from a low-cost UAV equipped with multi-sensor system. The basic procedures for the real-time ortho-rectification are: (1) Decompilation of video stream into individual frames; (2) Finding of interior camera orientation parameters; (3) Finding the relative exterior orientation parameters for each video frames with respect to each other; (4) Finding the absolute exterior orientation parameters, using self-calibration adjustment with the aid of a mathematical model. Each ortho-rectified video frame is then mosaicked together to produce a 2-D planimetric mapping, which can be compared with a well referenced existing digital map for the purpose of georeferencing and aerial surveillance. A test field located in Abuja, Nigeria was used for testing our method. Fifteen minutes video and telemetry data were collected using the UAV and the data collected were processed using the four-step ortho-rectification procedure. The results demonstrated that the geometric measurement of the control field from ortho-images are more reliable than those from original perspective photographs when used to pin point the exact location of targets on the video imagery acquired by the UAV. The 2-D planimetric accuracy when compared with the 6 control points measured by a GPS receiver is between 3 to 5 meters.Keywords: geo-referencing, ortho-rectification, video frame, self-calibration
Procedia PDF Downloads 478324 Understanding Hydrodynamic in Lake Victoria Basin in a Catchment Scale: A Literature Review
Authors: Seema Paul, John Mango Magero, Prosun Bhattacharya, Zahra Kalantari, Steve W. Lyon
Abstract:
The purpose of this review paper is to develop an understanding of lake hydrodynamics and the potential climate impact on the Lake Victoria (LV) catchment scale. This paper briefly discusses the main problems of lake hydrodynamics and its’ solutions that are related to quality assessment and climate effect. An empirical methodology in modeling and mapping have considered for understanding lake hydrodynamic and visualizing the long-term observational daily, monthly, and yearly mean dataset results by using geographical information system (GIS) and Comsol techniques. Data were obtained for the whole lake and five different meteorological stations, and several geoprocessing tools with spatial analysis are considered to produce results. The linear regression analyses were developed to build climate scenarios and a linear trend on lake rainfall data for a long period. A potential evapotranspiration rate has been described by the MODIS and the Thornthwaite method. The rainfall effect on lake water level observed by Partial Differential Equations (PDE), and water quality has manifested by a few nutrients parameters. The study revealed monthly and yearly rainfall varies with monthly and yearly maximum and minimum temperatures, and the rainfall is high during cool years and the temperature is high associated with below and average rainfall patterns. Rising temperatures are likely to accelerate evapotranspiration rates and more evapotranspiration is likely to lead to more rainfall, drought is more correlated with temperature and cloud is more correlated with rainfall. There is a trend in lake rainfall and long-time rainfall on the lake water surface has affected the lake level. The onshore and offshore have been concentrated by initial literature nutrients data. The study recommended that further studies should consider fully lake bathymetry development with flow analysis and its’ water balance, hydro-meteorological processes, solute transport, wind hydrodynamics, pollution and eutrophication these are crucial for lake water quality, climate impact assessment, and water sustainability.Keywords: climograph, climate scenarios, evapotranspiration, linear trend flow, rainfall event on LV, concentration
Procedia PDF Downloads 99323 Social Movements of Central-Eastern Europe: Examining Trends of Cooperation and Antagonism by Using Big Data
Authors: Reka Zsuzsanna Mathe
Abstract:
The globalization and the Europeanization have significantly contributed to a change in the role of the nation-states. The global economic crisis, the climate changes, and the recent refugee crisis, are just a few among many challenges that cannot be effectively addressed by the traditional role of the nation-states. One of the main roles of the states is to solve collective action problems, however due to their changing roles; apparently this is getting more and more difficult. Depending on political culture, collective action problems are solved either through cooperation or conflict. The political culture of Central and Eastern European (CEE) countries is marked by low civic participation and by a weak civil society. In this type of culture collective action problems are likely to be induced through conflict, rather than the democratic process of dialogue and any type of social change is probably to be introduced by social movements. Several studies have been conducted on the social movements of the CEE countries, yet, it is still not clear if the most significant social movements of the region tend to choose rather the cooperative or the conflictual way as action strategy. This study differentiates between a national and a European action field, having different social orders. The actors of the two fields are the broadly understood civil society members, conceptualized as social movements. This research tries to answer the following questions: a) What are the norms that best characterize the CEE countries’ social order? b) What type of actors would prefer a change and in which areas? c) Is there a significant difference between the main actors active in the national versus the European field? The main hypotheses are that there are conflicting norms defining the national and the European action field, and there is a significant difference between the action strategies adopted by social movements acting in the two different fields. In mapping the social order, the study uses data provided by the European Social Survey. Big data of the Global Data on Events, Location and Tone (GDELT) database offers information regarding the main social movements and their preferred type of action. The unit of the analysis is the so called ‘Visegrad 4’ countries: Poland, Czech Republic, Slovakia and Hungary and the research uses data starting from 2005 (after the European accession of these four countries) until May, 2017. According to the data, the main hypotheses were confirmed.Keywords: big data, Central and Eastern Europe, civil society, GDELT, social movements
Procedia PDF Downloads 161