Search results for: single error upset
436 Analyzing Social Media Discourses of Domestic Violence in Promoting Awareness and Support Seeking: An Exploratory Study
Authors: Sudha Subramani, Hua Wang
Abstract:
Domestic Violence (DV) against women is now recognized to be a serious and widespread problem worldwide. There is a growing concern that violence against women has a global public health impact, as well as a violation of human rights. From the existing statistical surveys, it is revealed that there exists a strong relationship between DV and health issues of women like bruising, lacerations, depression, anxiety, flashbacks, sleep disturbances, hyper-arousal, emotional distress, sexually transmitted diseases and so on. This social problem is still considered as behind the closed doors issue and stigmatized topic. Women conceal their sufferings from family and friends, as they experience a lack of trust in others, feelings of shame and embarrassment among the society. Hence, women survivors of DV experience some barriers in seeking the support of specialized services such as health care access, crisis support, and legal guidance. Fortunately, with the popularity of social media like Facebook and Twitter, people share their opinions and emotional feelings to seek the social and emotional support, for sympathetic encouragement, to show compassion and empathy among the public. Considering the DV, social media plays a predominant role in creating the awareness and promoting the support services to the public, as we live in the golden era of social media. The various professional people like the public health researchers, clinicians, psychologists, social workers, national family health organizations, lawyers, and victims or their family and friends share the unprecedentedly valuable information (personal opinions and experiences) in a single platform to improve the social welfare of the community. Though each tweet or post contains a less informational value, the consolidation of millions of messages can generate actionable knowledge and provide valuable insights about the public opinion in general. Hence, this paper reports on an exploratory analysis of the effectiveness of social media for unobtrusive assessment of attitudes and awareness towards DV. In this paper, mixed methods such as qualitative analysis and text mining approaches are used to understand the social media disclosures of DV through the lenses of opinion sharing, anonymity, and support seeking. The results of this study could be helpful to avoid the cost of wide scale surveys, while still maintaining appropriate research conditions is to leverage the abundance of data publicly available on the web. Also, this analysis with data enrichment and consolidation would be useful in assisting advocacy and national family health organizations to provide information about resources and support, raise awareness and counter common stigmatizing attitudes about DV.Keywords: domestic violence, social media, social stigma and support, women health
Procedia PDF Downloads 289435 Linkages between Innovation Policies and SMEs' Innovation Activities: Empirical Evidence from 15 Transition Countries
Authors: Anita Richter
Abstract:
Innovation is one of the key foundations of competitive advantage, generating growth and welfare worldwide. Consequently, all firms should innovate to bring new ideas to the market. Innovation is a vital growth driver, particularly for transition countries to move towards knowledge-based, high-income economies. However, numerous barriers, such as financial, regulatory or infrastructural constraints prevent, in particular, new and small firms in transition countries from innovating. Thus SMEs’ innovation output may benefit substantially from government support. This research paper aims to assess the effect of government interventions on innovation activities in SMEs in emerging countries. Until now academic research related to the innovation policies focused either on single country and/or high-income countries assessments and less on cross-country and/or low and middle-income countries. Therefore the paper seeks to close the research gap by providing empirical evidence from 8,500 firms in 15 transition countries (Eastern Europe, South Caucasus, South East Europe, Middle East and North Africa). Using firm-level data from the Business Environment and Enterprise Performance Survey of the World Bank and EBRD and policy data from the SME Policy Index of the OECD, the paper investigates how government interventions affect SME’s likelihood of investing in any technological and non-technological innovation. Using the Standard Linear Regression, the impact of government interventions on SMEs’ innovation output and R&D activities is measured. The empirical analysis suggests that a firm’s decision to invest into innovative activities is sensitive to government interventions. A firm’s likelihood to invest into innovative activities increases by 3% to 8%, if the innovation eco-system noticeably improves (measured by an increase of 1 level in the SME Policy Index). At the same time, a better eco-system encourages SMEs to invest more in R&D. Government reforms in establishing a dedicated policy framework (IP legislation), institutional infrastructure (science and technology parks, incubators) and financial support (public R&D grants, innovation vouchers) are particularly relevant to stimulate innovation performance in SMEs. Particular segments of the SME population, namely micro and manufacturing firms, are more likely to benefit from an increased innovation framework conditions. The marginal effects are particularly strong on product innovation, process innovation, and marketing innovation, but less on management innovation. In conclusion, government interventions supporting innovation will likely lead to higher innovation performance of SMEs. They increase productivity at both firm and country level, which is a vital step in transitioning towards knowledge-based market economies.Keywords: innovation, research and development, government interventions, economic development, small and medium-sized enterprises, transition countries
Procedia PDF Downloads 324434 Using the ISO 9705 Room Corner Test for Smoke Toxicity Quantification of Polyurethane
Authors: Gabrielle Peck, Ryan Hayes
Abstract:
Polyurethane (PU) foam is typically sold as acoustic foam that is often used as sound insulation in settings such as night clubs and bars. As a construction product, PU is tested by being glued to the walls and ceiling of the ISO 9705 room corner test room. However, when heat is applied to PU foam, it melts and burns as a pool fire due to it being a thermoplastic. The current test layout is unable to accurately measure mass loss and doesn’t allow for the material to burn as a pool fire without seeping out of the test room floor. The lack of mass loss measurement means gas yields pertaining to smoke toxicity analysis can’t be calculated, which makes data comparisons from any other material or test method difficult. Additionally, the heat release measurements are not representative of the actual measurements taken as a lot of the material seeps through the floor (when a tray to catch the melted material is not used). This research aimed to modify the ISO 9705 test to provide the ability to measure mass loss to allow for better calculation of gas yields and understanding of decomposition. It also aimed to accurately measure smoke toxicity in both the doorway and duct and enable dilution factors to be calculated. Finally, the study aimed to examine if doubling the fuel loading would force under-ventilated flaming. The test layout was modified to be a combination of the SBI (single burning item) test set up inside oof the ISO 9705 test room. Polyurethane was tested in two different ways with the aim of altering the ventilation condition of the tests. Test one was conducted using 1 x SBI test rig aiming for well-ventilated flaming. Test two was conducted using 2 x SBI rigs (facing each other inside the test room) (doubling the fuel loading) aiming for under-ventilated flaming. The two different configurations used were successful in achieving both well-ventilated flaming and under-ventilated flaming, shown by the measured equivalence ratios (measured using a phi meter designed and created for these experiments). The findings show that doubling the fuel loading will successfully force under-ventilated flaming conditions to be achieved. This method can therefore be used when trying to replicate post-flashover conditions in future ISO 9705 room corner tests. The radiative heat generated by the two SBI rigs facing each other facilitated a much higher overall heat release resulting in a more severe fire. The method successfully allowed for accurate measurement of smoke toxicity produced from the PU foam in terms of simple gases such as oxygen depletion, CO and CO2. Overall, the proposed test modifications improve the ability to measure the smoke toxicity of materials in different fire conditions on a large-scale.Keywords: flammability, ISO9705, large-scale testing, polyurethane, smoke toxicity
Procedia PDF Downloads 74433 Investigation of Cavitation in a Centrifugal Pump Using Synchronized Pump Head Measurements, Vibration Measurements and High-Speed Image Recording
Authors: Simon Caba, Raja Abou Ackl, Svend Rasmussen, Nicholas E. Pedersen
Abstract:
It is a challenge to directly monitor cavitation in a pump application during operation because of a lack of visual access to validate the presence of cavitation and its form of appearance. In this work, experimental investigations are carried out in an inline single-stage centrifugal pump with optical access. Hence, it gives the opportunity to enhance the value of CFD tools and standard cavitation measurements. Experiments are conducted using two impellers running in the same volute at 3000 rpm and the same flow rate. One of the impellers used is optimized for lower NPSH₃% by its blade design, whereas the other one is manufactured using a standard casting method. The cavitation is detected by pump performance measurements, vibration measurements and high-speed image recordings. The head drop and the pump casing vibration caused by cavitation are correlated with the visual appearance of the cavitation. The vibration data is recorded in an axial direction of the impeller using accelerometers recording at a sample rate of 131 kHz. The vibration frequency domain data (up to 20 kHz) and the time domain data are analyzed as well as the root mean square values. The high-speed recordings, focusing on the impeller suction side, are taken at 10,240 fps to provide insight into the flow patterns and the cavitation behavior in the rotating impeller. The videos are synchronized with the vibration time signals by a trigger signal. A clear correlation between cloud collapses and abrupt peaks in the vibration signal can be observed. The vibration peaks clearly indicate cavitation, especially at higher NPSHA values where the hydraulic performance is not affected. It is also observed that below a certain NPSHA value, the cavitation started in the inlet bend of the pump. Above this value, cavitation occurs exclusively on the impeller blades. The impeller optimized for NPSH₃% does show a lower NPSH₃% than the standard impeller, but the head drop starts at a higher NPSHA value and is more gradual. Instabilities in the head drop curve of the optimized impeller were observed in addition to a higher vibration level. Furthermore, the cavitation clouds on the suction side appear more unsteady when using the optimized impeller. The shape and location of the cavitation are compared to 3D fluid flow simulations. The simulation results are in good agreement with the experimental investigations. In conclusion, these investigations attempt to give a more holistic view on the appearance of cavitation by comparing the head drop, vibration spectral data, vibration time signals, image recordings and simulation results. Data indicates that a criterion for cavitation detection could be derived from the vibration time-domain measurements, which requires further investigation. Usually, spectral data is used to analyze cavitation, but these investigations indicate that the time domain could be more appropriate for some applications.Keywords: cavitation, centrifugal pump, head drop, high-speed image recordings, pump vibration
Procedia PDF Downloads 178432 Immunoprotective Role of Baker's Yeast (Saccharomyces cerevisiae) against Experimentally Induced Aflatoxicosis in Broiler Chicks
Authors: Zain Ul Abadeen, Muhammad Zargham Khan, Muhammad Kashif Saleemi, Ahrar Khan, Ijaz Javed Hassan, Aisha Khatoon, Qasim Altaf
Abstract:
Aflatoxins are secondary metabolites produced by toxigenic fungi, and there are four types of aflatoxins include AFB1, AFB2, AFG1 and AFG2. Aflatoxin B1 (AFB1) is considered as most toxic form. It is mainly responsible for the contamination of poultry feed and produces a condition called aflatoxicosis leads to immunosuppression in poultry birds. Saccharomyces cerevisiae is a single cell microorganism and acts as a source of growth factors, minerals and amino acids which improve the immunity and digestibility in poultry birds as probiotics. Saccharomyces cerevisiae is well recognized to cause the biological degradation of mycotoxins (toxin binder) because its cell wall contains β-glucans and mannans which specifically bind with aflatoxins and reduce their absorption or transfer them to some non-toxic compounds. The present study was designed to investigate the immunosuppressive effects of aflatoxins in broiler chicks and the reduction of severity of these effects by the use of Baker’s Yeast (Saccharomyces cerevisiae). One-day-old broiler chicks were procured from local hatchery and were divided into various groups (A-I). These groups were treated with different levels of AFB1 @ 400 µg/kg and 600 µg/kg along with different levels of Baker’s Yeast (Saccharomyces cerevisiae) 0.1% and 0.5 % in the feed. The total duration of the experiment was six weeks and different immunological parameters including the cellular immune response by injecting PHA-P (Phytohemagglutinin-P) in the skin of the birds, phagocytic function of mononuclear cells by Carbon clearance assay from blood samples and humoral immune response against intravenously injected sheep RBCs from the serum samples were determined. The birds from each group were slaughtered at the end of the experiment to determine the presence of gross lesions in the immune organs and these tissues were fixed in 10% neutral buffered formalin for histological investigations. The results showed that AFB1 intoxicated groups had reduced body weight gain, feed intake, organs weight and immunological responses compared to the control and Baker’s Yeast (Saccharomyces cerevisiae) treated groups. Different gross and histological degenerative changes were recorded in the immune organs of AFB1 intoxicated groups compared to control and Baker’s Yeast (Saccharomyces cerevisiae) treated groups. The present study concluded that Baker’s Yeast (Saccharomyces cerevisiae) addition in the feed helps to ameliorate the immunotoxigenic effects produced by AFB1 in broiler chicks.Keywords: aflatoxins, body weight gain, feed intake, immunological response, toxigenic effect
Procedia PDF Downloads 311431 CybeRisk Management in Banks: An Italian Case Study
Authors: E. Cenderelli, E. Bruno, G. Iacoviello, A. Lazzini
Abstract:
The financial sector is exposed to the risk of cyber-attacks like any other industrial sector. Furthermore, the topic of CybeRisk (cyber risk) has become particularly relevant given that Information Technology (IT) attacks have increased drastically in recent years, and cannot be stopped by single organizations requiring a response at international and national level. IT risk is never a matter purely for the IT manager, although he clearly plays a key role. A bank's risk management function requires a thorough understanding of the evolving risks as well as the tools and practical techniques available to address them. Upon the request of European and national legislation regarding CybeRisk in the financial system, banks are therefore called upon to strengthen the operational model for CybeRisk management. This will require an important change with a more intense collaboration with the structures that deal with information security for the development of an ad hoc system for the evaluation and control of this type of risk. The aim of the work is to propose a framework for the management and control of CybeRisk that will bridge the gap in the literature regarding the understanding and consideration of CybeRisk as an integral part of business management. The IT function has a strong relevance in the management of CybeRisk, which is perceived mainly as operational risk, but with a positive tendency on the part of risk management to the identification of CybeRisk assessment methods that are increasingly complete, quantitative and able to better describe the possible impacts on the business. The paper provides answers to the research questions: Is it possible to define a CybeRisk governance structure able to support the comparison between risk and security? How can the relationships between IT assets be integrated into a cyberisk assessment framework to guarantee a system of protection and risks control? From a methodological point of view, this research uses a case study approach. The choice of “Monte dei Paschi di Siena” was determined by the specific features of one of Italy’s biggest lenders. It is chosen to use an intensive research strategy: an in-depth study of reality. The case study methodology is an empirical approach to explore a complex and current phenomenon that develops over time. The use of cases has also the advantage of allowing the deepening of aspects concerning the "how" and "why" of contemporary events, on which the scholar has little control. The research bases on quantitative data and qualitative information obtained through semi-structured interviews of an open-ended nature and questionnaires to directors, members of the audit committee, risk, IT and compliance managers, and those responsible for internal audit function and anti-money laundering. The added value of the paper can be seen in the development of a framework based on a mapping of IT assets from which it is possible to identify their relationships for purposes of a more effective management and control of cyber risk.Keywords: bank, CybeRisk, information technology, risk management
Procedia PDF Downloads 231430 Acoustic Energy Harvesting Using Polyvinylidene Fluoride (PVDF) and PVDF-ZnO Piezoelectric Polymer
Authors: S. M. Giripunje, Mohit Kumar
Abstract:
Acoustic energy that exists in our everyday life and environment have been overlooked as a green energy that can be extracted, generated, and consumed without any significant negative impact to the environment. The harvested energy can be used to enable new technology like wireless sensor networks. Technological developments in the realization of truly autonomous MEMS devices and energy storage systems have made acoustic energy harvesting (AEH) an increasingly viable technology. AEH is the process of converting high and continuous acoustic waves from the environment into electrical energy by using an acoustic transducer or resonator. AEH is not popular as other types of energy harvesting methods since sound waves have lower energy density and such energy can only be harvested in very noisy environment. However, the energy requirements for certain applications are also correspondingly low and also there is a necessity to observe the noise to reduce noise pollution. So the ability to reclaim acoustic energy and store it in a usable electrical form enables a novel means of supplying power to relatively low power devices. A quarter-wavelength straight-tube acoustic resonator as an acoustic energy harvester is introduced with polyvinylidene fluoride (PVDF) and PVDF doped with ZnO nanoparticles, piezoelectric cantilever beams placed inside the resonator. When the resonator is excited by an incident acoustic wave at its first acoustic eigen frequency, an amplified acoustic resonant standing wave is developed inside the resonator. The acoustic pressure gradient of the amplified standing wave then drives the vibration motion of the PVDF piezoelectric beams, generating electricity due to the direct piezoelectric effect. In order to maximize the amount of the harvested energy, each PVDF and PVDF-ZnO piezoelectric beam has been designed to have the same structural eigen frequency as the acoustic eigen frequency of the resonator. With a single PVDF beam placed inside the resonator, the harvested voltage and power become the maximum near the resonator tube open inlet where the largest acoustic pressure gradient vibrates the PVDF beam. As the beam is moved to the resonator tube closed end, the voltage and power gradually decrease due to the decreased acoustic pressure gradient. Multiple piezoelectric beams PVDF and PVDF-ZnO have been placed inside the resonator with two different configurations: the aligned and zigzag configurations. With the zigzag configuration which has the more open path for acoustic air particle motions, the significant increases in the harvested voltage and power have been observed. Due to the interruption of acoustic air particle motion caused by the beams, it is found that placing PVDF beams near the closed tube end is not beneficial. The total output voltage of the piezoelectric beams increases linearly as the incident sound pressure increases. This study therefore reveals that the proposed technique used to harvest sound wave energy has great potential of converting free energy into useful energy.Keywords: acoustic energy, acoustic resonator, energy harvester, eigenfrequency, polyvinylidene fluoride (PVDF)
Procedia PDF Downloads 382429 Peak Constituent Fluxes from Small Arctic Rivers Generated by Late Summer Episodic Precipitation Events
Authors: Shawn G. Gallaher, Lilli E. Hirth
Abstract:
As permafrost thaws with the continued warming of the Alaskan North Slope, a progressively thicker active thaw layer is evidently releasing previously sequestered nutrients, metals, and particulate matter exposed to fluvial transport. In this study, we estimate material fluxes on the North Slope of Alaska during the 2019-2022 melt seasons. The watershed of the Alaskan North Slope can be categorized into three regions: mountains, tundra, and coastal plain. Precipitation and discharge data were collected from repeat visits to 14 sample sites for biogeochemical surface water samples, 7 point discharge measurements, 3 project deployed meteorology stations, and 2 U. S. Geological Survey (USGS) continuous discharge observation sites. The timing, intensity, and spatial distribution of precipitation determine the material flux composition in the Sagavanirktok and surrounding bodies of water, with geogenic constituents (e.g., dissolved inorganic carbon (DIC)) expected from mountain flushed events and biogenic constituents (e.g., dissolved organic compound (DOC)) expected from transitional tundra precipitation events. Project goals include connecting late summer precipitation events to peak discharge to determine the responses of the watershed to localized atmospheric forcing. Field study measurements showed widespread precipitation in August 2019, generating an increase in total suspended solids, dissolved organic carbon, and iron fluxes from the tundra, shifting the main-stem mountain river biogeochemistry toward tundra source characteristics typically only observed during the spring floods. Intuitively, a large-scale precipitation event (as defined by this study as exceeding 12.5 mm of precipitation on a single observation day) would dilute a body of water; however, in this study, concentrations increased with higher discharge responses on several occasions. These large-scale precipitation events continue to produce peak constituent fluxes as the thaw layer increases in depth and late summer precipitation increases, evidenced by 6 large-scale events in July 2022 alone. This increase in late summer events is in sharp contrast to the 3 or fewer large events in July in each of the last 10 years. Changes in precipitation intensity, timing, and location have introduced late summer peak constituent flux events previously confined to the spring freshet.Keywords: Alaska North Slope, arctic rivers, material flux, precipitation
Procedia PDF Downloads 73428 Examining the Overuse of Cystoscopy in the Evaluation of Lower Urinary Tract Symptoms in Men with Benign Prostatic Hyperplasia: A Prospective Study
Authors: Ilija Kelepurovski, Stefan Lazorovski, Pece Petkovski, Marian Anakievski, Svetlana Petkovska
Abstract:
Introduction: Benign prostatic hyperplasia (BPH) is a common condition that affects men over the age of 50 and is characterized by an enlarged prostate gland that can cause lower urinary tract symptoms (LUTS). Uroflowmetry and cystoscopy are two commonly used diagnostic tests to evaluate LUTS and diagnose BPH. While both tests can be useful, there is a risk of overusing cystoscopy and underusing uroflowmetry in the evaluation of LUTS. The aim of this study was to compare the use of uroflowmetry and cystoscopy in a prospective cohort of 100 patients with suspected BPH or other urinary tract conditions and to assess the diagnostic yield of each test. Materials and Methods: This was a prospective study of 100 male patients over the age of 50 with suspected BPH or other urinary tract conditions who underwent uroflowmetry and cystoscopy for the evaluation of LUTS at a single tertiary care center. Inclusion criteria included male patients over the age of 50 with suspected BPH or other urinary tract conditions, while exclusion criteria included previous urethral or bladder surgery, active urinary tract infection, and significant comorbidities. The primary outcome of the study was the frequency of cystoscopy in the evaluation of LUTS, and the secondary outcome was the diagnostic yield of each test. Results: Of the 100 patients included in the study, 86 (86%) were diagnosed with BPH and 14 (14%) had other urinary tract conditions. The mean age of the study population was 67 years. Uroflowmetry was performed on all 100 patients, while cystoscopy was performed on 70 (70%) of the patients. The diagnostic yield of uroflowmetry was high, with a clear diagnosis made in 92 (92%) of the patients. The diagnostic yield of cystoscopy was also high, with a clear diagnosis made in 63 (90%) of the patients who underwent the procedure. There was no statistically significant difference in the diagnostic yield of uroflowmetry and cystoscopy (p = 0.20). Discussion: Our study found that uroflowmetry is an effective and well-tolerated diagnostic tool for evaluating LUTS and diagnosing BPH, with a high diagnostic yield and low risk of complications. Cystoscopy is also a useful diagnostic tool, but it is more invasive and carries a small risk of complications such as bleeding or urinary tract infection. Both tests had a high diagnostic yield, suggesting that either test can provide useful information in the evaluation of LUTS. However, the fact that 70% of the study population underwent cystoscopy raises concerns about the potential overuse of this test in the evaluation of LUTS. This is especially relevant given the focus on patient-centered care and the need to minimize unnecessary or invasive procedures. Our findings underscore the importance of considering the clinical context and using evidence-based guidelines. Conclusion: In this prospective study of 100 patients with suspected BPH or other urinary tract conditions, we found that uroflowmetry and cystoscopy were both valuable diagnostic tools for the evaluation of LUTS. However, the potential overuse of cystoscopy in this population warrants further investigation and highlights the need for careful consideration of the optimal use of diagnostic tests in the evaluation of LUTS and the diagnosis of BPH. Further research is needed to better understand the relative roles of uroflowmetry and cystoscopy in the diagnostic workup of patients with LUTS, and to develop evidence-based guidelines for their appropriate use.Keywords: uroflowmetry, cystoscopy, LUTS, BPH
Procedia PDF Downloads 76427 Vicarious Cues in Portraying Emotion: Musicians' Self-Appraisal
Authors: W. Linthicum-Blackhorse, P. Martens
Abstract:
This present study seeks to discover attitudinal commonalities and differences within a musician population relative to the communication of emotion via music. We hypothesized that instrument type, as well as age and gender, would bear significantly on musicians’ opinions. A survey was administered to 178 participants; 152 were current music majors (mean age 20.3 years, 62 female) and 26 were adult participants in a community choir (mean age 54.0 years, 12 female). The adult participants were all vocalists, while student participants represented the full range of orchestral instruments. The students were grouped by degree program, (performance, music education, or other) and instrument type (voice, brass, woodwinds, strings, percussion). The survey asked 'How important are each of the following areas to you for portraying emotion in music?' Participants were asked to rate each of 15 items on a scale of 1 (not at all important) to 10 (very important). Participants were also instructed to leave blank any item that they did not understand. The 15 items were: dynamic contrast, overall volume, phrasing, facial expression, staging (placement), pitch accuracy, tempo changes, bodily movement, your mood, your attitude, vibrato, rubato, stage/room lighting, clothing type, and clothing color. Contrary to our hypothesis, there was no overall effect of gender or age, and neither did any single response item show a significant difference due to these subject parameters. Among the student participants, however, one-way ANOVA revealed a significant effect of degree program on the rated importance of four items: dynamic contrast, tempo changes, vibrato, and rubato. Significant effects of instrument type were found in the responses to eight items: facial expression, staging, body movement, vibrato, rubato, lighting, clothing type, and clothing color. Post hoc comparisons (Tukey) show that some variation follows from obvious differences between instrument types (e.g. string players are more concerned with vibrato than everyone but woodwind players; vocalists are significantly more concerned with facial expression than everyone but string players), but other differences could point to communal mindsets toward vicarious cues within instrument type. These mindsets could be global (e.g. brass players deeming body movement significantly less important than string players, being less often featured as soloists and appearing less often at the front of the stage) or local (e.g. string players being significantly more concerned than all other groups about both clothing color and type, perhaps due to the strongly-expressed opinions of specific teachers). Future work will attempt to identify the source of these self-appraisals, whether enculturated via explicit pedagogy, or whether absorbed from individuals' observations and performance experience.Keywords: performance, vicarious cues, communication, emotion
Procedia PDF Downloads 108426 Digital Twins: Towards an Overarching Framework for the Built Environment
Authors: Astrid Bagireanu, Julio Bros-Williamson, Mila Duncheva, John Currie
Abstract:
Digital Twins (DTs) have entered the built environment from more established industries like aviation and manufacturing, although there has never been a common goal for utilising DTs at scale. Defined as the cyber-physical integration of data between an asset and its virtual counterpart, DT has been identified in literature from an operational standpoint – in addition to monitoring the performance of a built asset. However, this has never been translated into how DTs should be implemented into a project and what responsibilities each project stakeholder holds in the realisation of a DT. What is needed is an approach to translate these requirements into actionable DT dimensions. This paper presents a foundation for an overarching framework specific to the built environment. For the purposes of this research, the UK widely used the Royal Institute of British Architects (RIBA) Plan of Work from 2020 is used as a basis for itemising project stages. The RIBA Plan of Work consists of eight stages designed to inform on the definition, briefing, design, coordination, construction, handover, and use of a built asset. Similar project stages are utilised in other countries; therefore, the recommendations from the interviews presented in this paper are applicable internationally. Simultaneously, there is not a single mainstream software resource that leverages DT abilities. This ambiguity meets an unparalleled ambition from governments and industries worldwide to achieve a national grid of interconnected DTs. For the construction industry to access these benefits, it necessitates a defined starting point. This research aims to provide a comprehensive understanding of the potential applications and ramifications of DT in the context of the built environment. This paper is an integral part of a larger research aimed at developing a conceptual framework for the Architecture, Engineering, and Construction (AEC) sector following a conventional project timeline. Therefore, this paper plays a pivotal role in providing practical insights and a tangible foundation for developing a stage-by-stage approach to assimilate the potential of DT within the built environment. First, the research focuses on a review of relevant literature, albeit acknowledging the inherent constraint of limited sources available. Secondly, a qualitative study compiling the views of 14 DT experts is presented, concluding with an inductive analysis of the interview findings - ultimately highlighting the barriers and strengths of DT in the context of framework development. As parallel developments aim to progress net-zero-centred design and improve project efficiencies across the built environment, the limited resources available to support DTs should be leveraged to propel the industry to reach its digitalisation era, in which AEC stakeholders have a fundamental role in understanding this, from the earliest stages of a project.Keywords: digital twins, decision-making, design, net-zero, built environment
Procedia PDF Downloads 120425 Life Cycle Assessment of a Parabolic Solar Cooker
Authors: Bastien Sanglard, Lou Magnat, Ligia Barna, Julian Carrey, Sébastien Lachaize
Abstract:
Cooking is a primary need for humans, several techniques being used around the globe based on different sources of energy: electricity, solid fuel (wood, coal...), fuel or liquefied petroleum gas. However, all of them leads to direct or indirect greenhouse gas emissions and sometimes health damage in household. Therefore, the solar concentrated power represent a great option to lower the damages because of a cleaner using phase. Nevertheless, the construction phase of the solar cooker still requires primary energy and materials, which leads to environmental impacts. The aims of this work is to analyse the ecological impacts of a commercialaluminium parabola and to compare it with other means of cooking, taking the boiling of 2 litres of water three times a day during 40 years as the functional unit. Life cycle assessment was performed using the software Umberto and the EcoInvent database. Calculations were realized over more than 13 criteria using two methods: the international panel on climate change method and the ReCiPe method. For the reflector itself, different aluminium provenances were compared, as well as the use of recycled aluminium. For the structure, aluminium was compared to iron (primary and recycled) and wood. Results show that climate impacts of the studied parabola was 0.0353 kgCO2eq/kWh when built with Chinese aluminium and can be reduced by 4 using aluminium from Canada. Assessment also showed that using 32% of recycled aluminium would reduce the impact by 1.33 and 1.43 compared to the use of primary Canadian aluminium and primary Chinese aluminium, respectively. The exclusive use of recycled aluminium lower the impact by 17. Besides, the use of iron (recycled or primary) or wood for the structure supporting the reflector significantly lowers the impact. The impact categories of the ReCiPe method show that the parabola made from Chinese aluminium has the heaviest impact - except for metal resource depletion - compared to aluminium from Canada, recycled aluminium or iron. Impact of solar cooking was then compared to gas stove and induction. The gas stove model was a cast iron tripod that supports the cooking pot, and the induction plate was as well a single spot plate. Results show the parabolic solar cooker has the lowest ecological impact over the 13 criteria of the ReCiPe method and over the global warming potential compared to the two other technologies. The climate impact of gas cooking is 0.628kgCO2/kWh when used with natural gas and 0.723 kgCO2/kWh when used with a bottle of gas. In each case, the main part of emissions came from gas burning. Induction cooking has a global warming potential of 0.12 kgCO2eq/kWh with the electricity mix of France, 96.3% of the impact being due to electricity production. Therefore, the electricity mix is a key factor for this impact: for instance, with the electricity mix of Germany and Poland, impacts are 0.81kgCO2eq/kWh and 1.39 kgCO2eq/kWh, respectively. Therefore, the parabolic solar cooker has a real ecological advantages compared to both gas stove and induction plate.Keywords: life cycle assessement, solar concentration, cooking, sustainability
Procedia PDF Downloads 183424 Investigation of Residual Stress Relief by in-situ Rolling Deposited Bead in Directed Laser Deposition
Authors: Ravi Raj, Louis Chiu, Deepak Marla, Aijun Huang
Abstract:
Hybridization of the directed laser deposition (DLD) process using an in-situ micro-roller to impart a vertical compressive load on the deposited bead at elevated temperatures can relieve tensile residual stresses incurred in the process. To investigate this stress relief mechanism and its relationship with the in-situ rolling parameters, a fully coupled dynamic thermo-mechanical model is presented in this study. A single bead deposition of Ti-6Al-4V alloy with an in-situ roller made of mild steel moving at a constant speed with a fixed nominal bead reduction is simulated using the explicit solver of the finite element software, Abaqus. The thermal model includes laser heating during the deposition process and the heat transfer between the roller and the deposited bead. The laser heating is modeled using a moving heat source with a Gaussian distribution, applied along the pre-formed bead’s surface using the VDFLUX Fortran subroutine. The bead’s cross-section is assumed to be semi-elliptical. The interfacial heat transfer between the roller and the bead is considered in the model. Besides, the roller is cooled internally using axial water flow, considered in the model using convective heat transfer. The mechanical model for the bead and substrate includes the effects of rolling along with the deposition process, and their elastoplastic material behavior is captured using the J2 plasticity theory. The model accounts for strain, strain rate, and temperature effects on the yield stress based on Johnson-Cook’s theory. Various aspects of this material behavior are captured in the FE software using the subroutines -VUMAT for elastoplastic behavior, VUHARD for yield stress, and VUEXPAN for thermal strain. The roller is assumed to be elastic and does not undergo any plastic deformation. Also, contact friction at the roller-bead interface is considered in the model. Based on the thermal results of the bead, the distance between the roller and the deposition nozzle (roller o set) can be determined to ensure rolling occurs around the beta-transus temperature for the Ti-6Al-4V alloy. It is identified that roller offset and the nominal bead height reduction are crucial parameters that influence the residual stresses in the hybrid process. The results obtained from a simulation at roller offset of 20 mm and nominal bead height reduction of 7% reveal that the tensile residual stresses decrease to about 52% due to in-situ rolling throughout the deposited bead. This model can be used to optimize the rolling parameters to minimize the residual stresses in the hybrid DLD process with in-situ micro-rolling.Keywords: directed laser deposition, finite element analysis, hybrid in-situ rolling, thermo-mechanical model
Procedia PDF Downloads 109423 Molecular Defects Underlying Genital Ambiguity in Egyptian Patients: A Systematic Review
Authors: Y. Z. Gad
Abstract:
Disorders of Sex Development (DSD) are defined as congenital conditions in which development of chromosomal, gonadal or anatomical sex is atypical. The DSD are relatively prevalent in Egypt. In spite of that, the relative rarity of the individual disease types or their molecular pathologies frequently resulted in reporting on single or few cases. This augmented the challenging nature of phenotype-genotype correlation in this disease group and its utilization in the management of such medical emergency. Through critical assessment of the published DSD reports, the current review aims at analyzing the clinical characteristics of the various DSD forms in relation to the underlying molecular pathologies. A systematic literature search was done in Pubmed, using relevant keywords (Egypt versus DSD, genital ambiguity or ambiguous genitalia, the old terms of 'intersex, hermaphroditism and pseudohermaphroditism', and a list of the DSD entities and their related genes). The search yielded 24 reports of molecular data in Egyptian patients presenting with ambiguous genitalia. However, only 21 publications fulfilled the criteria of inclusion of detailed clinical descriptions and definitive molecular diagnoses of individual patients. Curation of the data yielded a total of 53 cases that were ascertained from 40 families. Fifty-one patients present with ambiguous genitalia only while 2 had multiple congenital anomalies. Parental consanguinity was noted in 60% of cases. Sex of rearing at initial presentation was female in 75% and 60% in 46,XY and 46,XX DSD cases, respectively. The external genital phenotype in 2/3 of the 46,XY DSD cases showed moderate undermasculinization [Quigley scores 3 & 4] and 1/3 had severe presentations [scores 5 & 6]. For 46,XX subjects, 1 had severe virilization of the external genitalia while 8 had moderate phenotype. Hormonal data were inconclusive or contradictory to final diagnosis in a forth of cases. Collectively, 31 families [31/40, 77.5%] with 46,XY DSD had molecular defects in the genes, 5 alpha reductase 2 (SRD5A2) [12/31], 17 beta-hydroxysteroid dehydrogenase 3 [8/31], androgen receptor [7/31], Steroidogenic factor 1 [2/31], luteinizing hormone receptor [1/31], and fibroblast growth factor receptor 1 [1/31]. In a multiethnic study, 9 families afflicted with 46,XX DSD due to 11 beta hydroxylase (CYP11B1) deficiency were documented. Two recurrent mutations, G34R and N160D, in SRD5A2 were present, respectively, in 42 and 17% of cases. Similarly, 4 recurrent mutations resulted in 89% of the CYP11B1 presentations. In conclusion, this analysis highlights the importance of autosomal recessive inheritance and inbreeding among DSD presentations, the importance of founder effect in at least 2 disorders, the difficulties in relating the genotype with the indeterminate genital phenotype, the under-reporting of some DSD subtypes, and the notion that the reported mutational profiles among Egyptian DSD cases are relatively different from those reported in other ethnic groups.Keywords: disorders of sex development, genital ambiguity, mutation, molecular diagnosis, Egypt
Procedia PDF Downloads 136422 AI-Based Information System for Hygiene and Safety Management of Shared Kitchens
Authors: Jongtae Rhee, Sangkwon Han, Seungbin Ji, Junhyeong Park, Byeonghun Kim, Taekyung Kim, Byeonghyeon Jeon, Jiwoo Yang
Abstract:
The shared kitchen is a concept that transfers the value of the sharing economy to the kitchen. It is a type of kitchen equipped with cooking facilities that allows multiple companies or chefs to share time and space and use it jointly. These shared kitchens provide economic benefits and convenience, such as reduced investment costs and rent, but also increase the risk of safety management, such as cross-contamination of food ingredients. Therefore, to manage the safety of food ingredients and finished products in a shared kitchen where several entities jointly use the kitchen and handle various types of food ingredients, it is critical to manage followings: the freshness of food ingredients, user hygiene and safety and cross-contamination of cooking equipment and facilities. In this study, it propose a machine learning-based system for hygiene safety and cross-contamination management, which are highly difficult to manage. User clothing management and user access management, which are most relevant to the hygiene and safety of shared kitchens, are solved through machine learning-based methodology, and cutting board usage management, which is most relevant to cross-contamination management, is implemented as an integrated safety management system based on artificial intelligence. First, to prevent cross-contamination of food ingredients, we use images collected through a real-time camera to determine whether the food ingredients match a given cutting board based on a real-time object detection model, YOLO v7. To manage the hygiene of user clothing, we use a camera-based facial recognition model to recognize the user, and real-time object detection model to determine whether a sanitary hat and mask are worn. In addition, to manage access for users qualified to enter the shared kitchen, we utilize machine learning based signature recognition module. By comparing the pairwise distance between the contract signature and the signature at the time of entrance to the shared kitchen, access permission is determined through a pre-trained signature verification model. These machine learning-based safety management tasks are integrated into a single information system, and each result is managed in an integrated database. Through this, users are warned of safety dangers through the tablet PC installed in the shared kitchen, and managers can track the cause of the sanitary and safety accidents. As a result of system integration analysis, real-time safety management services can be continuously provided by artificial intelligence, and machine learning-based methodologies are used for integrated safety management of shared kitchens that allows dynamic contracts among various users. By solving this problem, we were able to secure the feasibility and safety of the shared kitchen business.Keywords: artificial intelligence, food safety, information system, safety management, shared kitchen
Procedia PDF Downloads 66421 Strategic Asset Allocation Optimization: Enhancing Portfolio Performance Through PCA-Driven Multi-Objective Modeling
Authors: Ghita Benayad
Abstract:
Asset allocation, which affects the long-term profitability of portfolios by distributing assets to fulfill a range of investment objectives, is the cornerstone of investment management in the dynamic and complicated world of financial markets. This paper offers a technique for optimizing strategic asset allocation with the goal of improving portfolio performance by addressing the inherent complexity and uncertainty of the market through the use of Principal Component Analysis (PCA) in a multi-objective modeling framework. The study's first section starts with a critical evaluation of conventional asset allocation techniques, highlighting how poorly they are able to capture the intricate relationships between assets and the volatile nature of the market. In order to overcome these challenges, the project suggests a PCA-driven methodology that isolates important characteristics influencing asset returns by decreasing the dimensionality of the investment universe. This decrease provides a stronger basis for asset allocation decisions by facilitating a clearer understanding of market structures and behaviors. Using a multi-objective optimization model, the project builds on this foundation by taking into account a number of performance metrics at once, including risk minimization, return maximization, and the accomplishment of predetermined investment goals like regulatory compliance or sustainability standards. This model provides a more comprehensive understanding of investor preferences and portfolio performance in comparison to conventional single-objective optimization techniques. While applying the PCA-driven multi-objective optimization model to historical market data, aiming to construct portfolios better under different market situations. As compared to portfolios produced from conventional asset allocation methodologies, the results show that portfolios optimized using the proposed method display improved risk-adjusted returns, more resilience to market downturns, and better alignment with specified investment objectives. The study also looks at the implications of this PCA technique for portfolio management, including the prospect that it might give investors a more advanced framework for navigating financial markets. The findings suggest that by combining PCA with multi-objective optimization, investors may obtain a more strategic and informed asset allocation that is responsive to both market conditions and individual investment preferences. In conclusion, this capstone project improves the field of financial engineering by creating a sophisticated asset allocation optimization model that integrates PCA with multi-objective optimization. In addition to raising concerns about the condition of asset allocation today, the proposed method of portfolio management opens up new avenues for research and application in the area of investment techniques.Keywords: asset allocation, portfolio optimization, principle component analysis, multi-objective modelling, financial market
Procedia PDF Downloads 44420 Tracing a Timber Breakthrough: A Qualitative Study of the Introduction of Cross-Laminated-Timber to the Student Housing Market in Norway
Authors: Marius Nygaard, Ona Flindall
Abstract:
The Palisaden student housing project was completed in August 2013 and was, with its eight floors, Norway’s tallest timber building at the time of completion. It was the first time cross-laminated-timber (CLT) was utilized at this scale in Norway. The project was the result of a concerted effort by a newly formed management company to establish CLT as a sustainable and financially competitive alternative to conventional steel and concrete systems. The introduction of CLT onto the student housing market proved so successful that by 2017 more than 4000 individual student residences will have been built using the same model of development and construction. The aim of this paper is to identify the key factors that enabled this breakthrough for CLT. It is based on an in-depth study of a series of housing projects and the role of the management company who both instigated and enabled this shift of CLT from the margin to the mainstream. Specifically, it will look at how a new building system was integrated into a marketing strategy that identified a market potential within the existing structure of the construction industry and within the economic restrictions inherent to student housing in Norway. It will show how a key player established a project model that changed both the patterns of cooperation and the information basis for decisions. Based on qualitative semi-structured interviews with managers, contractors and the interdisciplinary teams of consultants (architects, structural engineers, acoustical experts etc.) this paper will trace the introduction, expansion and evolution of CLT-based building systems in the student housing market. It will show how the project management firm’s position in the value chain enabled them to function both as a liaison between contractor and client, and between contractor and producer. A position that allowed them to improve the flow of information. This ensured that CLT was handled on equal terms to other structural solutions in the project specifications, enabling realistic pricing and risk evaluation. Secondly, this paper will describe and discuss how the project management firm established and interacted with a growing network of contractors, architects and engineers to pool expertise and broaden the knowledge base across Norway’s regional markets. Finally, it will examine the role of the client, the building typology, and the industrial and technological factors in achieving this breakthrough for CLT in the construction industry. This paper gives an in-depth view of the progression of a single case rather than a broad description of the state of the art of large-scale timber building in Norway. However, this type of study may offer insights that are important to the understanding not only of specific markets but also of how new technologies should be introduced in big and well-established industries.Keywords: cross-laminated-timber (CLT), industry breakthrough, student housing, timber market
Procedia PDF Downloads 221419 Rt-Pcr Negative COVID-19 Infection in a Bodybuilding Competitor Using Anabolic Steroids: A Case Report
Authors: Mariana Branco, Nahida Sobrino, Cristina Neves, Márcia Santos, Afonso Granja, João Rosa Oliveira, Joana Costa, Luísa Castro Leite
Abstract:
This case reports a COVID-19 infection in an unvaccinated adult man with no history of COVID-19 and no relevant clinical history besides anabolic steroid use, undergoing weaning with tamoxifen after a bodybuilding competition. The patient presented a 4cm cervical mass 3 weeks after COVID-19 infection in his cohabitants. He was otherwise asymptomatic and tested negative to multiple RT-PCR tests. Nevertheless, the IgG COVID-19 antibody was positive, suggesting the previous infection. This report raises a potential link between anabolic steroid use and atypical COVID-19 onset. Objectives: The goals of this paper are to raise a potential link between anabolic steroid use and atypical COVID-19 onset but also to report an uncommon case of COVID-19 infection with consecutive negative gold standard tests. Methodology: The authors used CARE guidelines for case report writing. Introduction: This case reports a COVID-19 infection case in an unvaccinated adult man, with multiple serial negative reverse transcription polymerase chain reaction (RT-PCR) test results, presenting with single cervical lymphadenopathy. Although the association between COVID-19 and lymphadenopathy is well established, there are no cases with this presentation, and consistently negative RT-PCR tests have been reported. Methodologies: The authors used CARE guidelines for case report writing. Case presentation: This case reports a 28-year-old Caucasian man with no previous history of COVID-19 infection or vaccination and no relevant clinical history besides anabolic steroid use undergoing weaning with tamoxifendue to participation in a bodybuilding competition. He visits his primary care physician because of a large (4 cm) cervical lump, present for 3 days prior to the consultation. There was a positive family history for COVID-19 infection 3 weeks prior to the visit, during which the patient cohabited with the infected family members. The patient never had any previous clinical manifestation of COVID-19 infection and, despite multiple consecutive RT-PCR testing, never tested positive. The patient was treated with an NSAID and a broad-spectrum antibiotic, with little to no effect. Imagiological testing was performed via a cervical ultrasound, followed by a needle biopsy for histologic analysis. Serologic testing for COVID-19 immunity was conducted, revealing a positive Anti-SARS-CoV-2 IgG (Spike S1) antibody, suggesting the previous infection, given the unvaccinated status of our patient Conclusion: In patients with a positive epidemiologic context and cervical lymphadenopathy, physicians should still consider COVID-19 infection as a differential diagnosis, despite negative PCR testing. This case also raises a potential link between anabolic steroid use and atypical COVID-19 onset, never before reported in scientific literature.Keywords: COVID-19, cervical lymphadenopathy, anabolic steroids, primary care
Procedia PDF Downloads 114418 Tailoring Quantum Oscillations of Excitonic Schrodinger’s Cats as Qubits
Authors: Amit Bhunia, Mohit Kumar Singh, Maryam Al Huwayz, Mohamed Henini, Shouvik Datta
Abstract:
We report [https://arxiv.org/abs/2107.13518] experimental detection and control of Schrodinger’s Cat like macroscopically large, quantum coherent state of a two-component Bose-Einstein condensate of spatially indirect electron-hole pairs or excitons using a resonant tunneling diode of III-V Semiconductors. This provides access to millions of excitons as qubits to allow efficient, fault-tolerant quantum computation. In this work, we measure phase-coherent periodic oscillations in photo-generated capacitance as a function of an applied voltage bias and light intensity over a macroscopically large area. Periodic presence and absence of splitting of excitonic peaks in the optical spectra measured by photocapacitance point towards tunneling induced variations in capacitive coupling between the quantum well and quantum dots. Observation of negative ‘quantum capacitance’ due to a screening of charge carriers by the quantum well indicates Coulomb correlations of interacting excitons in the plane of the sample. We also establish that coherent resonant tunneling in this well-dot heterostructure restricts the available momentum space of the charge carriers within this quantum well. Consequently, the electric polarization vector of the associated indirect excitons collective orients along the direction of applied bias and these excitons undergo Bose-Einstein condensation below ~100 K. Generation of interference beats in photocapacitance oscillation even with incoherent white light further confirm the presence of stable, long-range spatial correlation among these indirect excitons. We finally demonstrate collective Rabi oscillations of these macroscopically large, ‘multipartite’, two-level, coupled and uncoupled quantum states of excitonic condensate as qubits. Therefore, our study not only brings the physics and technology of Bose-Einstein condensation within the reaches of semiconductor chips but also opens up experimental investigations of the fundamentals of quantum physics using similar techniques. Operational temperatures of such two-component excitonic BEC can be raised further with a more densely packed, ordered array of QDs and/or using materials having larger excitonic binding energies. However, fabrications of single crystals of 0D-2D heterostructures using 2D materials (e.g. transition metal di-chalcogenides, oxides, perovskites etc.) having higher excitonic binding energies are still an open challenge for semiconductor optoelectronics. As of now, these 0D-2D heterostructures can already be scaled up for mass production of miniaturized, portable quantum optoelectronic devices using the existing III-V and/or Nitride based semiconductor fabrication technologies.Keywords: exciton, Bose-Einstein condensation, quantum computation, heterostructures, semiconductor Physics, quantum fluids, Schrodinger's Cat
Procedia PDF Downloads 178417 Phenomena-Based Approach for Automated Generation of Process Options and Process Models
Authors: Parminder Kaur Heer, Alexei Lapkin
Abstract:
Due to global challenges of increased competition and demand for more sustainable products/processes, there is a rising pressure on the industry to develop innovative processes. Through Process Intensification (PI) the existing and new processes may be able to attain higher efficiency. However, very few PI options are generally considered. This is because processes are typically analysed at a unit operation level, thus limiting the search space for potential process options. PI performed at more detailed levels of a process can increase the size of the search space. The different levels at which PI can be achieved is unit operations, functional and phenomena level. Physical/chemical phenomena form the lowest level of aggregation and thus, are expected to give the highest impact because all the intensification options can be described by their enhancement. The objective of the current work is thus, generation of numerous process alternatives based on phenomena, and development of their corresponding computer aided models. The methodology comprises: a) automated generation of process options, and b) automated generation of process models. The process under investigation is disintegrated into functions viz. reaction, separation etc., and these functions are further broken down into the phenomena required to perform them. E.g., separation may be performed via vapour-liquid or liquid-liquid equilibrium. A list of phenomena for the process is formed and new phenomena, which can overcome the difficulties/drawbacks of the current process or can enhance the effectiveness of the process, are added to the list. For instance, catalyst separation issue can be handled by using solid catalysts; the corresponding phenomena are identified and added. The phenomena are then combined to generate all possible combinations. However, not all combinations make sense and, hence, screening is carried out to discard the combinations that are meaningless. For example, phase change phenomena need the co-presence of the energy transfer phenomena. Feasible combinations of phenomena are then assigned to the functions they execute. A combination may accomplish a single or multiple functions, i.e. it might perform reaction or reaction with separation. The combinations are then allotted to the functions needed for the process. This creates a series of options for carrying out each function. Combination of these options for different functions in the process leads to the generation of superstructure of process options. These process options, which are formed by a list of phenomena for each function, are passed to the model generation algorithm in the form of binaries (1, 0). The algorithm gathers the active phenomena and couples them to generate the model. A series of models is generated for the functions, which are combined to get the process model. The most promising process options are then chosen subjected to a performance criterion, for example purity of product, or via a multi-objective Pareto optimisation. The methodology was applied to a two-step process and the best route was determined based on the higher product yield. The current methodology can identify, produce and evaluate process intensification options from which the optimal process can be determined. It can be applied to any chemical/biochemical process because of its generic nature.Keywords: Phenomena, Process intensification, Process models , Process options
Procedia PDF Downloads 230416 Prolactin and Its Abnormalities: Its Implications on the Male Reproductive Tract and Male Factor Infertility
Authors: Rizvi Hasan
Abstract:
Male factor infertility due to abnormalities in prolactin levels is encountered in a significant proportion. This was a case-control study carried out to determine the effects of prolactin abnormalities in normal males with infertility, recruiting 297 male infertile patients with informed written consent. All underwent a Basic Seminal Fluid Analysis (BSA) and endocrine profiles of FSH, LH, testosterone and prolactin (PRL) hormones using the random access chemiluminescent immunoassay method (normal range 2.5-17ng/ml). Age, weight, and height matched voluntary controls were recruited for comparison. None of the cases had anatomical, medical or surgical disorders related to infertility. Among the controls; mean age 33.2yrs ± 5.2, BMI 21.04 ± 1.39kgm-2, BSA 34×106, a number of children fathered 2±1, PRL 6.78 ± 2.92ng/ml. Of the 297 patients, 28 were hyperprolactinaemic while one was hypoprolactinaemic. All the hyperprolactinaemic patients had oligoasthenospermia, abnormal morphology and decreased viability. The serum testosterone levels were markedly lowered in 26 (92.86%) of the hyperprolactinaemic subjects. In the other 2 hyperprolactinaemic subjects and the single hypoprolactinaemic subject, the serum testosterone levels were normal. FSH and LH were normal in all patients. The 29 male patients with abnormalities in their serum PRL profiles were followed up for 12 months. The 28 patients suffering from hyperprolactinaemia were treated with oral bromocriptine in a dose of 2.5 mg twice daily. The hypoprolactinaemic patient defaulted treatment. From the follow-up, it was evident that 19 (67.86%) of the treated patients responded after 3 months of therapy while 4 (14.29%) showed improvement after approximately 6 months of bromocriptine therapy. One patient responded after 1 year of therapy while 2 patients showed improvements although not up to normal levels within the same period. Response to treatment was assessed by improvement in their BSA parameters. Prolactin abnormalities affect the male reproductive system and semen parameters necessitating further studies to ascertain the exact role of prolactin on the male reproductive tract. A parallel study was carried out incorporating 200 male white rats that were grouped and subjected to variations in their serum PRL levels. At the end of 100 days of treatment, these rats were subjected to morphological studies of their male reproductive tracts.Varying morphological changes depending on the levels of PRL changes induced were evident. Notable changes were arrest of spermatogenesis at the spermatid stage, a reduced testicular cellularity, a reduction in microvilli of the pseudostratified epithelial lining of the epididymis, while measurement of the tubular diameter showed a 30% reduction compared to normal tissue. There were no changes in the vas deferens, seminal vesicles, and the prostate. It is evident that both hyperprolactinaemia and hypoprolactinaemia have a direct effect on the morphology and function of the male reproductive tract. The morphological studies carried out on the groups of rats who were subjected to variations in their PRL levels could be the basis for infertility in male human beings.Keywords: male factor infertility, morphological studies, prolactin, seminal fluid analysis
Procedia PDF Downloads 344415 Dynamic EEG Desynchronization in Response to Vicarious Pain
Authors: Justin Durham, Chanda Rooney, Robert Mather, Mickie Vanhoy
Abstract:
The psychological construct of empathy is to understand a person’s cognitive perspective and experience the other person’s emotional state. Deciphering emotional states is conducive for interpreting vicarious pain. Observing others' physical pain activates neural networks related to the actual experience of pain itself. The study addresses empathy as a nonlinear dynamic process of simulation for individuals to understand the mental states of others and experience vicarious pain, exhibiting self-organized criticality. Such criticality follows from a combination of neural networks with an excitatory feedback loop generating bistability to resonate permutated empathy. Cortical networks exhibit diverse patterns of activity, including oscillations, synchrony and waves, however, the temporal dynamics of neurophysiological activities underlying empathic processes remain poorly understood. Mu rhythms are EEG oscillations with dominant frequencies of 8-13 Hz becoming synchronized when the body is relaxed with eyes open and when the sensorimotor system is in idle, thus, mu rhythm synchrony is expected to be highest in baseline conditions. When the sensorimotor system is activated either by performing or simulating action, mu rhythms become suppressed or desynchronize, thus, should be suppressed while observing video clips of painful injuries if previous research on mirror system activation holds. Twelve undergraduates contributed EEG data and survey responses to empathy and psychopathy scales in addition to watching consecutive video clips of sports injuries. Participants watched a blank, black image on a computer monitor before and after observing a video of consecutive sports injuries incidents. Each video condition lasted five-minutes long. A BIOPAC MP150 recorded EEG signals from sensorimotor and thalamocortical regions related to a complex neural network called the ‘pain matrix’. Physical and social pain are activated in this network to resonate vicarious pain responses to processing empathy. Five EEG single electrode locations were applied to regions measuring sensorimotor electrical activity in microvolts (μV) to monitor mu rhythms. EEG signals were sampled at a rate of 200 Hz. Mu rhythm desynchronization was measured via 8-13 Hz at electrode sites (F3 & F4). Data for each participant’s mu rhythms were analyzed via Fast Fourier Transformation (FFT) and multifractal time series analysis.Keywords: desynchronization, dynamical systems theory, electroencephalography (EEG), empathy, multifractal time series analysis, mu waveform, neurophysiology, pain simulation, social cognition
Procedia PDF Downloads 283414 Mechanical Testing of Composite Materials for Monocoque Design in Formula Student Car
Authors: Erik Vassøy Olsen, Hirpa G. Lemu
Abstract:
Inspired by the Formula-1 competition, IMechE (Institute of Mechanical Engineers) and Formula SAE (Society of Mechanical Engineers) organize annual competitions for University and College students worldwide to compete with a single-seat race car they have designed and built. The design of the chassis or the frame is a key component of the competition because the weight and stiffness properties are directly related with the performance of the car and the safety of the driver. In addition, a reduced weight of the chassis has a direct influence on the design of other components in the car. Among others, it improves the power to weight ratio and the aerodynamic performance. As the power output of the engine or the battery installed in the car is limited to 80 kW, increasing the power to weight ratio demands reduction of the weight of the chassis, which represents the major part of the weight of the car. In order to reduce the weight of the car, ION Racing team from the University of Stavanger, Norway, opted for a monocoque design. To ensure fulfilment of the above-mentioned requirements of the chassis, the monocoque design should provide sufficient torsional stiffness and absorb the impact energy in case of a possible collision. The study reported in this article is based on the requirements for Formula Student competition. As part of this study, diverse mechanical tests were conducted to determine the mechanical properties and performances of the monocoque design. Upon a comprehensive theoretical study of the mechanical properties of sandwich composite materials and the requirements of monocoque design in the competition rules, diverse tests were conducted including 3-point bending test, perimeter shear test and test for absorbed energy. The test panels were homemade and prepared with an equivalent size of the side impact zone of the monocoque, i.e. 275 mm x 500 mm so that the obtained results from the tests can be representative. Different layups of the test panels with identical core material and the same number of layers of carbon fibre were tested and compared. Influence of the core material thickness was also studied. Furthermore, analytical calculations and numerical analysis were conducted to check compliance to the stated rules for Structural Equivalency with steel grade SAE/AISI 1010. The test results were also compared with calculated results with respect to bending and torsional stiffness, energy absorption, buckling, etc. The obtained results demonstrate that the material composition and strength of the composite material selected for the monocoque design has equivalent structural properties as a welded frame and thus comply with the competition requirements. The developed analytical calculation algorithms and relations will be useful for future monocoque designs with different lay-ups and compositions.Keywords: composite material, Formula student, ION racing, monocoque design, structural equivalence
Procedia PDF Downloads 501413 Collaborative Management Approach for Logistics Flow Management of Cuban Medicine Supply Chain
Authors: Ana Julia Acevedo Urquiaga, Jose A. Acevedo Suarez, Ana Julia Urquiaga Rodriguez, Neyfe Sablon Cossio
Abstract:
Despite the progress made in logistics and supply chains fields, it is unavoidable the development of business models that use efficiently information to facilitate the integrated logistics flows management between partners. Collaborative management is an important tool for materializing the cooperation between companies, as a way to achieve the supply chain efficiency and effectiveness. The first face of this research was a comprehensive analysis of the collaborative planning on the Cuban companies. It is evident that they have difficulties in supply chains planning where production, supplies and replenishment planning are independent tasks, as well as logistics and distribution operations. Large inventories generate serious financial and organizational problems for entities, demanding increasing levels of working capital that cannot be financed. Problems were found in the efficient application of Information and Communication Technology on business management. The general objective of this work is to develop a methodology that allows the deployment of a planning and control system in a coordinated way on the medicine’s logistics system in Cuba. To achieve these objectives, several mechanisms of supply chain coordination, mathematical programming models, and other management techniques were analyzed to meet the requirements of collaborative logistics management in Cuba. One of the findings is the practical and theoretical inadequacies of the studied models to solve the current situation of the Cuban logistics systems management. To contribute to the tactical-operative management of logistics, the Collaborative Logistics Flow Management Model (CLFMM) is proposed as a tool for the balance of cycles, capacities, and inventories, always to meet the final customers’ demands in correspondence with the service level expected by these. The CLFMM has as center the supply chain planning and control system as a unique information system, which acts on the processes network. The development of the model is based on the empirical methods of analysis-synthesis and the study cases. Other finding is the demonstration of the use of a single information system to support the supply chain logistics management, allows determining the deadlines and quantities required in each process. This ensures that medications are always available to patients and there are no faults that put the population's health at risk. The simulation of planning and control with the CLFMM in medicines such as dipyrone and chlordiazepoxide, during 5 months of 2017, permitted to take measures to adjust the logistic flow, eliminate delayed processes and avoid shortages of the medicines studied. As a result, the logistics cycle efficiency can be increased to 91%, the inventory rotation would increase, and this results in a release of financial resources.Keywords: collaborative management, medicine logistic system, supply chain planning, tactical-operative planning
Procedia PDF Downloads 174412 Implementation of Ecological and Energy-Efficient Building Concepts
Authors: Robert Wimmer, Soeren Eikemeier, Michael Berger, Anita Preisler
Abstract:
A relatively large percentage of energy and resource consumption occurs in the building sector. This concerns the production of building materials, the construction of buildings and also the energy consumption during the use phase. Therefore, the overall objective of this EU LIFE project “LIFE Cycle Habitation” (LIFE13 ENV/AT/000741) is to demonstrate innovative building concepts that significantly reduce CO₂emissions, mitigate climate change and contain a minimum of grey energy over their entire life cycle. The project is being realised with the contribution of the LIFE financial instrument of the European Union. The ultimate goal is to design and build prototypes for carbon-neutral and “LIFE cycle”-oriented residential buildings and make energy-efficient settlements the standard of tomorrow in line with the EU 2020 objectives. To this end, a resource and energy-efficient building compound is being built in Böheimkirchen, Lower Austria, which includes 6 living units and a community area as well as 2 single family houses with a total usable floor surface of approximately 740 m². Different innovative straw bale construction types (load bearing and pre-fabricated non loadbearing modules) together with a highly innovative energy-supply system, which is based on the maximum use of thermal energy for thermal energy services, are going to be implemented. Therefore only renewable resources and alternative energies are used to generate thermal as well as electrical energy. This includes the use of solar energy for space heating, hot water and household appliances like dishwasher or washing machine, but also a cooking place for the community area operated with thermal oil as heat transfer medium on a higher temperature level. Solar collectors in combination with a biomass cogeneration unit and photovoltaic panels are used to provide thermal and electric energy for the living units according to the seasonal demand. The building concepts are optimised by support of dynamic simulations. A particular focus is on the production and use of modular prefabricated components and building parts made of regionally available, highly energy-efficient, CO₂-storing renewable materials like straw bales. The building components will be produced in collaboration by local SMEs that are organised in an efficient way. The whole building process and results are monitored and prepared for knowledge transfer and dissemination including a trial living in the residential units to test and monitor the energy supply system and to involve stakeholders into evaluation and dissemination of the applied technologies and building concepts. The realised building concepts should then be used as templates for a further modular extension of the settlement in a second phase.Keywords: energy-efficiency, green architecture, renewable resources, sustainable building
Procedia PDF Downloads 148411 Separating Landform from Noise in High-Resolution Digital Elevation Models through Scale-Adaptive Window-Based Regression
Authors: Anne M. Denton, Rahul Gomes, David W. Franzen
Abstract:
High-resolution elevation data are becoming increasingly available, but typical approaches for computing topographic features, like slope and curvature, still assume small sliding windows, for example, of size 3x3. That means that the digital elevation model (DEM) has to be resampled to the scale of the landform features that are of interest. Any higher resolution is lost in this resampling. When the topographic features are computed through regression that is performed at the resolution of the original data, the accuracy can be much higher, and the reported result can be adjusted to the length scale that is relevant locally. Slope and variance are calculated for overlapping windows, meaning that one regression result is computed per raster point. The number of window centers per area is the same for the output as for the original DEM. Slope and variance are computed by performing regression on the points in the surrounding window. Such an approach is computationally feasible because of the additive nature of regression parameters and variance. Any doubling of window size in each direction only takes a single pass over the data, corresponding to a logarithmic scaling of the resulting algorithm as a function of the window size. Slope and variance are stored for each aggregation step, allowing the reported slope to be selected to minimize variance. The approach thereby adjusts the effective window size to the landform features that are characteristic to the area within the DEM. Starting with a window size of 2x2, each iteration aggregates 2x2 non-overlapping windows from the previous iteration. Regression results are stored for each iteration, and the slope at minimal variance is reported in the final result. As such, the reported slope is adjusted to the length scale that is characteristic of the landform locally. The length scale itself and the variance at that length scale are also visualized to aid in interpreting the results for slope. The relevant length scale is taken to be half of the window size of the window over which the minimum variance was achieved. The resulting process was evaluated for 1-meter DEM data and for artificial data that was constructed to have defined length scales and added noise. A comparison with ESRI ArcMap was performed and showed the potential of the proposed algorithm. The resolution of the resulting output is much higher and the slope and aspect much less affected by noise. Additionally, the algorithm adjusts to the scale of interest within the region of the image. These benefits are gained without additional computational cost in comparison with resampling the DEM and computing the slope over 3x3 images in ESRI ArcMap for each resolution. In summary, the proposed approach extracts slope and aspect of DEMs at the lengths scales that are characteristic locally. The result is of higher resolution and less affected by noise than existing techniques.Keywords: high resolution digital elevation models, multi-scale analysis, slope calculation, window-based regression
Procedia PDF Downloads 126410 The Effect of Nanocomposite on the Release of Imipenem on Bacteria Causing Infections with Implants
Authors: Mohammad Hossein Pazandeh, Monir Doudi, Sona Rostampour Yasouri
Abstract:
—Results The prudent administration of antibiotics aims to avoid the side effects and the microbes' resistance to antibiotics. An approach developing methods of local administration of antibiotics is especially required for localized infections caused by bacterial colonization of medical devices or implant materials. Among the wide variety of materials used as drug delivery systems, bioactive glasses (BG) have large utilization in regenerative medicine . firstly, the production of bioactive glass/nickel oxide/tin dioxide nanocomposite using sol-gel method, and then, the controlled release of imipenem from the double metal oxide/bioactive glass nanocomposite, and finally, the investigation of the antibacterial property of the nanocomposite. against a number of implant-related infectious agents. In this study, BG/SnO2 and BG/NiO single systema with different metal oxide present and BG/NiO/SnO2 nanocomposites were synthesized by sol-gel as drug carriers for tetracycline and imepinem. These two antibiotics were widely used for osteomyelitis because of its favorable penetration and bactericidal effect on all the probable osteomyelitis pathogens. The antibacterial activity of synthesized samples were evaluated against Staphylococcus aureus, Escherichia coli, Pseudomonas aeruginosa as bacteria model using disk diffusion method. The BG modification using metal oxides results to antibacterial property of samples containing metal oxide with highest efficiency for nancomposite. bioactivity of all samples was assessed by determining the surface morphology, structural and composition changes using scanning electron microscopy (SEM), FTIR and X-ray diffraction (XRD) spectroscopy, respectively, after soaking in simulated body fluid (SBF) for 28 days. The hydroxyapatite formation was clearly observed as a bioactivity measurement. Then, BG nanocomposite sample was loaded using two antibiotics, separately and their release profiles were studied. The BG nancomposite sample was shown the slow and continuous drug releasing for a period of 72 hours which is desirable for a drug delivery system. The loaded antibiotic nanocomposite sample retaining antibacterial property and showing inactivation effect against bacteria under test. The modified bioactive glass forming hydroxyapatite with controlled release drug and effective against bacterial infections can be introduced as scaffolds for bone implants after clinical trials for biomedical applications . Considering the formation of biofilm by infectious bacteria after sticking on the surfaces of implants, medical devices, etc. Also, considering the complications of traditional methods, solving the problems caused by the above-mentioned microorganisms in technical and biomedical industries was one of the necessities of this research.Keywords: antibacterial, bioglass, drug delivery system, sol- gel
Procedia PDF Downloads 58409 Powder Assisted Sheet Forming to Fabricate Ti Capsule Magnetic Hyperthermia Implant
Authors: Keigo Nishitani, Kohei Mizuta Mizuta, Kazuyoshi Kurita, Yukinori Taniguchi
Abstract:
To establish mass production process of Ti capsule which has Fe powder inside as magnetic hyperthermia implant, we assumed that Ti thin sheet can be drawn into a φ1.0 mm die hole through the medium of Fe Powder and becomes outer shell of capsule. This study discusses mechanism of powder assisted deep drawing process by both of numerical simulation and experiment. Ti thin sheet blank was placed on die, and was covered by Fe powder layer without pressurizing. Then upper punch was indented on the Fe powder layer, and the blank can be drawn into die cavity as pressurized powder particles were extruded into die cavity from behind of the drawn blank. Distinct Element Method (DEM) has been used to demonstrate the process. To identify bonding parameters on Fe particles which are cohesion, tensile bond stress and inter particle friction angle, axial and diametrical compression failure test of Fe powder compact was conducted. Several density ratios of powder compacts in range of 0.70 - 0.85 were investigated and relationship between mean stress and equivalent stress was calculated with consideration of critical state line which rules failure criterion in consolidation of Fe powder. Since variation of bonding parameters with density ratio has been experimentally identified, and good agreement has been recognized between several failure tests and its simulation, demonstration of powder assisted sheet forming by using DEM becomes applicable. Results of simulation indicated that indent/drawing length of Ti thin sheet is promoted by smaller Fe particle size, larger indent punch diameter, lower friction coefficient between die surface and Ti sheet and certain degrees of die inlet taper angle. In the deep drawing test, we have made die-set with φ2.4 mm punch and φ1.0 mm die bore diameter. Pure Ti sheet with 100 μm thickness, annealed at 650 deg. C has been tested. After indentation, indented/drawn capsule has been observed by microscope, and its length was measured to discuss the feasibility of this capsulation process. Longer drawing length exists on progressive loading pass comparing with the case of single stroke loading. It is expected that progressive loading has an advantage of which extrusion of powder particle into die cavity with Ti sheet is promoted since powder particle layer can be rebuilt while the punch is withdrawn from the layer in each loading steps. This capsulation phenomenon is qualitatively demonstrated by DEM simulation. Finally, we have fabricated Ti capsule which has Fe powder inside for magnetic hyperthermia cancer care treatment. It is concluded that suggested method is possible to use the manufacturing of Ti capsule implant for magnetic hyperthermia cancer care.Keywords: metal powder compaction, metal forming, distinct element method, cancer care, magnetic hyperthermia
Procedia PDF Downloads 296408 Problem Solving in Mathematics Education: A Case Study of Nigerian Secondary School Mathematics Teachers’ Conceptions in Relation to Classroom Instruction
Authors: Carol Okigbo
Abstract:
Mathematical problem solving has long been accorded an important place in mathematics curricula at every education level in both advanced and emerging economies. Its classroom approaches have varied, such as teaching for problem-solving, teaching about problem-solving, and teaching mathematics through problem-solving. It requires engaging in tasks for which the solution methods are not eminent, making sense of problems and persevering in solving them by exhibiting processes, strategies, appropriate attitude, and adequate exposure. Teachers play important roles in helping students acquire competency in problem-solving; thus, they are expected to be good problem-solvers and have proper conceptions of problem-solving. Studies show that teachers’ conceptions influence their decisions about what to teach and how to teach. Therefore, how teachers view their roles in teaching problem-solving will depend on their pedagogical conceptions of problem-solving. If teaching problem-solving is a major component of secondary school mathematics instruction, as recommended by researchers and mathematics educators, then it is necessary to establish teachers’ conceptions, what they do, and how they approach problem-solving. This study is designed to determine secondary school teachers’ conceptions regarding mathematical problem solving, its current situation, how teachers’ conceptions relate to their demographics, as well as the interaction patterns in the mathematics classroom. There have been many studies of mathematics problem solving, some of which addressed teachers’ conceptions using single-method approaches, thereby presenting only limited views of this important phenomenon. To address the problem more holistically, this study adopted an integrated mixed methods approach which involved a quantitative survey, qualitative analysis of open-ended responses, and ethnographic observations of teachers in class. Data for the analysis came from a random sample of 327 secondary school mathematics teachers in two Nigerian states - Anambra State and Enugu State who completed a 45-item questionnaire. Ten of the items elicited demographic information, 11 items were open-ended questions, and 25 items were Likert-type questions. Of the 327 teachers who responded to the questionnaires, 37 were randomly selected and observed in their classes. Data analysis using ANOVA, t-tests, chi-square tests, and open coding showed that the teachers had different conceptions about problem-solving, which fall into three main themes: practice on exercises and word application problems, a process of solving mathematical problems, and a way of teaching mathematics. Teachers reported that no period is set aside for problem-solving; typically, teachers solve problems on the board, teach problem-solving strategies, and allow students time to struggle with problems on their own. The result shows a significant difference between male and female teachers’ conception of problems solving, a significant relationship among teachers’ conceptions and academic qualifications, and teachers who have spent ten years or more teaching mathematics were significantly different from the group with seven to nine years of experience in terms of their conceptions of problem-solving.Keywords: conceptions, education, mathematics, problem solving, teacher
Procedia PDF Downloads 75407 p-Type Multilayer MoS₂ Enabled by Plasma Doping for Ultraviolet Photodetectors Application
Authors: Xiao-Mei Zhang, Sian-Hong Tseng, Ming-Yen Lu
Abstract:
Two-dimensional (2D) transition metal dichalcogenides (TMDCs), such as MoS₂, have attracted considerable attention owing to the unique optical and electronic properties related to its 2D ultrathin atomic layer structure. MoS₂ is becoming prevalent in post-silicon digital electronics and in highly efficient optoelectronics due to its extremely low thickness and its tunable band gap (Eg = 1-2 eV). For low-power, high-performance complementary logic applications, both p- and n-type MoS₂ FETs (NFETs and PFETs) must be developed. NFETs with an electron accumulation channel can be obtained using unintentionally doped n-type MoS₂. However, the fabrication of MoS₂ FETs with complementary p-type characteristics is challenging due to the significant difficulty of injecting holes into its inversion channel. Plasma treatments with different species (including CF₄, SF₆, O₂, and CHF₃) have also been found to achieve the desired property modifications of MoS₂. In this work, we demonstrated a p-type multilayer MoS₂ enabled by selective-area doping using CHF₃ plasma treatment. Compared with single layer MoS₂, multilayer MoS₂ can carry a higher drive current due to its lower bandgap and multiple conduction channels. Moreover, it has three times the density of states at its minimum conduction band. Large-area growth of MoS₂ films on 300 nm thick SiO₂/Si substrate is carried out by thermal decomposition of ammonium tetrathiomolybdate, (NH₄)₂MoS₄, in a tube furnace. A two-step annealing process is conducted to synthesize MoS₂ films. For the first step, the temperature is set to 280 °C for 30 min in an N₂ rich environment at 1.8 Torr. This is done to transform (NH₄)₂MoS₄ into MoS₃. To further reduce MoS₃ into MoS₂, the second step of annealing is performed. For the second step, the temperature is set to 750 °C for 30 min in a reducing atmosphere consisting of 90% Ar and 10% H₂ at 1.8 Torr. The grown MoS₂ films are subjected to out-of-plane doping by CHF₃ plasma treatment using a Dry-etching system (ULVAC original NLD-570). The radiofrequency power of this dry-etching system is set to 100 W and the pressure is set to 7.5 mTorr. The final thickness of the treated samples is obtained by etching for 30 s. Back-gated MoS₂ PFETs were presented with an on/off current ratio in the order of 10³ and a field-effect mobility of 65.2 cm²V⁻¹s⁻¹. The MoS₂ PFETs photodetector exhibited ultraviolet (UV) photodetection capability with a rapid response time of 37 ms and exhibited modulation of the generated photocurrent by back-gate voltage. This work suggests the potential application of the mild plasma-doped p-type multilayer MoS₂ in UV photodetectors for environmental monitoring, human health monitoring, and biological analysis.Keywords: photodetection, p-type doping, multilayers, MoS₂
Procedia PDF Downloads 103