Search results for: ecological planning method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 22188

Search results for: ecological planning method

12858 Analysis of Post-vaccination Immunity in Children with Severe Chronic Diseases Receiving Immunosuppressive Therapy by Specific IgG Antibodies Definition Method

Authors: Marina G. Galitskaya, Svetlana G. Makarova, Andrey P. Fisenko.

Abstract:

Children on medication-induced immunosuppression are at high risk of developing severe course infectious diseases. Therefore, preventive vaccination is especially important for these children. However, due to the immunosuppressive effects of treatment for the underlying disease, the effectiveness of vaccination may decrease below the protective level. In a multidisciplinary children's medical center, post-vaccination immunity was studied in 79 children aged 4-17 years. The children were divided into 2 groups: Group 1 (38 children) with kidney pathology (Nephrotic Syndrome) and Group 2 (41 children) with inflammatory bowel diseases (Ulcerative Colitis, Crohn's Disease). Both groups of children were vaccinated according to the national vaccination calendar and received immunosuppressive therapy (prednisolone, methotrexate, cyclosporine, and other drugs) for at least 1 year. Using the enzyme-linked immunosorbent assay method, specific IgG antibodies to vaccine-preventable infections were determined: measles, rubella, mumps, diphtheria, pertussis, tetanus, and hepatitis B. The study showed the percentage of children with positive IgG values for vaccine-preventable infections. The highest percentage of children had protective antibody levels to measles (84.2% in children with nephrotic syndrome and 92.6% in those with inflammatory bowel disease) and rubella (71% and 80.4%, respectively). The lowest percentage of children with protective antibodies was for hepatitis B (5.2% and 29.2% respectively). Antibodies to mumps, diphtheria, pertussis, and tetanus were found not in all children (from 39,4% to 82,9%). The remaining percentage of children did not have detectable IgG antibodies to vaccine-preventable infections. Not all children, despite the previous vaccination, preserved antibodies to vaccine-controlled infections and remained unprotected by specific IgG antibodies. The issue of a booster vaccine dose should be considered in children without contraindications to vaccination. Children receiving long-term immunosuppressive therapy require an individual vaccination approach, including a specific definition of the performed vaccination.

Keywords: immunosuppressive therapy, inflammatory bowel diseases, nephrotic syndrome, post-vaccination immunity, specific antibodies, vaccine-preventable infections.

Procedia PDF Downloads 19
12857 Synthesis and Characterization of Cyclic PNC-28 Peptide, Residues 17–26 (ETFSDLWKLL), A Binding Domain of p53

Authors: Deepshikha Verma, V. N. Rajasekharan Pillai

Abstract:

The present study reports the synthesis of cyclic PNC-28 peptides with solid-phase peptide synthesis method. In the first step, we synthesize the linear PNC-28 Peptide and in the second step, we cyclize (N-to-C or head-to-tail cyclization) the linear PNC-28 peptide. The molecular formula of cyclic PNC-28 peptide is C64H88N12O16 and its m/z mass is ≈1233.64. Elemental analysis of cyclic PNC-28 is C, 59.99; H, 6.92; N, 13.12; O, 19.98. The characterization of LC-MS, CD, FT-IR, and 1HNMR has been done to confirm the successful synthesis and cyclization of linear PNC-28 peptides.

Keywords: CD, FTIR, 1HNMR, cyclic peptide

Procedia PDF Downloads 118
12856 Data Confidentiality in Public Cloud: A Method for Inclusion of ID-PKC Schemes in OpenStack Cloud

Authors: N. Nalini, Bhanu Prakash Gopularam

Abstract:

The term data security refers to the degree of resistance or protection given to information from unintended or unauthorized access. The core principles of information security are the confidentiality, integrity and availability, also referred as CIA triad. Cloud computing services are classified as SaaS, IaaS and PaaS services. With cloud adoption the confidential enterprise data are moved from organization premises to untrusted public network and due to this the attack surface has increased manifold. Several cloud computing platforms like OpenStack, Eucalyptus, Amazon EC2 offer users to build and configure public, hybrid and private clouds. While the traditional encryption based on PKI infrastructure still works in cloud scenario, the management of public-private keys and trust certificates is difficult. The Identity based Public Key Cryptography (also referred as ID-PKC) overcomes this problem by using publicly identifiable information for generating the keys and works well with decentralized systems. The users can exchange information securely without having to manage any trust information. Another advantage is that access control (role based access control policy) information can be embedded into data unlike in PKI where it is handled by separate component or system. In OpenStack cloud platform the keystone service acts as identity service for authentication and authorization and has support for public key infrastructure for auto services. In this paper, we explain OpenStack security architecture and evaluate the PKI infrastructure piece for data confidentiality. We provide method to integrate ID-PKC schemes for securing data while in transit and stored and explain the key measures for safe guarding data against security attacks. The proposed approach uses JPBC crypto library for key-pair generation based on IEEE P1636.3 standard and secure communication to other cloud services.

Keywords: data confidentiality, identity based cryptography, secure communication, open stack key stone, token scoping

Procedia PDF Downloads 365
12855 Minimizing Unscheduled Maintenance from an Aircraft and Rolling Stock Maintenance Perspective: Preventive Maintenance Model

Authors: Adel A. Ghobbar, Varun Raman

Abstract:

The Corrective maintenance of components and systems is a problem plaguing almost every industry in the world today. Train operators’ and the maintenance repair and overhaul subsidiary of the Dutch railway company is also facing this problem. A considerable portion of the maintenance activities carried out by the company are unscheduled. This, in turn, severely stresses and stretches the workforce and resources available. One possible solution is to have a robust preventive maintenance plan. The other possible solution is to plan maintenance based on real-time data obtained from sensor-based ‘Health and Usage Monitoring Systems.’ The former has been investigated in this paper. The preventive maintenance model developed for train operator will subsequently be extended, to tackle the unscheduled maintenance problem also affecting the aerospace industry. The extension of the model to the aerospace sector will be dealt with in the second part of the research, and it would, in turn, validate the soundness of the model developed. Thus, there are distinct areas that will be addressed in this paper, including the mathematical modelling of preventive maintenance and optimization based on cost and system availability. The results of this research will help an organization to choose the right maintenance strategy, allowing it to save considerable sums of money as opposed to overspending under the guise of maintaining high asset availability. The concept of delay time modelling was used to address the practical problem of unscheduled maintenance in this paper. The delay time modelling can be used to help with support planning for a given asset. The model was run using MATLAB, and the results are shown that the ideal inspection intervals computed using the extended from a minimal cost perspective were 29 days, and from a minimum downtime, perspective was 14 days. Risk matrix integration was constructed to represent the risk in terms of the probability of a fault leading to breakdown maintenance and its consequences in terms of maintenance cost. Thus, the choice of an optimal inspection interval of 29 days, resulted in a cost of approximately 50 Euros and the corresponding value of b(T) was 0.011. These values ensure that the risk associated with component X being maintained at an inspection interval of 29 days is more than acceptable. Thus, a switch in maintenance frequency from 90 days to 29 days would be optimal from the point of view of cost, downtime and risk.

Keywords: delay time modelling, unscheduled maintenance, reliability, maintainability, availability

Procedia PDF Downloads 122
12854 A Novel Approach of Secret Communication Using Douglas-Peucker Algorithm

Authors: R. Kiruthika, A. Kannan

Abstract:

Steganography is the problem of hiding secret messages in 'innocent – looking' public communication so that the presence of the secret message cannot be detected. This paper introduces a steganographic security in terms of computational in-distinguishability from a channel of probability distributions on cover messages. This method first splits the cover image into two separate blocks using Douglas – Peucker algorithm. The text message and the image will be hided in the Least Significant Bit (LSB) of the cover image.

Keywords: steganography, lsb, embedding, Douglas-Peucker algorithm

Procedia PDF Downloads 349
12853 Development of a Microfluidic Device for Low-Volume Sample Lysis

Authors: Abbas Ali Husseini, Ali Mohammad Yazdani, Fatemeh Ghadiri, Alper Şişman

Abstract:

We developed a microchip device that uses surface acoustic waves for rapid lysis of low level of cell samples. The device incorporates sharp-edge glass microparticles for improved performance. We optimized the lysis conditions for high efficiency and evaluated the device's feasibility for point-of-care applications. The microchip contains a 13-finger pair interdigital transducer with a 30-degree focused angle. It generates high-intensity acoustic beams that converge 6 mm away. The microchip operates at a frequency of 16 MHz, exciting Rayleigh waves with a 250 µm wavelength on the LiNbO3 substrate. Cell lysis occurs when Candida albicans cells and glass particles are placed within the focal area. The high-intensity surface acoustic waves induce centrifugal forces on the cells and glass particles, resulting in cell lysis through lateral forces from the sharp-edge glass particles. We conducted 42 pilot cell lysis experiments to optimize the surface acoustic wave-induced streaming. We varied electrical power, droplet volume, glass particle size, concentration, and lysis time. A regression machine-learning model determined the impact of each parameter on lysis efficiency. Based on these findings, we predicted optimal conditions: electrical signal of 2.5 W, sample volume of 20 µl, glass particle size below 10 µm, concentration of 0.2 µg, and a 5-minute lysis period. Downstream analysis successfully amplified a DNA target fragment directly from the lysate. The study presents an efficient microchip-based cell lysis method employing acoustic streaming and microparticle collisions within microdroplets. Integration of a surface acoustic wave-based lysis chip with an isothermal amplification method enables swift point-of-care applications.

Keywords: cell lysis, surface acoustic wave, micro-glass particle, droplet

Procedia PDF Downloads 65
12852 An Approach to Wind Turbine Modeling for Increasing Its Efficiency

Authors: Rishikesh Dingari, Sai Kiran Dornala

Abstract:

In this paper, a simple method of achieving maximum power by mechanical energy transmission device (METD) with integration to induction generator is proposed. METD functioning is explained and dynamic response of system to step input is plotted. Induction generator is being operated at self-excited mode with excitation capacitor at stator. Voltage and current are observed when linked to METD.

Keywords: mechanical energy transmitting device(METD), self-excited induction generator, wind turbine, hydraulic actuators

Procedia PDF Downloads 334
12851 Optical Flow Technique for Supersonic Jet Measurements

Authors: Haoxiang Desmond Lim, Jie Wu, Tze How Daniel New, Shengxian Shi

Abstract:

This paper outlines the development of a novel experimental technique in quantifying supersonic jet flows, in an attempt to avoid seeding particle problems frequently associated with particle-image velocimetry (PIV) techniques at high Mach numbers. Based on optical flow algorithms, the idea behind the technique involves using high speed cameras to capture Schlieren images of the supersonic jet shear layers, before they are subjected to an adapted optical flow algorithm based on the Horn-Schnuck method to determine the associated flow fields. The proposed method is capable of offering full-field unsteady flow information with potentially higher accuracy and resolution than existing point-measurements or PIV techniques. Preliminary study via numerical simulations of a circular de Laval jet nozzle successfully reveals flow and shock structures typically associated with supersonic jet flows, which serve as useful data for subsequent validation of the optical flow based experimental results. For experimental technique, a Z-type Schlieren setup is proposed with supersonic jet operated in cold mode, stagnation pressure of 8.2 bar and exit velocity of Mach 1.5. High-speed single-frame or double-frame cameras are used to capture successive Schlieren images. As implementation of optical flow technique to supersonic flows remains rare, the current focus revolves around methodology validation through synthetic images. The results of validation test offers valuable insight into how the optical flow algorithm can be further improved to improve robustness and accuracy. Details of the methodology employed and challenges faced will be further elaborated in the final conference paper should the abstract be accepted. Despite these challenges however, this novel supersonic flow measurement technique may potentially offer a simpler way to identify and quantify the fine spatial structures within the shock shear layer.

Keywords: Schlieren, optical flow, supersonic jets, shock shear layer

Procedia PDF Downloads 304
12850 Low Carbon Tourism Management: Strategies for Climate-Friendly Tourism of Koh Mak, Thailand

Authors: Panwad Wongthong, Thanan Apivantanaporn, Sutthiwan Amattayakul

Abstract:

Nature-based tourism is one of the fastest growing industries that can bring in economic benefits, improve quality of life and promote conservation of biodiversity and habitats. As tourism develops, substantial socio-economic and environmental costs become more explicit. Particularly in island destinations, the dynamic system and geographical limitations makes the intensity of tourism development and severity of the negative environmental impacts greater. The current contribution of the tourism sector to global climate change is established at approximately 5% of global anthropogenic CO2 emissions. In all scenarios, tourism is anticipated to grow substantially and to account for an increasingly large share of global greenhouse gas emissions. This has prompted an urgent call for more sustainable alternatives. This study selected a small island of Koh Mak in Thailand as a case study because of its reputation of being laid back, family oriented and rich in biodiversity. Importantly, it is a test platform for low carbon tourism development project supported by the Designated Areas for Sustainable Tourism Administration (DASTA) in collaboration with the Institute for Small and Medium Enterprises Development (ISMED). The study explores strategies for low carbon tourism management and assesses challenges and opportunities for Koh Mak to become a low carbon tourism destination. The goal is to identify suitable management approaches applicable for Koh Mak which may then be adapted to other small islands in Thailand and the region. Interventions/initiatives to increase energy efficiency in hotels and resorts; cut carbon emissions; reduce impacts on the environment; and promote conservation will be analyzed. Ways toward long-term sustainability of climate-friendly tourism will be recommended. Recognizing the importance of multi-stakeholder involvement in the tourism sector, findings from this study can reward Koh Mak tourism industry with a triple-win: cost savings and compliance with higher standards/markets; less waste, air emissions and effluents; and better capabilities of change, motivation of business owners, staff, tourists as well as residents. The consideration of climate change issues in the planning and implementation of tourism development is of great significance to protect the tourism sector from negative impacts.

Keywords: climate change, CO2 emissions, low carbon tourism, sustainable tourism management

Procedia PDF Downloads 271
12849 Modeling the Impact of Aquaculture in Wetland Ecosystems Using an Integrated Ecosystem Approach: Case Study of Setiu Wetlands, Malaysia

Authors: Roseliza Mat Alipiah, David Raffaelli, J. C. R. Smart

Abstract:

This research is a new approach as it integrates information from both environmental and social sciences to inform effective management of the wetlands. A three-stage research framework was developed for modelling the drivers and pressures imposed on the wetlands and their impacts to the ecosystem and the local communities. Firstly, a Bayesian Belief Network (BBN) was used to predict the probability of anthropogenic activities affecting the delivery of different key wetland ecosystem services under different management scenarios. Secondly, Choice Experiments (CEs) were used to quantify the relative preferences which key wetland stakeholder group (aquaculturists) held for delivery of different levels of these key ecosystem services. Thirdly, a Multi-Criteria Decision Analysis (MCDA) was applied to produce an ordinal ranking of the alternative management scenarios accounting for their impacts upon ecosystem service delivery as perceived through the preferences of the aquaculturists. This integrated ecosystem management approach was applied to a wetland ecosystem in Setiu, Terengganu, Malaysia which currently supports a significant level of aquaculture activities. This research has produced clear guidelines to inform policy makers considering alternative wetland management scenarios: Intensive Aquaculture, Conservation or Ecotourism, in addition to the Status Quo. The findings of this research are as follows: The BBN revealed that current aquaculture activity is likely to have significant impacts on water column nutrient enrichment, but trivial impacts on caged fish biomass, especially under the Intensive Aquaculture scenario. Secondly, the best fitting CE models identified several stakeholder sub-groups for aquaculturists, each with distinct sets of preferences for the delivery of key ecosystem services. Thirdly, the MCDA identified Conservation as the most desirable scenario overall based on ordinal ranking in the eyes of most of the stakeholder sub-groups. Ecotourism and Status Quo scenarios were the next most preferred and Intensive Aquaculture was the least desirable scenario. The methodologies developed through this research provide an opportunity for improving planning and decision making processes that aim to deliver sustainable management of wetland ecosystems in Malaysia.

Keywords: Bayesian belief network (BBN), choice experiments (CE), multi-criteria decision analysis (MCDA), aquaculture

Procedia PDF Downloads 275
12848 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks

Authors: Wang Yichen, Haruka Yamashita

Abstract:

In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.

Keywords: recurrent neural network, players lineup, basketball data, decision making model

Procedia PDF Downloads 117
12847 Highway Waste Management in Zambia Policy Preparedness and Remedies: The Case of Great East Road

Authors: Floyd Misheck Mwanza, Paul Boniface Majura

Abstract:

The paper looked at highways/ roadside waste generation, disposal and the consequent environmental impacts. The dramatic increase in vehicular and paved roads in the recent past in Zambia, has given rise to the indiscriminate disposal of litter that now poses a threat to health and the environment. Primary data was generated by carrying out oral interviews and field observations for holistic and in–depth assessment of the environment and the secondary data was obtained from desk review method, information on effects of roadside wastes on environment were obtained from relevant literatures. The interviews were semi structured and a purposive sampling method was adopted and analyzed descriptively. The results of the findings showed that population growth and unplanned road expansion has exceeded the expected limit in recent time with resultant poor system of roadside wastes disposal. Roadside wastes which contain both biodegradable and non-biodegradable roadside wastes are disposed at the shoulders of major highways in temporary dumpsites and are never collected by a road development agency (RDA). There is no organized highway to highway or street to street collection of the wastes in Zambia by the key organization the RDA. The study revealed that roadside disposal of roadside wastes has serious impacts on the environment. Some of these impacts include physical nuisance of the wastes to the environment, the waste dumps also serve as hideouts for rodents and snakes which are dangerous. Waste are blown around by wind making the environment filthy, most of the wastes are also been washed by overland flow during heavy downpour to block drainage channels and subsequently lead to flooding of the environment. Most of the non- biodegradable wastes contain toxic chemicals which have serious implications on the environmental sustainability and human health. The paper therefore recommends that Government/ RDA should come up with proper orientation and environmental laws should be put in place for the general public and also to provide necessary facilities and arrange for better methods of collection of wastes.

Keywords: biodegradable, disposal, environment, impacts

Procedia PDF Downloads 326
12846 Heavy Metal Contamination of Mining-Impacted Mangrove Sediments and Its Correlation with Vegetation and Sediment Attributes

Authors: Jumel Christian P. Nicha, Severino G. Salmo III

Abstract:

This study investigated the concentration of heavy metals (HM) in mangrove sediments of Lake Uacon, Zambales, Philippines. The relationship among the studied HM (Cr, Ni, Pb, Cu, Cd, Fe) and the mangrove vegetation and sediment characteristics were assessed. Fourteen sampling plots were designated across the lake (10 vegetated and 4 un-vegetated) based on distance from the mining effluents. In each plot, three sediment cores were collected at 20 cm depth. Among the dominant mangrove species recorded were (in order of dominance): Sonneratia alba, Rhizophora stylosa, Avicennia marina, Excoecaria agallocha and Bruguiera gymnorrhiza. Sediment samples were digested with aqua regia, and the HM concentrations were quantified using Atomic Absorption Spectroscopy (AAS). Results showed that HM concentrations were higher in the vegetated plots as compared to the un-vegetated sites. Vegetated sites had high Ni (mean: 881.71 mg/kg) and Cr (mean: 776.36 mg/kg) that exceeded the threshold values (cf. by the United States Environmental Protection Agency; USEPA). Fe, Pb, Cu and Cd had a mean concentration of 2597.92 mg/kg, 40.94 mg/kg, 36.81 mg/kg and 2.22 mg/kg respectively. Vegetation variables were not significantly correlated with HM concentration. However, the HM concentration was significantly correlated with sediment variables particularly pH, redox, particle size, nitrogen, phosphorus, moisture and organic matter contents. The Pollution Load Index (PLI) indicated moderate to high pollution in the lake. Risk assessment and management should be designed in order to mitigate the ecological risk posed by HM. The need of a regular monitoring scheme for lake and mangrove rehabilitation programs and management should be designed.

Keywords: heavy metals, mangrove vegetation, mining, Philippines, sediment

Procedia PDF Downloads 151
12845 Gene Expressions in Left Ventricle Heart Tissue of Rat after 150 Mev Proton Irradiation

Authors: R. Fardid, R. Coppes

Abstract:

Introduction: In mediastinal radiotherapy and to a lesser extend also in total-body irradiation (TBI) radiation exposure may lead to development of cardiac diseases. Radiation-induced heart disease is dose-dependent and it is characterized by a loss of cardiac function, associated with progressive heart cells degeneration. We aimed to determine the in-vivo radiation effects on fibronectin, ColaA1, ColaA2, galectin and TGFb1 gene expression levels in left ventricle heart tissues of rats after irradiation. Material and method: Four non-treatment adult Wistar rats as control group (group A) were selected. In group B, 4 adult Wistar rats irradiated to 20 Gy single dose of 150 Mev proton beam locally in heart only. In heart plus lung irradiate group (group C) 4 adult rats was irradiated by 50% of lung laterally plus heart radiation that mentioned in before group. At 8 weeks after radiation animals sacrificed and left ventricle heart dropped in liquid nitrogen for RNA extraction by Absolutely RNA® Miniprep Kit (Stratagen, Cat no. 400800). cDNA was synthesized using M-MLV reverse transcriptase (Life Technologies, Cat no. 28025-013). We used Bio-Rad machine (Bio Rad iQ5 Real Time PCR) for QPCR testing by relative standard curve method. Results: We found that gene expression of fibronectin in group C significantly increased compared to control group, but it was not showed significant change in group B compared to group A. The levels of gene expressions of Cola1 and Cola2 in mRNA did not show any significant changes between normal and radiation groups. Changes of expression of galectin target significantly increased only in group C compared to group A. TGFb1 expressions in group C more than group B showed significant enhancement compared to group A. Conclusion: In summary we can say that 20 Gy of proton exposure of heart tissue may lead to detectable damages in heart cells and may distribute function of them as a component of heart tissue structure in molecular level.

Keywords: gene expression, heart damage, proton irradiation, radiotherapy

Procedia PDF Downloads 479
12844 A Simple Chemical Precipitation Method of Titanium Dioxide Nanoparticles Using Polyvinyl Pyrrolidone as a Capping Agent and Their Characterization

Authors: V. P. Muhamed Shajudheen, K. Viswanathan, K. Anitha Rani, A. Uma Maheswari, S. Saravana Kumar

Abstract:

In this paper, a simple chemical precipitation route for the preparation of titanium dioxide nanoparticles, synthesized by using titanium tetra isopropoxide as a precursor and polyvinyl pyrrolidone (PVP) as a capping agent, is reported. The Differential Scanning Calorimetry (DSC) and Thermo Gravimetric Analysis (TGA) of the samples were recorded and the phase transformation temperature of titanium hydroxide, Ti(OH)4 to titanium oxide, TiO2 was investigated. The as-prepared Ti(OH)4 precipitate was annealed at 800°C to obtain TiO2 nanoparticles. The thermal, structural, morphological and textural characterizations of the TiO2 nanoparticle samples were carried out by different techniques such as DSC-TGA, X-Ray Diffraction (XRD), Fourier Transform Infra-Red spectroscopy (FTIR), Micro Raman spectroscopy, UV-Visible absorption spectroscopy (UV-Vis), Photoluminescence spectroscopy (PL) and Field Effect Scanning Electron Microscopy (FESEM) techniques. The as-prepared precipitate was characterized using DSC-TGA and confirmed the mass loss of around 30%. XRD results exhibited no diffraction peaks attributable to anatase phase, for the reaction products, after the solvent removal. The results indicate that the product is purely rutile. The vibrational frequencies of two main absorption bands of prepared samples are discussed from the results of the FTIR analysis. The formation of nanosphere of diameter of the order of 10 nm, has been confirmed by FESEM. The optical band gap was found by using UV-Visible spectrum. From photoluminescence spectra, a strong emission was observed. The obtained results suggest that this method provides a simple, efficient and versatile technique for preparing TiO2 nanoparticles and it has the potential to be applied to other systems for photocatalytic activity.

Keywords: TiO2 nanoparticles, chemical precipitation route, phase transition, Fourier Transform Infra-Red spectroscopy (FTIR), micro-Raman spectroscopy, UV-Visible absorption spectroscopy (UV-Vis), Photoluminescence Spectroscopy (PL) and Field Effect Scanning electron microscopy (FESEM)

Procedia PDF Downloads 308
12843 Nuclear Mitochondrial Pseudogenes in Anastrepha fraterculus Complex

Authors: Pratibha Srivastava, Ayyamperumal Jeyaprakash, Gary Steck, Jason Stanley, Leroy Whilby

Abstract:

Exotic, invasive tephritid fruit flies (Diptera: Tephritidae) are a major threat to fruit and vegetable industries in the United States. The establishment of pest fruit fly in the agricultural industries and produce severe ecological and economic impacts on agricultural diversification and trade. Detection and identification of these agricultural pests in a timely manner will facilitate the possibility of eradication from newly invaded areas. Identification of larval stages to species level is difficult, but is required to determine pest loads and their pathways into the United States. The aim of this study is the New World genus, Anastrepha which includes pests of major economic importance. Mitochondrial cytochrome c oxidase I (COI) gene sequences were amplified from Anastrepha fraterculus specimens collected from South America (Ecuador and Peru). Phylogenetic analysis was performed to characterize the Anastrepha fraterculus complex at a molecular level. During phylogenetics analysis numerous nuclear mitochondrial pseudogenes (numts) were discovered in different specimens. The numts are nonfunctional copies of the mtDNA present in the nucleus and are easily coamplified with the mitochondrial COI gene copy by using conserved universal primers. This is problematic for DNA Barcoding, which attempts to characterize all living organisms by using the COI gene. This study is significant for national quarantine use, as morphological diagnostics to separate larvae of the various members remain poorly developed.

Keywords: tephritid, Anastrepha fraterculus, COI, numts

Procedia PDF Downloads 221
12842 Diagnosis of Choledocholithiasis with Endosonography

Authors: A. Kachmazova, A. Shadiev, Y. Teterin, P. Yartcev

Abstract:

Introduction: Biliary calculi disease (LCS) still occupies the leading position among urgent diseases of the abdominal cavity, manifesting itself from asymptomatic course to life-threatening states. Nowadays arsenal of diagnostic methods for choledocholithiasis is quite wide: ultrasound, hepatobiliscintigraphy (HBSG), magnetic resonance imaging (MRI), endoscopic retrograde cholangiography (ERCP). Among them, transabdominal ultrasound (TA ultrasound) is the most accessible and routine diagnostic method. Nowadays ERCG is the "gold" standard in diagnosis and one-stage treatment of biliary tract obstruction. However, transpapillary techniques are accompanied by serious postoperative complications (postmanipulative pancreatitis (3-5%), endoscopic papillosphincterotomy bleeding (2%), cholangitis (1%)), the lethality being 0.4%. GBSG and MRI are also quite informative methods in the diagnosis of choledocholithiasis. Small size of concrements, their localization in intrapancreatic and retroduodenal part of common bile duct significantly reduces informativity of all diagnostic methods described above, that demands additional studying of this problem. Materials and Methods: 890 patients with the diagnosis of cholelithiasis (calculous cholecystitis) were admitted to the Sklifosovsky Scientific Research Institute of Hospital Medicine in the period from August, 2020 to June, 2021. Of them 115 people with mechanical jaundice caused by concrements in bile ducts. Results: Final EUS diagnosis was made in all patients (100,0%). In all patients in whom choledocholithiasis diagnosis was revealed or confirmed after EUS, ERCP was performed urgently (within two days from the moment of its detection) as the X-ray operation room was provided; it confirmed the presence of concrements. All stones were removed by lithoextraction using Dormia basket. The postoperative period in these patients had no complications. Conclusions: EUS is the most informative and safe diagnostic method, which allows to detect choledocholithiasis in patients with discrepancies between clinical-laboratory and instrumental methods of diagnosis in shortest time, that in its turn will help to decide promptly on the further tactics of patient treatment. We consider it reasonable to include EUS in the diagnostic algorithm for choledocholithiasis. Disclosure: Nothing to disclose.

Keywords: endoscopic ultrasonography, choledocholithiasis, common bile duct, concrement, ERCP

Procedia PDF Downloads 73
12841 Digital Structural Monitoring Tools @ADaPT for Cracks Initiation and Growth due to Mechanical Damage Mechanism

Authors: Faizul Azly Abd Dzubir, Muhammad F. Othman

Abstract:

Conventional structural health monitoring approach for mechanical equipment uses inspection data from Non-Destructive Testing (NDT) during plant shut down window and fitness for service evaluation to estimate the integrity of the equipment that is prone to crack damage. Yet, this forecast is fraught with uncertainty because it is often based on assumptions of future operational parameters, and the prediction is not continuous or online. Advanced Diagnostic and Prognostic Technology (ADaPT) uses Acoustic Emission (AE) technology and a stochastic prognostic model to provide real-time monitoring and prediction of mechanical defects or cracks. The forecast can help the plant authority handle their cracked equipment before it ruptures, causing an unscheduled shutdown of the facility. The ADaPT employs process historical data trending, finite element analysis, fitness for service, and probabilistic statistical analysis to develop a prediction model for crack initiation and growth due to mechanical damage. The prediction model is combined with live equipment operating data for real-time prediction of the remaining life span owing to fracture. ADaPT was devised at a hot combined feed exchanger (HCFE) that had suffered creep crack damage. The ADaPT tool predicts the initiation of a crack at the top weldment area by April 2019. During the shutdown window in April 2019, a crack was discovered and repaired. Furthermore, ADaPT successfully advised the plant owner to run at full capacity and improve output by up to 7% by April 2019. ADaPT was also used on a coke drum that had extensive fatigue cracking. The initial cracks are declared safe with ADaPT, with remaining crack lifetimes extended another five (5) months, just in time for another planned facility downtime to execute repair. The prediction model, when combined with plant information data, allows plant operators to continuously monitor crack propagation caused by mechanical damage for improved maintenance planning and to avoid costly shutdowns to repair immediately.

Keywords: mechanical damage, cracks, continuous monitoring tool, remaining life, acoustic emission, prognostic model

Procedia PDF Downloads 63
12840 Using a Mobile App to Foster Children Active Travel to School in Spain

Authors: P. Pérez-Martín, G. Pedrós, P. Martínez-Jiménez, M. Varo-Martínez

Abstract:

In recent decades, family habits related to children’s displacements to school have changed, increasing motorized travels against active modes. This entails a major negative impact on the urban environment, road safety in cities and the physical and psychological development of children. One of the more common actions used to reverse this trend is Walking School Bus (WSB), which consists of a predefined adult-scorted pedestrian route to school with several stops along the path where schoolchildren are collected. At Tirso de Molina School in Cordoba (Spain), a new ICT-based methodology to deploy WSB has been tested. A mobile app that allows the geoposition of the group, the notification of the arrival and real-time communication between the WSB participants have been presented to the families in order to organize and register the daily participation. After an initial survey to know the travel mode and the spatial distribution of the interested families, three WSB routes have been established and the families have been trained in the app usage. During nine weeks, 33 children have joined the WSB and their parents have accompanied the groups in turns. A high recurrence in the attendance has been registered. Through a final survey, participants have valued highly the tool and the methodology designed, emphasizing as most useful features of the mobile app: notifications system, chat and real-time monitoring. It has also been found that the tool has had a major impact on the degree of confidence of parents regarding the autonomous on foot displacement of their children to school. Moreover, 37,9% of the participant families have reported a total or partial modal shift from car to walking, and the benefits more reported are an increment of the parents available time and less problems in the travel to school daily organization. As a consequence, It has been proved the effectiveness of this user-centric innovative ICT-based methodology to reduce the levels of private car drop offs, minimize barriers of time constraints, volunteer recruitment, and parents’ safety concerns, while, at the same time, increase convenience and time savings for families. This pilot study can offer guidance for community coordinated actions and local authority interventions to support sustainable school travel outcomes.

Keywords: active travel, mobile app, sustainable mobility, urban transportation planning, walking school bus

Procedia PDF Downloads 323
12839 The Effect of MOOC-Based Distance Education in Academic Engagement and Its Components on Kerman University Students

Authors: Fariba Dortaj, Reza Asadinejad, Akram Dortaj, Atena Baziyar

Abstract:

The aim of this study was to determine the effect of distance education (based on MOOC) on the components of academic engagement of Kerman PNU. The research was quasi-experimental method that cluster sampling with an appropriate volume was used in this study (one class in experimental group and one class in controlling group). Sampling method is single-stage cluster sampling. The statistical society is students of Kerman Payam Noor University, which) were selected 40 of them as sample (20 students in the control group and 20 students in experimental group). To test the hypothesis, it was used the analysis of univariate and Co-covariance to offset the initial difference (difference of control) in the experimental group and the control group. The instrument used in this study is academic engagement questionnaire of Zerang (2012) that contains component of cognitive, behavioral and motivational engagement. The results showed that there is no significant difference between mean scores of academic components of academic engagement in experimental group and the control group on the post-test, after elimination of the pre-test. The adjusted mean scores of components of academic engagement in the experimental group were higher than the adjusted average of scores after the test in the control group. The use of technology-based education in distance education has been effective in increasing cognitive engagement, motivational engagement and behavioral engagement among students. Experimental variable with the effect size 0.26, predicted 26% of cognitive engagement component variance. Experimental variable with the effect size 0.47, predicted 47% of the motivational engagement component variance. Experimental variable with the effect size 0.40, predicted 40% of behavioral engagement component variance. So teaching with technology (MOOC) has a positive impact on increasing academic engagement and academic performance of students in educational technology. The results suggest that technology (MOOC) is used to enrich the teaching of other lessons of PNU.

Keywords: educational technology, distance education, components of academic engagement, mooc technology

Procedia PDF Downloads 135
12838 Monitoring Land Cover/Land Use Change in Rupandehi District by Optimising Remotely Sensed Image

Authors: Hritik Bhattarai

Abstract:

Land use and land cover play a crucial role in preserving and managing Earth's natural resources. Various factors, such as economic, demographic, social, cultural, technological, and environmental processes, contribute to changes in land use and land cover (LULC). Rupandehi District is significantly influenced by a combination of driving forces, including its geographical location, rapid population growth, economic opportunities, globalization, tourism activities, and political events. Urbanization and urban growth in the region have been occurring in an unplanned manner, with internal migration and natural population growth being the primary contributors. Internal migration, particularly from neighboring districts in the higher and lower Himalayan regions, has been high, leading to increased population growth and density. This study utilizes geospatial technology, specifically geographic information system (GIS), to analyze and illustrate the land cover and land use changes in the Rupandehi district for the years 2009 and 2019, using freely available Landsat images. The identified land cover categories include built-up area, cropland, Das-Gaja, forest, grassland, other woodland, riverbed, and water. The statistical analysis of the data over the 10-year period (2009-2019) reveals significant percentage changes in LULC. Notably, Das-Gaja shows a minimal change of 99.9%, while water and forest exhibit increases of 34.5% and 98.6%, respectively. Riverbed and built-up areas experience changes of 95.3% and 39.6%, respectively. Cropland and grassland, however, show concerning decreases of 102.6% and 140.0%, respectively. Other woodland also indicates a change of 50.6%. The most noteworthy trends are the substantial increase in water areas and built-up areas, leading to the degradation of agricultural and open spaces. This emphasizes the urgent need for effective urban planning activities to ensure the development of a sustainable city. While Das-Gaja seems unaffected, the decreasing trends in cropland and grassland, accompanied by the increasing built-up areas, are unsatisfactory. It is imperative for relevant authorities to be aware of these trends and implement proactive measures for sustainable urban development.

Keywords: land use and land cover, geospatial, urbanization, geographic information system, sustainable urban development

Procedia PDF Downloads 43
12837 Wireless FPGA-Based Motion Controller Design by Implementing 3-Axis Linear Trajectory

Authors: Kiana Zeighami, Morteza Ozlati Moghadam

Abstract:

Designing a high accuracy and high precision motion controller is one of the important issues in today’s industry. There are effective solutions available in the industry but the real-time performance, smoothness and accuracy of the movement can be further improved. This paper discusses a complete solution to carry out the movement of three stepper motors in three dimensions. The objective is to provide a method to design a fully integrated System-on-Chip (SOC)-based motion controller to reduce the cost and complexity of production by incorporating Field Programmable Gate Array (FPGA) into the design. In the proposed method the FPGA receives its commands from a host computer via wireless internet communication and calculates the motion trajectory for three axes. A profile generator module is designed to realize the interpolation algorithm by translating the position data to the real-time pulses. This paper discusses an approach to implement the linear interpolation algorithm, since it is one of the fundamentals of robots’ movements and it is highly applicable in motion control industries. Along with full profile trajectory, the triangular drive is implemented to eliminate the existence of error at small distances. To integrate the parallelism and real-time performance of FPGA with the power of Central Processing Unit (CPU) in executing complex and sequential algorithms, the NIOS II soft-core processor was added into the design. This paper presents different operating modes such as absolute, relative positioning, reset and velocity modes to fulfill the user requirements. The proposed approach was evaluated by designing a custom-made FPGA board along with a mechanical structure. As a result, a precise and smooth movement of stepper motors was observed which proved the effectiveness of this approach.

Keywords: 3-axis linear interpolation, FPGA, motion controller, micro-stepping

Procedia PDF Downloads 196
12836 Using the Bootstrap for Problems Statistics

Authors: Brahim Boukabcha, Amar Rebbouh

Abstract:

The bootstrap method based on the idea of exploiting all the information provided by the initial sample, allows us to study the properties of estimators. In this article we will present a theoretical study on the different methods of bootstrapping and using the technique of re-sampling in statistics inference to calculate the standard error of means of an estimator and determining a confidence interval for an estimated parameter. We apply these methods tested in the regression models and Pareto model, giving the best approximations.

Keywords: bootstrap, error standard, bias, jackknife, mean, median, variance, confidence interval, regression models

Procedia PDF Downloads 370
12835 Amrita Bose-Einstein Condensate Solution Formed by Gold Nanoparticles Laser Fusion and Atmospheric Water Generation

Authors: Montree Bunruanses, Preecha Yupapin

Abstract:

In this work, the quantum material called Amrita (elixir) is made from top-down gold into nanometer particles by fusing 99% gold with a laser and mixing it with drinking water using the atmospheric water (AWG) production system, which is made of water with air. The high energy laser power destroyed the four natural force bindings from gravity-weak-electromagnetic and strong coupling forces, where finally it was the purified Bose-Einstein condensate (BEC) states. With this method, gold atoms in the form of spherical single crystals with a diameter of 30-50 nanometers are obtained and used. They were modulated (activated) with a frequency generator into various matrix structures mixed with AWG water to be used in the upstream conversion (quantum reversible) process, which can be applied on humans both internally or externally by drinking or applying on the treated surfaces. Doing both space (body) and time (mind) will go back to the origin and start again from the coupling of space-time on both sides of time at fusion (strong coupling force) and push out (Big Bang) at the equilibrium point (singularity) occurs as strings and DNA with neutrinos as coupling energy. There is no distortion (purification), which is the point where time and space have not yet been determined, and there is infinite energy. Therefore, the upstream conversion is performed. It is reforming DNA to make it be purified. The use of Amrita is a method used for people who cannot meditate (quantum meditation). Various cases were applied, where the results show that the Amrita can make the body and the mind return to their pure origins and begin the downstream process with the Big Bang movement, quantum communication in all dimensions, DNA reformation, frequency filtering, crystal body forming, broadband quantum communication networks, black hole forming, quantum consciousness, body and mind healing, etc.

Keywords: quantum materials, quantum meditation, quantum reversible, Bose-Einstein condensate

Procedia PDF Downloads 55
12834 Electrochemical Inactivation of Toxic Cyanobacteria and Degradation of Cyanotoxins

Authors: Belal Bakheet, John Beardall, Xiwang Zhang, David McCarthy

Abstract:

The potential risks associated with toxic cyanobacteria have raised growing environmental and public health concerns leading to an increasing effort into researching ways to bring about their removal from water, together with destruction of their associated cyanotoxins. A variety of toxins are synthesized by cyanobacteria and include hepatotoxins, neurotoxins, and cytotoxins which can cause a range of symptoms in humans from skin irritation to serious liver and nerve damage. Therefore drinking water treatment processes should ensure the consumers’ safety by removing both cyanobacterial cells, and cyanotoxins from the water. Cyanobacterial cells and cyanotoxins presented challenges to the conventional water treatment systems; their accumulation within drinking water treatment plants has been reported leading to plants shut down. Thus, innovative and effective water purification systems to tackle cyanobacterial pollution are required. In recent years there has been increasing attention to the electrochemical oxidation process as a feasible alternative disinfection method which is able to generate in situ a variety of oxidants that would achieve synergistic effects in the water disinfection process and toxin degradation. By utilizing only electric current, the electrochemical process through electrolysis can produce reactive oxygen species such as hydroxyl radicals from the water, or other oxidants such as chlorine from chloride ions present in the water. From extensive physiological and morphological investigation of cyanobacterial cells during electrolysis, our results show that these oxidants have significant impact on cell inactivation, simultaneously with cyanotoxins removal without the need for chemicals addition. Our research aimed to optimize existing electrochemical oxidation systems and develop new systems to treat water containing toxic cyanobacteria and cyanotoxins. The research covers detailed mechanism study on oxidants production and cell inactivation in the treatment under environmental conditions. Overall, our study suggests that the electrochemical treatment process e is an effective method for removal of toxic cyanobacteria and cyanotoxins.

Keywords: toxic cyanobacteria, cyanotoxins, electrochemical process, oxidants

Procedia PDF Downloads 217
12833 Creating a Digital Map to Monitor the Care of People Living with HIV/Aids in Porto Alegre, Brazil: An Experience Report

Authors: Tiago Sigal Linhares, Ana Amélia Nascimento da Silva Bones, Juliana Miola, McArthur Alexander Barrow, Airton Tetelbom Stein

Abstract:

Introduction: As a result of increased globalization and changing migration trends, it is expected that a significant portion of People Living with HIV/AIDS (PLWHA) will change their place of residence over time. In order to provide better health care, monitor the HIV epidemic and plan urban public health care and policies, there is a growing need to formulate a strategy for monitoring PLWHA care, location and migration patterns. The Porto Alegre District is characterized by a high prevalence of PLWHA and is considered one of the epicenters of HIV epidemic in Latin America. Objectives: The aim of this study is to create a digital and easily editable map in order to create a visual representation of the location of PLWHA and to monitor their migration within the city and the country in an effort to promote longitudinal care. Methods: This Experience Report used Google Maps Map Creator to generate an active digital map showing the location and changes in residence of 165 PLWHA who received care at two Primary Health Care (PHC) clinics, which attended an estimated population of five thousand patients, in downtown Porto Alegre over the last four years. Their current addresses were discovered in the unified Brazilian health care system digital records (e-SUS) and updated on the map. Results: A digital map with PLWHA current residence location was created. It was possible to demonstrate visually areas with a large concentration of PLWHA and the migration of the population within the city as wells as other cities, regions and states. Conclusions: An easily reproducible and free map could aid in PLWHA monitoring, urban public health planning, target interventions and situational diagnosis. Moreover, a visual representation of PLWHA location and migration could help bring more attention and investments to areas with geographic inequities or higher prevalence of PLWHA. It also enables notification of local PHC units of monitored patients inside their area, which are in clinical risk or with treatment abandonment through active case findings, improving the care of PLWHA.

Keywords: health care, medical public health, theoretical and conceptual innovations, urban public health

Procedia PDF Downloads 105
12832 Urban Noise and Air Quality: Correlation between Air and Noise Pollution; Sensors, Data Collection, Analysis and Mapping in Urban Planning

Authors: Massimiliano Condotta, Paolo Ruggeri, Chiara Scanagatta, Giovanni Borga

Abstract:

Architects and urban planners, when designing and renewing cities, have to face a complex set of problems, including the issues of noise and air pollution which are considered as hot topics (i.e., the Clean Air Act of London and the Soundscape definition). It is usually taken for granted that these problems go by together because the noise pollution present in cities is often linked to traffic and industries, and these produce air pollutants as well. Traffic congestion can create both noise pollution and air pollution, because NO₂ is mostly created from the oxidation of NO, and these two are notoriously produced by processes of combustion at high temperatures (i.e., car engines or thermal power stations). We can see the same process for industrial plants as well. What have to be investigated – and is the topic of this paper – is whether or not there really is a correlation between noise pollution and air pollution (taking into account NO₂) in urban areas. To evaluate if there is a correlation, some low-cost methodologies will be used. For noise measurements, the OpeNoise App will be installed on an Android phone. The smartphone will be positioned inside a waterproof box, to stay outdoor, with an external battery to allow it to collect data continuously. The box will have a small hole to install an external microphone, connected to the smartphone, which will be calibrated to collect the most accurate data. For air, pollution measurements will be used the AirMonitor device, an Arduino board to which the sensors, and all the other components, are plugged. After assembling the sensors, they will be coupled (one noise and one air sensor) and placed in different critical locations in the area of Mestre (Venice) to map the existing situation. The sensors will collect data for a fixed period of time to have an input for both week and weekend days, in this way it will be possible to see the changes of the situation during the week. The novelty is that data will be compared to check if there is a correlation between the two pollutants using graphs that should show the percentage of pollution instead of the values obtained with the sensors. To do so, the data will be converted to fit on a scale that goes up to 100% and will be shown thru a mapping of the measurement using GIS methods. Another relevant aspect is that this comparison can help to choose which are the right mitigation solutions to be applied in the area of the analysis because it will make it possible to solve both the noise and the air pollution problem making only one intervention. The mitigation solutions must consider not only the health aspect but also how to create a more livable space for citizens. The paper will describe in detail the methodology and the technical solution adopted for the realization of the sensors, the data collection, noise and pollution mapping and analysis.

Keywords: air quality, data analysis, data collection, NO₂, noise mapping, noise pollution, particulate matter

Procedia PDF Downloads 202
12831 Air Dispersion Model for Prediction Fugitive Landfill Gaseous Emission Impact in Ambient Atmosphere

Authors: Moustafa Osman Mohammed

Abstract:

This paper will explore formation of HCl aerosol at atmospheric boundary layers and encourages the uptake of environmental modeling systems (EMSs) as a practice evaluation of gaseous emissions (“framework measures”) from small and medium-sized enterprises (SMEs). The conceptual model predicts greenhouse gas emissions to ecological points beyond landfill site operations. It focuses on incorporation traditional knowledge into baseline information for both measurement data and the mathematical results, regarding parameters influence model variable inputs. The paper has simplified parameters of aerosol processes based on the more complex aerosol process computations. The simple model can be implemented to both Gaussian and Eulerian rural dispersion models. Aerosol processes considered in this study were (i) the coagulation of particles, (ii) the condensation and evaporation of organic vapors, and (iii) dry deposition. The chemical transformation of gas-phase compounds is taken into account photochemical formulation with exposure effects according to HCl concentrations as starting point of risk assessment. The discussion set out distinctly aspect of sustainability in reflection inputs, outputs, and modes of impact on the environment. Thereby, models incorporate abiotic and biotic species to broaden the scope of integration for both quantification impact and assessment risks. The later environmental obligations suggest either a recommendation or a decision of what is a legislative should be achieved for mitigation measures of landfill gas (LFG) ultimately.

Keywords: air pollution, landfill emission, environmental management, monitoring/methods and impact assessment

Procedia PDF Downloads 303
12830 Contextual SenSe Model: Word Sense Disambiguation using Sense and Sense Value of Context Surrounding the Target

Authors: Vishal Raj, Noorhan Abbas

Abstract:

Ambiguity in NLP (Natural language processing) refers to the ability of a word, phrase, sentence, or text to have multiple meanings. This results in various kinds of ambiguities such as lexical, syntactic, semantic, anaphoric and referential am-biguities. This study is focused mainly on solving the issue of Lexical ambiguity. Word Sense Disambiguation (WSD) is an NLP technique that aims to resolve lexical ambiguity by determining the correct meaning of a word within a given context. Most WSD solutions rely on words for training and testing, but we have used lemma and Part of Speech (POS) tokens of words for training and testing. Lemma adds generality and POS adds properties of word into token. We have designed a novel method to create an affinity matrix to calculate the affinity be-tween any pair of lemma_POS (a token where lemma and POS of word are joined by underscore) of given training set. Additionally, we have devised an al-gorithm to create the sense clusters of tokens using affinity matrix under hierar-chy of POS of lemma. Furthermore, three different mechanisms to predict the sense of target word using the affinity/similarity value are devised. Each contex-tual token contributes to the sense of target word with some value and whichever sense gets higher value becomes the sense of target word. So, contextual tokens play a key role in creating sense clusters and predicting the sense of target word, hence, the model is named Contextual SenSe Model (CSM). CSM exhibits a noteworthy simplicity and explication lucidity in contrast to contemporary deep learning models characterized by intricacy, time-intensive processes, and chal-lenging explication. CSM is trained on SemCor training data and evaluated on SemEval test dataset. The results indicate that despite the naivety of the method, it achieves promising results when compared to the Most Frequent Sense (MFS) model.

Keywords: word sense disambiguation (wsd), contextual sense model (csm), most frequent sense (mfs), part of speech (pos), natural language processing (nlp), oov (out of vocabulary), lemma_pos (a token where lemma and pos of word are joined by underscore), information retrieval (ir), machine translation (mt)

Procedia PDF Downloads 88
12829 Assessing Overall Thermal Conductance Value of Low-Rise Residential Home Exterior Above-Grade Walls Using Infrared Thermography Methods

Authors: Matthew D. Baffa

Abstract:

Infrared thermography is a non-destructive test method used to estimate surface temperatures based on the amount of electromagnetic energy radiated by building envelope components. These surface temperatures are indicators of various qualitative building envelope deficiencies such as locations and extent of heat loss, thermal bridging, damaged or missing thermal insulation, air leakage, and moisture presence in roof, floor, and wall assemblies. Although infrared thermography is commonly used for qualitative deficiency detection in buildings, this study assesses its use as a quantitative method to estimate the overall thermal conductance value (U-value) of the exterior above-grade walls of a study home. The overall U-value of exterior above-grade walls in a home provides useful insight into the energy consumption and thermal comfort of a home. Three methodologies from the literature were employed to estimate the overall U-value by equating conductive heat loss through the exterior above-grade walls to the sum of convective and radiant heat losses of the walls. Outdoor infrared thermography field measurements of the exterior above-grade wall surface and reflective temperatures and emissivity values for various components of the exterior above-grade wall assemblies were carried out during winter months at the study home using a basic thermal imager device. The overall U-values estimated from each methodology from the literature using the recorded field measurements were compared to the nominal exterior above-grade wall overall U-value calculated from materials and dimensions detailed in architectural drawings of the study home. The nominal overall U-value was validated through calendarization and weather normalization of utility bills for the study home as well as various estimated heat loss quantities from a HOT2000 computer model of the study home and other methods. Under ideal environmental conditions, the estimated overall U-values deviated from the nominal overall U-value between ±2% to ±33%. This study suggests infrared thermography can estimate the overall U-value of exterior above-grade walls in low-rise residential homes with a fair amount of accuracy.

Keywords: emissivity, heat loss, infrared thermography, thermal conductance

Procedia PDF Downloads 298