Search results for: real volume
2000 Evaluating the Benefits of Intelligent Acoustic Technology in Classrooms: A Case Study
Authors: Megan Burfoot, Ali GhaffarianHoseini, Nicola Naismith, Amirhosein GhaffarianHoseini
Abstract:
Intelligent Acoustic Technology (IAT) is a novel architectural device used in buildings to automatically vary the acoustic conditions of space. IAT is realized by integrating two components: Variable Acoustic Technology (VAT) and an intelligent system. The VAT passively alters the RT by changing the total sound absorption in a room. In doing so, the Reverberation Time (RT) is changed and thus, the sound strength and clarity are altered. The intelligent system detects sound waves in real-time to identify the aural situation, and the RT is adjusted accordingly based on pre-programmed algorithms. IAT - the synthesis of these two components - can dramatically improve acoustic comfort, as the acoustic condition is automatically optimized for any detected aural situation. This paper presents an evaluation of the improvements of acoustic comfort in an existing tertiary classroom located at Auckland University of Technology in New Zealand. This is a pilot case study, the first of its’ kind attempting to quantify the benefits of IAT. Naturally, the potential acoustic improvements from IAT can be actualized by only installing the VAT component of IAT and by manually adjusting it rather than utilizing an intelligent system. Such a simplified methodology is adopted for this case study to understand the potential significance of IAT without adopting a time and cost-intensive strategy. For this study, the VAT is built by overlaying reflective, rotating louvers over sound absorption panels. RT's are measured according to international standards before and after installing VAT in the classroom. The louvers are manually rotated in increments by the experimenter and further RT measurements are recorded. The results are compared with recommended guidelines and reference values from national standards for spaces intended for speech and communication. The results obtained from the measurements are used to quantify the potential improvements in classroom acoustic comfort, where IAT to be used. This evaluation reveals the current existence of poor acoustic conditions in the classroom caused by high RT's. The poor acoustics are also largely attributed to the classrooms’ inability to vary acoustic parameters for changing aural situations. The classroom experiences one static acoustic state, neglecting to recognize the nature of classrooms as flexible, dynamic spaces. Evidently, when using VAT the classroom is prescribed with a wide range of RTs it can achieve. Namely, acoustic requirements for varying teaching approaches are satisfied, and acoustic comfort is improved. By quantifying the benefits of using VAT, it can confidently suggest these same benefits are achieved with IAT. Nevertheless, it is encouraged that future studies continue this line of research toward the eventual development of IAT and its’ acceptance into mainstream architecture.Keywords: acoustic comfort, classroom acoustics, intelligent acoustics, variable acoustics
Procedia PDF Downloads 1891999 Risk Assessment of Flood Defences by Utilising Condition Grade Based Probabilistic Approach
Authors: M. Bahari Mehrabani, Hua-Peng Chen
Abstract:
Management and maintenance of coastal defence structures during the expected life cycle have become a real challenge for decision makers and engineers. Accurate evaluation of the current condition and future performance of flood defence structures is essential for effective practical maintenance strategies on the basis of available field inspection data. Moreover, as coastal defence structures age, it becomes more challenging to implement maintenance and management plans to avoid structural failure. Therefore, condition inspection data are essential for assessing damage and forecasting deterioration of ageing flood defence structures in order to keep the structures in an acceptable condition. The inspection data for flood defence structures are often collected using discrete visual condition rating schemes. In order to evaluate future condition of the structure, a probabilistic deterioration model needs to be utilised. However, existing deterioration models may not provide a reliable prediction of performance deterioration for a long period due to uncertainties. To tackle the limitation, a time-dependent condition-based model associated with a transition probability needs to be developed on the basis of condition grade scheme for flood defences. This paper presents a probabilistic method for predicting future performance deterioration of coastal flood defence structures based on condition grading inspection data and deterioration curves estimated by expert judgement. In condition-based deterioration modelling, the main task is to estimate transition probability matrices. The deterioration process of the structure related to the transition states is modelled according to Markov chain process, and a reliability-based approach is used to estimate the probability of structural failure. Visual inspection data according to the United Kingdom Condition Assessment Manual are used to obtain the initial condition grade curve of the coastal flood defences. The initial curves then modified in order to develop transition probabilities through non-linear regression based optimisation algorithms. The Monte Carlo simulations are then used to evaluate the future performance of the structure on the basis of the estimated transition probabilities. Finally, a case study is given to demonstrate the applicability of the proposed method under no-maintenance and medium-maintenance scenarios. Results show that the proposed method can provide an effective predictive model for various situations in terms of available condition grading data. The proposed model also provides useful information on time-dependent probability of failure in coastal flood defences.Keywords: condition grading, flood defense, performance assessment, stochastic deterioration modelling
Procedia PDF Downloads 2331998 I, Me and the Bot: Forming a Theory of Symbolic Interactivity with a Chatbot
Authors: Felix Liedel
Abstract:
The rise of artificial intelligence has numerous and far-reaching consequences. In addition to the obvious consequences for entire professions, the increasing interaction with chatbots also has a wide range of social consequences and implications. We are already increasingly used to interacting with digital chatbots, be it in virtual consulting situations, creative development processes or even in building personal or intimate virtual relationships. A media-theoretical classification of these phenomena has so far been difficult, partly because the interactive element in the exchange with artificial intelligence has undeniable similarities to human-to-human communication but is not identical to it. The proposed study, therefore, aims to reformulate the concept of symbolic interaction in the tradition of George Herbert Mead as symbolic interactivity in communication with chatbots. In particular, Mead's socio-psychological considerations will be brought into dialog with the specific conditions of digital media, the special dispositive situation of chatbots and the characteristics of artificial intelligence. One example that illustrates this particular communication situation with chatbots is so-called consensus fiction: In face-to-face communication, we use symbols on the assumption that they will be interpreted in the same or a similar way by the other person. When briefing a chatbot, it quickly becomes clear that this is by no means the case: only the bot's response shows whether the initial request corresponds to the sender's actual intention. This makes it clear that chatbots do not just respond to requests. Rather, they function equally as projection surfaces for their communication partners but also as distillations of generalized social attitudes. The personalities of the chatbot avatars result, on the one hand, from the way we behave towards them and, on the other, from the content we have learned in advance. Similarly, we interpret the response behavior of the chatbots and make it the subject of our own actions with them. In conversation with the virtual chatbot, we enter into a dialog with ourselves but also with the content that the chatbot has previously learned. In our exchanges with chatbots, we, therefore, interpret socially influenced signs and behave towards them in an individual way according to the conditions that the medium deems acceptable. This leads to the emergence of situationally determined digital identities that are in exchange with the real self but are not identical to it: In conversation with digital chatbots, we bring our own impulses, which are brought into permanent negotiation with a generalized social attitude by the chatbot. This also leads to numerous media-ethical follow-up questions. The proposed approach is a continuation of my dissertation on moral decision-making in so-called interactive films. In this dissertation, I attempted to develop a concept of symbolic interactivity based on Mead. Current developments in artificial intelligence are now opening up new areas of application.Keywords: artificial intelligence, chatbot, media theory, symbolic interactivity
Procedia PDF Downloads 521997 Inhibitory Mechanism of Ag and Fe Colloidal Nanoparticles on P. aeruginosa and E.coli Growth
Authors: Fatemeh Moradian, Razieh Ghorbani, Poria Biparva
Abstract:
Growing resistance of microorganisms to potent antibiotics has renewed a great interest towards investigating bactericidal properties of nanoparticles and their Nano composites as an alternative. The use of metal nanoparticles to combat bacterial infections is one of the most wide spread applications of nanotechnology in the field of antibacterial. Nanomaterials have unique properties compared to their bulk counterparts. In this report, we demonstrate the antimicrobial activity of zerovalent Iron(ZVI) and Ag(silver) nanoparticles against Gram-negative bacteria E.coli(DH5α) and Pseudomonas aeruginosa. At first ZVI and Ag nanoparticles were synthesized by chemical reduction method and using scanning electron microscopy (SEM) the nanoparticle size determined. Different concentrations of Ag and ZVI nanoparticles were added to bacteria on nutrient agar medium. Minimum inhibitory concentration (MIC) of Ag and Fe nanoparticles for P. aeruginosa were 5µM and 1µg as well as for E.coli were 6µM. and 10 µg, respectively. Among the two nanoparticles, ZVI showed that the greatest antimicrobial activity against E.coli and Ag nanoparticle on P.aeruginosa. Results suggested that the bactericidal effect of metal nanoparticles has been attributed to their small size as well as high surface to volume ratio and NPs could be used as an effective antibacterial material.Keywords: bactericidal properties, MIC, nanoparticle, SEM
Procedia PDF Downloads 6001996 Analysis of Non-Conventional Roundabout Performance in Mixed Traffic Conditions
Authors: Guneet Saini, Shahrukh, Sunil Sharma
Abstract:
Traffic congestion is the most critical issue faced by those in the transportation profession today. Over the past few years, roundabouts have been recognized as a measure to promote efficiency at intersections globally. In developing countries like India, this type of intersection still faces a lot of issues, such as bottleneck situations, long queues and increased waiting times, due to increasing traffic which in turn affect the performance of the entire urban network. This research is a case study of a non-conventional roundabout, in terms of geometric design, in a small town in India. These types of roundabouts should be analyzed for their functionality in mixed traffic conditions, prevalent in many developing countries. Microscopic traffic simulation is an effective tool to analyze traffic conditions and estimate various measures of operational performance of intersections such as capacity, vehicle delay, queue length and Level of Service (LOS) of urban roadway network. This study involves analyzation of an unsymmetrical non-circular 6-legged roundabout known as “Kala Aam Chauraha” in a small town Bulandshahr in Uttar Pradesh, India using VISSIM simulation package which is the most widely used software for microscopic traffic simulation. For coding in VISSIM, data are collected from the site during morning and evening peak hours of a weekday and then analyzed for base model building. The model is calibrated on driving behavior and vehicle parameters and an optimal set of calibrated parameters is obtained followed by validation of the model to obtain the base model which can replicate the real field conditions. This calibrated and validated model is then used to analyze the prevailing operational traffic performance of the roundabout which is then compared with a proposed alternative to improve efficiency of roundabout network and to accommodate pedestrians in the geometry. The study results show that the alternative proposed is an advantage over the present roundabout as it considerably reduces congestion, vehicle delay and queue length and hence, successfully improves roundabout performance without compromising on pedestrian safety. The study proposes similar designs for modification of existing non-conventional roundabouts experiencing excessive delays and queues in order to improve their efficiency especially in the case of developing countries. From this study, it can be concluded that there is a need to improve the current geometry of such roundabouts to ensure better traffic performance and safety of drivers and pedestrians negotiating the intersection and hence this proposal may be considered as a best fit.Keywords: operational performance, roundabout, simulation, VISSIM
Procedia PDF Downloads 1391995 Universe at Zero Second and the Creation Process of the First Particle from the Absolute Void
Authors: Shivan Sirdy
Abstract:
In this study, we discuss the properties of absolute void space or the universe at zero seconds, and how these properties play a vital role in creating a mechanism in which the very first particle gets created simultaneously everywhere. We find the limit in which when the absolute void volume reaches will lead to the collapse that leads to the creation of the first particle. This discussion is made following the elementary dimensions theory study that was peer-reviewed at the end of 2020; everything in the universe is made from four elementary dimensions, these dimensions are the three spatial dimensions (X, Y, and Z) and the Void resistance as the factor of change among the four. Time itself was not considered as the fourth dimension. Rather time corresponds to a factor of change, and during the research, it was found out that the Void resistance is the factor of change in the absolute Void space, where time is a hypothetical concept that represents changes during certain events compared to a constant change rate event. Therefore, time does exist, but as a factor of change as the Void resistance: Time= factor of change= Void resistance.Keywords: elementary dimensions, absolute void, time alternative, early universe, universe at zero second, Void resistant, Hydrogen atom, Hadron field, Lepton field
Procedia PDF Downloads 2021994 New Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator
Authors: Wedad Albalawi
Abstract:
The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques, and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then, dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is an arbitrary nonempty closed subset of the real numbers. Then, the dynamic inequalities on time scales have received a lot of attention in the literature and has become a major field in pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on Hardy and Coposon inequalities, using Steklov operator on time scale in double integrals to obtain special cases of time-scale inequalities of Hardy and Copson on high dimensions. The advantage of this study is that it uses the one-dimensional classical Hardy inequality to obtain higher dimensional on time scale versions that will be applied in the solution of the Cauchy problem for the wave equation. In addition, the obtained inequalities have various applications involving discontinuous domains such as bug populations, phytoremediation of metals, wound healing, maximization problems. The proof can be done by introducing restriction on the operator in several cases. The concepts in time scale version such as time scales calculus will be used that allows to unify and extend many problems from the theories of differential and of difference equations. In addition, using chain rule, and some properties of multiple integrals on time scales, some theorems of Fubini and the inequality of H¨older.Keywords: time scales, inequality of hardy, inequality of coposon, steklov operator
Procedia PDF Downloads 951993 Early Detection of Lymphedema in Post-Surgery Oncology Patients
Authors: Sneha Noble, Rahul Krishnan, Uma G., D. K. Vijaykumar
Abstract:
Breast-Cancer related Lymphedema is a major problem that affects many women. Lymphedema is the swelling that generally occurs in the arms or legs caused by the removal of or damage to lymph nodes as a part of cancer treatment. Treating it at the earliest possible stage is the best way to manage the condition and prevent it from leading to pain, recurrent infection, reduced mobility, and impaired function. So, this project aims to focus on the multi-modal approaches to identify the risks of Lymphedema in post-surgical oncology patients and prevent it at the earliest. The Kinect IR Sensor is utilized to capture the images of the body and after image processing techniques, the region of interest is obtained. Then, performing the voxelization method will provide volume measurements in pre-operative and post-operative periods in patients. The formation of a mathematical model will help in the comparison of values. Clinical pathological data of patients will be investigated to assess the factors responsible for the development of lymphedema and its risks.Keywords: Kinect IR sensor, Lymphedema, voxelization, lymph nodes
Procedia PDF Downloads 1381992 Numerical Investigation of the Effect of Number of Waves on Heat Transfer in a Wavy Wall Enclosure
Authors: Ali Reza Tahavvor, Saeed Hosseini, Afshin Karimzadeh Fard
Abstract:
In this paper the effect of wall waviness of side walls in a two-dimensional wavy enclosure is numerically investigated. Two vertical wavy walls and straight top wall are kept isothermal and the bottom wall temperature is higher and spatially varying with cosinusoidal temperature distribution. A computational code based on Finite-volume approach is used to solve governing equations and SIMPLE method is used for pressure velocity coupling. Test is performed for several different numbers of undulations. The Prandtl number was kept constant and the Ra number denotes that the flow is laminar. Temperature and velocity fields are determined. Therefore, according to the obtained results a correlation is proposed for average Nusselt number as a function of number of side wall waves. The results indicate that the Nusselt number is highly affected by number of waves and increasing it decreases the wavy walls Nusselt number; although the Nusselt number is not highly affected by surface waviness when the number of undulations is below one.Keywords: cavity, natural convection, Nusselt number, wavy wall
Procedia PDF Downloads 4701991 A Study of the Prevalence of Trichinellosis in Domestic and Wild Animals for the Region of Sofia, Bulgaria
Authors: Valeria Dilcheva, Svetlozara Petkova, Ivelin Vladov
Abstract:
Nemathodes of the genus Trichinella are zoonotic parasites with a cosmopolitan distribution. More than 100 species of mammals, birds and reptiles are involved in the natural cycle of this nematode. At present, T. spiralis, T. pseudospiralis, and T. britovi have been found in Bulgaria. The existence of natural wildlife and domestic reservoirs of Trichinella spp. can be a serious threat to human health. Three trichinella isolates caused human trichinella infection outbreaks from three regions of Sofia City Province were used for the research: sample No. 1 - Ratus norvegicus, sample No. 2 – domestic pig (Sus scrofa domestica), sample No. 3 - domestic pig (Sus scrofa domestica). Trichinella larvae of the studied species were isolated via digestive method (pepsin, hydrochloric acid, water) at 37ºC by standard procedure and were determined by gender (male and female) based on their morphological characteristics. As a reference trichinella species were used: T. spiralis, T. pseudospiralis, T. nativa and T. britovi. Single male and female larvae of the three isolates were crossed with single male and female larvae of the reference trichinella species as well as reciprocally. As a result of cross-breeding, offspring of muscular larvae with T. spiralis and T. britovi were obtained, while in experiments with T. pseudospiralis and T. nativa, trichinella larvae were not found in the laboratory mice. The results obtained in the control groups indicate that the trichinella larvae used from the isolates and the four trichinella species are infective. Also, the infective ability of the F1 offspring from the successful cross-breeding between isolates and reference species was investigated. Through the data obtained in the experiment was found that isolates No. 1 and No. 2 belong to the species T. spiralis, and isolate No. 3 belongs to the species T. britovi. The results were confirmed by PCR and real-time PCR analysis. Thus the presence and circulation of the species T. spiralis and T. britovi in Bulgaria was confirmed. Probably the rodents (rats) are involved in the distribution of T. spiralis in urban environment. The species T. britovi found in a domestic pig speaks of some contact with wild animals for which T. britovi is characteristic. The probable reason is that a large number of farmers in Bulgaria practice the free-range breeding of domestic pigs. Part of the farmers also used as food for domestic pigs waste products from the game (foxes, jackals, bears, wolves) and probably thus the infection was obtained. The distribution range of trichinella species in Bulgaria is not strictly outlined. It is believed that T. spiralis is most common in domestic animals and T. britovi and T. pseudospiralis are characteristic of wildlife. To answer the question whether wild and synanthropic animals are infected with the same or different trichinella species, which species predominate in nature and what their distribution among different hosts is, further research is required.Keywords: cross-breeding, Sofia, trichinellosis, Trichinella britovi, Trichinella spiralis
Procedia PDF Downloads 1891990 Performance of Self-Compacting Mortars Containing Foam Glass Granulate
Authors: Brahim Safi, Djamila Aboutaleb, Mohammed Saidi, Abdelbaki Benmounah, Fahima Benbrahim
Abstract:
The inorganic wastes are currently used in the manufacture of concretes as mineral additions by cement substitution or as fine/coarse aggregates by replacing traditional aggregates. In this respect, this study aims to valorize the mineral wastes in particular glass wastes to produce granulated foam glass (as fine aggregates). Granulated foam glasses (GFG) were prepared from the glass powder (glass cullet) and foaming agent (limestone) according to applied manufacturing of GFG (at a heat treatment 850 ° C for 20min). After, self-compacting mortars were elaborated with fine aggregate (sand) and other variant mortars with granulated foam glass at volume ratio (0, 30, 50 and 100 %). Rheological characterization tests (fluidity) and physic-mechanical (density, porosity /absorption of water and mechanical tests) were carried out on studied mortars. The results obtained show that a slightly decreasing of compressive strength of mortars having lightness very important for building construction.Keywords: glass wastes, lightweight aggregate, mortar, fluidity, density, mechanical strength
Procedia PDF Downloads 2281989 Building an Opinion Dynamics Model from Experimental Data
Authors: Dino Carpentras, Paul J. Maher, Caoimhe O'Reilly, Michael Quayle
Abstract:
Opinion dynamics is a sub-field of agent-based modeling that focuses on people’s opinions and their evolutions over time. Despite the rapid increase in the number of publications in this field, it is still not clear how to apply these models to real-world scenarios. Indeed, there is no agreement on how people update their opinion while interacting. Furthermore, it is not clear if different topics will show the same dynamics (e.g., more polarized topics may behave differently). These problems are mostly due to the lack of experimental validation of the models. Some previous studies started bridging this gap in the literature by directly measuring people’s opinions before and after the interaction. However, these experiments force people to express their opinion as a number instead of using natural language (and then, eventually, encoding it as numbers). This is not the way people normally interact, and it may strongly alter the measured dynamics. Another limitation of these studies is that they usually average all the topics together, without checking if different topics may show different dynamics. In our work, we collected data from 200 participants on 5 unpolarized topics. Participants expressed their opinions in natural language (“agree” or “disagree”). We also measured the certainty of their answer, expressed as a number between 1 and 10. However, this value was not shown to other participants to keep the interaction based on natural language. We then showed the opinion (and not the certainty) of another participant and, after a distraction task, we repeated the measurement. To make the data compatible with opinion dynamics models, we multiplied opinion and certainty to obtain a new parameter (here called “continuous opinion”) ranging from -10 to +10 (using agree=1 and disagree=-1). We firstly checked the 5 topics individually, finding that all of them behaved in a similar way despite having different initial opinions distributions. This suggested that the same model could be applied for different unpolarized topics. We also observed that people tend to maintain similar levels of certainty, even when they changed their opinion. This is a strong violation of what is suggested from common models, where people starting at, for example, +8, will first move towards 0 instead of directly jumping to -8. We also observed social influence, meaning that people exposed with “agree” were more likely to move to higher levels of continuous opinion, while people exposed with “disagree” were more likely to move to lower levels. However, we also observed that the effect of influence was smaller than the effect of random fluctuations. Also, this configuration is different from standard models, where noise, when present, is usually much smaller than the effect of social influence. Starting from this, we built an opinion dynamics model that explains more than 80% of data variance. This model was also able to show the natural conversion of polarization from unpolarized states. This experimental approach offers a new way to build models grounded on experimental data. Furthermore, the model offers new insight into the fundamental terms of opinion dynamics models.Keywords: experimental validation, micro-dynamics rule, opinion dynamics, update rule
Procedia PDF Downloads 1091988 Convectory Policing-Reconciling Historic and Contemporary Models of Police Service Delivery
Authors: Mark Jackson
Abstract:
Description: This paper is based on an theoretical analysis of the efficacy of the dominant model of policing in western jurisdictions. Those results are then compared with a similar analysis of a traditional reactive model. It is found that neither model provides for optimal delivery of services. Instead optimal service can be achieved by a synchronous hybrid model, termed the Convectory Policing approach. Methodology and Findings: For over three decades problem oriented policing (PO) has been the dominant model for western police agencies. Initially based on the work of Goldstein during the 1970s the problem oriented framework has spawned endless variants and approaches, most of which embrace a problem solving rather than a reactive approach to policing. This has included the Area Policing Concept (APC) applied in many smaller jurisdictions in the USA, the Scaled Response Policing Model (SRPM) currently under trial in Western Australia and the Proactive Pre-Response Approach (PPRA) which has also seen some success. All of these, in some way or another, are largely based on a model that eschews a traditional reactive model of policing. Convectory Policing (CP) is an alternative model which challenges the underpinning assumptions which have seen proliferation of the PO approach in the last three decades and commences by questioning the economics on which PO is based. It is argued that in essence, the PO relies on an unstated, and often unrecognised assumption that resources will be available to meet demand for policing services, while at the same time maintaining the capacity to deploy staff to develop solutions to the problems which were ultimately manifested in those same calls for service. The CP model relies on the observations from a numerous western jurisdictions to challenge the validity of that underpinning assumption, particularly in fiscally tight environment. In deploying staff to pursue and develop solutions to underpinning problems, there is clearly an opportunity cost. Those same staff cannot be allocated to alternative duties while engaged in a problem solution role. At the same time, resources in use responding to calls for service are unavailable, while committed to that role, to pursue solutions to the problems giving rise to those same calls for service. The two approaches, reactive and PO are therefore dichotomous. One cannot be optimised while the other is being pursued. Convectory Policing is a pragmatic response to the schism between the competing traditional and contemporary models. If it is not possible to serve either model with any real rigour, it becomes necessary to taper an approach to deliver specific outcomes against which success or otherwise might be measured. CP proposes that a structured roster-driven approach to calls for service, combined with the application of what is termed a resource-effect response capacity has the potential to resolve the inherent conflict between traditional and models of policing and the expectations of the community in terms of community policing based problem solving models.Keywords: policing, reactive, proactive, models, efficacy
Procedia PDF Downloads 4831987 Sedimentation and Morphology of the Kura River-Deltaic System in the Southern Caucasus under Anthropogenic and Sea-Level Controls
Authors: Elmira Aliyeva, Dadash Huseynov, Robert Hoogendoorn, Salomon Kroonenberg
Abstract:
The Kura River is the major water artery in the Southern Caucasus; it is a third river in the Caspian Sea basin in terms of length and size of the catchment area, the second in terms of the water budget, and the first in the volume of sediment load. Understanding of major controls on the Kura fluvial- deltaic system is valuable for efficient management of the highly populated river basin and coastal zone. We have studied grain size of sediments accumulated in the river channels and delta and dated by 210Pb method, astrophotographs, old topographic and geological maps, and archive data. At present time sediments are supplied by the Kura River to the Caspian Sea through three distributary channels oriented north-east, south-east, and south-west. The river is dominated by the suspended load - mud, silt, very fine sand. Coarse sediments are accumulated in the distributaries, levees, point bar, and delta front. The annual suspended sediment budget in the time period 1934-1952 before construction of the Mingechavir water reservoir in 1953 in the Kura River midstream area was 36 mln.t/yr. From 1953 to 1964, the suspended load has dropped to 12 mln.t/yr. After regulation of the Kura River discharge the volume of suspended load transported via north-eastern channel reduced from 35% of the total sediment amount to 4%, and through the main south-eastern channel increased from 65% to 96% with further fall to 56% due to creation of new south-western channel in 1964. Between 1967-1976 the annual sediment budget of the Kura River reached 22,5 mln. t/yr. From 1977 to 1986, the sediment load carried by the Kura River dropped to 17,6 mln.t/yr. The historical data show that between 1860 and 1907, during relatively stable Caspian Sea level two channels - N and SE, appear to have distributed an equal amount of sediments as seen from the bilateral geometry of the delta. In the time period 1907-1929, two new channels - E and NE, appeared. The growth of three delta lobes - N, NE, and SE, and rapid progradation of the delta has occurred on the background of the Caspian Sea level rise as a result of very high sediment supply. Since 1929 the Caspian Sea level decline was followed by the progradation of the delta occurring along the SE channel. The eastern and northern channels have been silted up. The slow rate of progradation at its initial stage was caused by the artificial reduction in the sediment budget. However, the continuous sea-level fall has brought to this river bed gradient increase, high erosional rate, increase in the sediment supply, and more rapid progradation. During the subsequent sea-level rise after 1977 accompanied by the decrease in the sediment budget, the southern part of the delta has turned into a complex of small, shallow channels oriented to the south. The data demonstrate that behaviour of the Kura fluvial – deltaic system and variations in the sediment budget besides anthropogenic regulation are strongly governed by the Caspian Sea level very rapid changes.Keywords: anthropogenic control on sediment budget, Caspian sea-level variations, Kura river sediment load, morphology of the Kura river delta, sedimentation in the Kura river delta
Procedia PDF Downloads 1541986 Behaviour of Lightweight Expanded Clay Aggregate Concrete Exposed to High Temperatures
Authors: Lenka Bodnárová, Rudolf Hela, Michala Hubertová, Iveta Nováková
Abstract:
This paper is concerning the issues of behaviour of lightweight expanded clay aggregates concrete exposed to high temperature. Lightweight aggregates from expanded clay are produced by firing of row material up to temperature 1050°C. Lightweight aggregates have suitable properties in terms of volume stability, when exposed to temperatures up to 1050°C, which could indicate their suitability for construction applications with higher risk of fire. The test samples were exposed to heat by using the standard temperature-time curve ISO 834. Negative changes in resulting mechanical properties, such as compressive strength, tensile strength, and flexural strength were evaluated. Also visual evaluation of the specimen was performed. On specimen exposed to excessive heat, an explosive spalling could be observed, due to evaporation of considerable amount of unbounded water from the inner structure of the concrete.Keywords: expanded clay aggregate, explosive spalling, high temperature, lightweight concrete, temperature-time curve ISO 834
Procedia PDF Downloads 4471985 Optical and Near-UV Spectroscopic Properties of Low-Redshift Jetted Quasars in the Main Sequence in the Main Sequence Context
Authors: Shimeles Terefe Mengistue, Ascensión Del Olmo, Paola Marziani, Mirjana Pović, María Angeles Martínez-Carballo, Jaime Perea, Isabel M. Árquez
Abstract:
Quasars have historically been classified into two distinct classes, radio-loud (RL) and radio-quiet (RQ), taking into account the presence and absence of relativistic radio jets, respectively. The absence of spectra with a high S/N ratio led to the impression that all quasars (QSOs) are spectroscopically similar. Although different attempts were made to unify these two classes, there is a long-standing open debate involving the possibility of a real physical dichotomy between RL and RQ quasars. In this work, we present new high S/N spectra of 11 extremely powerful jetted quasars with radio-to-optical flux density ratio > 1000 that concomitantly cover the low-ionization emission of Mgii𝜆2800 and Hbeta𝛽 as well as the Feii blends in the redshift range 0.35 < z < 1, observed at Calar Alto Observatory (Spain). This work aims to quantify broad emission line differences between RL and RQ quasars by using the four-dimensional eigenvector 1 (4DE1) parameter space and its main sequence (MS) and to check the effect of powerful radio ejection on the low ionization broad emission lines. Emission lines are analysed by making two complementary approaches, a multicomponent non-linear fitting to account for the individual components of the broad emission lines and by analysing the full profile of the lines through parameters such as total widths, centroid velocities at different fractional intensities, asymmetry, and kurtosis indices. It is found that broad emission lines show large reward asymmetry both in Hbeta𝛽 and Mgii2800A. The location of our RL sources in a UV plane looks similar to the optical one, with weak Feii UV emission and broad Mgii2800A. We supplement the 11 sources with large samples from previous work to gain some general inferences. The result shows, compared to RQ, our extreme RL quasars show larger median Hbeta full width at half maximum (FWHM), weaker Feii emission, larger 𝑀BH, lower 𝐿bol/𝐿Edd, and a restricted space occupation in the optical and UV MS planes. The differences are more elusive when the comparison is carried out by restricting the RQ population to the region of the MS occupied by RL quasars, albeit an unbiased comparison matching 𝑀BH and 𝐿bol/𝐿Edd suggests that the most powerful RL quasars show the highest redward asymmetries in Hbeta.Keywords: galaxies, active, line, profiles, quasars, emission lines, supermassive black holes
Procedia PDF Downloads 591984 Optimizing Road Transportation Network Considering the Durability Factors
Authors: Yapegue Bayogo, Ahmadou Halassi Dicko, Brahima Songore
Abstract:
In developing countries, the road transportation system occupies an important place because of its flexibility and the low prices of infrastructure and rolling stock. While road transport is necessary for economic development, the movement of people and their goods, it is urgent to use transportation systems that minimize carbon emissions in order to ensure sustainable development. One of the main objectives of OEDC and the Word Bank is to ensure sustainable economic’ development. This paper aims to develop a road transport network taking into account environmental impacts. The methodology adopted consists of formulating a model optimizing the flow of goods and then collecting information relating to the transport of products. Our model was tested with data on product transport in CMDT areas in the Republic of Mali. The results of our study indicate that emissions from the transport sector can be significantly reduced by minimizing the traffic volume. According to our study, optimizing the transportation network, we benefit from a significant amount of tons of CO₂.Keywords: road transport, transport sustainability, pollution, flexibility, optimized network
Procedia PDF Downloads 1501983 Detailed Quantum Circuit Design and Evaluation of Grover's Algorithm for the Bounded Degree Traveling Salesman Problem Using the Q# Language
Authors: Wenjun Hou, Marek Perkowski
Abstract:
The Traveling Salesman problem is famous in computing and graph theory. In short, it asks for the Hamiltonian cycle of the least total weight in a given graph with N nodes. All variations on this problem, such as those with K-bounded-degree nodes, are classified as NP-complete in classical computing. Although several papers propose theoretical high-level designs of quantum algorithms for the Traveling Salesman Problem, no quantum circuit implementation of these algorithms has been created up to our best knowledge. In contrast to previous papers, the goal of this paper is not to optimize some abstract complexity measures based on the number of oracle iterations, but to be able to evaluate the real circuit and time costs of the quantum computer. Using the emerging quantum programming language Q# developed by Microsoft, which runs quantum circuits in a quantum computer simulation, an implementation of the bounded-degree problem and its respective quantum circuit were created. To apply Grover’s algorithm to this problem, a quantum oracle was designed, evaluating the cost of a particular set of edges in the graph as well as its validity as a Hamiltonian cycle. Repeating the Grover algorithm with an oracle that finds successively lower cost each time allows to transform the decision problem to an optimization problem, finding the minimum cost of Hamiltonian cycles. N log₂ K qubits are put into an equiprobablistic superposition by applying the Hadamard gate on each qubit. Within these N log₂ K qubits, the method uses an encoding in which every node is mapped to a set of its encoded edges. The oracle consists of several blocks of circuits: a custom-written edge weight adder, node index calculator, uniqueness checker, and comparator, which were all created using only quantum Toffoli gates, including its special forms, which are Feynman and Pauli X. The oracle begins by using the edge encodings specified by the qubits to calculate each node that this path visits and adding up the edge weights along the way. Next, the oracle uses the calculated nodes from the previous step and check that all the nodes are unique. Finally, the oracle checks that the calculated cost is less than the previously-calculated cost. By performing the oracle an optimal number of times, a correct answer can be generated with very high probability. The oracle of the Grover Algorithm is modified using the recalculated minimum cost value, and this procedure is repeated until the cost cannot be further reduced. This algorithm and circuit design have been verified, using several datasets, to generate correct outputs.Keywords: quantum computing, quantum circuit optimization, quantum algorithms, hybrid quantum algorithms, quantum programming, Grover’s algorithm, traveling salesman problem, bounded-degree TSP, minimal cost, Q# language
Procedia PDF Downloads 1901982 Evotrader: Bitcoin Trading Using Evolutionary Algorithms on Technical Analysis and Social Sentiment Data
Authors: Martin Pellon Consunji
Abstract:
Due to the rise in popularity of Bitcoin and other crypto assets as a store of wealth and speculative investment, there is an ever-growing demand for automated trading tools, such as bots, in order to gain an advantage over the market. Traditionally, trading in the stock market was done by professionals with years of training who understood patterns and exploited market opportunities in order to gain a profit. However, nowadays a larger portion of market participants are at minimum aided by market-data processing bots, which can generally generate more stable signals than the average human trader. The rise in trading bot usage can be accredited to the inherent advantages that bots have over humans in terms of processing large amounts of data, lack of emotions of fear or greed, and predicting market prices using past data and artificial intelligence, hence a growing number of approaches have been brought forward to tackle this task. However, the general limitation of these approaches can still be broken down to the fact that limited historical data doesn’t always determine the future, and that a lot of market participants are still human emotion-driven traders. Moreover, developing markets such as those of the cryptocurrency space have even less historical data to interpret than most other well-established markets. Due to this, some human traders have gone back to the tried-and-tested traditional technical analysis tools for exploiting market patterns and simplifying the broader spectrum of data that is involved in making market predictions. This paper proposes a method which uses neuro evolution techniques on both sentimental data and, the more traditionally human-consumed, technical analysis data in order to gain a more accurate forecast of future market behavior and account for the way both automated bots and human traders affect the market prices of Bitcoin and other cryptocurrencies. This study’s approach uses evolutionary algorithms to automatically develop increasingly improved populations of bots which, by using the latest inflows of market analysis and sentimental data, evolve to efficiently predict future market price movements. The effectiveness of the approach is validated by testing the system in a simulated historical trading scenario, a real Bitcoin market live trading scenario, and testing its robustness in other cryptocurrency and stock market scenarios. Experimental results during a 30-day period show that this method outperformed the buy and hold strategy by over 260% in terms of net profits, even when taking into consideration standard trading fees.Keywords: neuro-evolution, Bitcoin, trading bots, artificial neural networks, technical analysis, evolutionary algorithms
Procedia PDF Downloads 1231981 Topological Language for Classifying Linear Chord Diagrams via Intersection Graphs
Authors: Michela Quadrini
Abstract:
Chord diagrams occur in mathematics, from the study of RNA to knot theory. They are widely used in theory of knots and links for studying the finite type invariants, whereas in molecular biology one important motivation to study chord diagrams is to deal with the problem of RNA structure prediction. An RNA molecule is a linear polymer, referred to as the backbone, that consists of four types of nucleotides. Each nucleotide is represented by a point, whereas each chord of the diagram stands for one interaction for Watson-Crick base pairs between two nonconsecutive nucleotides. A chord diagram is an oriented circle with a set of n pairs of distinct points, considered up to orientation preserving diffeomorphisms of the circle. A linear chord diagram (LCD) is a special kind of graph obtained cutting the oriented circle of a chord diagram. It consists of a line segment, called its backbone, to which are attached a number of chords with distinct endpoints. There is a natural fattening on any linear chord diagram; the backbone lies on the real axis, while all the chords are in the upper half-plane. Each linear chord diagram has a natural genus of its associated surface. To each chord diagram and linear chord diagram, it is possible to associate the intersection graph. It consists of a graph whose vertices correspond to the chords of the diagram, whereas the chord intersections are represented by a connection between the vertices. Such intersection graph carries a lot of information about the diagram. Our goal is to define an LCD equivalence class in terms of identity of intersection graphs, from which many chord diagram invariants depend. For studying these invariants, we introduce a new representation of Linear Chord Diagrams based on a set of appropriate topological operators that permits to model LCD in terms of the relations among chords. Such set is composed of: crossing, nesting, and concatenations. The crossing operator is able to generate the whole space of linear chord diagrams, and a multiple context free grammar able to uniquely generate each LDC starting from a linear chord diagram adding a chord for each production of the grammar is defined. In other words, it allows to associate a unique algebraic term to each linear chord diagram, while the remaining operators allow to rewrite the term throughout a set of appropriate rewriting rules. Such rules define an LCD equivalence class in terms of the identity of intersection graphs. Starting from a modelled RNA molecule and the linear chord, some authors proposed a topological classification and folding. Our LCD equivalence class could contribute to the RNA folding problem leading to the definition of an algorithm that calculates the free energy of the molecule more accurately respect to the existing ones. Such LCD equivalence class could be useful to obtain a more accurate estimate of link between the crossing number and the topological genus and to study the relation among other invariants.Keywords: chord diagrams, linear chord diagram, equivalence class, topological language
Procedia PDF Downloads 2011980 Evaluation of Dynamic Log Files for Different Dose Rates in IMRT Plans
Authors: Saad Bin Saeed, Fayzan Ahmed, Shahbaz Ahmed, Amjad Hussain
Abstract:
The aim of this study is to evaluate dynamic log files (Dynalogs) at different dose rates by dose-volume histograms (DVH) and used as a (QA) procedure of IMRT. Seven patients of phase one head and neck cancer with similar OAR`s are selected randomly. Reference plans of dose rate 300 and 600 MU/Min with prescribed dose of 50Gy in 25 fractions for each patient is made. Dynalogs produced by delivery of reference plans processed by in-house MATLAB program which produces new field files contain actual positions of multi-leaf collimators (MLC`s) instead of planned positions in reference plans. Copies of reference plans are used to import new field files generated by MATLAB program and renamed as Dyn.plan. After dose calculations of Dyn.plans for different dose rates, DVH, and multiple linear regression tools are used to evaluate reference and Dyn.plans. The results indicate good agreement of correlation between different dose rate plans. The maximum dose difference among PTV and OAR`s are found to be less than 5% and 9% respectively. The study indicates the potential of dynalogs to be used as patient-specific QA of IMRT at different dose rate.Keywords: IMRT, dynalogs, dose rate, DVH
Procedia PDF Downloads 5351979 Use of Cobalt Graphene in Place of Platnium in Catalytic Converter
Authors: V. Srinivasan, S. M. Sriram Nandan
Abstract:
Today in the modern world the most important problem faced by the mankind is increasing the pollution in a very high rate. It affects the ecosystem of the environment and also aids to increase the greenhouse effect. The exhaust gases from the automobile is the major cause of a pollution. Automobiles have increased to a large number which has increased the pollution of our world to an alarming rate. There are two methods of controlling the pollution namely, pre-pollution control method and post-pollution control method. This paper is based on controlling the emission by post-pollution control method. The ratio of surface area of nanoparticles to the volume of the nanoparticles is inversely proportional to the radius of the nanoparticles. So decreasing the radius, this ratio is leading resulting in an increased rate of reaction and thus the concentration of the pollution is decreased. To achieve this objective, use of cobalt-graphene element is proposed. The proposed method is mainly to decrease the cost of platinum as it is expensive. This has a longer life than the platinum-based catalysts.Keywords: automobile emissions, catalytic converter, cobalt-graphene, replacement of platinum
Procedia PDF Downloads 3891978 Menopause Hormone Therapy: An Insight into Knowledge and Attitudes of Obstetricians and Gynecologists in Singapore
Authors: Tan Hui Ying Renee, Stella Rizalina Sasha, Farah Safdar Husain
Abstract:
Introduction: Menopause Hormone Therapy (MHT) is an effective drug indicated for the treatment of menopausal symptoms and as replacement therapy in women who undergo premature menopause. In 2020, less than 8.8% of perimenopausal Singaporean women are on hormonal therapy, as compared to the Western population, where up to 50% may be on MHT. Factors associated with MHT utilization have been studied from patient characteristics, but the impact of locally prescribing physicians resulting in low MHT utilization has yet to be evaluated. The aim of the study is to determine the level of knowledge physicians in the Obstetrics and Gynaecology specialty have and their attitudes toward MHT. We believe that knowledge of MHT is lacking and that negative attitudes towards MHT may influence its use and undermine the benefits MHT may have for women. This paper is a part of a larger study on Singaporean physicians’ knowledge and attitudes towards MHT. Methods: This is a cross-sectional study intended to assess the knowledge and attitudes of physicians toward Menopausal Hormone Therapy. An anonymous questionnaire was disseminated via institutional internal circulations to optimize reach to physicians who may prescribe MHT, particularly in the fields of Gynaecology, Family Medicine and Endocrinology. Responses were completed voluntarily. Physicians had the option for each question to declare that they were ‘unsure’ or that the question was ‘beyond their expertise’. 21 knowledge questions tested factual recall on indications, contraindications, and risks of MHT. The remaining 6 questions were clinical scenarios crafted with the intention of testing specific principles related to the use of MHT. These questions received face validation from experts in the field. 198 responses were collected, 79 of which were from physicians in the Obstetrics and Gynaecology specialty. The data will be statistically analyzed to investigate areas that can be improved to increase the overall benefits of MHT for the Singaporean population. Results: Preliminary results show that the prevailing factors that limit Singaporean gynecologists and obstetricians from prescribing MHT are a lack of knowledge of MHT and a lack of confidence in prescribing MHT. Risks and indications of MHT were not well known by many physicians, with the majority of the questions having more than 25% incorrect and ‘unsure’ as their reply. The clinical scenario questions revealed significant shortcomings in knowledge on how to navigate real-life challenges in MHT use, with 2 of 6 questions with more than 50% incorrect or ‘beyond their expertise’ as their reply. The portion of the questionnaire that investigated the attitudes of physicians showed that though a large majority believed MHT to be an effective drug, only 40.5% were confident in prescribing it. Conclusion: Physicians in the Obstetrics and Gynaecology specialty lack knowledge and confidence in MHT. Therefore, it is imperative to formulate solutions on both the individual and institutional levels to fill these gaps and ensure that MHT is used appropriately and prescribed to the patients who need it.Keywords: menopause, menopause hormone therapy, physician factors, obstetrics and gynecology, menopausal symptoms, Singapore
Procedia PDF Downloads 391977 Fabrication of Poly(Ethylene Oxide)/Chitosan/Indocyanine Green Nanoprobe by Co-Axial Electrospinning Method for Early Detection
Authors: Zeynep R. Ege, Aydin Akan, Faik N. Oktar, Betul Karademir, Oguzhan Gunduz
Abstract:
Early detection of cancer could save human life and quality in insidious cases by advanced biomedical imaging techniques. Designing targeted detection system is necessary in order to protect of healthy cells. Electrospun nanofibers are efficient and targetable nanocarriers which have important properties such as nanometric diameter, mechanical properties, elasticity, porosity and surface area to volume ratio. In the present study, indocyanine green (ICG) organic dye was stabilized and encapsulated in polymer matrix which polyethylene oxide (PEO) and chitosan (CHI) multilayer nanofibers via co-axial electrospinning method at one step. The co-axial electrospun nanofibers were characterized as morphological (SEM), molecular (FT-IR), and entrapment efficiency of Indocyanine Green (ICG) (confocal imaging). Controlled release profile of PEO/CHI/ICG nanofiber was also evaluated up to 40 hours.Keywords: chitosan, coaxial electrospinning, controlled releasing, drug delivery, indocyanine green, polyethylene oxide
Procedia PDF Downloads 1691976 Entropy Generation Analyze Due to the Steady Natural Convection of Newtonian Fluid in a Square Enclosure
Authors: T. T. Naas, Y. Lasbet, C. Kezrane
Abstract:
The thermal control in many systems is widely accomplished applying mixed convection process due to its low cost, reliability and easy maintenance. Typical applications include the aircraft electronic equipment, rotating-disc heat exchangers, turbo machinery, and nuclear reactors, etc. Natural convection in an inclined square enclosure heated via wall heater has been studied numerically. Finite volume method is used for solving momentum and energy equations in the form of stream function–vorticity. The right and left walls are kept at a constant temperature, while the other parts are adiabatic. The range of the inclination angle covers a whole revolution. The method is validated for a vertical cavity. A general power law dependence of the Nusselt number with respect to the Rayleigh number with the coefficient and exponent as functions of the inclination angle is presented. For a fixed Rayleigh number, the inclination angle increases or decreases is found.Keywords: natural convection in enclosure, inclined enclosure, Nusselt number, entropy generation analyze
Procedia PDF Downloads 2601975 Accessibility of Social Justice through Social Security in Indian Organisations: Analysis Based on Workforce
Authors: Neelima Rashmi Lakra
Abstract:
India was among one of the highly developed economy up to 1850 due to its cottage industries. During the end of the 18th century, modern industrial enterprises began with the first cotton mill in Bombay, the jute mill near Calcutta and the coal mine in Raniganj. This was counted as the real beginning of industry in 1854 in India. Prior to this period people concentrated only to agriculture, menial service or handicraft, and the introduction of industries exposed them to the disciplines of factory which was very tedious for them. With increasing number of factories been setup adding on to mining and introduction of railway, World War Period (1914-19), Second World War Period (1939-45) and the Great Depression (1929-33) there were visible change in the nature of work for the people, which resulted in outburst of strike for various reasons in these factories. Here, with India’s independence there was emergence of public sector industries and labour legislations were introduced. Meanwhile, trade unions came to notice to the rescue of the oppressed but failed to continue till long. Soon after, with the New Economic Policy organisations came across to face challenges to perform their best, where social justice for the workmen was in question. On these backdrops, studies were found discussing the central human capabilities which could be addressed through Social Security schemes. Therefore, this study was taken up to look at the reforms and legislations mainly meant for the welfare of the labour. This paper will contribute to the large number of Indian population who are serving in public sectors in India since the introduction of industries and will complement the issue of social justice through social security measures among this huge crowd serving the nation. The objectives of the study include; to find out what labour Legislations have already been existing in India, the role of Trade Union Movement, to look at the effects of New Economic Policy on these reforms and its effects and measures taken for the workforce employed in the public sectors and finally, if these measures fulfil the social justice aspects for the larger society on whole. The methodology followed collection of data from books, journal articles, reports, company reports and manuals focusing mainly on Indian studies and the data was analysed following content analysis method. The findings showed the measures taken for Social Security, but there were also reflections of very few particular additions or amendments to these Acts and provisions with the onset of New Liberalisation Policy. Therefore, the study concluded examining the social justice aspects in the context of a developing economy and discussing the recommendations.Keywords: public sectors, social justice, social security schemes, trade union movement
Procedia PDF Downloads 4501974 Effect of Fresh Concrete Curing Methods on Its Compressive Strength
Authors: Xianghe Dai, Dennis Lam, Therese Sheehan, Naveed Rehman, Jie Yang
Abstract:
Concrete is one of the most used construction materials that may be made onsite as fresh concrete and then placed in formwork to produce the desired shapes of structures. It has been recognized that the raw materials and mix proportion of concrete dominate the mechanical characteristics of hardened concrete, and the curing method and environment applied to the concrete in early stages of hardening will significantly influence the concrete properties, such as compressive strength, durability, permeability etc. In construction practice, there are various curing methods to maintain the presence of mixing water throughout the early stages of concrete hardening. They are also beneficial to concrete in hot weather conditions as they provide cooling and prevent the evaporation of water. Such methods include ponding or immersion, spraying or fogging, saturated wet covering etc. Also there are various curing methods that may be implemented to decrease the level of water lost which belongs to the concrete surface, such as putting a layer of impervious paper, plastic sheeting or membrane on the concrete to cover it. In the concrete material laboratory, accelerated strength gain methods supply the concrete with heat and additional moisture by applying live steam, coils that are subject to heating or pads that have been warmed electrically. Currently when determining the mechanical parameters of a concrete, the concrete is usually sampled from fresh concrete on site and then cured and tested in laboratories where standardized curing procedures are adopted. However, in engineering practice, curing procedures in the construction sites after the placing of concrete might be very different from the laboratory criteria, and this includes some standard curing procedures adopted in the laboratory that can’t be applied on site. Sometimes the contractor compromises the curing methods in order to reduce construction costs etc. Obviously the difference between curing procedures adopted in the laboratory and those used on construction sites might over- or under-estimate the real concrete quality. This paper presents the effect of three typical curing methods (air curing, water immersion curing, plastic film curing) and of maintaining concrete in steel moulds on the compressive strength development of normal concrete. In this study, Portland cement with 30% fly ash was used and different curing periods, 7 days, 28 days and 60 days were applied. It was found that the highest compressive strength was observed from concrete samples to which 7-day water immersion curing was applied and from samples maintained in steel moulds up to the testing date. The research results implied that concrete used as infill in steel tubular members might develop a higher strength than predicted by design assumptions based on air curing methods. Wrapping concrete with plastic film as a curing method might delay the concrete strength development in the early stages. Water immersion curing for 7 days might significantly increase the concrete compressive strength.Keywords: compressive strength, air curing, water immersion curing, plastic film curing, maintaining in steel mould, comparison
Procedia PDF Downloads 2931973 Personalized Climate Change Advertising: The Role of Augmented Reality (A.R.) Technology in Encouraging Users for Climate Change Action
Authors: Mokhlisur Rahman
Abstract:
The growing consensus among scientists and world leaders indicates that immediate action should be considered regarding the climate change phenomenon. However, climate change is no more a global issue but a personal one. Thus, individual participation is necessary to address such a significant issue. Studies show that individuals who perceive climate change as a personal issue are more likely to act toward it. This abstract presents augmented reality (A.R.) technology in the social media platform Facebook video advertising. The idea involves creating a video advertisement that enables users to interact with the video by navigating its features and experiencing the result uniquely and engagingly. This advertisement uses A.R. to bring changes, such as people making changes in real-life scenarios by simple clicks on the video and hearing an instant rewarding fact about their choices. The video shows three options: room, lawn, and driveway. Users select one option and engage in interaction based on while holding the camera in their personal spaces: Suppose users select the first option, room, and hold their camera toward spots such as by the windows, balcony, corners, and even walls. In that case, the A.R. offers users different plants appropriate for those unoccupied spaces in the room. Users can change the options of the plants and see which space at their house deserves a plant that makes it more natural. When a user adds a natural element to the video, the video content explains a piece of beneficiary information about how the user contributes to the world more to be livable and why it is necessary. With the help of A.R., if users select the second option, lawn, and hold their camera toward their lawn, the options are various small trees for their lawn to make it more environmentally friendly and decorative. The video plays a beneficiary explanation here too. Suppose users select the third option, driveway, and hold their camera toward their driveway. In that case, the A.R. video option offers unique recycle bin designs using A.I. measurement of spaces. The video plays audio information on anthropogenic contribution to greenhouse gas emission. IoT embeds tracking code in the video ad on Facebook, which stores the exact number of views in the cloud for data analysis. An online survey at the end collects short qualitative answers. This study helps understand the number of users involved and willing to change their behavior; It makes personalized advertising in social media. Considering the current state of climate change, the urgency for action is increasing. This ad increases the chance to make direct connections with individuals and gives a sense of personal responsibility for climate change to actKeywords: motivations, climate, iot, personalized-advertising, action
Procedia PDF Downloads 731972 Motivation of Doctors and its Impact on the Quality of Working Life
Authors: E. V. Fakhrutdinova, K. R. Maksimova, P. B. Chursin
Abstract:
At the present stage of the society progress the health care is an integral part of both the economic system and social, while in the second case the medicine is a major component of a number of basic and necessary social programs. Since the foundation of the health system are highly qualified health professionals, it is logical proposition that increase of doctor`s professionalism improves the effectiveness of the system as a whole. Professionalism of the doctor is a collection of many components, essential role played by such personal-psychological factors as honesty, willingness and desire to help people, and motivation. A number of researchers consider motivation as an expression of basic human needs that have passed through the “filter” which is a worldview and values learned in the process of socialization by the individual, to commit certain actions designed to achieve the expected result. From this point of view a number of researchers propose the following classification of highly skilled employee’s needs: 1. the need for confirmation the competence (setting goals that meet the professionalism and receipt of positive emotions in their decision), 2. The need for independence (the ability to make their own choices in contentious situations arising in the process carry out specialist functions), 3. The need for ownership (in the case of health care workers, to the profession and accordingly, high in the eyes of the public status of the doctor). Nevertheless, it is important to understand that in a market economy a significant motivator for physicians (both legal and natural persons) is to maximize its own profits. In the case of health professionals duality motivational structure creates an additional contrast, as in the public mind the image of the ideal physician; usually a altruistically minded person thinking is not primarily about their own benefit, and to assist others. In this context, the question of the real motivation of health workers deserves special attention. The survey conducted by the American researcher Harrison Terni for the magazine "Med Tech" in 2010 revealed the opinion of more than 200 medical students starting courses, and the primary motivation in a profession choice is "desire to help people", only 15% said that they want become a doctor, "to earn a lot". From the point of view of most of the classical theories of motivation this trend can be called positive, as intangible incentives are more effective. However, it is likely that over time the opinion of the respondents may change in the direction of mercantile motives. Thus, it is logical to assume that well-designed system of motivation of doctor`s labor should be based on motivational foundations laid during training in higher education.Keywords: motivation, quality of working life, health system, personal-psychological factors, motivational structure
Procedia PDF Downloads 3581971 Response Surface Methodology for the Optimization of Radioactive Wastewater Treatment with Chitosan-Argan Nutshell Beads
Authors: Fatima Zahra Falah, Touria El. Ghailassi, Samia Yousfi, Ahmed Moussaif, Hasna Hamdane, Mouna Latifa Bouamrani
Abstract:
The management and treatment of radioactive wastewater pose significant challenges to environmental safety and public health. This study presents an innovative approach to optimizing radioactive wastewater treatment using a novel biosorbent: chitosan-argan nutshell beads. By employing Response Surface Methodology (RSM), we aimed to determine the optimal conditions for maximum removal efficiency of radioactive contaminants. Chitosan, a biodegradable and non-toxic biopolymer, was combined with argan nutshell powder to create composite beads. The argan nutshell, a waste product from argan oil production, provides additional adsorption sites and mechanical stability to the biosorbent. The beads were characterized using Fourier Transform Infrared Spectroscopy (FTIR), Scanning Electron Microscopy (SEM), and X-ray Diffraction (XRD) to confirm their structure and composition. A three-factor, three-level Box-Behnken design was utilized to investigate the effects of pH (3-9), contact time (30-150 minutes), and adsorbent dosage (0.5-2.5 g/L) on the removal efficiency of radioactive isotopes, primarily focusing on cesium-137. Batch adsorption experiments were conducted using synthetic radioactive wastewater with known concentrations of these isotopes. The RSM analysis revealed that all three factors significantly influenced the adsorption process. A quadratic model was developed to describe the relationship between the factors and the removal efficiency. The model's adequacy was confirmed through analysis of variance (ANOVA) and various diagnostic plots. Optimal conditions for maximum removal efficiency were pH 6.8, a contact time of 120 minutes, and an adsorbent dosage of 0.8 g/L. Under these conditions, the experimental removal efficiency for cesium-137 was 94.7%, closely matching the model's predictions. Adsorption isotherms and kinetics were also investigated to elucidate the mechanism of the process. The Langmuir isotherm and pseudo-second-order kinetic model best described the adsorption behavior, indicating a monolayer adsorption process on a homogeneous surface. This study demonstrates the potential of chitosan-argan nutshell beads as an effective and sustainable biosorbent for radioactive wastewater treatment. The use of RSM allowed for the efficient optimization of the process parameters, potentially reducing the time and resources required for large-scale implementation. Future work will focus on testing the biosorbent's performance with real radioactive wastewater samples and investigating its regeneration and reusability for long-term applications.Keywords: adsorption, argan nutshell, beads, chitosan, mechanism, optimization, radioactive wastewater, response surface methodology
Procedia PDF Downloads 35