Search results for: tuning of process parameters
4128 The Impression of Adaptive Capacity of the Rural Community in the Indian Himalayan Region: A Way Forward for Sustainable Livelihood Development
Authors: Rommila Chandra, Harshika Choudhary
Abstract:
The value of integrated, participatory, and community based sustainable development strategies is eminent, but in practice, it still remains fragmentary and often leads to short-lived results. Despite the global presence of climate change, its impacts are felt differently by different communities based on their vulnerability. The developing countries have the low adaptive capacity and high dependence on environmental variables, making them highly susceptible to outmigration and poverty. We need to understand how to enable these approaches, taking into account the various governmental and non-governmental stakeholders functioning at different levels, to deliver long-term socio-economic and environmental well-being of local communities. The research assessed the financial and natural vulnerability of Himalayan networks, focusing on their potential to adapt to various changes, through accessing their perceived reactions and local knowledge. The evaluation was conducted by testing indices for vulnerability, with a major focus on indicators for adaptive capacity. Data for the analysis were collected from the villages around Govind National Park and Wildlife Sanctuary, located in the Indian Himalayan Region. The villages were stratified on the basis of connectivity via road, thus giving two kinds of human settlements connected and isolated. The study focused on understanding the complex relationship between outmigration and the socio-cultural sentiments of local people to not abandon their land, assessing their adaptive capacity for livelihood opportunities, and exploring their contribution that integrated participatory methodologies can play in delivering sustainable development. The result showed that the villages having better road connectivity, access to market, and basic amenities like health and education have a better understanding about the climatic shift, natural hazards, and a higher adaptive capacity for income generation in comparison to the isolated settlements in the hills. The participatory approach towards environmental conservation and sustainable use of natural resources were seen more towards the far-flung villages. The study helped to reduce the gap between local understanding and government policies by highlighting the ongoing adaptive practices and suggesting precautionary strategies for the community studied based on their local conditions, which differ on the basis of connectivity and state of development. Adaptive capacity in this study has been taken as the externally driven potential of different parameters, leading to a decrease in outmigration and upliftment of the human environment that could lead to sustainable livelihood development in the rural areas of Himalayas.Keywords: adaptive capacity, Indian Himalayan region, participatory, sustainable livelihood development
Procedia PDF Downloads 1184127 Constructing Masculinity through Images: Content Analysis of Lifestyle Magazines in Croatia
Authors: Marija Lončar, Zorana Šuljug Vučica, Magdalena Nigoević
Abstract:
Diverse social, cultural and economic trends and changes in contemporary societies influence the ways masculinity is represented in a variety of media. Masculinity is constructed within media images as a dynamic process that changes slowly over time and is shaped by various social factors. In many societies, dominant masculinity is still associated with authority, heterosexuality, marriage, professional and financial success, ethnic dominance and physical strength. But contemporary media depict men in ways that suggest a change in the approach to media images. The number of media images of men, which promote men’s identity through their body, have increased. With the male body more scrutinized and commodified, it is necessary to highlight how the body is represented and which visual elements are crucial since the body has an important role in the construction of masculinities. The study includes content analysis of male body images in the advertisements of different men’s and women’s lifestyle magazines available in Croatia. The main aim was to explore how masculinities are currently being portrayed through body regarding age, physical appearance, fashion, touch and gaze. The findings are also discussed in relation to female images since women are central in many of the processes constructing masculinities and according to the recent conceptualization of masculinity. Although the construction of male images varies through body features, almost all of them convey the message that men’s identity could be managed through manipulation and by enhancing the appearance. Furthermore, they suggest that men should engage in “bodywork” through advertised products, activities and/or practices, in order to achieve their preferred social image.Keywords: body images, content analysis, lifestyle magazines, masculinity
Procedia PDF Downloads 2454126 Characterization of 2,4,6-Trinitrotoluene (Tnt)-Metabolizing Bacillus Cereus Sp TUHP2 Isolated from TNT-Polluted Soils in the Vellore District, Tamilnadu, India
Authors: S. Hannah Elizabeth, A. Panneerselvam
Abstract:
Objective: The main objective was to evaluate the degradative properties of Bacillus cereus sp TUHP2 isolated from TNT-Polluted soils in the Vellore District, Tamil Nadu, India. Methods: Among the 3 bacterial genera isolated from different soil samples, one potent TNT degrading strain Bacillus cereus sp TUHP2 was identified. The morphological, physiological and the biochemical properties of the strain Bacillus cereus sp TUHP2 was confirmed by conventional methods and genotypic characterization was carried out using 16S r-DNA partial gene amplification and sequencing. The broken down by products of DNT in the extract was determined by Gas Chromatogram- Mass spectrometry (GC-MS). Supernatant samples from the broth studied at 24 h interval were analyzed by HPLC analysis and the effect on various nutritional and environmental factors were analysed and optimized for the isolate. Results: Out of three isolates one strain TUHP2 were found to have potent efficiency to degrade TNT and revealed the genus Bacillus. 16S rDNA gene sequence analysis showed highest homology (98%) with Bacillus cereus and was assigned as Bacillus cereus sp TUHP2. Based on the energy of the predicted models, the secondary structure predicted by MFE showed the more stable structure with a minimum energy. Products of TNT Transformation showed colour change in the medium during cultivation. TNT derivates such as 2HADNT and 4HADNT were detected by HPLC chromatogram and 2ADNT, 4ADNT by GC/MS analysis. Conclusion: Hence this study presents the clear evidence for the biodegradation process of TNT by strain Bacillus cereus sp TUHP2.Keywords: bioremediation, biodegradation, biotransformation, sequencing
Procedia PDF Downloads 4624125 Success Factors for Innovations in SME Networks
Authors: J. Gochermann
Abstract:
Due to complex markets and products, and increasing need to innovate, cooperation between small and medium size enterprises arose during the last decades, which are not prior driven by process optimization or sales enhancement. Especially small and medium sized enterprises (SME) collaborate increasingly in innovation and knowledge networks to enhance their knowledge and innovation potential, and to find strategic partners for product and market development. These networks are characterized by dual objectives, the superordinate goal of the total network, and the specific objectives of the network members, which can cause target conflicts. Moreover, most SMEs do not have structured innovation processes and they are not accustomed to collaborate in complex innovation projects in an open network structure. On the other hand, SMEs have suitable characteristics for promising networking. They are flexible and spontaneous, they have flat hierarchies, and the acting people are not anonymous. These characteristics indeed distinguish them from bigger concerns. Investigation of German SME networks have been done to identify success factors for SME innovation networks. The fundamental network principles, donation-return and confidence, could be confirmed and identified as basic success factors. Further factors are voluntariness, adequate number of network members, quality of communication, neutrality and competence of the network management, as well as reliability and obligingness of the network services. Innovation and knowledge networks with an appreciable number of members from science and technology institutions need also active sense-making to bring different disciplines into successful collaboration. It has also been investigated, whether and how the involvement in an innovation network impacts the innovation structure and culture inside the member companies. The degree of reaction grows with time and intensity of commitment.Keywords: innovation and knowledge networks, SME, success factors, innovation structure and culture
Procedia PDF Downloads 2834124 Management Problems in a Patient With Long-term Undiagnosed Permanent Hypoparathyroidism
Authors: Babarina Maria, Andropova Margarita
Abstract:
Introduction: Hypoparathyroidism (HypoPT) is a rare endocrine disorder with an estimated prevalence of 0.25 per 1000 individuals. The most common cause of HypoPT is the loss of active parathyroid tissue following thyroid or parathyroid surgery. Sometimes permanent postoperative HypoPT occures, manifested by hypocalcemia in combination with low levels of PTH during 6 months or more after surgery. Cognitive impairments in patients with hypocalcemia due to chronic HypoPT are observed, and this can lead to problems and challenges in everyday living: memory loss and impaired concentration, that may be the cause of poor compliance. Clinical case: Patient K., 66 years old, underwent thyroidectomy in 2013 (at the age of 55) because of papillary thyroid cancer T1NxMx, histopathology findings confirmed the diagnosis. 5 years after the surgery, she was followed up on an outpatient basis, TSH levelsonly were monitored, and the dose of levothyroxine was adjusted. In 2018 due to, increasing complaints include tingling and cramps in the arms and legs, memory loss, sleep disorder, fatigue, anxiety, hair loss, muscle pain, tachycardia, positive Chvostek, and Trousseau signs were diagnosed during examination, also in blood analyses: total Ca 1.86 mmol/l (2.15-2.55), Ca++ 0.96 mmol/l (1.12-1.3), P 1.55 mmol/l (0.74-1.52), Mg 0.79 mmol/l (0.66-1.07) - chronic postoperative HypoPT was diagnosed. Therapy was initiated: alfacalcidol 0.5 mcg per day, calcium carbonate 2000 mg per day, cholecalciferol 1000 IU per day, magnesium orotate 3000 mg per day. During the case follow-up, hypocalcemia, hyperphosphatemia persisted, hypercalciuria15.7 mmol/day (2.5-6.5) was diagnosed. Dietary recommendations were given because of the high content of phosphorus rich foods, and therapy was adjusted: the dose of alfacalcidol was increased to 2.5 mcg per day, and the dose of calcium carbonate was reduced to 1500 mg per day. As part of the screening for complications of hypoPT, data for cataracts, Fahr syndrome, nephrocalcinosis, and kidney stone disease were not obtained. However, HypoPT compensation was not achieved, and therefore hydrochlorothiazide 25 mg was initiated, the dose of alfacalcidol was increased to 3 mcg per day, calcium carbonate to 3000 mg per day, magnesium orotate and cholecalciferol were continued at the same doses. Therapeutic goals were achieved: calcium phosphate product <4.4 mmol2/l2, there were no episodes of hypercalcemia, twenty-four-hour urinary calcium excretion was significantly reduced. Conclusion: Timely prescription, careful explanation of drugs usage rules, and monitoring and maintaining blood and urine parameters within the target contribute to the prevention of HypoPT complications development and life-threatening events.Keywords: hypoparathyroidism, hypocalcemia, hyperphosphatemia, hypercalciuria
Procedia PDF Downloads 1084123 The Feasibility of Anaerobic Digestion at 45⁰C
Authors: Nuruol S. Mohd, Safia Ahmed, Rumana Riffat, Baoqiang Li
Abstract:
Anaerobic digestion at mesophilic and thermophilic temperatures have been widely studied and evaluated by numerous researchers. Limited extensive research has been conducted on anaerobic digestion in the intermediate zone of 45°C, mainly due to the notion that limited microbial activity occurs within this zone. The objectives of this research were to evaluate the performance and the capability of anaerobic digestion at 45°C in producing class A biosolids, in comparison to a mesophilic and thermophilic anaerobic digestion system operated at 35°C and 55°C, respectively. In addition to that, the investigation on the possible inhibition factors affecting the performance of the digestion system at this temperature will be conducted as well. The 45°C anaerobic digestion systems were not able to achieve comparable methane yield and high-quality effluent compared to the mesophilic system, even though the systems produced biogas with about 62-67% methane. The 45°C digesters suffered from high acetate accumulation, but sufficient buffering capacity was observed as the pH, alkalinity and volatile fatty acids (VFA)-to-alkalinity ratio were within recommended values. The accumulation of acetate observed in 45°C systems were presumably due to the high temperature which contributed to high hydrolysis rate. Consequently, it produced a large amount of toxic salts that combined with the substrate making them not readily available to be consumed by methanogens. Acetate accumulation, even though contributed to 52 to 71% reduction in acetate degradation process, could not be considered as completely inhibitory. Additionally, at 45°C, no ammonia inhibition was observed and the digesters were able to achieve volatile solids (VS) reduction of 47.94±4.17%. The pathogen counts were less than 1,000 MPN/g total solids, thus, producing Class A biosolids.Keywords: 45°C anaerobic digestion, acetate accumulation, class A biosolids, salt toxicity
Procedia PDF Downloads 3044122 Occurrence of Foreign Matter in Food: Applied Identification Method - Association of Official Agricultural Chemists (AOAC) and Food and Drug Administration (FDA)
Authors: E. C. Mattos, V. S. M. G. Daros, R. Dal Col, A. L. Nascimento
Abstract:
The aim of this study is to present the results of a retrospective survey on the foreign matter found in foods analyzed at the Adolfo Lutz Institute, from July 2001 to July 2015. All the analyses were conducted according to the official methods described on Association of Official Agricultural Chemists (AOAC) for the micro analytical procedures and Food and Drug Administration (FDA) for the macro analytical procedures. The results showed flours, cereals and derivatives such as baking and pasta products were the types of food where foreign matters were found more frequently followed by condiments and teas. Fragments of stored grains insects, its larvae, nets, excrement, dead mites and rodent excrement were the most foreign matter found in food. Besides, foreign matters that can cause a physical risk to the consumer’s health such as metal, stones, glass, wood were found but rarely. Miscellaneous (shell, sand, dirt and seeds) were also reported. There are a lot of extraneous materials that are considered unavoidable since are something inherent to the product itself, such as insect fragments in grains. In contrast, there are avoidable extraneous materials that are less tolerated because it is preventable with the Good Manufacturing Practice. The conclusion of this work is that although most extraneous materials found in food are considered unavoidable it is necessary to keep the Good Manufacturing Practice throughout the food processing as well as maintaining a constant surveillance of the production process in order to avoid accidents that may lead to occurrence of these extraneous materials in food.Keywords: extraneous materials, food contamination, foreign matter, surveillance
Procedia PDF Downloads 3594121 Surfactant-Modified Chitosan Beads: An Efficient and Cost Effective Material for Adsorptive Removal of Lead from Aqueous Solutions
Authors: Preeti Pal, Anjali Pal
Abstract:
Chitosan is an effective sorbent for removal of contaminants from wastewater. However, the ability of pure chitosan is specific because of its cationic charge. It causes repulsion in the removal process of various cationic charged molecules. The present study has been carried out for the successful removal of Pb²⁺ ions from aqueous solution by modified chitosan beads. Surface modification of chitosan (CS) beads was performed by using the anionic surfactant (AS), sodium dodecyl sulfate (SDS). Micelle aggregation property of SDS has been utilized for the formation of bilayer over the CS beads to produce surfactant modified chitosan (SMCS) beads. Prepared adsorbents were characterized by Fourier transform infrared spectroscopy (FTIR) and scanning electron microscopy (SEM) in order to find out their composition and surface morphology. SMCS beads, when compared to the pure CS beads, showed three times higher adsorption. This higher adsorption is believed to be due to the adsolubilization of Pb²⁺ ions on SDS bilayer. This bilayer provides more adsorption sites for quick and effective removal of Pb²⁺ ions from the aqueous phase. Moreover, the kinetic and adsorption isotherm models were employed to the obtained data for the description of the lead adsorption processes. It was found that the removal kinetics follows pseudo-second order model. Adsorption isotherm data fits well to the Langmuir model. The maximum adsorption capacity obtained is 100 mg/g at the dosage of 0.675 g/L for 50 mg/L of Pb²⁺. The adsorption capacity is subject to increase with increasing the Pb²⁺ ions concentration in the solution. The results indicated that the prepared hydrogel beads are efficient adsorbent for removal of Pb²⁺ ions from the aqueous medium.Keywords: adsolubilisation, anionic surfactant, bilayer, chitosan, Pb²⁺
Procedia PDF Downloads 2404120 A Statistical Analysis on the Comparison of First and Second Waves of COVID-19 and Importance of Early Actions in Public Health for Third Wave in India
Authors: Maitri Dave
Abstract:
Coronaviruses (CoV) is such infectious virus which has impacted globally in a more dangerous manner causing severe lung problems and leaving behind more serious diseases among the people. This pandemic has affected globally and created severe respiratory problems, and damaged the lungs. India has reported the first case of COVID-19 in January 2020. The first wave of COVID-19 took place from April to September of 2020. Soon after, a second peak is also noticed in the month of March 2021, which in turn becomes more dangerous due to a lack of supply of medical equipment. It created resource deficiency globally, specifically in India, where some necessary life-saving equipment like ventilators and oxygenators were not sufficient to cater to the demand-supply ratio effectively. Through carefully examining such a situation, India began to execute the process of vaccination in the month of January 2021 and successfully administered 25,46,71,259 doses of vaccines till now, which is only 15.5% of the total population while only 3.6% of the total population is fully vaccinated. India has authorized the British Oxford–AstraZeneca vaccine (Covishield), the Indian BBV152 (Covaxin) vaccine, and the Russian Sputnik V vaccine for emergency use. In the present study, we have collected all the data state wisely of both first and second wave and analyzed them using MS Excel Version 2019 and SPSS Statistics Version 26. Following the trends, we have predicted the characteristics of the upcoming third wave of COVID-19 and recommended some strategies, early actions, and measures that can be taken by the public health system in India to combat the third wave more effectively.Keywords: COVID-19, vaccination, Covishiled, Coronavirus
Procedia PDF Downloads 2164119 Using Artificial Intelligence Technology to Build the User-Oriented Platform for Integrated Archival Service
Authors: Lai Wenfang
Abstract:
Tthis study will describe how to use artificial intelligence (AI) technology to build the user-oriented platform for integrated archival service. The platform will be launched in 2020 by the National Archives Administration (NAA) in Taiwan. With the progression of information communication technology (ICT) the NAA has built many systems to provide archival service. In order to cope with new challenges, such as new ICT, artificial intelligence or blockchain etc. the NAA will try to use the natural language processing (NLP) and machine learning (ML) skill to build a training model and propose suggestions based on the data sent to the platform. NAA expects the platform not only can automatically inform the sending agencies’ staffs which records catalogues are against the transfer or destroy rules, but also can use the model to find the details hidden in the catalogues and suggest NAA’s staff whether the records should be or not to be, to shorten the auditing time. The platform keeps all the users’ browse trails; so that the platform can predict what kinds of archives user could be interested and recommend the search terms by visualization, moreover, inform them the new coming archives. In addition, according to the Archives Act, the NAA’s staff must spend a lot of time to mark or remove the personal data, classified data, etc. before archives provided. To upgrade the archives access service process, the platform will use some text recognition pattern to black out automatically, the staff only need to adjust the error and upload the correct one, when the platform has learned the accuracy will be getting higher. In short, the purpose of the platform is to deduct the government digital transformation and implement the vision of a service-oriented smart government.Keywords: artificial intelligence, natural language processing, machine learning, visualization
Procedia PDF Downloads 1744118 Design and Implementation of Low-code Model-building Methods
Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu
Abstract:
This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment
Procedia PDF Downloads 314117 Synthesis of Montmorillonite/CuxCd1-xS Nanocomposites and Their Application to the Photodegradation of Methylene Blue
Authors: H. Boukhatem, L. Djouadi, H. Khalaf, R. M. Navarro, F. V. Ganzalez
Abstract:
Synthetic organic dyes are used in various industries, such as textile industry, leather tanning industry, paper production, hair dye production, etc. Wastewaters containing these dyes may be harmful to the environment and living organisms. Therefore, it is very important to remove or degrade these dyes before discharging them into the environment. In addition to standard technologies for the degradation and/or removal of dyes, several new specific technologies, the so-called advanced oxidation processes (AOPs), have been developed to eliminate dangerous compounds from polluted waters. AOPs are all characterized by the same chemical feature: production of radicals (•OH) through a multistep process, although different reaction systems are used. These radicals show little selectivity of attack and are able to oxidize various organic pollutants due to their high oxidative capacity (reduction potential of HO• Eo = 2.8 V). Heterogeneous photocatalysis, as one of the AOPs, could be effective in the oxidation/degradation of organic dyes. A major advantage of using heterogeneous photocatalysis for this purpose is the total mineralization of organic dyes, which results in CO2, H2O and corresponding mineral acids. In this study, nanomaterials based on montmorillonite and CuxCd1-xS with different Cu concentration (0.3 < x < 0.7) were utilized for the degradation of the commercial cationic textile dye Methylene blue (MB), used as a model pollutant. The synthesized nanomaterials were characterized by fourier transform infrared (FTIR) and thermogravimetric-differential thermal analysis (TG–DTA). Test results of photocatalysis of methylene blue under UV-Visible irradiation show that the photoactivity of nanomaterials montmorillonite/ CuxCd1-xS increases with the increasing of Cu concentration. The kinetics of the degradation of the MB dye was described with the Langmuir–Hinshelwood (L–H) kinetic model.Keywords: heterogeneous photocatalysis, methylene blue, montmorillonite, nanomaterial
Procedia PDF Downloads 3744116 Design of UV Based Unicycle Robot to Disinfect Germs and Communicate With Multi-Robot System
Authors: Charles Koduru, Parth Patel, M. Hassan Tanveer
Abstract:
In this paper, the communication between a team of robots is used to sanitize an environment with germs is proposed. We introduce capabilities from a team of robots (most likely heterogeneous), a wheeled robot named ROSbot 2.0 that consists of a mounted LiDAR and Kinect sensor, and a modified prototype design of a unicycle-drive Roomba robot called the UV robot. The UV robot consists of ultrasonic sensors to avoid obstacles and is equipped with an ultraviolet light system to disinfect and kill germs, such as bacteria and viruses. In addition, the UV robot is equipped with disinfectant spray to target hidden objects that ultraviolet light is unable to reach. Using the sensors from the ROSbot 2.0, the robot will create a 3-D model of the environment which will be used to factor how the ultraviolet robot will disinfect the environment. Together this proposed system is known as the RME assistive robot device or RME system, which communicates between a navigation robot and a germ disinfecting robot operated by a user. The RME system includes a human-machine interface that allows the user to control certain features of each robot in the RME assistive robot device. This method allows the cleaning process to be done at a more rapid and efficient pace as the UV robot disinfects areas just by moving around in the environment while using the ultraviolet light system to kills germs. The RME system can be used in many applications including, public offices, stores, airports, hospitals, and schools. The RME system will be beneficial even after the COVID-19 pandemic. The Kennesaw State University will continue the research in the field of robotics, engineering, and technology and play its role to serve humanity.Keywords: multi robot system, assistive robots, COVID-19 pandemic, ultraviolent technology
Procedia PDF Downloads 1864115 Comparison of Methodologies to Compute the Probabilistic Seismic Hazard Involving Faults and Associated Uncertainties
Authors: Aude Gounelle, Gloria Senfaute, Ludivine Saint-Mard, Thomas Chartier
Abstract:
The long-term deformation rates of faults are not fully captured by Probabilistic Seismic Hazard Assessment (PSHA). PSHA that use catalogues to develop area or smoothed-seismicity sources is limited by the data available to constraint future earthquakes activity rates. The integration of faults in PSHA can at least partially address the long-term deformation. However, careful treatment of fault sources is required, particularly, in low strain rate regions, where estimated seismic hazard levels are highly sensitive to assumptions concerning fault geometry, segmentation and slip rate. When integrating faults in PSHA various constraints on earthquake rates from geologic and seismologic data have to be satisfied. For low strain rate regions where such data is scarce it would be especially challenging. Faults in PSHA requires conversion of the geologic and seismologic data into fault geometries, slip rates and then into earthquake activity rates. Several approaches exist for translating slip rates into earthquake activity rates. In the most frequently used approach, the background earthquakes are handled using a truncated approach, in which earthquakes with a magnitude lower or equal to a threshold magnitude (Mw) occur in the background zone, with a rate defined by the rate in the earthquake catalogue. Although magnitudes higher than the threshold are located on the fault with a rate defined using the average slip rate of the fault. As high-lighted by several research, seismic events with magnitudes stronger than the selected magnitude threshold may potentially occur in the background and not only at the fault, especially in regions of slow tectonic deformation. It also has been known that several sections of a fault or several faults could rupture during a single fault-to-fault rupture. It is then essential to apply a consistent modelling procedure to allow for a large set of possible fault-to-fault ruptures to occur aleatory in the hazard model while reflecting the individual slip rate of each section of the fault. In 2019, a tool named SHERIFS (Seismic Hazard and Earthquake Rates in Fault Systems) was published. The tool is using a methodology to calculate the earthquake rates in a fault system where the slip-rate budget of each fault is conversed into rupture rates for all possible single faults and faultto-fault ruptures. The objective of this paper is to compare the SHERIFS method with one other frequently used model to analyse the impact on the seismic hazard and through sensibility studies better understand the influence of key parameters and assumptions. For this application, a simplified but realistic case study was selected, which is in an area of moderate to hight seismicity (South Est of France) and where the fault is supposed to have a low strain.Keywords: deformation rates, faults, probabilistic seismic hazard, PSHA
Procedia PDF Downloads 664114 Object Recognition System Operating from Different Type Vehicles Using Raspberry and OpenCV
Authors: Maria Pavlova
Abstract:
In our days, it is possible to put the camera on different vehicles like quadcopter, train, airplane and etc. The camera also can be the input sensor in many different systems. That means the object recognition like non separate part of monitoring control can be key part of the most intelligent systems. The aim of this paper is to focus of the object recognition process during vehicles movement. During the vehicle’s movement the camera takes pictures from the environment without storage in Data Base. In case the camera detects a special object (for example human or animal), the system saves the picture and sends it to the work station in real time. This functionality will be very useful in emergency or security situations where is necessary to find a specific object. In another application, the camera can be mounted on crossroad where do not have many people and if one or more persons come on the road, the traffic lights became the green and they can cross the road. In this papers is presented the system has solved the aforementioned problems. It is presented architecture of the object recognition system includes the camera, Raspberry platform, GPS system, neural network, software and Data Base. The camera in the system takes the pictures. The object recognition is done in real time using the OpenCV library and Raspberry microcontroller. An additional feature of this library is the ability to display the GPS coordinates of the captured objects position. The results from this processes will be sent to remote station. So, in this case, we can know the location of the specific object. By neural network, we can learn the module to solve the problems using incoming data and to be part in bigger intelligent system. The present paper focuses on the design and integration of the image recognition like a part of smart systems.Keywords: camera, object recognition, OpenCV, Raspberry
Procedia PDF Downloads 2184113 The Visual Side of Islamophobia: A Social-Semiotic Analysis
Authors: Carmen Aguilera-Carnerero
Abstract:
Islamophobia, the unfounded hostility towards Muslims and Islam, has been deeply studied in the last decades from different perspectives ranging from anthropology, sociology, media studies, and linguistics. In the past few years, we have witnessed how the birth of social media has transformed formerly passive audiences into an active group that not only receives and digests information but also creates and comments publicly on any event of their interest. In this way, average citizens now have been entitled with the power of becoming potential opinion leaders. This rise of social media in the last years gave way to a different way of Islamophobia, the so called ‘cyberIslamophobia’. Considerably less attention, however, has been given to the study of islamophobic images that accompany the texts in social media. This paper attempts to analyse a corpus of 300 images of islamophobic nature taken from social media (from Twitter and Facebook) from the years 2014-2017 to see: a) how hate speech is visually constructed, b) how cyberislamophobia is articulated through images and whether there are differences/similarities between the textual and the visual elements, c) the impact of those images in the audience and their reaction to it and d) whether visual cyberislamophobia has undergone any process of permeating popular culture (for example, through memes) and its real impact. To carry out this task, we have used Critical Discourse Analysis as the most suitable theoretical framework that analyses and criticizes the dominant discourses that affect inequality, injustice, and oppression. The analysis of images was studied according to the theoretical framework provided by the visual framing theory and the visual design grammar to conclude that memes are subtle but very powerful tools to spread Islamophobia and foster hate speech under the guise of humour within popular culture.Keywords: cyberIslamophobia, visual grammar, social media, popular culture
Procedia PDF Downloads 1674112 Anti-Corruption, an Important Challenge for the Construction Industry!
Authors: Ahmed Stifi, Sascha Gentes, Fritz Gehbauer
Abstract:
The construction industry is perhaps one of the oldest industry of the world. The ancient monuments like the egyptian pyramids, the temples of Greeks and Romans like Parthenon and Pantheon, the robust bridges, old Roman theatres, the citadels and many more are the best testament to that. The industry also has a symbiotic relationship with other . Some of the heavy engineering industry provide construction machineries, chemical industry develop innovative construction materials, finance sector provides fund solutions for complex construction projects and many more. Construction Industry is not only mammoth but also very complex in nature. Because of the complexity, construction industry is prone to various tribulations which may have the propensity to hamper its growth. The comparitive study of this industry with other depicts that it is associated with a state of tardiness and delay especially when we focus on the managerial aspects and the study of triple constraint (time, cost and scope). While some institutes says the complexity associated with it as a major reason, others like lean construction, refers to the wastes produced across the construction process as the prime reason. This paper introduces corruption as one of the prime factors for such delays.To support this many international reports and studies are available depicting that construction industry is one of the most corrupt sectors worldwide, and the corruption can take place throught the project cycle comprising project selection, planning, design, funding, pre-qualification, tendering, execution, operation and maintenance, and even through the reconstrction phase. It also happens in many forms such as bribe, fraud, extortion, collusion, embezzlement and conflict of interest and the self-sufficient. As a solution to cope the corruption in construction industry, the paper introduces the integrity as a key factor and build a new integrity framework to develop and implement an integrity management system for construction companies and construction projects.Keywords: corruption, construction industry, integrity, lean construction
Procedia PDF Downloads 3774111 Robust Inference with a Skew T Distribution
Authors: M. Qamarul Islam, Ergun Dogan, Mehmet Yazici
Abstract:
There is a growing body of evidence that non-normal data is more prevalent in nature than the normal one. Examples can be quoted from, but not restricted to, the areas of Economics, Finance and Actuarial Science. The non-normality considered here is expressed in terms of fat-tailedness and asymmetry of the relevant distribution. In this study a skew t distribution that can be used to model a data that exhibit inherent non-normal behavior is considered. This distribution has tails fatter than a normal distribution and it also exhibits skewness. Although maximum likelihood estimates can be obtained by solving iteratively the likelihood equations that are non-linear in form, this can be problematic in terms of convergence and in many other respects as well. Therefore, it is preferred to use the method of modified maximum likelihood in which the likelihood estimates are derived by expressing the intractable non-linear likelihood equations in terms of standardized ordered variates and replacing the intractable terms by their linear approximations obtained from the first two terms of a Taylor series expansion about the quantiles of the distribution. These estimates, called modified maximum likelihood estimates, are obtained in closed form. Hence, they are easy to compute and to manipulate analytically. In fact the modified maximum likelihood estimates are equivalent to maximum likelihood estimates, asymptotically. Even in small samples the modified maximum likelihood estimates are found to be approximately the same as maximum likelihood estimates that are obtained iteratively. It is shown in this study that the modified maximum likelihood estimates are not only unbiased but substantially more efficient than the commonly used moment estimates or the least square estimates that are known to be biased and inefficient in such cases. Furthermore, in conventional regression analysis, it is assumed that the error terms are distributed normally and, hence, the well-known least square method is considered to be a suitable and preferred method for making the relevant statistical inferences. However, a number of empirical researches have shown that non-normal errors are more prevalent. Even transforming and/or filtering techniques may not produce normally distributed residuals. Here, a study is done for multiple linear regression models with random error having non-normal pattern. Through an extensive simulation it is shown that the modified maximum likelihood estimates of regression parameters are plausibly robust to the distributional assumptions and to various data anomalies as compared to the widely used least square estimates. Relevant tests of hypothesis are developed and are explored for desirable properties in terms of their size and power. The tests based upon modified maximum likelihood estimates are found to be substantially more powerful than the tests based upon least square estimates. Several examples are provided from the areas of Economics and Finance where such distributions are interpretable in terms of efficient market hypothesis with respect to asset pricing, portfolio selection, risk measurement and capital allocation, etc.Keywords: least square estimates, linear regression, maximum likelihood estimates, modified maximum likelihood method, non-normality, robustness
Procedia PDF Downloads 3974110 Public-Private Partnership for Critical Infrastructure Resilience
Authors: Anjula Negi, D. T. V. Raghu Ramaswamy, Rajneesh Sareen
Abstract:
Road infrastructure is emphatically one of the top most critical infrastructure to the Indian economy. Road network in the country of around 3.3 million km is the second largest in the world. Nationwide statistics released by Ministry of Road, Transport and Highways reveal that every minute an accident happens and one death every 3.7 minutes. This reported scale in terms of safety is a matter of grave concern, and economically represents a national loss of 3% to the GDP. Union Budget 2016-17 has allocated USD 12 billion annually for development and strengthening of roads, an increase of 56% from last year. Thus, highlighting the importance of roads as critical infrastructure. National highway alone represent only 1.7% of the total road linkages, however, carry over 40% of traffic. Further, trends analysed from 2002 -2011 on national highways, indicate that in less than a decade, a 22 % increase in accidents have been reported, but, 68% increase in death fatalities. Paramount inference is that accident severity has increased with time. Over these years many measures to increase road safety, lessening damage to physical assets, reducing vulnerabilities leading to a build-up for resilient road infrastructure have been taken. In the context of national highway development program, policy makers proposed implementation of around 20 % of such road length on PPP mode. These roads were taken up on high-density traffic considerations and for qualitative implementation. In order to understand resilience impacts and safety parameters, enshrined in various PPP concession agreements executed with the private sector partners, such highway specific projects would be appraised. This research paper would attempt to assess such safety measures taken and the possible reasons behind an increase in accident severity through these PPP case study projects. Delving further on safety features to understand policy measures adopted in these cases and an introspection on reasons of severity, whether an outcome of increased speeds, faulty road design and geometrics, driver negligence, or due to lack of discipline in following lane traffic with increased speed. Assessment exercise would study these aspects hitherto to PPP and post PPP project structures, based on literature review and opinion surveys with sectoral experts. On the way forward, it is understood that the Ministry of Road, Transport and Highway’s estimate for strengthening the national highway network is USD 77 billion within next five years. The outcome of this paper would provide an understanding of resilience measures adopted, possible options for accessible and safe road network and its expansion to policy makers for possible policy initiatives and funding allocation in securing critical infrastructure.Keywords: national highways, policy, PPP, safety
Procedia PDF Downloads 2574109 DIF-JACKET: a Thermal Protective Jacket for Firefighters
Authors: Gilda Santos, Rita Marques, Francisca Marques, João Ribeiro, André Fonseca, João M. Miranda, João B. L. M. Campos, Soraia F. Neves
Abstract:
Every year, an unacceptable number of firefighters are seriously burned during firefighting operations, with some of them eventually losing their life. Although thermal protective clothing research and development has been searching solutions to minimize firefighters heat load and skin burns, currently commercially available solutions focus in solving isolated problems, for example, radiant heat or water-vapor resistance. Therefore, episodes of severe burns and heat strokes are still frequent. Taking this into account, a consortium composed by Portuguese entities has joined synergies to develop an innovative protective clothing system by following a procedure based on the application of numerical models to optimize the design and using a combinationof protective clothing components disposed in different layers. Recently, it has been shown that Phase Change Materials (PCMs) can contribute to the reduction of potential heat hazards in fire extinguish operations, and consequently, their incorporation into firefighting protective clothing has advantages. The greatest challenge is to integrate these materials without compromising garments ergonomics and, at the same time, accomplishing the International Standard of protective clothing for firefighters – laboratory test methods and performance requirements for wildland firefighting clothing. The incorporation of PCMs into the firefighter's protective jacket will result in the absorption of heat from the fire and consequently increase the time that the firefighter can be exposed to it. According to the project studies and developments, to favor a higher use of the PCM storage capacityand to take advantage of its high thermal inertia more efficiently, the PCM layer should be closer to the external heat source. Therefore, in this stage, to integrate PCMs in firefighting clothing, a mock-up of a vest specially designed to protect the torso (back, chest and abdomen) and to be worn over a fire-resistant jacketwas envisaged. Different configurations of PCMs, as well as multilayer approaches, were studied using suitable joining technologies such as bonding, ultrasound, and radiofrequency. Concerning firefighter’s protective clothing, it is important to balance heat protection and flame resistance with comfort parameters, namely, thermaland water-vapor resistances. The impact of the most promising solutions regarding thermal comfort was evaluated to refine the performance of the global solutions. Results obtained with experimental bench scale model and numerical simulation regarding the integration of PCMs in a vest designed as protective clothing for firefighters will be presented.Keywords: firefighters, multilayer system, phase change material, thermal protective clothing
Procedia PDF Downloads 1634108 Heritage Impact Assessment Policy within Western Balkans, Albania
Authors: Anisa Duraj
Abstract:
As usually acknowledged, cultural heritage is the weakest component in EIA studies. The role of heritage impact assessment (HIA) in development projects is not often accounted for, and in those cases where it is, HIA is considered as a reactive response and not as a solutions provider. Because of continuous development projects, in most cases, heritage is unconsidered and often put under threat. Cultural protection and development challenges ask for prudent legal regulation and appropriate policy implementation. The challenges become even more peculiar in underdeveloped countries or endangered areas, which are generally characterized by numerous legal constraints. Therefore, the need for strategic proposals for HIA is of high importance. In order to trigger HIA as a proactive operation in the IA process and make sure to cover cultural heritage in the whole EIA framework, an appropriate system of evaluation of impacts should be provided. To obtain the required results for HIA, this last must be part of a regional policy, which will address and guide development projects toward a proper evaluation of their impacts affecting heritage. In order to get a clearer picture of existing gabs but also new possibilities for HIA, this paper will focus on the Western Balkans region and the undergoing changes that it faces. Concerning continuous development pressure in the region and within the aspiration of the Western Balkans countries to join the European Union (EU) as member states, attention should be paid to new development policies under the EU directives for conducting EIAs, and accurate support is required for the restructuration of existing policies as well as for the implementation of the UN Agenda for SDGs. In the framework of new emerging needs, if HIA is taken into account, the outcome would be an inclusive regional program that would help to overcome marginality issues of spaces and people.Keywords: cultural heritage, impact assessment, SDGs, urban development, western Balkans, regional policy, HIA, EIA
Procedia PDF Downloads 1144107 Madness in Susanna Kaysen’s Girl, Interrupted: A Focouldian Reading
Authors: Somaye Sabetnia
Abstract:
This paper is accomplished to probe Susanna Kaysen’s memoir Girl, Interrupted in the light of Michel Foucault’s theory of madness comprehensively set forth in his History of Madness (1961). It is an endeavor to analysis this novel based on Foucault’s idea of madness. In his archeological study of madness, Foucault introduces a way to perceive madness and its association with dominant discourses. He argues that the concept of madness is constructed within the social context, and different institutions affect its definition. Furthermore, he takes into consideration how each era treats madness, and affirms that in modern times, people considered mad are exiled out of cities, confined in madhouses, and later in clinics where they are treated with drugs. Set after World War II, the novel under observation highlights women’s conditions in which they were becoming a housewife or following their own desires; in fact, choosing the second one results in labeling mad. The protagonist of novel is labeled 'mad,' and is hence impelled to go to asylums where so-called patients are under the vigilant surveillance of the authorities to go through the process of 'normalization.' To discern how she is considered 'mad,' this article probes the dominant discourse of the time when the stories take place to provide a better understanding of madness under the impact of social, cultural, and political conditions. It examines how a so-called mad considered 'Other' and treated after being confined by the disciplinary system of the asylum in a panoptic world. In addition to, it describes the aim of treatment is to punish and control a patient not to cure. This article aims to indicate that Susanna Kaysen tries to picture what is defined as women’s madness is the result of the patriarchal society of the post-war America as well as the mental illness has nothing to do with blood; it is rather the result of the social inequality of the age.Keywords: clinical treatment, disciplining and punishment, dominant discourse, normalization, other, panoptic world, reason vs. unreason
Procedia PDF Downloads 3214106 Characterizing Surface Machining-Induced Local Deformation Using Electron Backscatter Diffraction
Authors: Wenqian Zhang, Xuelin Wang, Yujin Hu, Siyang Wang
Abstract:
The subsurface layer of a component plays a significant role in its service performance. Any surface mechanical process during fabrication can introduce a deformed layer near the surface, which can be related to the microstructure alteration and strain hardening, and affects the mechanical properties and corrosion resistance of the material. However, there exists a great difficulty in determining the subsurface deformation induced by surface machining. In this study, electron backscatter diffraction (EBSD) was used to study the deformed layer of surface milled 316 stainless steel. The microstructure change was displayed by the EBSD maps and characterized by misorientation variation. The results revealed that the surface milling resulted in heavily nonuniform deformations in the subsurface layer and even in individual grains. The direction of the predominant grain deformation was about 30-60 deg to the machined surface. Moreover, a local deformation rate (LDR) was proposed to quantitatively evaluate the local deformation degree. Both of the average and maximum LDRs were utilized to characterize the deformation trend along the depth direction. It was revealed that the LDR had a strong correlation with the development of grain and sub-grain boundaries. In this work, a scan step size of 1.2 μm was chosen for the EBSD measurement. A LDR higher than 18 deg/μm indicated a newly developed grain boundary, while a LDR ranged from 2.4 to 18 deg/μm implied the generation of a sub-grain boundary. And a lower LDR than 2.4 deg/μm could only introduce a slighter deformation and no sub-grain boundary was produced. According to the LDR analysis with the evolution of grain or sub grain boundaries, the deformed layer could be classified into four zones: grain broken layer, seriously deformed layer, slightly deformed layer and non-deformed layer.Keywords: surface machining, EBSD, subsurface layer, local deformation
Procedia PDF Downloads 3314105 Effect of Fresh Concrete Curing Methods on Its Compressive Strength
Authors: Xianghe Dai, Dennis Lam, Therese Sheehan, Naveed Rehman, Jie Yang
Abstract:
Concrete is one of the most used construction materials that may be made onsite as fresh concrete and then placed in formwork to produce the desired shapes of structures. It has been recognized that the raw materials and mix proportion of concrete dominate the mechanical characteristics of hardened concrete, and the curing method and environment applied to the concrete in early stages of hardening will significantly influence the concrete properties, such as compressive strength, durability, permeability etc. In construction practice, there are various curing methods to maintain the presence of mixing water throughout the early stages of concrete hardening. They are also beneficial to concrete in hot weather conditions as they provide cooling and prevent the evaporation of water. Such methods include ponding or immersion, spraying or fogging, saturated wet covering etc. Also there are various curing methods that may be implemented to decrease the level of water lost which belongs to the concrete surface, such as putting a layer of impervious paper, plastic sheeting or membrane on the concrete to cover it. In the concrete material laboratory, accelerated strength gain methods supply the concrete with heat and additional moisture by applying live steam, coils that are subject to heating or pads that have been warmed electrically. Currently when determining the mechanical parameters of a concrete, the concrete is usually sampled from fresh concrete on site and then cured and tested in laboratories where standardized curing procedures are adopted. However, in engineering practice, curing procedures in the construction sites after the placing of concrete might be very different from the laboratory criteria, and this includes some standard curing procedures adopted in the laboratory that can’t be applied on site. Sometimes the contractor compromises the curing methods in order to reduce construction costs etc. Obviously the difference between curing procedures adopted in the laboratory and those used on construction sites might over- or under-estimate the real concrete quality. This paper presents the effect of three typical curing methods (air curing, water immersion curing, plastic film curing) and of maintaining concrete in steel moulds on the compressive strength development of normal concrete. In this study, Portland cement with 30% fly ash was used and different curing periods, 7 days, 28 days and 60 days were applied. It was found that the highest compressive strength was observed from concrete samples to which 7-day water immersion curing was applied and from samples maintained in steel moulds up to the testing date. The research results implied that concrete used as infill in steel tubular members might develop a higher strength than predicted by design assumptions based on air curing methods. Wrapping concrete with plastic film as a curing method might delay the concrete strength development in the early stages. Water immersion curing for 7 days might significantly increase the concrete compressive strength.Keywords: compressive strength, air curing, water immersion curing, plastic film curing, maintaining in steel mould, comparison
Procedia PDF Downloads 2934104 Determines the Continuity of Void in Underground Mine Tunnel Using Ground Penetrating Radar
Authors: Farid Adisaputra Gumilang
Abstract:
Kucing Liar Underground Mine is a future mine of PT Freeport Indonesia PTFI that is currently being developed. In the development process, problems were found when blasting the tunnels; there were overbreak, and void occur caused by geological contact or poor rock conditions. Geotechnical engineers must evaluate not only the remnant capacity of ground support systems but also investigate the depth of rock mass yield within pillars. To prevent the potential hazard caused by void zones, geotechnical engineers must ensure the planned drift is mined in the best location where people can work safely. GPR, or Ground penetrating radar, is a geophysical method that can image the subsurface. This non-destructive method uses electromagnetic radiation and detects the reflected signals from subsurface structures. The GPR survey measurements are conducted 48 meters along the drift that has a poor ground condition with 150MHz antenna with several angles (roof, wall, and floor). Concern grounds are determined by the continuity of reflector/low reflector in the radargram section. Concern grounds are determined by the continuity of reflector/low reflector in the radargram section. In this paper, processing data using instantaneous amplitude to identify the void zone. In order to have a good interpretation and result, it combines with the geological information and borehole camera data, so the calibrated GPR data allows the geotechnical engineer to determine the safe location to change the drift location.Keywords: underground mine, ground penetrating radar, reflectivity, borehole camera
Procedia PDF Downloads 834103 Salinity Reduction from Saharan Brackish Water by Fluoride Removal on Activated Natural Materials: A Comparative Study
Authors: Amina Ramadni, Safia Taleb, André Dératani
Abstract:
The present study presents, firstly, to characterize the physicochemical quality of brackish groundwater of the Terminal Complex (TC) from the region of Eloued-souf and to investigate the presence of fluoride, and secondly, to study the comparison of adsorbing power of three materials, such as (activated alumina AA, sodium clay SC and hydroxyapatite HAP) against the groundwater in the region of Eloued-souf. To do this, a sampling campaign over 16 wells and consumer taps was undertaken. The results show that the groundwater can be characterized by very high fluoride content and excessive mineralization that require in some cases, specific treatment before supply. The study of adsorption revealed removal efficiencies fluoride by three adsorbents, maximum adsorption is achieved after 45 minutes at 90%, 83.4% and 73.95%, and with an adsorbed fluoride content of 0.22 mg/L, 0.318 mg/L and 0.52 mg/L for AA, HAP and SC, respectively. The acidity of the medium significantly affects the removal fluoride. Results deducted from the adsorption isotherms also showed that the retention follows the Langmuir model. The adsorption tests by adsorbent materials show that the physicochemical characteristics of brackish water are changed after treatment. The adsorption mechanism is an exchange between the OH- ions and fluoride ions. Three materials are proving to be effective adsorbents for fluoride removal that could be developed into a viable technology to help reduce the salinity of the Saharan hyper-fluorinated waters. Finally, a comparison between the results obtained from the different adsorbents allowed us to conclude that the defluoridation by AA is the process of choice for many waters of the region of Eloued-souf, because it was shown to be a very interesting and promising technique.Keywords: fluoride removal, hydrochemical characterization of groundwater, natural materials, nanofiltration
Procedia PDF Downloads 2184102 Task Scheduling and Resource Allocation in Cloud-based on AHP Method
Authors: Zahra Ahmadi, Fazlollah Adibnia
Abstract:
Scheduling of tasks and the optimal allocation of resources in the cloud are based on the dynamic nature of tasks and the heterogeneity of resources. Applications that are based on the scientific workflow are among the most widely used applications in this field, which are characterized by high processing power and storage capacity. In order to increase their efficiency, it is necessary to plan the tasks properly and select the best virtual machine in the cloud. The goals of the system are effective factors in scheduling tasks and resource selection, which depend on various criteria such as time, cost, current workload and processing power. Multi-criteria decision-making methods are a good choice in this field. In this research, a new method of work planning and resource allocation in a heterogeneous environment based on the modified AHP algorithm is proposed. In this method, the scheduling of input tasks is based on two criteria of execution time and size. Resource allocation is also a combination of the AHP algorithm and the first-input method of the first client. Resource prioritization is done with the criteria of main memory size, processor speed and bandwidth. What is considered in this system to modify the AHP algorithm Linear Max-Min and Linear Max normalization methods are the best choice for the mentioned algorithm, which have a great impact on the ranking. The simulation results show a decrease in the average response time, return time and execution time of input tasks in the proposed method compared to similar methods (basic methods).Keywords: hierarchical analytical process, work prioritization, normalization, heterogeneous resource allocation, scientific workflow
Procedia PDF Downloads 1454101 Treatment of Isopropyl Alcohol in Aqueous Solutions by VUV-Based AOPs within a Laminar-Falling-Film-Slurry Type Photoreactor
Authors: Y. S. Shen, B. H. Liao
Abstract:
This study aimed to develop the design equation of a laminar-falling-film-slurry (LFFS) type photoreactor for the treatment of organic wastewaters containing isopropyl alcohol (IPA) by VUV-based advanced oxidation processes (AOPs). The photoreactor design equations were established by combining with the chemical kinetics of the photocatalytic system, light absorption model within the photoreactor, and was used to predict the decomposition of IPA in aqueous solutions in the photoreactors of different geometries at various operating conditions (volumetric flow rate, oxidants, catalysts, solution pH values, UV light intensities, and initial concentration of pollutants) to verify its rationality and feasibility. By the treatment of the LFFS-VUV only process, it was found that the decomposition rates of IPA in aqueous solutions increased with the increase of volumetric flow rate, VUV light intensity, dosages of TiO2 and H2O2. The removal efficiencies of IPA by photooxidation processes were in the order: VUV/H2O2>VUV/TiO2/H2O2>VUV/TiO2>VUV only. In VUV, VUV/H2O2, VUV/TiO2/H2O2 processes, integrating with the reaction kinetic equations of IPA, the mass conservation equation and the linear light source model, the photoreactor design equation can reasonably to predict reaction behaviors of IPA at various operating conditions and to describe the concentration distribution profiles of IPA within photoreactors.The results of this research can be useful basis for the future application of the homogeneous and heterogeneous VUV-based advanced oxidation processes.Keywords: isopropyl alcohol, photoreactor design, VUV, AOPs
Procedia PDF Downloads 3774100 The Studies of the Sorption Capabilities of the Porous Microspheres with Lignin
Authors: M. Goliszek, M. Sobiesiak, O. Sevastyanova, B. Podkoscielna
Abstract:
Lignin is one of three main constituents of biomass together with cellulose and hemicellulose. It is a complex biopolymer, which contains a large number of functional groups, including aliphatic and aromatic hydroxyl groups, carbohylic groups and methoxy groups in its structure, that is why it shows potential capacities for process of sorption. Lignin is a highly cross-linked polymer with a three-dimentional structure which can provide large surface area and pore volumes. It can also posses better dispersion, diffusion and mass transfer behavior in a field of the removal of, e.g., heavy-metal-ions or aromatic pollutions. In this work emulsion-suspension copolymerization method, to synthesize the porous microspheres of divinylbenzene (DVB), styrene (St) and lignin was used. There are also microspheres without the addition of lignin for comparison. Before the copolymerization, modification lignin with methacryloyl chloride, to improve its reactivity with other monomers was done. The physico-chemical properties of the obtained microspheres, e.g., pore structures (adsorption-desorption measurements), thermal properties (DSC), tendencies to swell and the actual shapes were also studied. Due to well-developed porous structure and the presence of functional groups our materials may have great potential in sorption processes. To estimate the sorption capabilities of the microspheres towards phenol and its chlorinated derivatives the off-line SPE (solid-phase extraction) method is going to be applied. This method has various advantages, including low-cost, easy to use and enables the rapid measurements for a large number of chemicals. The efficiency of the materials in removing phenols from aqueous solution and in desorption processes will be evaluated.Keywords: microspheres, lignin, sorption, solid-phase extraction
Procedia PDF Downloads 1834099 Drive Sharing with Multimodal Interaction: Enhancing Safety and Efficiency
Authors: Sagar Jitendra Mahendrakar
Abstract:
Exploratory testing is a dynamic and adaptable method of software quality assurance that is frequently praised for its ability to find hidden flaws and improve the overall quality of the product. Instead of using preset test cases, exploratory testing allows testers to explore the software application dynamically. This is in contrast to scripted testing methodologies, which primarily rely on tester intuition, creativity, and adaptability. There are several tools and techniques that can aid testers in the exploratory testing process which we will be discussing in this talk.Tests of this kind are able to find bugs of this kind that are harder to find during structured testing or that other testing methods may have overlooked.The purpose of this abstract is to examine the nature and importance of exploratory testing in modern software development methods. It explores the fundamental ideas of exploratory testing, highlighting the value of domain knowledge and tester experience in spotting possible problems that may escape the notice of traditional testing methodologies. Throughout the software development lifecycle, exploratory testing promotes quick feedback loops and continuous improvement by giving testers the ability to make decisions in real time based on their observations. This abstract also clarifies the unique features of exploratory testing, like its non-linearity and capacity to replicate user behavior in real-world settings. Testers can find intricate bugs, usability problems, and edge cases in software through impromptu exploration that might go undetected. Exploratory testing's flexible and iterative structure fits in well with agile and DevOps processes, allowing for a quicker time to market without sacrificing the quality of the final product.Keywords: exploratory, testing, automation, quality
Procedia PDF Downloads 52