Search results for: edge detector
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1220

Search results for: edge detector

260 Climate Change Effects of Vehicular Carbon Monoxide Emission from Road Transportation in Part of Minna Metropolis, Niger State, Nigeria

Authors: H. M. Liman, Y. M. Suleiman A. A. David

Abstract:

Poor air quality often considered one of the greatest environmental threats facing the world today is caused majorly by the emission of carbon monoxide into the atmosphere. The principal air pollutant is carbon monoxide. One prominent source of carbon monoxide emission is the transportation sector. Not much was known about the emission levels of carbon monoxide, the primary pollutant from the road transportation in the study area. Therefore, this study assessed the levels of carbon monoxide emission from road transportation in the Minna, Niger State. The database shows the carbon monoxide data collected. MSA Altair gas alert detector was used to take the carbon monoxide emission readings in Parts per Million for the peak and off-peak periods of vehicular movement at the road intersections. Their Global Positioning System (GPS) coordinates were recorded in the Universal Transverse Mercator (UTM). Bar chart graphs were plotted by using the emissions level of carbon dioxide as recorded on the field against the scientifically established internationally accepted safe limit of 8.7 Parts per Million of carbon monoxide in the atmosphere. Further statistical analysis was also carried out on the data recorded from the field using the Statistical Package for Social Sciences (SPSS) software and Microsoft excel to show the variance of the emission levels of each of the parameters in the study area. The results established that emissions’ level of atmospheric carbon monoxide from the road transportation in the study area exceeded the internationally accepted safe limits of 8.7 parts per million. In addition, the variations in the average emission levels of CO between the four parameters showed that morning peak is having the highest average emission level of 24.5PPM followed by evening peak with 22.84PPM while morning off peak is having 15.33 and the least is evening off peak 12.94PPM. Based on these results, recommendations made for poor air quality mitigation via carbon monoxide emissions reduction from transportation include Introduction of the urban mass transit would definitely reduce the number of traffic on the roads, hence the emissions from several vehicles that would have been on the road. This would also be a cheaper means of transportation for the masses and Encouraging the use of vehicles using alternative sources of energy like solar, electric and biofuel will also result in less emission levels as the these alternative energy sources other than fossil fuel originated diesel and petrol vehicles do not emit especially carbon monoxide.

Keywords: carbon monoxide, climate change emissions, road transportation, vehicular

Procedia PDF Downloads 375
259 On Elastic Anisotropy of Fused Filament Fabricated Acrylonitrile Butadiene Styrene Structures

Authors: Joseph Marae Djouda, Ashraf Kasmi, François Hild

Abstract:

Fused filament fabrication is one of the most widespread additive manufacturing techniques because of its low-cost implementation. Its initial development was based on part fabrication with thermoplastic materials. The influence of the manufacturing parameters such as the filament orientation through the nozzle, the deposited layer thickness, or the speed deposition on the mechanical properties of the parts has been widely experimentally investigated. It has been recorded the remarkable variations of the anisotropy in the function of the filament path during the fabrication process. However, there is a lack in the development of constitutive models describing the mechanical properties. In this study, integrated digital image correlation (I-DIC) is used for the identification of mechanical constitutive parameters of two configurations of ABS samples: +/-45° and so-called “oriented deposition.” In this last, the filament was deposited in order to follow the principal strain of the sample. The identification scheme based on the gap reduction between simulation and the experiment directly from images recorded from a single sample (single edge notched tension specimen) is developed. The macroscopic and mesoscopic analysis are conducted from images recorded in both sample surfaces during the tensile test. The elastic and elastoplastic models in isotropic and orthotropic frameworks have been established. It appears that independently of the sample configurations (filament orientation during the fabrication), the elastoplastic isotropic model gives the correct description of the behavior of samples. It is worth noting that in this model, the number of constitutive parameters is limited to the one considered in the elastoplastic orthotropic model. This leads to the fact that the anisotropy of the architectured 3D printed ABS parts can be neglected in the establishment of the macroscopic behavior description.

Keywords: elastic anisotropy, fused filament fabrication, Acrylonitrile butadiene styrene, I-DIC identification

Procedia PDF Downloads 126
258 AI-Driven Solutions for Optimizing Master Data Management

Authors: Srinivas Vangari

Abstract:

In the era of big data, ensuring the accuracy, consistency, and reliability of critical data assets is crucial for data-driven enterprises. Master Data Management (MDM) plays a crucial role in this endeavor. This paper investigates the role of Artificial Intelligence (AI) in enhancing MDM, focusing on how AI-driven solutions can automate and optimize various stages of the master data lifecycle. By integrating AI (Quantitative and Qualitative Analysis) into processes such as data creation, maintenance, enrichment, and usage, organizations can achieve significant improvements in data quality and operational efficiency. Quantitative analysis is employed to measure the impact of AI on key metrics, including data accuracy, processing speed, and error reduction. For instance, our study demonstrates an 18% improvement in data accuracy and a 75% reduction in duplicate records across multiple systems post-AI implementation. Furthermore, AI’s predictive maintenance capabilities reduced data obsolescence by 22%, as indicated by statistical analyses of data usage patterns over a 12-month period. Complementing this, a qualitative analysis delves into the specific AI-driven strategies that enhance MDM practices, such as automating data entry and validation, which resulted in a 28% decrease in manual errors. Insights from case studies highlight how AI-driven data cleansing processes reduced inconsistencies by 25% and how AI-powered enrichment strategies improved data relevance by 24%, thus boosting decision-making accuracy. The findings demonstrate that AI significantly enhances data quality and integrity, leading to improved enterprise performance through cost reduction, increased compliance, and more accurate, real-time decision-making. These insights underscore the value of AI as a critical tool in modern data management strategies, offering a competitive edge to organizations that leverage its capabilities.

Keywords: artificial intelligence, master data management, data governance, data quality

Procedia PDF Downloads 17
257 Study of Polyphenol Profile and Antioxidant Capacity in Italian Ancient Apple Varieties by Liquid Chromatography

Authors: A. M. Tarola, R. Preti, A. M. Girelli, P. Campana

Abstract:

Safeguarding, studying and enhancing biodiversity play an important and indispensable role in re-launching agriculture. The ancient local varieties are therefore a precious resource for genetic and health improvement. In order to protect biodiversity through the recovery and valorization of autochthonous varieties, in this study we analyzed 12 samples of four ancient apple cultivars representative of Friuli Venezia Giulia, selected by local farmers who work on a project for the recovery of ancient apple cultivars. The aim of this study is to evaluate the polyphenolic profile and the antioxidant capacity that characterize the organoleptic and functional qualities of this fruit species, besides having beneficial properties for health. In particular, for each variety, the following compounds were analyzed, both in the skins and in the pulp: gallic acid, catechin, chlorogenic acid, epicatechin, caffeic acid, coumaric acid, ferulic acid, rutin, phlorizin, phloretin and quercetin to highlight any differences in the edible parts of the apple. The analysis of individual phenolic compounds was performed by High Performance Liquid Chromatography (HPLC) coupled with a diode array UV detector (DAD), the antioxidant capacity was estimated using an in vitro essay based on a Free Radical Scavenging Method and the total phenolic compounds was determined using the Folin-Ciocalteau method. From the results, it is evident that the catechins are the most present polyphenols, reaching a value of 140-200 μg/g in the pulp and of 400-500 μg/g in the skin, with the prevalence of epicatechin. Catechins and phlorizin, a dihydrohalcone typical of apples, are always contained in larger quantities in the peel. Total phenolic compounds content was positively correlated with antioxidant activity in apple pulp (r2 = 0,850) and peel (r2 = 0,820). Comparing the results, differences between the varieties analyzed and between the edible parts (pulp and peel) of the apple were highlighted. In particular, apple peel is richer in polyphenolic compounds than pulp and flavonols are exclusively present in the peel. In conclusion, polyphenols, being antioxidant substances, have confirmed the benefits of fruit in the diet, especially as a prevention and treatment for degenerative diseases. They demonstrated to be also a good marker for the characterization of different apple cultivars. The importance of protecting biodiversity in agriculture was also highlighted through the exploitation of native products and ancient varieties of apples now forgotten.

Keywords: apple, biodiversity, polyphenols, antioxidant activity, HPLC-DAD, characterization

Procedia PDF Downloads 136
256 Mobile App Architecture in 2023: Build Your Own Mobile App

Authors: Mounir Filali

Abstract:

Companies use many innovative ways to reach their customers to stay ahead of the competition. Along with the growing demand for innovative business solutions is the demand for new technology. The most noticeable area of demand for business innovations is the mobile application industry. Recently, companies have recognized the growing need to integrate proprietary mobile applications into their suite of services; Companies have realized that developing mobile apps gives them a competitive edge. As a result, many have begun to rapidly develop mobile apps to stay ahead of the competition. Mobile application development helps companies meet the needs of their customers. Mobile apps also help businesses to take advantage of every potential opportunity to generate leads that convert into sales. Mobile app download growth statistics with the recent rise in demand for business-related mobile apps, there has been a similar rise in the range of mobile app solutions being offered. Today, companies can use the traditional route of the software development team to build their own mobile applications. However, there are also many platform-ready "low-code and no-code" mobile apps available to choose from. These mobile app development options have more streamlined business processes. This helps them be more responsive to their customers without having to be coding experts. Companies must have a basic understanding of mobile app architecture to attract and maintain the interest of mobile app users. Mobile application architecture refers to the buildings or structural systems and design elements that make up a mobile application. It also includes the technologies, processes, and components used during application development. The underlying foundation of all applications consists of all elements of the mobile application architecture, developing a good mobile app architecture requires proper planning and strategic design. The technology framework or platform on the back end and user-facing side of a mobile application is part of the mobile architecture of the application. In-application development Software programmers loosely refer to this set of mobile architecture systems and processes as the "technology stack".

Keywords: mobile applications, development, architecture, technology

Procedia PDF Downloads 102
255 Calibration of 2D and 3D Optical Measuring Instruments in Industrial Environments at Submillimeter Range

Authors: Alberto Mínguez-Martínez, Jesús de Vicente y Oliva

Abstract:

Modern manufacturing processes have led to the miniaturization of systems and, as a result, parts at the micro-and nanoscale are produced. This trend seems to become increasingly important in the near future. Besides, as a requirement of Industry 4.0, the digitalization of the models of production and processes makes it very important to ensure that the dimensions of newly manufactured parts meet the specifications of the models. Therefore, it is possible to reduce the scrap and the cost of non-conformities, ensuring the stability of the production at the same time. To ensure the quality of manufactured parts, it becomes necessary to carry out traceable measurements at scales lower than one millimeter. Providing adequate traceability to the SI unit of length (the meter) to 2D and 3D measurements at this scale is a problem that does not have a unique solution in industrial environments. Researchers in the field of dimensional metrology all around the world are working on this issue. A solution for industrial environments, even if it is not complete, will enable working with some traceability. At this point, we believe that the study of the surfaces could provide us with a first approximation to a solution. Among the different options proposed in the literature, the areal topography methods may be the most relevant because they could be compared to those measurements performed using Coordinate Measuring Machines (CMM’s). These measuring methods give (x, y, z) coordinates for each point, expressing it in two different ways, either expressing the z coordinate as a function of x, denoting it as z(x), for each Y-axis coordinate, or as a function of the x and y coordinates, denoting it as z (x, y). Between others, optical measuring instruments, mainly microscopes, are extensively used to carry out measurements at scales lower than one millimeter because it is a non-destructive measuring method. In this paper, the authors propose a calibration procedure for the scales of optical measuring instruments, particularizing for a confocal microscope, using material standards easy to find and calibrate in metrology and quality laboratories in industrial environments. Confocal microscopes are measuring instruments capable of filtering the out-of-focus reflected light so that when it reaches the detector, it is possible to take pictures of the part of the surface that is focused. Varying and taking pictures at different Z levels of the focus, a specialized software interpolates between the different planes, and it could reconstruct the surface geometry into a 3D model. As it is easy to deduce, it is necessary to give traceability to each axis. As a complementary result, the roughness Ra parameter will be traced to the reference. Although the solution is designed for a confocal microscope, it may be used for the calibration of other optical measuring instruments by applying minor changes.

Keywords: industrial environment, confocal microscope, optical measuring instrument, traceability

Procedia PDF Downloads 155
254 Influence of Thermal Damage on the Mechanical Strength of Trimmed CFRP

Authors: Guillaume Mullier, Jean François Chatelain

Abstract:

Carbon Fiber Reinforced Plastics (CFRPs) are widely used for advanced applications, in particular in aerospace, automotive and wind energy industries. Once cured to near net shape, CFRP parts need several finishing operations such as trimming, milling or drilling in order to accommodate fastening hardware and meeting the final dimensions. The present research aims to study the effect of the cutting temperature in trimming on the mechanical strength of high performance CFRP laminates used for aeronautics applications. The cutting temperature is of great importance when dealing with trimming of CFRP. Temperatures higher than the glass-transition temperature (Tg) of the resin matrix are highly undesirable: they cause degradation of the matrix in the trimmed edges area, which can severely affect the mechanical performance of the entire component. In this study, a 9.50 mm diameter CVD diamond coated carbide tool with six flutes was used to trim 24-plies CFRP laminates. A 300 m/min cutting speed and 1140 mm/min feed rate were used in the experiments. The tool was heated prior to trimming using a blowtorch, for temperatures ranging from 20°C to 300°C. The temperature at the cutting edge was measured using embedded K-Type thermocouples. Samples trimmed for different cutting temperatures, below and above Tg, were mechanically tested using three-points bending short-beam loading configurations. New cutting tools as well as worn cutting tools were utilized for the experiments. The experiments with the new tools could not prove any correlation between the length of cut, the cutting temperature and the mechanical performance. Thus mechanical strength was constant, regardless of the cutting temperature. However, for worn tools, producing a cutting temperature rising up to 450°C, thermal damage of the resin was observed. The mechanical tests showed a reduced mean resistance in short beam configuration, while the resistance in three point bending decreases with increase of the cutting temperature.

Keywords: composites, trimming, thermal damage, surface quality

Procedia PDF Downloads 321
253 Normalized P-Laplacian: From Stochastic Game to Image Processing

Authors: Abderrahim Elmoataz

Abstract:

More and more contemporary applications involve data in the form of functions defined on irregular and topologically complicated domains (images, meshs, points clouds, networks, etc). Such data are not organized as familiar digital signals and images sampled on regular lattices. However, they can be conveniently represented as graphs where each vertex represents measured data and each edge represents a relationship (connectivity or certain affinities or interaction) between two vertices. Processing and analyzing these types of data is a major challenge for both image and machine learning communities. Hence, it is very important to transfer to graphs and networks many of the mathematical tools which were initially developed on usual Euclidean spaces and proven to be efficient for many inverse problems and applications dealing with usual image and signal domains. Historically, the main tools for the study of graphs or networks come from combinatorial and graph theory. In recent years there has been an increasing interest in the investigation of one of the major mathematical tools for signal and image analysis, which are Partial Differential Equations (PDEs) variational methods on graphs. The normalized p-laplacian operator has been recently introduced to model a stochastic game called tug-of-war-game with noise. Part interest of this class of operators arises from the fact that it includes, as particular case, the infinity Laplacian, the mean curvature operator and the traditionnal Laplacian operators which was extensiveley used to models and to solve problems in image processing. The purpose of this paper is to introduce and to study a new class of normalized p-Laplacian on graphs. The introduction is based on the extension of p-harmonious function introduced in as discrete approximation for both infinity Laplacian and p-Laplacian equations. Finally, we propose to use these operators as a framework for solving many inverse problems in image processing.

Keywords: normalized p-laplacian, image processing, stochastic game, inverse problems

Procedia PDF Downloads 512
252 Graph Clustering Unveiled: ClusterSyn - A Machine Learning Framework for Predicting Anti-Cancer Drug Synergy Scores

Authors: Babak Bahri, Fatemeh Yassaee Meybodi, Changiz Eslahchi

Abstract:

In the pursuit of effective cancer therapies, the exploration of combinatorial drug regimens is crucial to leverage synergistic interactions between drugs, thereby improving treatment efficacy and overcoming drug resistance. However, identifying synergistic drug pairs poses challenges due to the vast combinatorial space and limitations of experimental approaches. This study introduces ClusterSyn, a machine learning (ML)-powered framework for classifying anti-cancer drug synergy scores. ClusterSyn employs a two-step approach involving drug clustering and synergy score prediction using a fully connected deep neural network. For each cell line in the training dataset, a drug graph is constructed, with nodes representing drugs and edge weights denoting synergy scores between drug pairs. Drugs are clustered using the Markov clustering (MCL) algorithm, and vectors representing the similarity of drug pairs to each cluster are input into the deep neural network for synergy score prediction (synergy or antagonism). Clustering results demonstrate effective grouping of drugs based on synergy scores, aligning similar synergy profiles. Subsequently, neural network predictions and synergy scores of the two drugs on others within their clusters are used to predict the synergy score of the considered drug pair. This approach facilitates comparative analysis with clustering and regression-based methods, revealing the superior performance of ClusterSyn over state-of-the-art methods like DeepSynergy and DeepDDS on diverse datasets such as Oniel and Almanac. The results highlight the remarkable potential of ClusterSyn as a versatile tool for predicting anti-cancer drug synergy scores.

Keywords: drug synergy, clustering, prediction, machine learning., deep learning

Procedia PDF Downloads 79
251 Frequency Interpretation of a Wave Function, and a Vertical Waveform Treated as A 'Quantum Leap'

Authors: Anthony Coogan

Abstract:

Born’s probability interpretation of wave functions would have led to nearly identical results had he chosen a frequency interpretation instead. Logically, Born may have assumed that only one electron was under consideration, making it nonsensical to propose a frequency wave. Author’s suggestion: the actual experimental results were not of a single electron; rather, they were groups of reflected x-ray photons. The vertical waveform used by Scrhödinger in his Particle in the Box Theory makes sense if it was intended to represent a quantum leap. The author extended the single vertical panel to form a bar chart: separate panels would represent different energy levels. The proposed bar chart would be populated by reflected photons. Expansion of basic ideas: Part of Scrhödinger’s ‘Particle in the Box’ theory may be valid despite negative criticism. The waveform used in the diagram is vertical, which may seem absurd because real waves decay at a measurable rate, rather than instantaneously. However, there may be one notable exception. Supposedly, following from the theory, the Uncertainty Principle was derived – may a Quantum Leap not be represented as an instantaneous waveform? The great Scrhödinger must have had some reason to suggest a vertical waveform if the prevalent belief was that they did not exist. Complex wave forms representing a particle are usually assumed to be continuous. The actual observations made were x-ray photons, some of which had struck an electron, been reflected, and then moved toward a detector. From Born’s perspective, doing similar work the years in question 1926-7, he would also have considered a single electron – leading him to choose a probability distribution. Probability Distributions appear very similar to Frequency Distributions, but the former are considered to represent the likelihood of future events. Born’s interpretation of the results of quantum experiments led (or perhaps misled) many researchers into claiming that humans can influence events just by looking at them, e.g. collapsing complex wave functions by 'looking at the electron to see which slit it emerged from', while in reality light reflected from the electron moved in the observer’s direction after the electron had moved away. Astronomers may say that they 'look out into the universe' but are actually using logic opposed to the views of Newton and Hooke and many observers such as Romer, in that light carries information from a source or reflector to an observer, rather the reverse. Conclusion: Due to the controversial nature of these ideas, especially its implications about the nature of complex numbers used in applications in science and engineering, some time may pass before any consensus is reached.

Keywords: complex wave functions not necessary, frequency distributions instead of wave functions, information carried by light, sketch graph of uncertainty principle

Procedia PDF Downloads 199
250 Multi-Functional Metal Oxides as Gas Sensors, Photo-Catalysts and Bactericides

Authors: Koyar Rane

Abstract:

Nano- to submicron size particles of narrow particle size distribution of semi-conducting TiO₂, ZnO, NiO, CuO, Fe₂O₃ have been synthesized by novel hydrazine method and tested for their gas sensing, photocatalytic and bactericidal activities and the behavior found to be enhanced when the oxides in the thin film forms, that obtained in a specially built spray pyrolysis reactor. Hydrazine method is novel in the sense, say, the UV absorption edge of the white pigment grade wide band gap (~3.2eV) TiO₂ and ZnO shifted to the visible region turning into yellowish particles, indicating modification occurring the band structure. The absorption in the visible region makes these oxides visible light sensitive photocatalysis in degrading pollutants, especially the organic dyes which otherwise increase the chemical oxygen demand of the drinking water, enabling the process feasible not under the harsh energetic UV radiation regime. The electromagnetic radiations on irradiation produce electron-hole pairs Semiconductor + hν → e⁻ + h⁺ The electron-hole pairs thus produced form Reactive Oxygen Species, ROS, on the surface of the semiconductors, O₂(adsorbed)+e⁻ → O₂• - superoxide ion OH-(surface)+h⁺ →•OH - Hydroxyl radical The ROS attack the organic material and micro-organisms. Our antibacterial studies indicate the metal oxides control the Biological Oxygen Demand (BOD) of drinking water which had beyond the safe level normally found in the municipal supply. Metal oxides in the thin film form show overall enhanced properties and the films are reusable. The results of the photodegradation and antibactericidal studies are discussed. Gas sensing studies too have been done to find the versatility of the multifunctional metal oxides.

Keywords: hydrazine method, visible light sensitive, photo-degradation of dyes, water/airborne pollutant

Procedia PDF Downloads 163
249 Bamboo as the Frontier for Economically Sustainable Solution to Flood Control and Human Wildlife Conflict

Authors: Nirman Kumar Ojha

Abstract:

Bamboo plantation can be integrated for natural embankment against flood and live fencing against wild animals, at the same time provide economic opportunity for the poor farmers as a sustainable solution and adaptation alternative. 2010 flood in the Rui River completely inundated fields of four VDCs in Madi, Chitwan National Park with extensive bank erosion. The main aim of this action research was to identify an economically sustainable natural embankment against flood and also providing wildlife friendly fencing to reduce human-wildlife conflict. Community people especially poor farmers were trained for soil testing, land identification, plantation, and the harvesting regime, nursery set up and intercropping along with bamboo plantation on the edge of the river bank in order to reduce or minimize soil erosion. Results show that farmers are able to establish cost efficient and economically sustainable river embankment with bamboo plantation also creating a fence for wildlife which has also promoted bamboo cultivation and conservation. This action research has amalgamated flood control and wildlife control with the livelihood of the farmers which otherwise would cost huge resource. Another major impact of the bamboo plantation is its role in climate change and its adaptation process reducing degradation and improving vegetation cover contributing to landscape management. Based on this study, we conclude that bamboo plantation in Madi, Chitwan promoted the livelihood of the poor farmers providing a sustainable economic solution to reduce bank erosion, human-wildlife conflict and contributes to landscape management.

Keywords: climate change and conservation, economic opportunity, flood control, national park

Procedia PDF Downloads 278
248 Mode II Fracture Toughness of Hybrid Fiber Reinforced Concrete

Authors: H. S. S Abou El-Mal, A. S. Sherbini, H. E. M. Sallam

Abstract:

Mode II fracture toughness (KIIc) of fiber reinforced concrete has been widely investigated under various patterns of testing geometries. The effect of fiber type, concrete matrix properties, and testing mechanisms were extensively studied. The area of hybrid fiber addition shows a lake of reported research data. In this paper an experimental investigation of hybrid fiber embedded in high strength concrete matrix is reported. Three different types of fibers; namely steel (S), glass (G), and polypropylene (PP) fibers were mixed together in four hybridization patterns, (S/G), (S/PP), (G/PP), (S/G/PP) with constant cumulative volume fraction (Vf) of 1.5%. The concrete matrix properties were kept the same for all hybrid fiber reinforced concrete patterns. In an attempt to estimate a fairly accepted value of fracture toughness KIIc, four testing geometries and loading types are employed in this investigation. Four point shear, Brazilian notched disc, double notched cube, and double edge notched specimens are investigated in a trial to avoid the limitations and sensitivity of each test regarding geometry, size effect, constraint condition, and the crack length to specimen width ratio a/w. The addition of all hybridization patterns of fiber reduced the compressive strength and increased mode II fracture toughness in pure mode II tests. Mode II fracture toughness of concrete KIIc decreased with the increment of a/w ratio for all concretes and test geometries. Mode II fracture toughness KIIc is found to be sensitive to the hybridization patterns of fiber. The (S/PP) hybridization pattern showed higher values than all other patterns, while the (S/G/PP) showed insignificant enhancement on mode II fracture toughness (KIIc). Four point shear (4PS) test set up reflects the most reliable values of mode II fracture toughness KIIc of concrete. Mode II fracture toughness KIIc of concrete couldn’t be assumed as a real material property.

Keywords: fiber reinforced concrete, Hybrid fiber, Mode II fracture toughness, testing geometry

Procedia PDF Downloads 326
247 Environmental Photodegradation of Tralkoxydim Herbicide and Its Formulation in Natural Waters

Authors: María José Patiño-Ropero, Manuel Alcamí, Al Mokhtar Lamsabhi, José Luis Alonso-Prados, Pilar Sandín-España

Abstract:

Tralkoxydim, commercialized under different trade names, among them Splendor® (25% active ingredient), is a cyclohexanedione herbicide used in wheat and barley fields for the post-emergence control of annual winter grass weeds. Due to their physicochemical properties, herbicides belonging to this family are known to be susceptible to reaching natural waters, where different degradation pathways can take place. Photolysis represents one of the main routes of abiotic degradation of these herbicides in water. This transformation pathway can lead to the formation of unknown by-products, which could be more toxic and/or persistent than the active substances themselves. Therefore, there is a growing need to understand the science behind such dissipation routes, which is key to estimating the persistence of these compounds and ensuring the accurate assessment of environmental behavior. However, to our best knowledge, any information regarding the photochemical behavior of tralkoxydim under natural conditions in an aqueous environment has not been available till now in the literature. This work has focused on investigating the photochemical behavior of tralkoxydim herbicide and its commercial formulation (Splendor®) in the ultrapure, river and spring water using simulated solar radiation. Besides, the evolution of detected degradation products formed in the samples has been studied. A reversed-phase HPLC-DAD (high-performance liquid chromatography with diode array detector) method was developed to evaluate the kinetic evolution and to obtain the half-lives. In both cases, the degradation rates of active ingredient tralkoxydim in natural waters were lower than in ultrapure water following the order; river water < spring water < ultrapure water, and with first-order half-life values of 5.1 h, 2.7 h and 1.1 h, respectively. These findings indicate that the photolytical behavior of active ingredients is largely affected by the water composition, and these components can exert an internal filter effect. In addition, tralkoxydim herbicide and its formulation showed the same half-lives for each one of the types of water studied, showing that the presence of adjuvants in the commercial formulation has not any effect on the degradation rates of the active ingredient. HPLC-MS (high-performance liquid chromatography with mass spectrometry) experiments were performed to study the by-products deriving from the photodegradation of tralkoxydim in water. Accordingly, three compounds were tentatively identified. These results provide a better understanding of the tralkoxydim herbicide behavior in natural waters and its fate in the environment.

Keywords: by-products, natural waters, photodegradation, tralkoxydim herbicide

Procedia PDF Downloads 92
246 Survey of Indoor Radon/Thoron Concentrations in High Lung Cancer Incidence Area in India

Authors: Zoliana Bawitlung, P. C. Rohmingliana, L. Z. Chhangte, Remlal Siama, Hming Chungnunga, Vanram Lawma, L. Hnamte, B. K. Sahoo, B. K. Sapra, J. Malsawma

Abstract:

Mizoram state has the highest lung cancer incidence rate in India due to its high-level consumption of tobacco and its products which is supplemented by the food habits. While smoking is mainly responsible for this incidence, the effect of inhalation of indoor radon gas cannot be discarded as the hazardous nature of this radioactive gas and its progenies on human population have been well-established worldwide where the radiation damage to bronchial cells eventually can be the second leading cause of lung cancer next to smoking. It is also known that the effect of radiation, however, small may be the concentration, cannot be neglected as they can bring about the risk of cancer incidence. Hence, estimation of indoor radon concentration is important to give a useful reference against radiation effects as well as establishing its safety measures and to create a baseline for further case-control studies. The indoor radon/thoron concentrations in Mizoram had been measured in 41 dwellings selected on the basis of spot gamma background radiation and construction type of the houses during 2015-2016. The dwellings were monitored for one year, in 4 months cycles to indicate seasonal variations, for the indoor concentration of radon gas and its progenies, outdoor gamma dose, and indoor gamma dose respectively. A time-integrated method using Solid State Nuclear Track Detector (SSNTD) based single entry pin-hole dosimeters were used for measurement of indoor Radon/Thoron concentration. Gamma dose measurements for indoor as well as outdoor were carried out using Geiger Muller survey meters. Seasonal variation of indoor radon/ thoron concentration was monitored. The results show that the annual average radon concentrations varied from 54.07 – 144.72 Bq/m³ with an average of 90.20 Bq/m³ and the annual average thoron concentration varied from 17.39 – 54.19 Bq/m³ with an average of 35.91 Bq/m³ which are below the permissible limit. The spot survey of gamma background radiation level varies between 9 to 24 µR/h inside and outside the dwellings throughout Mizoram which are all within acceptable limits. From the above results, there is no direct indication that radon/thoron is responsible for the high lung cancer incidence in the area. In order to find epidemiological evidence of natural radiations to high cancer incidence in the area, one may need to conduct a case-control study which is beyond this scope. However, the derived data of measurement will provide baseline data for further studies.

Keywords: background gamma radiation, indoor radon/thoron, lung cancer, seasonal variation

Procedia PDF Downloads 143
245 The Relationship between Spindle Sound and Tool Performance in Turning

Authors: N. Seemuang, T. McLeay, T. Slatter

Abstract:

Worn tools have a direct effect on the surface finish and part accuracy. Tool condition monitoring systems have been developed over a long period and used to avoid a loss of productivity resulting from using a worn tool. However, the majority of tool monitoring research has applied expensive sensing systems not suitable for production. In this work, the cutting sound in turning machine was studied using microphone. Machining trials using seven cutting conditions were conducted until the observable flank wear width (FWW) on the main cutting edge exceeded 0.4 mm. The cutting inserts were removed from the tool holder and the flank wear width was measured optically. A microphone with built-in preamplifier was used to record the machining sound of EN24 steel being face turned by a CNC lathe in a wet cutting condition using constant surface speed control. The sound was sampled at 50 kS/s and all sound signals recorded from microphone were transformed into the frequency domain by FFT in order to establish the frequency content in the audio signature that could be then used for tool condition monitoring. The extracted feature from audio signal was compared to the flank wear progression on the cutting inserts. The spectrogram reveals a promising feature, named as ‘spindle noise’, which emits from the main spindle motor of turning machine. The spindle noise frequency was detected at 5.86 kHz of regardless of cutting conditions used on this particular CNC lathe. Varying cutting speed and feed rate have an influence on the magnitude of power spectrum of spindle noise. The magnitude of spindle noise frequency alters in conjunction with the tool wear progression. The magnitude increases significantly in the transition state between steady-state wear and severe wear. This could be used as a warning signal to prepare for tool replacement or adapt cutting parameters to extend tool life.

Keywords: tool wear, flank wear, condition monitoring, spindle noise

Procedia PDF Downloads 338
244 Optimal Design of Wind Turbine Blades Equipped with Flaps

Authors: I. Kade Wiratama

Abstract:

As a result of the significant growth of wind turbines in size, blade load control has become the main challenge for large wind turbines. Many advanced techniques have been investigated aiming at developing control devices to ease blade loading. Amongst them, trailing edge flaps have been proven as effective devices for load alleviation. The present study aims at investigating the potential benefits of flaps in enhancing the energy capture capabilities rather than blade load alleviation. A software tool is especially developed for the aerodynamic simulation of wind turbines utilising blades equipped with flaps. As part of the aerodynamic simulation of these wind turbines, the control system must be also simulated. The simulation of the control system is carried out via solving an optimisation problem which gives the best value for the controlling parameter at each wind turbine run condition. Developing a genetic algorithm optimisation tool which is especially designed for wind turbine blades and integrating it with the aerodynamic performance evaluator, a design optimisation tool for blades equipped with flaps is constructed. The design optimisation tool is employed to carry out design case studies. The results of design case studies on wind turbine AWT 27 reveal that, as expected, the location of flap is a key parameter influencing the amount of improvement in the power extraction. The best location for placing a flap is at about 70% of the blade span from the root of the blade. The size of the flap has also significant effect on the amount of enhancement in the average power. This effect, however, reduces dramatically as the size increases. For constant speed rotors, adding flaps without re-designing the topology of the blade can improve the power extraction capability as high as of about 5%. However, with re-designing the blade pretwist the overall improvement can be reached as high as 12%.

Keywords: flaps, design blade, optimisation, simulation, genetic algorithm, WTAero

Procedia PDF Downloads 337
243 Performance of Rural and Urban Adult Participants on Neuropsychological Tests in Zambia

Authors: Happy Zulu

Abstract:

Neuropsychological examination is an important way of formally assessing brain function. While there is so much documentation about the influence that some factors, such as age and education, have on neuropsychological tests (NP), not so much has been done to assess the influence that residency (rural/urban) may have. The specific objectives of this study were to establish if there is a significant difference in mean test scores on NP tests between rural and urban participants and to assess which tests on the Zambia Neurobehavioural Test Battery (ZNTB) are more affected by the participants‘ residency (rural/urban) and to determine the extent to which education, gender, and age predict test performance on NP tests for rural and urban participants. The participants (324) were drawn from both urban and rural areas of Zambia (Rural = 152 and Urban = 172). However, only 234 participants (Rural = 152 and Urban 82) were used for all the analyses in this particular study. The 234 participants were used as the actual proportion of the rural vs urban population in Zambia was 65% : 35%, respectively (CSO, 2003). The rural-urban ratio for the participants that were captured during the data collection process was 152 : 172, respectively. Thus, all the rural participants (152) were included and 90 of the 172 urban participants were randomly excluded so that the rural/urban ratio reached the desired 65% : 35 % which was the required ideal statistic for appropriate representation of the actual population in Zambia. Data on NP tests were analyzed from 234 participants, rural (N=152) reflecting 65% and urban (N=82) reflecting 35%. T-tests indicated that urban participants had superior performances in all the seven NP test domains, and all the mean differences in all these domains were found to be statistically significant. Residency had a large or moderate effect in five domains, while its effect size was small only in two of the domains. A standard multiple regression revealed that education, age and residency as predictor variables made a significant contribution to variance in performance on various domains of the ZNTB. However, the gender of participants was not a major factor in determining one‘s performance on neuropsychological tests. This particular report is part of an ongoing, larger, cutting-edge study aimed at formulating the normative data for Zambia with regard to performance on neuropsychological tests. This is necessary for appropriate, effective, and efficient assessment or diagnosis of various neurocognitive and neurobehavioural deficits that a number of people may currently be suffering from. It has been shown in this study that it is vital to make careful analyses of the variables that may be associated with one‘s performance on neuropsychological tests.

Keywords: neuropsychology, neurobehavioural, residency, Zambia

Procedia PDF Downloads 55
242 Optimizing Organizational Performance: The Critical Role of Headcount Budgeting in Strategic Alignment and Financial Stability

Authors: Shobhit Mittal

Abstract:

Headcount budgeting stands as a pivotal element in organizational financial management, extending beyond traditional budgeting to encompass strategic resource allocation for workforce-related expenses. This process is integral to maintaining financial stability and fostering a productive workforce, requiring a comprehensive analysis of factors such as market trends, business growth projections, and evolving workforce skill requirements. It demands a collaborative approach, primarily involving Human Resources (HR) and finance departments, to align workforce planning with an organization's financial capabilities and strategic objectives. The dynamic nature of headcount budgeting necessitates continuous monitoring and adjustment in response to economic fluctuations, business strategy shifts, technological advancements, and market dynamics. Its significance in talent management is also highlighted, aligning financial planning with talent acquisition and retention strategies to ensure a competitive edge in the market. The consequences of incorrect headcount budgeting are explored, showing how it can lead to financial strain, operational inefficiencies, and hindered strategic objectives. Examining case studies like IBM's strategic workforce rebalancing and Microsoft's shift for long-term success, the importance of aligning headcount budgeting with organizational goals is underscored. These examples illustrate that effective headcount budgeting transcends its role as a financial tool, emerging as a strategic element crucial for an organization's success. This necessitates continuous refinement and adaptation to align with evolving business goals and market conditions, highlighting its role as a key driver in organizational success and sustainability.

Keywords: strategic planning, fiscal budget, headcount planning, resource allocation, financial management, decision-making, operational efficiency, risk management, headcount budget

Procedia PDF Downloads 50
241 Foucault and the Archaeology of Transhumanism

Authors: Michel Foucault, Friedrich Nietzsche, Max More, Natasha Vita-More, Francesca Ferrando

Abstract:

During the early part of his intellectual and academic career (1950s and 1960s), Michel Foucault developed an interest for what we can call the ‘anthropological question’, or how our modernity deals with human nature from an epistemological standpoint. The great originality of Foucault’s thought here lies in the fact that he approaches this question not from the perspective of this ‘sovereign subject’ (that has characterized the History of Western thought) he wishes to disclose and ‘denounce’, but rather at the level of discourses and the way they constitute who we are, so to speak. This led him, in turn, to formulate a series of though-provoking statements during his so-called ‘archaeological period’ of the 1960s concerning what we call ‘man’ in the West, such as that he is an ‘invention of recent date’ (as a proper object of concern and reflection), and, perhaps more importantly, that he might disappear in the near future, ‘like a face drawn in sand at the edge of the sea’. Foucault is following on the footsteps of Nietzsche in that regard, who had famously announced in the 19th ce. the ‘death of God’ and the need for the future generations to surpass (so to speak) the traditional ‘Christian-centred’ Western conception of the human. While Foucault exposed such insights more than half a century ago, they appear to be more actual than ever today with the development and rise in popularity of intellectual movements such as Transhumanism and Posthumanism, which seek to question and propose an alternative to the concepts of ‘man’ or ‘human nature’ in our culture. They rely for that on the same assumption as Foucault and Nietzsche that those concepts (and the meaning we attribute to them) have become ‘obsolete’ as it is and thus must be overcome (at a conceptual, but also a more practical level). Hence, those movements not only echo the important Foucauldian reflection of the 1950s and 1960s on the ‘anthropological question’ but seem to have been literally announced by it, so to speak. The aim of this paper will therefore be to show the relevance of Foucault (and in particular his archaeological method) in understanding the nature of Transhumanism (and Posthumanism), for instance, by analysing and assessing it as a form of discourse that is literally reshaping the way we understand ourselves as human beings in our (post)modern age, drawing for that on a number of key texts including from the early productions of Foucault.

Keywords: foucault, nietzsche, archaeology, transhumanism, posthumanism

Procedia PDF Downloads 70
240 Developing a Research Culture in the Faculty of Engineering and Information Technology at the Central University of Technology, Free State: Implications for Knowledge Management

Authors: Mpho Agnes Mbeo, Patient Rambe

Abstract:

The thirteenth year of the Central University of Technology, Free State’s (CUT) transition from a vocational and professional training orientation institution (i.e. a technikon) into a university with a strong research focus has neither been a smooth nor an easy one. At the heart of this transition was the need to transform the psychological faculties of academic and research staffs compliment who were accustomed to training graduates for industrial placement. The lack of a culture of research that fully embraces a strong ethos of conducting world-class research needed to be addressed. The induction and socialisation of academic staff into the development and execution of cutting-edge research also required the provision of research support and the creation of a conducive academic environment for research, both for emerging and non-research active academics. Drawing on ten cases, comprising four heads of departments, three prolific established researchers, and three emerging researchers, this study explores the challenges faced in establishing a strong research culture at the university. Furthermore, it gives an account of the extent to which the current research interventions have addressed the perceivably “missing research culture”, and the implications of these interventions for knowledge management. Evidence suggests that the endowment of an ideal institutional research environment (comprising strong internet networks, persistent connectivity on and off campus), research peer mentorship, and growing publication outputs should be matched by a coherent research incentive culture and strong research leadership. This is critical to building new knowledge and entrenching knowledge management founded on communities of practice and scholarly networking through the documentation and communication of research findings. The study concludes that the multiple policy documents set for the different domains of research may be creating pressure on researchers to engage research activities and increase output at the expense of research quality.

Keywords: Central University of Technology, performance, publication, research culture, university

Procedia PDF Downloads 173
239 Investigation of Geothermal Gradient of the Niger Delta from Recent Studies

Authors: Adedapo Jepson Olumide, Kurowska Ewa, K. Schoeneich, Ikpokonte A. Enoch

Abstract:

In this paper, subsurface temperature measured from continuous temperature logs were used to determine the geothermal gradient of NigerDelta sedimentary basin. The measured temperatures were corrected to the true subsurface temperatures by applying the American Association of Petroleum Resources (AAPG) correction factor, borehole temperature correction factor with La Max’s correction factor and Zeta Utilities borehole correction factor. Geothermal gradient in this basin ranges from 1.20C to 7.560C/100m. Six geothermal anomalies centres were observed at depth in the southern parts of the Abakaliki anticlinorium around Onitsha, Ihiala, Umuaha area and named A1 to A6 while two more centre appeared at depth of 3500m and 4000m named A7 and A8 respectively. Anomaly A1 describes the southern end of the Abakaliki anticlinorium and extends southwards, anomaly A2 to A5 were found associated with a NW-SE structural alignment of the Calabar hinge line with structures describing the edge of the Niger Delta basin with the basement block of the Oban massif. Anomaly A6 locates in the south-eastern part of the basin offshore while A7 and A8 are located in the south western part of the basin offshore. At the average exploratory depth of 3500m, the geothermal gradient values for these anomalies A1, A2, A3, A4, A5, A6, A7, and A8 are 6.50C/100m, 1.750C/100m, 7.50C/100m, 1.250C/100m, 6.50C/100m, 5.50C/100m, 60C/100m, and 2.250C/100m respectively. Anomaly A8 area may yield higher thermal value at greater depth than 3500m. These results show that anomalies areas of A1, A3, A5, A6 and A7 are potentially prospective and explorable for geothermal energy using abandoned oil wells in the study area. Anomalies A1, A3.A5, A6 occur at areas where drilled boreholes were not exploitable for oil and gas but for the remaining areas where wells are so exploitable there appears no geothermal anomaly. Geothermal energy is environmentally friendly, clean and reversible.

Keywords: temperature logs, geothermal gradient anomalies, alternative energy, Niger delta basin

Procedia PDF Downloads 278
238 Postoperative Budesonide Nasal Irrigation vs Normal Saline Irrigation for Chronic Rhinosinusitis: A Systematic Review and Meta-Analysis

Authors: Rakan Hassan M. Alzahrani, Ziyad Alzahrani, Bader Bashrahil, Abdulrahman Elyasi, Abdullah a Ghaddaf, Rayan Alzahrani, Mohammed Alkathlan, Nawaf Alghamdi, Dakheelallah Almutairi

Abstract:

Background: Corticosteroid irrigations, which regularly involve the off-label use of budesonide mixed with normal saline in high volume Sino-nasal irrigations, have been more commonly used in the management of post-operative chronic rhinosinusitis (CRS). Objective: This article attempted to measure the efficacy of post-operative budesonide nasal irrigation compared to normal saline-alone nasal irrigation in the management of chronic rhinosinusitis (CRS) through a systematic review and meta-analysis of randomized controlled trials (RCTs). Methods: The databases PubMed, Embase, and Cochrane Central Register of Controlled Trials were searched by two independent authors. Only RCTs comparing budesonide irrigation to normal saline alone irrigation for CRS with or without polyposis after functional endoscopic sinus surgery (FESS) were eligible. A random effect analysis model of the reported CRS-related quality of life (QOL) measures and the objective endoscopic assessment scales of the disease was done. Results: Only 6 RCTs met the eligibility criteria, with a total number of participants of 356. Compared to normal saline irrigation, budesonide nasal irrigation showed statically significant improvements in both the CRS-related quality of life (QOL) and the endoscopic findings (MD= -4.22 confidence interval [CI]: -5.63, -2.82 [P < 0.00001]), (SMD= -0.50 confidence interval [CI]: -0.93, -0.06 [P < 0.03]) respectively. Conclusion: Both intervention arms showed improvements in CRS-related QOL and endoscopic findings in post-FESS chronic rhinosinusitis with or without polyposis. However, budesonide irrigation seems to have a slight edge over conventional normal saline irrigation with no reported serious side effects, including hypothalamic-pituitary-adrenal (HPA) axis suppression.

Keywords: Budesonide, chronic rhinosinusitis, corticosteroids, nasal irrigation, normal saline

Procedia PDF Downloads 78
237 Information Tree: Establishment of Lifestyle-Based IT Visual Model

Authors: Chiung-Hui Chen

Abstract:

Traditional service channel is losing its edge due to emerging service technology. To establish interaction with the clients, the service industry is using effective mechanism to give clients direct access to services with emerging technologies. Thus, as service science receives attention, special and unique consumption pattern evolves; henceforth, leading to new market mechanism and influencing attitudes toward life and consumption patterns. The market demand for customized services is thus valued due to the emphasis of personal value, and is gradually changing the demand and supply relationship in the traditional industry. In respect of interior design service, in the process of traditional interior design, a designer converts to a concrete form the concept generated from the ideas and needs dictated by a user (client), by using his/her professional knowledge and drawing tool. The final product is generated through iterations of communication and modification, which is a very time-consuming process. Although this process has been accelerated with the help of computer graphics software today, repeated discussions and confirmations with users are still required to complete the task. In consideration of what is addressed above a space user’s life model is analyzed with visualization technique to create an interaction system modeled after interior design knowledge. The space user document intuitively personal life experience in a model requirement chart, allowing a researcher to analyze interrelation between analysis documents, identify the logic and the substance of data conversion. The repeated data which is documented are then transformed into design information for reuse and sharing. A professional interior designer may sort out the correlation among user’s preference, life pattern and design specification, thus deciding the critical design elements in the process of service design.

Keywords: information design, life model-based, aesthetic computing, communication

Procedia PDF Downloads 298
236 The Attitudinal Effects of Dental Hygiene Students When Changing Conventional Practices of Preventive Therapy in the Dental Hygiene Curriculum

Authors: Shawna Staud, Mary Kaye Scaramucci

Abstract:

Objective: Rubber cup polishing has been a traditional method of preventative therapy in dental hygiene treatment. Newer methods such as air polishing have changed the way dental hygiene care is provided, yet this technique has not been embraced by students in the program nor by practitioners in the workforce. Students entering the workforce tend to follow office protocol and are limited in confidence to introduce technologies learned in the curriculum. This project was designed to help students gain confidence in newer skills and encourage private practice settings to adopt newer technologies for patient care. Our program recently introduced air polishing earlier in the program before the rubber cup technique to determine if students would embrace the technology to become leading-edge professionals when they enter the marketplace. Methods: The class of 2022 was taught the traditional method of polishing in the first-year curriculum and air polishing in the second-year curriculum. The class of 2023 will be taught the air polishing method in the first-year curriculum and the traditional method of polishing in the second-year curriculum. Pre- and post-graduation survey data will be collected from both cohorts. Descriptive statistics and pre and post-paired t-tests with alpha set at .05 to compare pre and post-survey results will be used to assess data. Results: This study is currently in progress, with a completion date of October 2023. The class of 2022 completed the pre-graduation survey in the spring of 2022. The post-gradation survey will be sent out in October 2022. The class of 2023 cohort will be surveyed in the spring of 2023 and October 2023. Conclusion: Our hypothesis is students who are taught air polishing first will be more inclined to adopt that skill in private practice, thereby embracing newer technology and improving oral health care.

Keywords: luggage handling system at world’s largest pilgrimage center

Procedia PDF Downloads 102
235 Miniaturization of Germanium Photo-Detectors by Using Micro-Disk Resonator

Authors: Haifeng Zhou, Tsungyang Liow, Xiaoguang Tu, Eujin Lim, Chao Li, Junfeng Song, Xianshu Luo, Ying Huang, Lianxi Jia, Lianwee Luo, Kim Dowon, Qing Fang, Mingbin Yu, Guoqiang Lo

Abstract:

Several Germanium photodetectors (PD) built on silicon micro-disks are fabricated on the standard Si photonics multiple project wafers (MPW) and demonstrated to exhibit very low dark current, satisfactory operation bandwidth and moderate responsivity. Among them, a vertical p-i-n Ge PD based on a 2.0 µm-radius micro-disk has a dark current of as low as 35 nA, compared to a conventional PD current of 1 µA with an area of 100 µm2. The operation bandwidth is around 15 GHz at a reverse bias of 1V. The responsivity is about 0.6 A/W. Microdisk is a striking planar structure in integrated optics to enhance light-matter interaction and construct various photonics devices. The disk geometries feature in strongly and circularly confining light into an ultra-small volume in the form of whispering gallery modes. A laser may benefit from a microdisk in which a single mode overlaps the gain materials both spatially and spectrally. Compared to microrings, micro-disk removes the inner boundaries to enable even better compactness, which also makes it very suitable for some scenarios that electrical connections are needed. For example, an ultra-low power (≈ fJ) athermal Si modulator has been demonstrated with a bit rate of 25Gbit/s by confining both photons and electrically-driven carriers into a microscale volume.In this work, we study Si-based PDs with Ge selectively grown on a microdisk with the radius of a few microns. The unique feature of using microdisk for Ge photodetector is that mode selection is not important. In the applications of laser or other passive optical components, microdisk must be designed very carefully to excite the fundamental mode in a microdisk in that essentially the microdisk usually supports many higher order modes in the radial directions. However, for detector applications, this is not an issue because the local light absorption is mode insensitive. Light power carried by all modes are expected to be converted into photo-current. Another benefit of using microdisk is that the power circulation inside avoids any introduction of the reflector. A complete simulation model with all involved materials taken into account is established to study the promise of microdisk structures for photodetector by using finite difference time domain (FDTD) method. By viewing from the current preliminary data, the directions to further improve the device performance are also discussed.

Keywords: integrated optical devices, silicon photonics, micro-resonator, photodetectors

Procedia PDF Downloads 407
234 Tourist Behavior Towards Blockchain-Based Payments

Authors: A. Šapkauskienė, A. Mačerinskienė, R. Andrulienė, R. Bruzgė, S. Masteika, K. Driaunys

Abstract:

The COVID-19 pandemic has affected not only world markets and economies but also the daily lives of customers and their payment habits. The pandemic has accelerated the digital transformation, so the role of technology will become even more important post-COVID. Although the popularity of cryptocurrencies has reached unprecedented heights, there are still obstacles, such as a lack of consumer experience and distrust of these technologies, so exploring the role of cryptocurrency and blockchain in the context of international travel becomes extremely important. Research on tourists’ intentions to use cryptocurrencies for payment purposes is limited due to the small number of research studies. To fill this research gap, an exploratory study based on the analysis of survey data was conducted. The purpose of the research is to explore how the behavior of tourists has changed making their financial transactions when paying for the tourism services in order to determine the intention to pay in cryptocurrencies. Behavioral intention can be examined as a dependent variable that is useful for the study of the acceptance of blockchain as cutting-edge technology. Therefore, this study examines the intention of travelers to use cryptocurrencies in electronic payments for tourism services. Several studies have shown that the intention to accept payments in a cryptocurrency is affected by the perceived usefulness of these payments and the perceived ease of use. The findings deepen our understanding of the readiness of service users to apply for blockchain-based payment in the tourism sector. The tourism industry has to focus not only on the technology but on consumers who can use cryptocurrencies, creating new possibilities and increasing business competitiveness. Based on research results, suggestions are made to guide future research on the use of cryptocurrencies by tourists in the tourism industry. Therefore, in line with the rapid expansion of virtual currency users, market capitalization, and payment in cryptographic currencies, it is necessary to explore the possibilities of implementing a blockchain-based system aiming to promote the use of services in the tourism sector as the most affected by the pandemic.

Keywords: behavioral intention, blockchain-based payment, cryptocurrency, tourism

Procedia PDF Downloads 105
233 Development of an Appropriate Method for the Determination of Multiple Mycotoxins in Pork Processing Products by UHPLC-TCFLD

Authors: Jason Gica, Yi-Hsieng Samuel Wu, Deng-Jye Yang, Yi-Chen Chen

Abstract:

Mycotoxins, harmful secondary metabolites produced by certain fungi species, pose significant risks to animals and humans worldwide. Their stable properties lead to contamination during grain harvesting, transportation, and storage, as well as in processed food products. The prevalence of mycotoxin contamination has attracted significant attention due to its adverse impact on food safety and global trade. The secondary contamination pathway from animal products has been identified as an important route of exposure, posing health risks for livestock and humans consuming contaminated products. Pork, one of the highly consumed meat products in Taiwan according to the National Food Consumption Database, plays a critical role in the nation's diet and economy. Given its substantial consumption, pork processing products are a significant component of the food supply chain and a potential source of mycotoxin contamination. This study is paramount for formulating effective regulations and strategies to mitigate mycotoxin-related risks in the food supply chain. By establishing a reliable analytical method, this research contributes to safeguarding public health and enhancing the quality of pork processing products. The findings will serve as valuable guidance for policymakers, food industries, and consumers to ensure a safer food supply chain in the face of emerging mycotoxin challenges. An innovative and efficient analytical approach is proposed using Ultra-High Performance Liquid Chromatography coupled with Temperature Control Fluorescence Detector Light (UHPLC-TCFLD) to determine multiple mycotoxins in pork meat samples due to its exceptional capacity to detect multiple mycotoxins at the lowest levels of concentration, making it highly sensitive and reliable for comprehensive mycotoxin analysis. Additionally, its ability to simultaneously detect multiple mycotoxins in a single run significantly reduces the time and resources required for analysis, making it a cost-effective solution for monitoring mycotoxin contamination in pork processing products. The research aims to optimize the efficient mycotoxin QuEChERs extraction method and rigorously validate its accuracy and precision. The results will provide crucial insights into mycotoxin levels in pork processing products.

Keywords: multiple-mycotoxin analysis, pork processing products, QuEChERs, UHPLC-TCFLD, validation

Procedia PDF Downloads 68
232 High-Pressure Polymorphism of 4,4-Bipyridine Hydrobromide

Authors: Michalina Aniola, Andrzej Katrusiak

Abstract:

4,4-Bipyridine is an important compound often used in chemical practice and more recently frequently applied for designing new metal organic framework (MoFs). Here we present a systematic high-pressure study of its hydrobromide salt. 4,4-Bipyridine hydrobromide monohydrate, 44biPyHBrH₂O, at ambient-pressure is orthorhombic, space group P212121 (phase a). Its hydrostatic compression shows that it is stable to 1.32 GPa at least. However, the recrystallization above 0.55 GPa reveals a new hidden b-phase (monoclinic, P21/c). Moreover, when the 44biPyHBrH2O is heated to high temperature the chemical reactions of this compound in methanol solution can be observed. High-pressure experiments were performed using a Merrill-Bassett diamond-anvil cell (DAC), modified by mounting the anvils directly on the steel supports, and X-ray diffraction measurements were carried out on a KUMA and Excalibur diffractometer equipped with an EOS CCD detector. At elevated pressure, the crystal of 44biPyHBrH₂O exhibits several striking and unexpected features. No signs of instability of phase a were detected to 1.32 GPa, while phase b becomes stable at above 0.55 GPa, as evidenced by its recrystallizations. Phases a and b of 44biPyHBrH2O are partly isostructural: their unit-cell dimensions and the arrangement of ions and water molecules are similar. In phase b the HOH-Br- chains double the frequency of their zigzag motifs, compared to phase a, and the 44biPyH+ cations change their conformation. Like in all monosalts of 44biPy determined so far, in phase a the pyridine rings are twisted by about 30 degrees about bond C4-C4 and in phase b they assume energy-unfavorable planar conformation. Another unusual feature of 44biPyHBrH2O is that all unit-cell parameters become longer on the transition from phase a to phase b. Thus the volume drop on the transition to high-pressure phase b totally depends on the shear strain of the lattice. Higher temperature triggers chemical reactions of 44biPyHBrH2O with methanol. When the saturated methanol solution compound precipitated at 0.1 GPa and temperature of 423 K was required to dissolve all the sample, the subsequent slow recrystallization at isochoric conditions resulted in disalt 4,4-bipyridinium dibromide. For the 44biPyHBrH2O sample sealed in the DAC at 0.35 GPa, then dissolved at isochoric conditions at 473 K and recrystallized by slow controlled cooling, a reaction of N,N-dimethylation took place. It is characteristic that in both high-pressure reactions of 44biPyHBrH₂O the unsolvated disalt products were formed and that free base 44biPy and H₂O remained in the solution. The observed reactions indicate that high pressure destabilized ambient-pressure salts and favors new products. Further studies on pressure-induced reactions are carried out in order to better understand the structural preferences induced by pressure.

Keywords: conformation, high-pressure, negative area compressibility, polymorphism

Procedia PDF Downloads 246
231 Sequential and Combinatorial Pre-Treatment Strategy of Lignocellulose for the Enhanced Enzymatic Hydrolysis of Spent Coffee Waste

Authors: Rajeev Ravindran, Amit K. Jaiswal

Abstract:

Waste from the food-processing industry is produced in large amount and contains high levels of lignocellulose. Due to continuous accumulation throughout the year in large quantities, it creates a major environmental problem worldwide. The chemical composition of these wastes (up to 75% of its composition is contributed by polysaccharide) makes it inexpensive raw material for the production of value-added products such as biofuel, bio-solvents, nanocrystalline cellulose and enzymes. In order to use lignocellulose as the raw material for the microbial fermentation, the substrate is subjected to enzymatic treatment, which leads to the release of reducing sugars such as glucose and xylose. However, the inherent properties of lignocellulose such as presence of lignin, pectin, acetyl groups and the presence of crystalline cellulose contribute to recalcitrance. This leads to poor sugar yields upon enzymatic hydrolysis of lignocellulose. A pre-treatment method is generally applied before enzymatic treatment of lignocellulose that essentially removes recalcitrant components in biomass through structural breakdown. Present study is carried out to find out the best pre-treatment method for the maximum liberation of reducing sugars from spent coffee waste (SPW). SPW was subjected to a range of physical, chemical and physico-chemical pre-treatment followed by a sequential, combinatorial pre-treatment strategy is also applied on to attain maximum sugar yield by combining two or more pre-treatments. All the pre-treated samples were analysed for total reducing sugar followed by identification and quantification of individual sugar by HPLC coupled with RI detector. Besides, generation of any inhibitory compounds such furfural, hydroxymethyl furfural (HMF) which can hinder microbial growth and enzyme activity is also monitored. Results showed that ultrasound treatment (31.06 mg/L) proved to be the best pre-treatment method based on total reducing content followed by dilute acid hydrolysis (10.03 mg/L) while galactose was found to be the major monosaccharide present in the pre-treated SPW. Finally, the results obtained from the study were used to design a sequential lignocellulose pre-treatment protocol to decrease the formation of enzyme inhibitors and increase sugar yield on enzymatic hydrolysis by employing cellulase-hemicellulase consortium. Sequential, combinatorial treatment was found better in terms of total reducing yield and low content of the inhibitory compounds formation, which could be due to the fact that this mode of pre-treatment combines several mild treatment methods rather than formulating a single one. It eliminates the need for a detoxification step and potential application in the valorisation of lignocellulosic food waste.

Keywords: lignocellulose, enzymatic hydrolysis, pre-treatment, ultrasound

Procedia PDF Downloads 366