Search results for: automated teller machines (atm)
1162 Mesoporous Nanocomposites for Sustained Release Applications
Authors: Daniela Istrati, Alina Morosan, Maria Stanca, Bogdan Purcareanu, Adrian Fudulu, Laura Olariu, Alice Buteica, Ion Mindrila, Rodica Cristescu, Dan Eduard Mihaiescu
Abstract:
Our present work is related to the synthesis, characterization and applications of new nanocomposite materials based on silica mesoporous nanocompozites systems. The nanocomposite support was obtained by using a specific step–by–step multilayer structure buildup synthetic route, characterized by XRD (X-Ray Difraction), TEM (Transmission Electron Microscopy), FT-IR (Fourier Transform-Infra Red Spectrometry), BET (Brunauer–Emmett–Teller method) and loaded with Salvia officinalis plant extract obtained by a hydro-alcoholic extraction route. The sustained release of the target compounds was studied by a modified LC method, proving low release profiles, as expected for the high specific surface area support. The obtained results were further correlated with the in vitro / in vivo behavior of the nanocomposite material and recommending the silica mesoporous nanocomposites as good candidates for biomedical applications. Acknowledgements: This study has been funded by the Research Project PN-III-P2-2.1-PTE-2016-0160, 49-PTE / 2016 (PROZECHIMED) and Project Number PN-III-P4-ID-PCE-2016-0884 / 2017.Keywords: biomedical, mesoporous, nanocomposites, natural products, sustained release
Procedia PDF Downloads 2181161 The Social Psychology of Illegal Game Room Addiction in the Historic Chinatown District of Honolulu, Hawaii: Illegal Compulsive Gambling, Chinese-Polynesian Organized Crime Syndicates, Police Corruption, and Loan Sharking Rings
Authors: Gordon James Knowles
Abstract:
Historically the Chinatown district in Sandwich Islands has been plagued with the traditional vice crimes of illegal drugs, gambling, and prostitution since the early 1800s. However, a new form of psychologically addictive arcade style table gambling machines has become the dominant form of illegal revenue made in Honolulu, Hawaii. This study attempts to document the drive, desire, or will to play and wager with arcade style video gaming and understand the role of illegal game rooms in facilitating pathological gambling addiction. Indicators of police corruption by Chinese organized crime syndicates related to protection rackets, bribery, and pay-offs were revealed. Information fusion from a police science and sociological intelligence perspective indicates insurgent warfare is being waged on the streets of Honolulu by the People’s Republic of China. This state-sponsored communist terrorism in the Hawaiian Islands used “contactless” irregular warfare entailing: (1) the deployment of psychologically addictive gambling machines, (2) the distribution of the physically addictive fentanyl drug as a lethal chemical weapon, and (3) psychological warfare by circulating pro-China anti-American propaganda newspapers targeted at the small island populace.Keywords: Chinese and Polynesian organized crime, china daily newspaper, electronic arcade style table games, gaming technology addiction, illegal compulsive gambling, and police intelligence
Procedia PDF Downloads 741160 Preparation and Characterization of Modified ZnO Incorporated into Mesoporous MCM-22 Catalysts and Their Catalytic Performances of Crude Jatropha Oil to Biodiesel
Authors: Bashir Abubakar Abdulkadir, Anita Ramli, Lim Jun Wei, Yoshimitsu Uemura
Abstract:
In this study, the ZnO/MCM-22 catalyst with different ZnO loading were prepared using conventional wet impregnation process and the catalyst activity was tested for biodiesel production from Jatropha oil. The effects of reaction parameters with regards to catalyst activity were investigated. The synthesized catalysts samples were then characterized by X-ray diffraction (XRD) for crystal phase, Brunauer–Emmett–Teller (BET) for surface area, pore volume and pore size, Field Emission Scanning electron microscope attached to energy dispersive x-ray (FESEM/EDX) for morphology and elemental composition and TPD (NH3 and CO2) for basic and acidic properties of the catalyst. The XRD spectra couple with the EDX result shows the presence of ZnO in the catalyst confirming the positive intercalation of the metal oxide into the mesoporous MCM-22. The synthesized catalyst was confirmed to be mesoporous according to BET findings. Also, the catalysts can be considered as a bifunctional catalyst based on TPD outcomes. Transesterification results showed that the synthesized catalyst was highly efficient and effective to be used for biodiesel production from low grade oil such as Jatropha oil and other industrial application where the high fatty acid methyl ester (FAMEs) yield was achieved at moderate reaction conditions. It was also discovered that the catalyst can be used more than five (5) runs with little deactivation confirming the catalyst to be highly active and stable to the heat of reaction.Keywords: MCM-22, synthesis, transesterification, ZnO
Procedia PDF Downloads 2081159 2D Convolutional Networks for Automatic Segmentation of Knee Cartilage in 3D MRI
Authors: Ananya Ananya, Karthik Rao
Abstract:
Accurate segmentation of knee cartilage in 3-D magnetic resonance (MR) images for quantitative assessment of volume is crucial for studying and diagnosing osteoarthritis (OA) of the knee, one of the major causes of disability in elderly people. Radiologists generally perform this task in slice-by-slice manner taking 15-20 minutes per 3D image, and lead to high inter and intra observer variability. Hence automatic methods for knee cartilage segmentation are desirable and are an active field of research. This paper presents design and experimental evaluation of 2D convolutional neural networks based fully automated methods for knee cartilage segmentation in 3D MRI. The architectures are validated based on 40 test images and 60 training images from SKI10 dataset. The proposed methods segment 2D slices one by one, which are then combined to give segmentation for whole 3D images. Proposed methods are modified versions of U-net and dilated convolutions, consisting of a single step that segments the given image to 5 labels: background, femoral cartilage, tibia cartilage, femoral bone and tibia bone; cartilages being the primary components of interest. U-net consists of a contracting path and an expanding path, to capture context and localization respectively. Dilated convolutions lead to an exponential expansion of receptive field with only a linear increase in a number of parameters. A combination of modified U-net and dilated convolutions has also been explored. These architectures segment one 3D image in 8 – 10 seconds giving average volumetric Dice Score Coefficients (DSC) of 0.950 - 0.962 for femoral cartilage and 0.951 - 0.966 for tibia cartilage, reference being the manual segmentation.Keywords: convolutional neural networks, dilated convolutions, 3 dimensional, fully automated, knee cartilage, MRI, segmentation, U-net
Procedia PDF Downloads 2611158 Copyright Clearance for Artificial Intelligence Training Data: Challenges and Solutions
Authors: Erva Akin
Abstract:
– The use of copyrighted material for machine learning purposes is a challenging issue in the field of artificial intelligence (AI). While machine learning algorithms require large amounts of data to train and improve their accuracy and creativity, the use of copyrighted material without permission from the authors may infringe on their intellectual property rights. In order to overcome copyright legal hurdle against the data sharing, access and re-use of data, the use of copyrighted material for machine learning purposes may be considered permissible under certain circumstances. For example, if the copyright holder has given permission to use the data through a licensing agreement, then the use for machine learning purposes may be lawful. It is also argued that copying for non-expressive purposes that do not involve conveying expressive elements to the public, such as automated data extraction, should not be seen as infringing. The focus of such ‘copy-reliant technologies’ is on understanding language rules, styles, and syntax and no creative ideas are being used. However, the non-expressive use defense is within the framework of the fair use doctrine, which allows the use of copyrighted material for research or educational purposes. The questions arise because the fair use doctrine is not available in EU law, instead, the InfoSoc Directive provides for a rigid system of exclusive rights with a list of exceptions and limitations. One could only argue that non-expressive uses of copyrighted material for machine learning purposes do not constitute a ‘reproduction’ in the first place. Nevertheless, the use of machine learning with copyrighted material is difficult because EU copyright law applies to the mere use of the works. Two solutions can be proposed to address the problem of copyright clearance for AI training data. The first is to introduce a broad exception for text and data mining, either mandatorily or for commercial and scientific purposes, or to permit the reproduction of works for non-expressive purposes. The second is that copyright laws should permit the reproduction of works for non-expressive purposes, which opens the door to discussions regarding the transposition of the fair use principle from the US into EU law. Both solutions aim to provide more space for AI developers to operate and encourage greater freedom, which could lead to more rapid innovation in the field. The Data Governance Act presents a significant opportunity to advance these debates. Finally, issues concerning the balance of general public interests and legitimate private interests in machine learning training data must be addressed. In my opinion, it is crucial that robot-creation output should fall into the public domain. Machines depend on human creativity, innovation, and expression. To encourage technological advancement and innovation, freedom of expression and business operation must be prioritised.Keywords: artificial intelligence, copyright, data governance, machine learning
Procedia PDF Downloads 831157 Innovative Technologies for Aeration and Feeding of Fish in Aquaculture with Minimal Impact on the Environment
Authors: Vasile Caunii, Andreea D. Serban, Mihaela Ivancia
Abstract:
The paper presents a new approach in terms of the circular economy of technologies for feeding and aeration of accumulations and water basins for fish farming and aquaculture. Because fish is and will be one of the main foods on the planet, the use of bio-eco-technologies is a priority for all producers. The technologies proposed in the paper want to reduce by a substantial percentage the costs of operation of ponds and water accumulation, using non-polluting technologies with minimal impact on the environment. The paper proposes two innovative, intelligent systems, fully automated that use a common platform, completely eco-friendly. One system is intended to aerate the water of the fish pond, and the second is intended to feed the fish by dispersing an optimal amount of fodder, depending on population size, age and habits. Both systems use a floating platform, regenerative energy sources, are equipped with intelligent and innovative systems, and in addition to fully automated operation, significantly reduce the costs of aerating water accumulations (natural or artificial) and feeding fish. The intelligent system used for feeding, in addition, to reduce operating costs, optimizes the amount of food, thus preventing water pollution and the development of bacteria, microorganisms. The advantages of the systems are: increasing the yield of fish production, these are green installations, with zero pollutant emissions, can be arranged anywhere on the water surface, depending on the user's needs, can operate autonomously or remotely controlled, if there is a component failure, the system provides the operator with accurate data on the issue, significantly reducing maintenance costs, transmit data about the water physical and chemical parameters.Keywords: bio-eco-technologies, economy, environment, fish
Procedia PDF Downloads 1501156 Investigation of Boll Properties on Cotton Picker Machine Performance
Authors: Shahram Nowrouzieh, Abbas Rezaei Asl, Mohamad Ali Jafari
Abstract:
Cotton, as a strategic crop, plays an important role in providing human food and clothing need, because of its oil, protein, and fiber. Iran has been one of the largest cotton producers in the world in the past, but unfortunately, for economic reasons, its production is reduced now. One of the ways to reduce the cost of cotton production is to expand the mechanization of cotton harvesting. Iranian farmers do not accept the function of cotton harvesters. One reason for this lack of acceptance of cotton harvesting machines is the number of field losses on these machines. So, the majority of cotton fields are harvested by hand. Although the correct setting of the harvesting machine is very important in the cotton losses, the morphological properties of the cotton plant also affect the performance of cotton harvesters. In this study, the effect of some cotton morphological properties such as the height of the cotton plant, number, and length of sympodial and monopodial branches, boll dimensions, boll weight, number of carpels and bracts angle were evaluated on the performance of cotton picker. In this research, the efficiency of John Deere 9920 spindle Cotton picker is investigated on five different Iranian cotton cultivars. The results indicate that there was a significant difference between the five cultivars in terms of machine harvest efficiency. Golestan cultivar showed the best cotton harvester performance with an average of 87.6% of total harvestable seed cotton and Khorshid cultivar had the least cotton harvester performance. The principal component analysis showed that, at 50.76% probability, the cotton picker efficiency is affected by the bracts angle positively and by boll dimensions, the number of carpels and the height of cotton plants negatively. The seed cotton remains (in the plant and on the ground) after harvester in PCA scatter plot were in the same zone with boll dimensions and several carpels.Keywords: cotton, bract, harvester, carpel
Procedia PDF Downloads 1351155 Machine Learning Techniques in Bank Credit Analysis
Authors: Fernanda M. Assef, Maria Teresinha A. Steiner
Abstract:
The aim of this paper is to compare and discuss better classifier algorithm options for credit risk assessment by applying different Machine Learning techniques. Using records from a Brazilian financial institution, this study uses a database of 5,432 companies that are clients of the bank, where 2,600 clients are classified as non-defaulters, 1,551 are classified as defaulters and 1,281 are temporarily defaulters, meaning that the clients are overdue on their payments for up 180 days. For each case, a total of 15 attributes was considered for a one-against-all assessment using four different techniques: Artificial Neural Networks Multilayer Perceptron (ANN-MLP), Artificial Neural Networks Radial Basis Functions (ANN-RBF), Logistic Regression (LR) and finally Support Vector Machines (SVM). For each method, different parameters were analyzed in order to obtain different results when the best of each technique was compared. Initially the data were coded in thermometer code (numerical attributes) or dummy coding (for nominal attributes). The methods were then evaluated for each parameter and the best result of each technique was compared in terms of accuracy, false positives, false negatives, true positives and true negatives. This comparison showed that the best method, in terms of accuracy, was ANN-RBF (79.20% for non-defaulter classification, 97.74% for defaulters and 75.37% for the temporarily defaulter classification). However, the best accuracy does not always represent the best technique. For instance, on the classification of temporarily defaulters, this technique, in terms of false positives, was surpassed by SVM, which had the lowest rate (0.07%) of false positive classifications. All these intrinsic details are discussed considering the results found, and an overview of what was presented is shown in the conclusion of this study.Keywords: artificial neural networks (ANNs), classifier algorithms, credit risk assessment, logistic regression, machine Learning, support vector machines
Procedia PDF Downloads 1031154 Robust Segmentation of Salient Features in Automatic Breast Ultrasound (ABUS) Images
Authors: Lamees Nasser, Yago Diez, Robert Martí, Joan Martí, Ibrahim Sadek
Abstract:
Automated 3D breast ultrasound (ABUS) screening is a novel modality in medical imaging because of its common characteristics shared with other ultrasound modalities in addition to the three orthogonal planes (i.e., axial, sagittal, and coronal) that are useful in analysis of tumors. In the literature, few automatic approaches exist for typical tasks such as segmentation or registration. In this work, we deal with two problems concerning ABUS images: nipple and rib detection. Nipple and ribs are the most visible and salient features in ABUS images. Determining the nipple position plays a key role in some applications for example evaluation of registration results or lesion follow-up. We present a nipple detection algorithm based on color and shape of the nipple, besides an automatic approach to detect the ribs. In point of fact, rib detection is considered as one of the main stages in chest wall segmentation. This approach consists of four steps. First, images are normalized in order to minimize the intensity variability for a given set of regions within the same image or a set of images. Second, the normalized images are smoothed by using anisotropic diffusion filter. Next, the ribs are detected in each slice by analyzing the eigenvalues of the 3D Hessian matrix. Finally, a breast mask and a probability map of regions detected as ribs are used to remove false positives (FP). Qualitative and quantitative evaluation obtained from a total of 22 cases is performed. For all cases, the average and standard deviation of the root mean square error (RMSE) between manually annotated points placed on the rib surface and detected points on rib borders are 15.1188 mm and 14.7184 mm respectively.Keywords: Automated 3D Breast Ultrasound, Eigenvalues of Hessian matrix, Nipple detection, Rib detection
Procedia PDF Downloads 3301153 STML: Service Type-Checking Markup Language for Services of Web Components
Authors: Saqib Rasool, Adnan N. Mian
Abstract:
Web components are introduced as the latest standard of HTML5 for writing modular web interfaces for ensuring maintainability through the isolated scope of web components. Reusability can also be achieved by sharing plug-and-play web components that can be used as off-the-shelf components by other developers. A web component encapsulates all the required HTML, CSS and JavaScript code as a standalone package which must be imported for integrating a web component within an existing web interface. It is then followed by the integration of web component with the web services for dynamically populating its content. Since web components are reusable as off-the-shelf components, these must be equipped with some mechanism for ensuring their proper integration with web services. The consistency of a service behavior can be verified through type-checking. This is one of the popular solutions for improving the quality of code in many programming languages. However, HTML does not provide type checking as it is a markup language and not a programming language. The contribution of this work is to introduce a new extension of HTML called Service Type-checking Markup Language (STML) for adding support of type checking in HTML for JSON based REST services. STML can be used for defining the expected data types of response from JSON based REST services which will be used for populating the content within HTML elements of a web component. Although JSON has five data types viz. string, number, boolean, object and array but STML is made to supports only string, number and object. This is because of the fact that both object and array are considered as string, when populated in HTML elements. In order to define the data type of any HTML element, developer just needs to add the custom STML attributes of st-string, st-number and st-boolean for string, number and boolean respectively. These all annotations of STML are used by the developer who is writing a web component and it enables the other developers to use automated type-checking for ensuring the proper integration of their REST services with the same web component. Two utilities have been written for developers who are using STML based web components. One of these utilities is used for automated type-checking during the development phase. It uses the browser console for showing the error description if integrated web service is not returning the response with expected data type. The other utility is a Gulp based command line utility for removing the STML attributes before going in production. This ensures the delivery of STML free web pages in the production environment. Both of these utilities have been tested to perform type checking of REST services through STML based web components and results have confirmed the feasibility of evaluating service behavior only through HTML. Currently, STML is designed for automated type-checking of integrated REST services but it can be extended to introduce a complete service testing suite based on HTML only, and it will transform STML from Service Type-checking Markup Language to Service Testing Markup Language.Keywords: REST, STML, type checking, web component
Procedia PDF Downloads 2541152 NaOH/Pumice and LiOH/Pumice as Heterogeneous Solid Base Catalysts for Biodiesel Production from Soybean Oil: An Optimization Study
Authors: Joy Marie Mora, Mark Daniel De Luna, Tsair-Wang Chung
Abstract:
Transesterification reaction of soybean oil with methanol was carried out to produce fatty acid methyl esters (FAME) using calcined alkali metal (Na and Li) supported by pumice silica as the solid base catalyst. Pumice silica catalyst was activated by loading alkali metal ions to its surface via an ion-exchange method. Response surface methodology (RSM) in combination with Box-Behnken design (BBD) was used to optimize the operating parameters in biodiesel production, namely: reaction temperature, methanol to oil molar ratio, reaction time, and catalyst concentration. Using the optimized sets of parameters, FAME yields using sodium and lithium silicate catalysts were 98.80% and 98.77%, respectively. A pseudo-first order kinetic equation was applied to evaluate the kinetic parameters of the reaction. The prepared catalysts were characterized by several techniques such as X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FTIR), Brunauer-Emmett-Teller (BET) sorptometer, and scanning electron microscopy (SEM). In addition, the reusability of the catalysts was successfully tested in two subsequent cycles.Keywords: alkali metal, biodiesel, Box-Behnken design, heterogeneous catalyst, kinetics, optimization, pumice, transesterification
Procedia PDF Downloads 3061151 TiO2 Nanowires as Efficient Heterogeneous Photocatalysts for Waste-Water Treatment
Authors: Gul Afreen, Sreedevi Upadhyayula, Mahendra K. Sunkara
Abstract:
One-dimensional (1D) nanostructures like nanowires, nanotubes, and nanorods find variety of practical application owing to their unique physico-chemical properties. In this work, TiO2 nanowires were synthesized by direct oxidation of titanium particles in a unique microwave plasma jet reactor. The prepared TiO2 nanowires manifested the flexible features, and were characterized by using X-ray diffraction, Brunauer-Emmett-Teller (BET) surface area analyzer, UV-Visible and FTIR spectrophotometers, Scanning electron microscope, and Transmission electron microscope. Further, the photodegradation efficiency of these nanowires were tested against toxic organic dye like methylene blue (MB) and the results were compared with the commercial TiO2. It was found that TiO2 nanowires exhibited superior photocatalytic performance (89%) as compared to commercial TiO2 (75%) after 60 min of reaction. This is attributed to the lower recombination rate and increased interfacial charge transfer in TiO2 nanowire. Pseudo-first order kinetic modelling performed with the experimental results revealed that the rate constant of photodegradation in case of TiO2 nanowire was 1.3 times higher than that of commercial TiO2. Superoxide radical (O2˙−) was found to be the major contributor in the photodegradation mechanism. Based on the trapping experiments, a plausible mechanism of the photocatalytic reaction is discussed.Keywords: heterogeneous catalysis, photodegradation, reactive oxygen species, TiO₂ nanowires
Procedia PDF Downloads 1441150 Produce Large Surface Area Activated Carbon from Biomass for Water Treatment
Authors: Rashad Al-Gaashani
Abstract:
The physicochemical activation method was used to produce high-quality activated carbon (AC) with a large surface area of about 2000 m2/g from low-cost and abundant biomass wastes in Qatar, namely date seeds. X-Ray diffraction (XRD), scanning electron spectroscopy (SEM), energy dispersive X-Ray spectroscopy (EDS), and Brunauer-Emmett-Teller (BET) surface area analysis was used to evaluate the AC samples. AC produced from date seeds has a wide range of pores available, including micro- and nano-pores. This type of AC with a well-developed pore structure may be very attractive for different applications, including air and water purification from micro and nano pollutants. Heavy metals iron (III) and copper (II) ions were removed from wastewater using the AC produced using a batch adsorption technique. The AC produced from date seeds biomass wastes shows high removal of heavy metals such as iron (III) ions (100%) and copper (II) ions (97.25%). The highest removal of copper (II) ions (100%) with AC produced from date seeds was found at pH 8, whereas the lowest removal (22.63%) occurred at pH 2. The effect of adsorption time, adsorbent dose, and pH on the removal of heavy metals was studied.Keywords: activated carbon, date seeds, biomass, heavy metals removal, water treatment
Procedia PDF Downloads 761149 Cable De-Commissioning of Legacy Accelerators at CERN
Authors: Adya Uluwita, Fernando Pedrosa, Georgi Georgiev, Christian Bernard, Raoul Masterson
Abstract:
CERN is an international organisation funded by 23 countries that provide the particle physics community with excellence in particle accelerators and other related facilities. Founded in 1954, CERN has a wide range of accelerators that allow groundbreaking science to be conducted. Accelerators bring particles to high levels of energy and make them collide with each other or with fixed targets, creating specific conditions that are of high interest to physicists. A chain of accelerators is used to ramp up the energy of particles and eventually inject them into the largest and most recent one: the Large Hadron Collider (LHC). Among this chain of machines is, for instance the Proton Synchrotron, which was started in 1959 and is still in operation. These machines, called "injectors”, keep evolving over time, as well as the related infrastructure. Massive decommissioning of obsolete cables started in 2015 at CERN in the frame of the so-called "injectors de-cabling project phase 1". Its goal was to replace aging cables and remove unused ones, freeing space for new cables necessary for upgrades and consolidation campaigns. To proceed with the de-cabling, a project co-ordination team was assembled. The start of this project led to the investigation of legacy cables throughout the organisation. The identification of cables stacked over half a century proved to be arduous. Phase 1 of the injectors de-cabling was implemented for 3 years with success after overcoming some difficulties. Phase 2, started 3 years later, focused on improving safety and structure with the introduction of a quality assurance procedure. This paper discusses the implementation of this quality assurance procedure throughout phase 2 of the project and the transition between the two phases. Over hundreds of kilometres of cable were removed in the injectors complex at CERN from 2015 to 2023.Keywords: CERN, de-cabling, injectors, quality assurance procedure
Procedia PDF Downloads 931148 The Logistics Equation and Fractal Dimension in Escalators Operations
Authors: Ali Albadri
Abstract:
The logistics equation has never been used or studied in scientific fields outside the field of ecology. It has never been used to understand the behavior of a dynamic system of mechanical machines, like an escalator. We have studied the compatibility of the logistic map against real measurements from an escalator. This study has proven that there is good compatibility between the logistics equation and the experimental measurements. It has discovered the potential of a relationship between the fractal dimension and the non-linearity parameter, R, in the logistics equation. The fractal dimension increases as the R parameter (non-linear parameter) increases. It implies that the fractal dimension increases as the phase of the life span of the machine move from the steady/stable phase to the periodic double phase to a chaotic phase. The fractal dimension and the parameter R can be used as a tool to verify and check the health of machines. We have come up with a theory that there are three areas of behaviors, which they can be classified during the life span of a machine, a steady/stable stage, a periodic double stage, and a chaotic stage. The level of attention to the machine differs depending on the stage that the machine is in. The rate of faults in a machine increases as the machine moves through these three stages. During the double period and the chaotic stages, the number of faults starts to increase and become less predictable. The rate of predictability improves as our monitoring of the changes in the fractal dimension and the parameter R improves. The principles and foundations of our theory in this work have and will have a profound impact on the design of systems, on the way of operation of systems, and on the maintenance schedules of the systems. The systems can be mechanical, electrical, or electronic. The discussed methodology in this paper will give businesses the chance to be more careful at the design stage and planning for maintenance to control costs. The findings in this paper can be implied and used to correlate the three stages of a mechanical system to more in-depth mechanical parameters like wear and fatigue life.Keywords: logistcs map, bifurcation map, fractal dimension, logistics equation
Procedia PDF Downloads 1081147 Immature Palm Tree Detection Using Morphological Filter for Palm Counting with High Resolution Satellite Image
Authors: Nur Nadhirah Rusyda Rosnan, Nursuhaili Najwa Masrol, Nurul Fatiha MD Nor, Mohammad Zafrullah Mohammad Salim, Sim Choon Cheak
Abstract:
Accurate inventories of oil palm planted areas are crucial for plantation management as this would impact the overall economy and production of oil. One of the technological advancements in the oil palm industry is semi-automated palm counting, which is replacing conventional manual palm counting via digitizing aerial imagery. Most of the semi-automated palm counting method that has been developed was limited to mature palms due to their ideal canopy size represented by satellite image. Therefore, immature palms were often left out since the size of the canopy is barely visible from satellite images. In this paper, an approach using a morphological filter and high-resolution satellite image is proposed to detect immature palm trees. This approach makes it possible to count the number of immature oil palm trees. The method begins with an erosion filter with an appropriate window size of 3m onto the high-resolution satellite image. The eroded image was further segmented using watershed segmentation to delineate immature palm tree regions. Then, local minimum detection was used because it is hypothesized that immature oil palm trees are located at the local minimum within an oil palm field setting in a grayscale image. The detection points generated from the local minimum are displaced to the center of the immature oil palm region and thinned. Only one detection point is left that represents a tree. The performance of the proposed method was evaluated on three subsets with slopes ranging from 0 to 20° and different planting designs, i.e., straight and terrace. The proposed method was able to achieve up to more than 90% accuracy when compared with the ground truth, with an overall F-measure score of up to 0.91.Keywords: immature palm count, oil palm, precision agriculture, remote sensing
Procedia PDF Downloads 761146 Leadership in the Era of AI: Growing Organizational Intelligence
Authors: Mark Salisbury
Abstract:
The arrival of artificially intelligent avatars and the automation they bring is worrying many of us, not only for our livelihood but for the jobs that may be lost to our kids. We worry about what our place will be as human beings in this new economy where much of it will be conducted online in the metaverse – in a network of 3D virtual worlds – working with intelligent machines. The Future of Leadership was written to address these fears and show what our place will be – the right place – in this new economy of AI avatars, automation, and 3D virtual worlds. But to be successful in this new economy, our job will be to bring wisdom to our workplace and the marketplace. And we will use AI avatars and 3D virtual worlds to do it. However, this book is about more than AI and the avatars that we will work with in the metaverse. It’s about building Organizational intelligence (OI) -- the capability of an organization to comprehend and create knowledge relevant to its purpose; in other words, it is the intellectual capacity of the entire organization. To increase organizational intelligence requires a new kind of knowledge worker, a wisdom worker, that requires a new kind of leadership. This book begins your story for how to become a leader of wisdom workers and be successful in the emerging wisdom economy. After this presentation, conference participants will be able to do the following: Recognize the characteristics of the new generation of wisdom workers and how they differ from their predecessors. Recognize that new leadership methods and techniques are needed to lead this new generation of wisdom workers. Apply personal and professional values – personal integrity, belief in something larger than yourself, and keeping the best interest of others in mind – to improve your work performance and lead others. Exhibit an attitude of confidence, courage, and reciprocity of sharing knowledge to increase your productivity and influence others. Leverage artificial intelligence to accelerate your ability to learn, augment your decision-making, and influence others.Utilize new technologies to communicate with human colleagues and intelligent machines to develop better solutions more quickly.Keywords: metaverse, generative artificial intelligence, automation, leadership, organizational intelligence, wisdom worker
Procedia PDF Downloads 441145 Narrative Study to Resilience and Adversity's Response
Authors: Yun Hang Stanley Cheung
Abstract:
In recent years, many educators and entrepreneurs have often suggested that students’ and workers’ ability of the adversity response is very important, it would affect problem-solving strategies and ultimate success in their career or life. The meaning of resilience is discussed as the process of bouncing back and the ability to adapt well in adversity’s response, being resilient does not mean to live without any stress and difficulty, but to grow and thrive under pressure. The purpose of this study is to describe the process of resilience and adversity’s response. The use of the narrative inquiry aims for understanding the experiential process of adversity response, and the problem-solving strategies (such as emotion control, motivation, decisions making process), as well as making the experience become life story, which may be evaluated by its teller and its listeners. The narrative study describes the researcher’s self-experience of adversity’s response to the recovery of the seriously burnt injury from a hill fire at his 12 years old, as well as the adversities and obstacles related to the tragedy after the physical recovery. Sense-Making Theory and McCormack’s Lenses were used for constructive perspective and data analyzing. To conclude, this study has described the life story of fighting the adversities, also, those narratives come out some suggestions, which point out positive thinking is necessary to build up resilience and the ability of immediate adversity response. Also, some problem-solving strategies toward adversities are discussed, which are helpful for resilience education for youth and young adult.Keywords: adversity response, life story, narrative inquiry, resilience
Procedia PDF Downloads 3121144 Three-Dimensional Carbon Foams for the Application as Electrode Material in Energy Storage Systems
Authors: H. Beisch, J. Marx, S. Garlof, R. Shvets, I. I. Grygorchak, A. Kityk, B. Fiedler
Abstract:
Carbon materials, especially three-dimensional carbon foams, show very high potential in the application as electrode material for energy storage systems such as batteries and supercapacitors with unique fast charging and discharging times. Regarding their high specific surface areas (SSA) high specific capacities can be reached. Globugraphite is a newly developed carbon foam with an interconnected globular carbon morphology. Especially, this foam has a statistically distributed hierarchical pore structure resulting from the manufacturing process based on sintered ceramic templates which are synthetized during a final chemical vapor deposition (CVD) process. For morphology characterization scanning electron (SEM) and transmission electron microscopy (TEM) is used. In addition, the SSA is carried out by nitrogen adsorption combined with the Brunauer–Emmett–Teller (BET) theory. Electrochemical measurements in organic and inorganic electrolyte provide high energy densities and power densities resulting from ion absorption by forming an electrochemical double layer. All values are summarized in a Ragone Diagram. Finally, power densities up to 833 W/kg and energy densities up to 48 Wh/kg could be achieved. The corresponding SSA is between 376 m²/g and 859 m²/g. For organic electrolyte a specific capacity of 71 F/g at a density of 20 mg/cm³ was achieved.Keywords: BET, CVD process, electron microscopy, Ragone diagram
Procedia PDF Downloads 1741143 A Fermatean Fuzzy MAIRCA Approach for Maintenance Strategy Selection of Process Plant Gearbox Using Sustainability Criteria
Authors: Soumava Boral, Sanjay K. Chaturvedi, Ian Howard, Kristoffer McKee, V. N. A. Naikan
Abstract:
Due to strict regulations from government to enhance the possibilities of sustainability practices in industries, and noting the advances in sustainable manufacturing practices, it is necessary that the associated processes are also sustainable. Maintenance of large scale and complex machines is a pivotal task to maintain the uninterrupted flow of manufacturing processes. Appropriate maintenance practices can prolong the lifetime of machines, and prevent associated breakdowns, which subsequently reduces different cost heads. Selection of the best maintenance strategies for such machines are considered as a burdensome task, as they require the consideration of multiple technical criteria, complex mathematical calculations, previous fault data, maintenance records, etc. In the era of the fourth industrial revolution, organizations are rapidly changing their way of business, and they are giving their utmost importance to sensor technologies, artificial intelligence, data analytics, automations, etc. In this work, the effectiveness of several maintenance strategies (e.g., preventive, failure-based, reliability centered, condition based, total productive maintenance, etc.) related to a large scale and complex gearbox, operating in a steel processing plant is evaluated in terms of economic, social, environmental and technical criteria. As it is not possible to obtain/describe some criteria by exact numerical values, these criteria are evaluated linguistically by cross-functional experts. Fuzzy sets are potential soft-computing technique, which has been useful to deal with linguistic data and to provide inferences in many complex situations. To prioritize different maintenance practices based on the identified sustainable criteria, multi-criteria decision making (MCDM) approaches can be considered as potential tools. Multi-Attributive Ideal Real Comparative Analysis (MAIRCA) is a recent addition in the MCDM family and has proven its superiority over some well-known MCDM approaches, like TOPSIS (Technique for Order Preference by Similarity to Ideal Solution) and ELECTRE (ELimination Et Choix Traduisant la REalité). It has a simple but robust mathematical approach, which is easy to comprehend. On the other side, due to some inherent drawbacks of Intuitionistic Fuzzy Sets (IFS) and Pythagorean Fuzzy Sets (PFS), recently, the use of Fermatean Fuzzy Sets (FFSs) has been proposed. In this work, we propose the novel concept of FF-MAIRCA. We obtain the weights of the criteria by experts’ evaluation and use them to prioritize the different maintenance practices according to their suitability by FF-MAIRCA approach. Finally, a sensitivity analysis is carried out to highlight the robustness of the approach.Keywords: Fermatean fuzzy sets, Fermatean fuzzy MAIRCA, maintenance strategy selection, sustainable manufacturing, MCDM
Procedia PDF Downloads 1381142 Produce High-Quality Activated Carbon with a Large Surface Area from Date Seeds Biomass for Water Treatment
Authors: Rashad Al-Gaashani, Viktor Kochkodan, Jenny Lawler
Abstract:
Physico-chemical activation method wasused to produce high-quality activated carbon (AC) with a large surface area of about 2000 m2/g from low-cost and abundant biomasswastes in Qatar, namely date seeds. X-Ray diffraction (XRD), scanning electron spectroscopy (SEM), energy dispersive X-Ray spectroscopy (EDS), and Brunauer-Emmett-Teller (BET) surface area analysis were used to evaluate the AC samples. AC produced from date seeds have a wide range of the pores available, including micro- andnano-pores. This type of AC with a well-developed pore structure may be very attractive for different applications, including air and water purification from micro and nano pollutants. Heavy metalsiron (III) and copper (II) ions were removed from wastewater using the AC producedusinga batch adsorption technique. The AC produced from date seeds biomass wastes show high removal of heavy metals such as iron (III) ions (100%) and copper (II) ions (97.25%). The highest removal of copper (II) ions (100%) with AC produced from date seeds was found at pH 8, whereas the lowest removal (22.63%) occurred at pH 2. The effect of adsorption time, adsorbent dose, pH on the removal of heavy metalswere studied.Keywords: activated carbon, date seeds, biomass, heavy metals removal, water treatment
Procedia PDF Downloads 1021141 Human-Machine Cooperation in Facial Comparison Based on Likelihood Scores
Authors: Lanchi Xie, Zhihui Li, Zhigang Li, Guiqiang Wang, Lei Xu, Yuwen Yan
Abstract:
Image-based facial features can be classified into category recognition features and individual recognition features. Current automated face recognition systems extract a specific feature vector of different dimensions from a facial image according to their pre-trained neural network. However, to improve the efficiency of parameter calculation, an algorithm generally reduces the image details by pooling. The operation will overlook the details concerned much by forensic experts. In our experiment, we adopted a variety of face recognition algorithms based on deep learning, compared a large number of naturally collected face images with the known data of the same person's frontal ID photos. Downscaling and manual handling were performed on the testing images. The results supported that the facial recognition algorithms based on deep learning detected structural and morphological information and rarely focused on specific markers such as stains and moles. Overall performance, distribution of genuine scores and impostor scores, and likelihood ratios were tested to evaluate the accuracy of biometric systems and forensic experts. Experiments showed that the biometric systems were skilled in distinguishing category features, and forensic experts were better at discovering the individual features of human faces. In the proposed approach, a fusion was performed at the score level. At the specified false accept rate, the framework achieved a lower false reject rate. This paper contributes to improving the interpretability of the objective method of facial comparison and provides a novel method for human-machine collaboration in this field.Keywords: likelihood ratio, automated facial recognition, facial comparison, biometrics
Procedia PDF Downloads 1301140 3D Carbon Structures (Globugraphite) with Hierarchical Pore Morphology for the Application in Energy Storage Systems
Authors: Hubert Beisch, Janik Marx, Svenja Garlof, Roman Shvets, Ivan Grygorchak, Andriy Kityk, Bodo Fiedler
Abstract:
Three-dimensional carbon materials can be used as electrode materials for energy storage systems such as batteries and supercapacitors. Fast charging and discharging times are realizable without reducing the performance due to aging processes. Furthermore high specific surface area (SSA) of three-dimensional carbon structures leads to high specific capacities. One newly developed carbon foam is Globugraphite. This interconnected globular carbon morphology with statistically distributed hierarchical pores is manufactured by a chemical vapor deposition (CVD) process from ceramic templates resulting from a sintering process. Via scanning electron (SEM) and transmission electron microscopy (TEM), the morphology is characterized. Moreover, the SSA was measured by the Brunauer–Emmett–Teller (BET) theory. Measurements of Globugraphite in an organic and inorganic electrolyte show high energy densities and power densities resulting from ion absorption by forming an electrochemical double layer. A comparison of the specific values is summarized in a Ragone diagram. Energy densities up to 48 Wh/kg and power densities to 833 W/kg could be achieved for an SSA from 376 m²/g to 859 m²/g. For organic electrolyte, a specific capacity of 100 F/g at a density of 20 mg/cm³ was achieved.Keywords: BET, carbon foam, CVD process, electrochemical cell, Ragone diagram, SEM, TEM
Procedia PDF Downloads 2341139 Evaluation of Automated Analyzers of Polycyclic Aromatic Hydrocarbons and Black Carbon in a Coke Oven Plant by Comparison with Analytical Methods
Authors: L. Angiuli, L. Trizio, R. Giua, A. Digilio, M. Tutino, P. Dambruoso, F. Mazzone, C. M. Placentino
Abstract:
In the winter of 2014 a series of measurements were performed to evaluate the behavior of real-time PAHs and black carbon analyzers in a coke oven plant located in Taranto, a city of Southern Italy. Data were collected both insides than outside the plant, at air quality monitoring sites. Contemporary measures of PM2.5 and PM1 were performed. Particle-bound PAHs were measured by two methods: (1) aerosol photoionization using an Ecochem PAS 2000 analyzer, (2) PM2.5 and PM1 quartz filter collection and analysis by gas chromatography/mass spectrometry (GC/MS). Black carbon was determined both in real-time by Magee Aethalometer AE22 analyzer than by semi-continuous Sunset Lab EC/OC instrument. Detected PM2.5 and PM1 levels were higher inside than outside the plant while PAHs real-time values were higher outside than inside. As regards PAHs, inside the plant Ecochem PAS 2000 revealed concentrations not significantly different from those determined on the filter during low polluted days, but at increasing concentrations the automated instrument underestimated PAHs levels. At the external site, Ecochem PAS 2000 real-time concentrations were steadily higher than those on the filter. In the same way, real-time black carbon values were constantly lower than EC concentrations obtained by Sunset EC/OC in the inner site, while outside the plant real-time values were comparable to Sunset EC values. Results showed that in a coke plant real-time analyzers of PAHs and black carbon in the factory configuration provide qualitative information, with no accuracy and leading to the underestimation of the concentration. A site specific calibration is needed for these instruments before their installation in high polluted sites.Keywords: black carbon, coke oven plant, PAH, PAS, aethalometer
Procedia PDF Downloads 3441138 Phenomena-Based Approach for Automated Generation of Process Options and Process Models
Authors: Parminder Kaur Heer, Alexei Lapkin
Abstract:
Due to global challenges of increased competition and demand for more sustainable products/processes, there is a rising pressure on the industry to develop innovative processes. Through Process Intensification (PI) the existing and new processes may be able to attain higher efficiency. However, very few PI options are generally considered. This is because processes are typically analysed at a unit operation level, thus limiting the search space for potential process options. PI performed at more detailed levels of a process can increase the size of the search space. The different levels at which PI can be achieved is unit operations, functional and phenomena level. Physical/chemical phenomena form the lowest level of aggregation and thus, are expected to give the highest impact because all the intensification options can be described by their enhancement. The objective of the current work is thus, generation of numerous process alternatives based on phenomena, and development of their corresponding computer aided models. The methodology comprises: a) automated generation of process options, and b) automated generation of process models. The process under investigation is disintegrated into functions viz. reaction, separation etc., and these functions are further broken down into the phenomena required to perform them. E.g., separation may be performed via vapour-liquid or liquid-liquid equilibrium. A list of phenomena for the process is formed and new phenomena, which can overcome the difficulties/drawbacks of the current process or can enhance the effectiveness of the process, are added to the list. For instance, catalyst separation issue can be handled by using solid catalysts; the corresponding phenomena are identified and added. The phenomena are then combined to generate all possible combinations. However, not all combinations make sense and, hence, screening is carried out to discard the combinations that are meaningless. For example, phase change phenomena need the co-presence of the energy transfer phenomena. Feasible combinations of phenomena are then assigned to the functions they execute. A combination may accomplish a single or multiple functions, i.e. it might perform reaction or reaction with separation. The combinations are then allotted to the functions needed for the process. This creates a series of options for carrying out each function. Combination of these options for different functions in the process leads to the generation of superstructure of process options. These process options, which are formed by a list of phenomena for each function, are passed to the model generation algorithm in the form of binaries (1, 0). The algorithm gathers the active phenomena and couples them to generate the model. A series of models is generated for the functions, which are combined to get the process model. The most promising process options are then chosen subjected to a performance criterion, for example purity of product, or via a multi-objective Pareto optimisation. The methodology was applied to a two-step process and the best route was determined based on the higher product yield. The current methodology can identify, produce and evaluate process intensification options from which the optimal process can be determined. It can be applied to any chemical/biochemical process because of its generic nature.Keywords: Phenomena, Process intensification, Process models , Process options
Procedia PDF Downloads 2321137 Optimized Electron Diffraction Detection and Data Acquisition in Diffraction Tomography: A Complete Solution by Gatan
Authors: Saleh Gorji, Sahil Gulati, Ana Pakzad
Abstract:
Continuous electron diffraction tomography, also known as microcrystal electron diffraction (MicroED) or three-dimensional electron diffraction (3DED), is a powerful technique, which in combination with cryo-electron microscopy (cryo-ED), can provide atomic-scale 3D information about the crystal structure and composition of different classes of crystalline materials such as proteins, peptides, and small molecules. Unlike the well-established X-ray crystallography method, 3DED does not require large single crystals and can collect accurate electron diffraction data from crystals as small as 50 – 100 nm. This is a critical advantage as growing larger crystals, as required by X-ray crystallography methods, is often very difficult, time-consuming, and expensive. In most cases, specimens studied via 3DED method are electron beam sensitive, which means there is a limitation on the maximum amount of electron dose one can use to collect the required data for a high-resolution structure determination. Therefore, collecting data using a conventional scintillator-based fiber coupled camera brings additional challenges. This is because of the inherent noise introduced during the electron-to-photon conversion in the scintillator and transfer of light via the fibers to the sensor, which results in a poor signal-to-noise ratio and requires a relatively higher and commonly specimen-damaging electron dose rates, especially for protein crystals. As in other cryo-EM techniques, damage to the specimen can be mitigated if a direct detection camera is used which provides a high signal-to-noise ratio at low electron doses. In this work, we have used two classes of such detectors from Gatan, namely the K3® camera (a monolithic active pixel sensor) and Stela™ (that utilizes DECTRIS hybrid-pixel technology), to address this problem. The K3 is an electron counting detector optimized for low-dose applications (like structural biology cryo-EM), and Stela is also a counting electron detector but optimized for diffraction applications with high speed and high dynamic range. Lastly, data collection workflows, including crystal screening, microscope optics setup (for imaging and diffraction), stage height adjustment at each crystal position, and tomogram acquisition, can be one of the other challenges of the 3DED technique. Traditionally this has been all done manually or in a partly automated fashion using open-source software and scripting, requiring long hours on the microscope (extra cost) and extensive user interaction with the system. We have recently introduced Latitude® D in DigitalMicrograph® software, which is compatible with all pre- and post-energy-filter Gatan cameras and enables 3DED data acquisition in an automated and optimized fashion. Higher quality 3DED data enables structure determination with higher confidence, while automated workflows allow these to be completed considerably faster than before. Using multiple examples, this work will demonstrate how to direct detection electron counting cameras enhance 3DED results (3 to better than 1 Angstrom) for protein and small molecule structure determination. We will also show how Latitude D software facilitates collecting such data in an integrated and fully automated user interface.Keywords: continuous electron diffraction tomography, direct detection, diffraction, Latitude D, Digitalmicrograph, proteins, small molecules
Procedia PDF Downloads 1071136 Eosinopenia: Marker for Early Diagnosis of Enteric Fever
Authors: Swati Kapoor, Rajeev Upreti, Monica Mahajan, Abhaya Indrayan, Dinesh Srivastava
Abstract:
Enteric Fever is caused by gram negative bacilli Salmonella typhi and paratyphi. It is associated with high morbidity and mortality worldwide. Timely initiation of treatment is a crucial step for prevention of any complications. Cultures of body fluids are diagnostic, but not always conclusive or practically feasible in most centers. Moreover, the results of cultures delay the treatment initiation. Serological tests lack diagnostic value. The blood counts can offer a promising option in diagnosis. A retrospective study to find out the relevance of leucopenia and eosinopenia was conducted on 203 culture proven enteric fever patients and 159 culture proven non-enteric fever patients in a tertiary care hospital in New Delhi. The patient details were retrieved from the electronic medical records section of the hospital. Absolute eosinopenia was considered as absolute eosinophil count (AEC) of less than 40/mm³ (normal level: 40-400/mm³) using LH-750 Beckman Coulter Automated machine. Leucopoenia was defined as total leucocyte count (TLC) of less than 4 X 10⁹/l. Blood cultures were done using BacT/ALERT FA plus automated blood culture system before first antibiotic dose was given. Case and control groups were compared using Pearson Chi square test. It was observed that absolute eosinophil count (AEC) of 0-19/mm³ was a significant finding (p < 0.001) in enteric fever patients, whereas leucopenia was not a significant finding (p=0.096). Using Receiving Operating Characteristic (ROC) curves, it was observed that patients with both AEC < 14/mm³ and TCL < 8 x 10⁹/l had 95.6% chance of being diagnosed as enteric fever and only 4.4% chance of being diagnosed as non-enteric fever. This result was highly significant with p < 0.001. This is a very useful association of AEC and TLC found in enteric fever patients of this study which can be used for the early initiation of treatment in clinically suspected enteric fever patients.Keywords: absolute eosinopenia, absolute eosinophil count, enteric fever, leucopenia, total leucocyte count
Procedia PDF Downloads 1721135 Transformer-Driven Multi-Category Classification for an Automated Academic Strand Recommendation Framework
Authors: Ma Cecilia Siva
Abstract:
This study introduces a Bidirectional Encoder Representations from Transformers (BERT)-based machine learning model aimed at improving educational counseling by automating the process of recommending academic strands for students. The framework is designed to streamline and enhance the strand selection process by analyzing students' profiles and suggesting suitable academic paths based on their interests, strengths, and goals. Data was gathered from a sample of 200 grade 10 students, which included personal essays and survey responses relevant to strand alignment. After thorough preprocessing, the text data was tokenized, label-encoded, and input into a fine-tuned BERT model set up for multi-label classification. The model was optimized for balanced accuracy and computational efficiency, featuring a multi-category classification layer with sigmoid activation for independent strand predictions. Performance metrics showed an F1 score of 88%, indicating a well-balanced model with precision at 80% and recall at 100%, demonstrating its effectiveness in providing reliable recommendations while reducing irrelevant strand suggestions. To facilitate practical use, the final deployment phase created a recommendation framework that processes new student data through the trained model and generates personalized academic strand suggestions. This automated recommendation system presents a scalable solution for academic guidance, potentially enhancing student satisfaction and alignment with educational objectives. The study's findings indicate that expanding the data set, integrating additional features, and refining the model iteratively could improve the framework's accuracy and broaden its applicability in various educational contexts.Keywords: tokenized, sigmoid activation, transformer, multi category classification
Procedia PDF Downloads 91134 An Automated Magnetic Dispersive Solid-Phase Extraction Method for Detection of Cocaine in Human Urine
Authors: Feiyu Yang, Chunfang Ni, Rong Wang, Yun Zou, Wenbin Liu, Chenggong Zhang, Fenjin Sun, Chun Wang
Abstract:
Cocaine is the most frequently used illegal drug globally, with the global annual prevalence of cocaine used ranging from 0.3% to 0.4 % of the adult population aged 15–64 years. Growing consumption trend of abused cocaine and drug crimes are a great concern, therefore urine sample testing has become an important noninvasive sampling whereas cocaine and its metabolites (COCs) are usually present in high concentrations and relatively long detection windows. However, direct analysis of urine samples is not feasible because urine complex medium often causes low sensitivity and selectivity of the determination. On the other hand, presence of low doses of analytes in urine makes an extraction and pretreatment step important before determination. Especially, in gathered taking drug cases, the pretreatment step becomes more tedious and time-consuming. So developing a sensitive, rapid and high-throughput method for detection of COCs in human body is indispensable for law enforcement officers, treatment specialists and health officials. In this work, a new automated magnetic dispersive solid-phase extraction (MDSPE) sampling method followed by high performance liquid chromatography-mass spectrometry (HPLC-MS) was developed for quantitative enrichment of COCs from human urine, using prepared magnetic nanoparticles as absorbants. The nanoparticles were prepared by silanizing magnetic Fe3O4 nanoparticles and modifying them with divinyl benzene and vinyl pyrrolidone, which possesses the ability for specific adsorption of COCs. And this kind of magnetic particle facilitated the pretreatment steps by electromagnetically controlled extraction to achieve full automation. The proposed device significantly improved the sampling preparation efficiency with 32 samples in one batch within 40mins. Optimization of the preparation procedure for the magnetic nanoparticles was explored and the performances of magnetic nanoparticles were characterized by scanning electron microscopy, vibrating sample magnetometer and infrared spectra measurements. Several analytical experimental parameters were studied, including amount of particles, adsorption time, elution solvent, extraction and desorption kinetics, and the verification of the proposed method was accomplished. The limits of detection for the cocaine and cocaine metabolites were 0.09-1.1 ng·mL-1 with recoveries ranging from 75.1 to 105.7%. Compared to traditional sampling method, this method is time-saving and environmentally friendly. It was confirmed that the proposed automated method was a kind of highly effective way for the trace cocaine and cocaine metabolites analyses in human urine.Keywords: automatic magnetic dispersive solid-phase extraction, cocaine detection, magnetic nanoparticles, urine sample testing
Procedia PDF Downloads 2041133 Revolutionizing Autonomous Trucking Logistics with Customer Relationship Management Cloud
Authors: Sharda Kumari, Saiman Shetty
Abstract:
Autonomous trucking is just one of the numerous significant shifts impacting fleet management services. The Society of Automotive Engineers (SAE) has defined six levels of vehicle automation that have been adopted internationally, including by the United States Department of Transportation. On public highways in the United States, organizations are testing driverless vehicles with at least Level 4 automation which indicates that a human is present in the vehicle and can disable automation, which is usually done while the trucks are not engaged in highway driving. However, completely driverless vehicles are presently being tested in the state of California. While autonomous trucking can increase safety, decrease trucking costs, provide solutions to trucker shortages, and improve efficiencies, logistics, too, requires advancements to keep up with trucking innovations. Given that artificial intelligence, machine learning, and automated procedures enable people to do their duties in other sectors with fewer resources, CRM (Customer Relationship Management) can be applied to the autonomous trucking business to provide the same level of efficiency. In a society witnessing significant digital disruptions, fleet management is likewise being transformed by technology. Utilizing strategic alliances to enhance core services is an effective technique for capitalizing on innovations and delivering enhanced services. Utilizing analytics on CRM systems improves cost control of fuel strategy, fleet maintenance, driver behavior, route planning, road safety compliance, and capacity utilization. Integration of autonomous trucks with automated fleet management, yard/terminal management, and customer service is possible, thus having significant power to redraw the lines between the public and private spheres in autonomous trucking logistics.Keywords: autonomous vehicles, customer relationship management, customer experience, autonomous trucking, digital transformation
Procedia PDF Downloads 108