Search results for: key frame extraction
2328 Protocol for Dynamic Load Distributed Low Latency Web-Based Augmented Reality and Virtual Reality
Authors: Rohit T. P., Sahil Athrij, Sasi Gopalan
Abstract:
Currently, the content entertainment industry is dominated by mobile devices. As the trends slowly shift towards Augmented/Virtual Reality applications the computational demands on these devices are increasing exponentially and we are already reaching the limits of hardware optimizations. This paper proposes a software solution to this problem. By leveraging the capabilities of cloud computing we can offload the work from mobile devices to dedicated rendering servers that are way more powerful. But this introduces the problem of latency. This paper introduces a protocol that can achieve high-performance low latency Augmented/Virtual Reality experience. There are two parts to the protocol, 1) In-flight compression The main cause of latency in the system is the time required to transmit the camera frame from client to server. The round trip time is directly proportional to the amount of data transmitted. This can therefore be reduced by compressing the frames before sending. Using some standard compression algorithms like JPEG can result in minor size reduction only. Since the images to be compressed are consecutive camera frames there won't be a lot of changes between two consecutive images. So inter-frame compression is preferred. Inter-frame compression can be implemented efficiently using WebGL but the implementation of WebGL limits the precision of floating point numbers to 16bit in most devices. This can introduce noise to the image due to rounding errors, which will add up eventually. This can be solved using an improved interframe compression algorithm. The algorithm detects changes between frames and reuses unchanged pixels from the previous frame. This eliminates the need for floating point subtraction thereby cutting down on noise. The change detection is also improved drastically by taking the weighted average difference of pixels instead of the absolute difference. The kernel weights for this comparison can be fine-tuned to match the type of image to be compressed. 2) Dynamic Load distribution Conventional cloud computing architectures work by offloading as much work as possible to the servers, but this approach can cause a hit on bandwidth and server costs. The most optimal solution is obtained when the device utilizes 100% of its resources and the rest is done by the server. The protocol balances the load between the server and the client by doing a fraction of the computing on the device depending on the power of the device and network conditions. The protocol will be responsible for dynamically partitioning the tasks. Special flags will be used to communicate the workload fraction between the client and the server and will be updated in a constant interval of time ( or frames ). The whole of the protocol is designed so that it can be client agnostic. Flags are available to the client for resetting the frame, indicating latency, switching mode, etc. The server can react to client-side changes on the fly and adapt accordingly by switching to different pipelines. The server is designed to effectively spread the load and thereby scale horizontally. This is achieved by isolating client connections into different processes.Keywords: 2D kernelling, augmented reality, cloud computing, dynamic load distribution, immersive experience, mobile computing, motion tracking, protocols, real-time systems, web-based augmented reality application
Procedia PDF Downloads 742327 Face Shield Design with Additive Manufacturing Practice Combating COVID-19 Pandemic
Authors: May M. Youssef
Abstract:
This article introduces a design, for additive manufacturing technology, face shield as Personal Protective Equipment from the respiratory viruses such as coronavirus 2. The face shields help to reduce ocular exposure and play a vital role in diverting away from the respiratory COVID-19 air droplets around the users' face. The proposed face shield comprises three assembled polymer parts. The frame with a transparency overhead projector sheet visor is suitable for frontline health care workers and ordinary citizens. The frame design allows tightening the shield around the user’s head and permits rubber elastic straps to be used if required. That ergonomically designed with a unique face mask support used in case of wearing extra protective mask was created using computer aided design (CAD) software package. The finite element analysis (FEA) structural verification of the proposed design is performed by an advanced simulation technique. Subsequently, the prototype model was fabricated by a 3D printing using Fused Deposition Modeling (FDM) as a globally developed face shield product. This study provides a different face shield designs for global production, which showed to be suitable and effective toward supply chain shortages and frequent needs of personal protective goods during coronavirus disease and similar viruses.Keywords: additive manufacturing, Coronavirus-19, face shield, personal protective equipment, 3D printing
Procedia PDF Downloads 2012326 New Method for the Determination of Montelukast in Human Plasma by Solid Phase Extraction Using Liquid Chromatography Tandem Mass Spectrometry
Authors: Vijayalakshmi Marella, NageswaraRaoPilli
Abstract:
This paper describes a simple, rapid and sensitive liquid chromatography / tandem mass spectrometry assay for the determination of montelukast in human plasma using montelukast d6 as an internal standard. Analyte and the internal standard were extracted from 50 µL of human plasma via solid phase extraction technique without evaporation, drying and reconstitution steps. The chromatographic separation was achieved on a C18 column by using a mixture of methanol and 5mM ammonium acetate (80:20, v/v) as the mobile phase at a flow rate of 0.8 mL/min. Good linearity results were obtained during the entire course of validation. Method validation was performed as per FDA guidelines and the results met the acceptance criteria. A run time of 2.5 min for each sample made it possible to analyze more number of samples in short time, thus increasing the productivity. The proposed method was found to be applicable to clinical studies.Keywords: Montelukast, tandem mass spectrometry, montelukast d6, FDA guidelines
Procedia PDF Downloads 3152325 Formulation of Suppositories Using Allanblackia Floribunda Butter as a Base
Authors: Mary Konadu
Abstract:
The rectal route for drug administration is becoming attractive to drug formulators because it can avoid hepatic first-pass effects, decrease gastrointestinal side effects and avoid undesirable effects of meals on drug absorption. Suppositories have been recognized as an alternative to the oral route in situations such as when the patient is comatose, unable to swallow, or when the drug produces nausea or vomiting. Effective drug delivery with appropriate pharmaceutical excipient is key in the production of clinically useful preparations. The high cost of available excipients coupled with other disadvantages have led to the exploration of potential excipients from natural sources. Allanblackia floribunda butter, a naturally occurring lipid, is used for medicinal, culinary, and cosmetic purposes. Different extraction methods (solvent (hexane) extraction, traditional/hot water extraction, and cold/screw press extraction) were employed to extract the oil. The different extracts of A. floribunda oil were analyzed for their physicochemical properties and mineral content. The oil was used as a base to formulate Paracetamol and Diclofenac suppositories. Quality control test were carried out on the formulated suppositories. The %age oil yield for hexane extract, hot water extract, and cold press extract were 50.40 ±0.00, 37.36±0.00, and 20.48±0.00, respectively. The acid value, saponification value, iodine value and free fatty acid were 1.159 ± 0.065, 208.51 ± 8.450, 49.877 ± 0.690 and 0.583 ± 0.032 respectively for hexane extract; 3.480 ± 0.055, 204.672±2.863, 49.04 ± 0.76 and 1.747 ± 0.028 respectively for hot water/traditional extract; 4.43 ± 0.055, 192.05±1.56, 49.96 ± 0.29 and 2.23 ± 0.03 respectively for cold press extract. Calcium, sodium, magnesium, potassium, and iron were minerals found to be present in the A. floribunda butter extracts. The uniformity of weight, hardness, disintegration time, and uniformity of content were found to be within the acceptable range. The melting point ranges for all the suppositories were found to be satisfactory. The cumulative drug release (%) of the suppositories at 45 minutes was 90.19±0.00 (Hot water extract), 93.75±0.00 (Cold Pres Extract), and 98.16±0.00 (Hexane Extract) for Paracetamol suppositories. Diclofenac sodium suppositories had a cumulative %age release of 81.60±0.00 (Hot water Extract), 95.33±0.00 (Cold Press Extract), and 99.20±0.00 (Hexane Extract). The physicochemical parameters obtained from this study shows that Allanblackia floribunda seed oil is edible and can be used as a suppository base. The suppository formulation was successful, and the quality control tests conformed to Pharmacopoeia standard.Keywords: allanblackia foribunda, paracetamol, diclofenac, suppositories
Procedia PDF Downloads 1222324 A Quality Index Optimization Method for Non-Invasive Fetal ECG Extraction
Authors: Lucia Billeci, Gennaro Tartarisco, Maurizio Varanini
Abstract:
Fetal cardiac monitoring by fetal electrocardiogram (fECG) can provide significant clinical information about the healthy condition of the fetus. Despite this potentiality till now the use of fECG in clinical practice has been quite limited due to the difficulties in its measuring. The recovery of fECG from the signals acquired non-invasively by using electrodes placed on the maternal abdomen is a challenging task because abdominal signals are a mixture of several components and the fetal one is very weak. This paper presents an approach for fECG extraction from abdominal maternal recordings, which exploits the characteristics of pseudo-periodicity of fetal ECG. It consists of devising a quality index (fQI) for fECG and of finding the linear combinations of preprocessed abdominal signals, which maximize these fQI (quality index optimization - QIO). It aims at improving the performances of the most commonly adopted methods for fECG extraction, usually based on maternal ECG (mECG) estimating and canceling. The procedure for the fECG extraction and fetal QRS (fQRS) detection is completely unsupervised and based on the following steps: signal pre-processing; maternal ECG (mECG) extraction and maternal QRS detection; mECG component approximation and canceling by weighted principal component analysis; fECG extraction by fQI maximization and fetal QRS detection. The proposed method was compared with our previously developed procedure, which obtained the highest at the Physionet/Computing in Cardiology Challenge 2013. That procedure was based on removing the mECG from abdominal signals estimated by a principal component analysis (PCA) and applying the Independent component Analysis (ICA) on the residual signals. Both methods were developed and tuned using 69, 1 min long, abdominal measurements with fetal QRS annotation of the dataset A provided by PhysioNet/Computing in Cardiology Challenge 2013. The QIO-based and the ICA-based methods were compared in analyzing two databases of abdominal maternal ECG available on the Physionet site. The first is the Abdominal and Direct Fetal Electrocardiogram Database (ADdb) which contains the fetal QRS annotations thus allowing a quantitative performance comparison, the second is the Non-Invasive Fetal Electrocardiogram Database (NIdb), which does not contain the fetal QRS annotations so that the comparison between the two methods can be only qualitative. In particular, the comparison on NIdb was performed defining an index of quality for the fetal RR series. On the annotated database ADdb the QIO method, provided the performance indexes Sens=0.9988, PPA=0.9991, F1=0.9989 overcoming the ICA-based one, which provided Sens=0.9966, PPA=0.9972, F1=0.9969. The comparison on NIdb was performed defining an index of quality for the fetal RR series. The index of quality resulted higher for the QIO-based method compared to the ICA-based one in 35 records out 55 cases of the NIdb. The QIO-based method gave very high performances with both the databases. The results of this study foresees the application of the algorithm in a fully unsupervised way for the implementation in wearable devices for self-monitoring of fetal health.Keywords: fetal electrocardiography, fetal QRS detection, independent component analysis (ICA), optimization, wearable
Procedia PDF Downloads 2802323 Electrochemical Recovery of Lithium from Geothermal Brines
Authors: Sanaz Mosadeghsedghi, Mathew Hudder, Mohammad Ali Baghbanzadeh, Charbel Atallah, Seyedeh Laleh Dashtban Kenari, Konstantin Volchek
Abstract:
Lithium has recently been extensively used in lithium-ion batteries (LIBs) for electric vehicles and portable electronic devices. The conventional evaporative approach to recover and concentrate lithium is extremely slow and may take 10-24 months to concentrate lithium from dilute sources, such as geothermal brines. To response to the increasing industrial lithium demand, alternative extraction and concentration technologies should be developed to recover lithium from brines with low concentrations. In this study, a combination of electrocoagulation (EC) and electrodialysis (ED) was evaluated for the recovery of lithium from geothermal brines. The brine samples in this study, collected in Western Canada, had lithium concentrations of 50-75 mg/L on a background of much higher (over 10,000 times) concentrations of sodium. This very high sodium-to-lithium ratio poses challenges to the conventional direct-lithium extraction processes which employ lithium-selective adsorbents. EC was used to co-precipitate lithium using a sacrificial aluminium electrode. The precipitate was then dissolved, and the leachate was treated using ED to separate and concentrate lithium from other ions. The focus of this paper is on the study of ED, including a two-step ED process that included a mono-valent selective stage to separate lithium from multi-valent cations followed by a bipolar ED stage to convert lithium chloride (LiCl) to LiOH product. Eventually, the ED cell was reconfigured using mono-valent cation exchange with the bipolar membranes to combine the two ED steps in one. Using this process at optimum conditions, over 95% of the co-existing cations were removed and the purity of lithium increased to over 90% in the final product.Keywords: electrochemical separation, electrocoagulation, electrodialysis, lithium extraction
Procedia PDF Downloads 932322 Development of Deep Neural Network-Based Strain Values Prediction Models for Full-Scale Reinforced Concrete Frames Using Highly Flexible Sensing Sheets
Authors: Hui Zhang, Sherif Beskhyroun
Abstract:
Structural Health monitoring systems (SHM) are commonly used to identify and assess structural damage. In terms of damage detection, SHM needs to periodically collect data from sensors placed in the structure as damage-sensitive features. This includes abnormal changes caused by the strain field and abnormal symptoms of the structure, such as damage and deterioration. Currently, deploying sensors on a large scale in a building structure is a challenge. In this study, a highly stretchable strain sensors are used in this study to collect data sets of strain generated on the surface of full-size reinforced concrete (RC) frames under extreme cyclic load application. This sensing sheet can be switched freely between the test bending strain and the axial strain to achieve two different configurations. On this basis, the deep neural network prediction model of the frame beam and frame column is established. The training results show that the method can accurately predict the strain value and has good generalization ability. The two deep neural network prediction models will also be deployed in the SHM system in the future as part of the intelligent strain sensor system.Keywords: strain sensing sheets, deep neural networks, strain measurement, SHM system, RC frames
Procedia PDF Downloads 992321 Luminescent Properties of Plastic Scintillator with Large Area Photonic Crystal Prepared by a Combination of Nanoimprint Lithography and Atomic Layer Deposition
Authors: Jinlu Ruan, Liang Chen, Bo Liu, Xiaoping Ouyang, Zhichao Zhu, Zhongbing Zhang, Shiyi He, Mengxuan Xu
Abstract:
Plastic scintillators play an important role in the measurement of a mixed neutron/gamma pulsed radiation, neutron radiography and pulse shape discrimination technology. In some research, these luminescent properties are necessary that photons produced by the interactions between a plastic scintillator and radiations can be detected as much as possible by the photoelectric detectors and more photons can be emitted from the scintillators along a specific direction where detectors are located. Unfortunately, a majority of these photons produced are trapped in the plastic scintillators due to the total internal reflection (TIR), because there is a significant light-trapping effect when the incident angle of internal scintillation light is larger than the critical angle. Some of these photons trapped in the scintillator may be absorbed by the scintillator itself and the others are emitted from the edges of the scintillator. This makes the light extraction of plastic scintillators very low. Moreover, only a small portion of the photons emitted from the scintillator easily can be detected by detectors effectively, because the distribution of the emission directions of this portion of photons exhibits approximate Lambertian angular profile following a cosine emission law. Therefore, enhancing the light extraction efficiency and adjusting the emission angular profile become the keys for improving the number of photons detected by the detectors. In recent years, photonic crystal structures have been covered on inorganic scintillators to enhance the light extraction efficiency and adjust the angular profile of scintillation light successfully. However, that, preparation methods of photonic crystals will deteriorate performance of plastic scintillators and even destroy the plastic scintillators, makes the investigation on preparation methods of photonic crystals for plastic scintillators and luminescent properties of plastic scintillators with photonic crystal structures inadequate. Although we have successfully made photonic crystal structures covered on the surface of plastic scintillators by a modified self-assembly technique and achieved a great enhance of light extraction efficiency without evident angular-dependence for the angular profile of scintillation light, the preparation of photonic crystal structures with large area (the diameter is larger than 6cm) and perfect periodic structure is still difficult. In this paper, large area photonic crystals on the surface of scintillators were prepared by nanoimprint lithography firstly, and then a conformal layer with material of high refractive index on the surface of photonic crystal by atomic layer deposition technique in order to enhance the stability of photonic crystal structures and increase the number of leaky modes for improving the light extraction efficiency. The luminescent properties of the plastic scintillator with photonic crystals prepared by the mentioned method are compared with those of the plastic scintillator without photonic crystal. The results indicate that the number of photons detected by detectors is increased by the enhanced light extraction efficiency and the angular profile of scintillation light exhibits evident angular-dependence for the scintillator with photonic crystals. The mentioned preparation of photonic crystals is beneficial to scintillation detection applications and lays an important technique foundation for the plastic scintillators to meet special requirements under different application backgrounds.Keywords: angular profile, atomic layer deposition, light extraction efficiency, plastic scintillator, photonic crystal
Procedia PDF Downloads 2002320 Geometric Intuition and Formalism in Passing from Indivisibles to Infinitesimals: Pascal and Leibniz
Authors: Remus Titiriga
Abstract:
The paper focuses on Pascal's indivisibles evolving to Leibniz's infinitesimals. It starts with parallel developments by the two savants in Combinatorics (triangular numbers for Pascal and harmonic triangles for Leibniz) and their implication in determining the sum of mathematical series. It follows with a focus on the geometrical contributions of Pascal. He considered the cycloid and other mechanical curves the epitome of geometric comprehensibility in a series of challenging problems he posed to the mathematical world. Pascal provided the solutions in 1658, in a volume published under the pseudonym of Dettonville, using indivisibles and ratios between curved and straight lines. In the third part, the research follows the impact of this volume on Leibniz as the initial impetus for the elaboration of modern calculus as an algorithmic method disjoint of geometrical intuition. Then paper analyses the further steps and proves that Leibniz's developments relate to his philosophical frame (the search for a characteristic Universalis, the consideration of principle of continuity or the rule of sufficient reason) different from Pascal's and impacting mathematical problems and their solutions. At this stage in Leibniz's evolution, the infinitesimals replaced the indivisibles proper. The last part of the paper starts with speculation around "What if?". Could Pascal, if he lived more, accomplish the same feat? The document uses Pascal's reconstructed philosophical frame to formulate a positive answer. It also proposes to teach calculus with indivisibles and infinitesimals mimicking Pascal and Leibniz's achievements.Keywords: indivisibles, infinitesimals, characteristic triangle, the principle of continuity
Procedia PDF Downloads 1292319 Using the Smith-Waterman Algorithm to Extract Features in the Classification of Obesity Status
Authors: Rosa Figueroa, Christopher Flores
Abstract:
Text categorization is the problem of assigning a new document to a set of predetermined categories, on the basis of a training set of free-text data that contains documents whose category membership is known. To train a classification model, it is necessary to extract characteristics in the form of tokens that facilitate the learning and classification process. In text categorization, the feature extraction process involves the use of word sequences also known as N-grams. In general, it is expected that documents belonging to the same category share similar features. The Smith-Waterman (SW) algorithm is a dynamic programming algorithm that performs a local sequence alignment in order to determine similar regions between two strings or protein sequences. This work explores the use of SW algorithm as an alternative to feature extraction in text categorization. The dataset used for this purpose, contains 2,610 annotated documents with the classes Obese/Non-Obese. This dataset was represented in a matrix form using the Bag of Word approach. The score selected to represent the occurrence of the tokens in each document was the term frequency-inverse document frequency (TF-IDF). In order to extract features for classification, four experiments were conducted: the first experiment used SW to extract features, the second one used unigrams (single word), the third one used bigrams (two word sequence) and the last experiment used a combination of unigrams and bigrams to extract features for classification. To test the effectiveness of the extracted feature set for the four experiments, a Support Vector Machine (SVM) classifier was tuned using 20% of the dataset. The remaining 80% of the dataset together with 5-Fold Cross Validation were used to evaluate and compare the performance of the four experiments of feature extraction. Results from the tuning process suggest that SW performs better than the N-gram based feature extraction. These results were confirmed by using the remaining 80% of the dataset, where SW performed the best (accuracy = 97.10%, weighted average F-measure = 97.07%). The second best was obtained by the combination of unigrams-bigrams (accuracy = 96.04, weighted average F-measure = 95.97) closely followed by the bigrams (accuracy = 94.56%, weighted average F-measure = 94.46%) and finally unigrams (accuracy = 92.96%, weighted average F-measure = 92.90%).Keywords: comorbidities, machine learning, obesity, Smith-Waterman algorithm
Procedia PDF Downloads 2972318 Multimodal Analysis of News Magazines' Front-Page Portrayals of the US, Germany, China, and Russia
Authors: Alena Radina
Abstract:
On the global stage, national image is shaped by historical memory of wars and alliances, government ideology and particularly media stereotypes which represent countries in positive or negative ways. News magazine covers are a key site for national representation. The object of analysis in this paper is the portrayals of the US, Germany, China, and Russia in the front pages and cover stories of “Time”, “Der Spiegel”, “Beijing Review”, and “Expert”. Political comedy helps people learn about current affairs even if politics is not their area of interest, and thus satire indirectly sets the public agenda. Coupled with satirical messages, cover images and the linguistic messages embedded in the covers become persuasive visual and verbal factors, known to drive about 80% of magazine sales. Preliminary analysis identified satirical elements in magazine covers, which are known to influence and frame understandings and attract younger audiences. Multimodal and transnational comparative framing analyses lay the groundwork to investigate why journalists, editors and designers deploy certain frames rather than others. This research investigates to what degree frames used in covers correlate with frames within the cover stories and what these framings can tell us about media professionals’ representations of their own and other nations. The study sample includes 32 covers consisting of two covers representing each of the four chosen countries from the four magazines. The sampling framework considers two time periods to compare countries’ representation with two different presidents, and between men and women when present. The countries selected for analysis represent each category of the international news flows model: the core nations are the US and Germany; China is a semi-peripheral country; and Russia is peripheral. Examining textual and visual design elements on the covers and images in the cover stories reveals not only what editors believe visually attracts the reader’s attention to the magazine but also how the magazines frame and construct national images and national leaders. The cover is the most powerful editorial and design page in a magazine because images incorporate less intrusive framing tools. Thus, covers require less cognitive effort of audiences who may therefore be more likely to accept the visual frame without question. Analysis of design and linguistic elements in magazine covers helps to understand how media outlets shape their audience’s perceptions and how magazines frame global issues. While previous multimodal research of covers has focused mostly on lifestyle magazines or newspapers, this paper examines the power of current affairs magazines’ covers to shape audience perception of national image.Keywords: framing analysis, magazine covers, multimodality, national image, satire
Procedia PDF Downloads 1022317 Open-Loop Vector Control of Induction Motor with Space Vector Pulse Width Modulation Technique
Authors: Karchung, S. Ruangsinchaiwanich
Abstract:
This paper presents open-loop vector control method of induction motor with space vector pulse width modulation (SVPWM) technique. Normally, the closed loop speed control is preferred and is believed to be more accurate. However, it requires a position sensor to track the rotor position which is not desirable to use it for certain workspace applications. This paper exhibits the performance of three-phase induction motor with the simplest control algorithm without the use of a position sensor nor an estimation block to estimate rotor position for sensorless control. The motor stator currents are measured and are transformed to synchronously rotating (d-q-axis) frame by use of Clarke and Park transformation. The actual control happens in this frame where the measured currents are compared with the reference currents. The error signal is fed to a conventional PI controller, and the corrected d-q voltage is generated. The controller outputs are transformed back to three phase voltages and are fed to SVPWM block which generates PWM signal for the voltage source inverter. The open loop vector control model along with SVPWM algorithm is modeled in MATLAB/Simulink software and is experimented and validated in TMS320F28335 DSP board.Keywords: electric drive, induction motor, open-loop vector control, space vector pulse width modulation technique
Procedia PDF Downloads 1472316 Generalized Additive Model Approach for the Chilean Hake Population in a Bio-Economic Context
Authors: Selin Guney, Andres Riquelme
Abstract:
The traditional bio-economic method for fisheries modeling uses some estimate of the growth parameters and the system carrying capacity from a biological model for the population dynamics (usually a logistic population growth model) which is then analyzed as a traditional production function. The stock dynamic is transformed into a revenue function and then compared with the extraction costs to estimate the maximum economic yield. In this paper, the logistic population growth model for the population is combined with a forecast of the abundance and location of the stock by using a generalized additive model approach. The paper focuses on the Chilean hake population. This method allows for the incorporation of climatic variables and the interaction with other marine species, which in turn will increase the reliability of the estimates and generate better extraction paths for different conservation objectives, such as the maximum biological yield or the maximum economic yield.Keywords: bio-economic, fisheries, GAM, production
Procedia PDF Downloads 2522315 Best Timing for Capturing Satellite Thermal Images, Asphalt, and Concrete Objects
Authors: Toufic Abd El-Latif Sadek
Abstract:
The asphalt object represents the asphalted areas like roads, and the concrete object represents the concrete areas like concrete buildings. The efficient extraction of asphalt and concrete objects from one satellite thermal image occurred at a specific time, by preventing the gaps in times which give the close and same brightness values between asphalt and concrete, and among other objects. So that to achieve efficient extraction and then better analysis. Seven sample objects were used un this study, asphalt, concrete, metal, rock, dry soil, vegetation, and water. It has been found that, the best timing for capturing satellite thermal images to extract the two objects asphalt and concrete from one satellite thermal image, saving time and money, occurred at a specific time in different months. A table is deduced shows the optimal timing for capturing satellite thermal images to extract effectively these two objects.Keywords: asphalt, concrete, satellite thermal images, timing
Procedia PDF Downloads 3222314 COVID-19 Detection from Computed Tomography Images Using UNet Segmentation, Region Extraction, and Classification Pipeline
Authors: Kenan Morani, Esra Kaya Ayana
Abstract:
This study aimed to develop a novel pipeline for COVID-19 detection using a large and rigorously annotated database of computed tomography (CT) images. The pipeline consists of UNet-based segmentation, lung extraction, and a classification part, with the addition of optional slice removal techniques following the segmentation part. In this work, a batch normalization was added to the original UNet model to produce lighter and better localization, which is then utilized to build a full pipeline for COVID-19 diagnosis. To evaluate the effectiveness of the proposed pipeline, various segmentation methods were compared in terms of their performance and complexity. The proposed segmentation method with batch normalization outperformed traditional methods and other alternatives, resulting in a higher dice score on a publicly available dataset. Moreover, at the slice level, the proposed pipeline demonstrated high validation accuracy, indicating the efficiency of predicting 2D slices. At the patient level, the full approach exhibited higher validation accuracy and macro F1 score compared to other alternatives, surpassing the baseline. The classification component of the proposed pipeline utilizes a convolutional neural network (CNN) to make final diagnosis decisions. The COV19-CT-DB dataset, which contains a large number of CT scans with various types of slices and rigorously annotated for COVID-19 detection, was utilized for classification. The proposed pipeline outperformed many other alternatives on the dataset.Keywords: classification, computed tomography, lung extraction, macro F1 score, UNet segmentation
Procedia PDF Downloads 1312313 Statistical Modeling for Permeabilization of a Novel Yeast Isolate for β-Galactosidase Activity Using Organic Solvents
Authors: Shweta Kumari, Parmjit S. Panesar, Manab B. Bera
Abstract:
The hydrolysis of lactose using β-galactosidase is one of the most promising biotechnological applications, which has wide range of potential applications in food processing industries. However, due to intracellular location of the yeast enzyme, and expensive extraction methods, the industrial applications of enzymatic hydrolysis processes are being hampered. The use of permeabilization technique can help to overcome the problems associated with enzyme extraction and purification of yeast cells and to develop the economically viable process for the utilization of whole cell biocatalysts in food industries. In the present investigation, standardization of permeabilization process of novel yeast isolate was carried out using a statistical model approach known as Response Surface Methodology (RSM) to achieve maximal b-galactosidase activity. The optimum operating conditions for permeabilization process for optimal β-galactosidase activity obtained by RSM were 1:1 ratio of toluene (25%, v/v) and ethanol (50%, v/v), 25.0 oC temperature and treatment time of 12 min, which displayed enzyme activity of 1.71 IU /mg DW.Keywords: β-galactosidase, optimization, permeabilization, response surface methodology, yeast
Procedia PDF Downloads 2562312 Benefits of PRP in Third Molar Surgery - A Review of the Literature
Authors: Nitesh Kumar, Adel Elrasheed, Antonio Gagliardilugo
Abstract:
Introduction and aims: PRP has been increasing in popularity over the past decade. It is used in many facets of medicine and dentistry such as osteoarthritis, hair loss, skin rejunavation, healing of tendons after injury. Due to the increasing popularity of PRP in third molar surgery in dentistry, this study aims to identify the role of platelet rich plasma and its function in third molar surgery. Methodology: Three databases were chosen to source the articles for review: pubmed, science direct, and Cochrane. The keywords “platelet rich plasma”, “third molar extraction” and “wisdom tooth extraction” and literature review were used to search for relevant articles. Articles that were not in English were omitted and only systematic reviews relevant to the study were collected. All systematic reviews abstracts pertinent to the study were read by two reviewers to avoid bias. Results/statistics: 20 review articles were obtained of which 13 fulfilled the criteria. The Amstar tool validified the strength of these review articles. There is strong evidence in the literature that PRP in third molar surgery decreases post op pain, swelling and recovery time. 20 review articles were obtained of which 13 fulfilled the criteria. The Amstar tool validified the strength of these review articles. There is strong evidence in the literature that PRP in third molar surgery decreases post op pain, swelling and recovery time. Conclusions/clinical relevance: Platelet rich plasma plays a crucial role in patient recovery following the extraction of third molars and should be considered and offered as a routine part of third molar therapy.Keywords: PRP, third molar, extractions, wisdom teeth
Procedia PDF Downloads 622311 A Facile Nanocomposite of Graphene Oxide Reinforced Chitosan/Poly-Nitroaniline Polymer as a Highly Efficient Adsorbent for Extracting Polycyclic Aromatic Hydrocarbons from Tea Samples
Authors: Adel M. Al-Shutairi, Ahmed H. Al-Zahrani
Abstract:
Tea is a popular beverage drunk by millions of people throughout the globe. Tea has considerable health advantages, in-cluding antioxidant, antibacterial, antiviral, chemopreventive, and anticarcinogenic properties. As a result of environmental pollution (atmospheric deposition) and the production process, tealeaves may also include a variety of dangerous substances, such as polycyclic aromatic hydrocarbons (PAHs). In this study, graphene oxide reinforced chitosan/poly-nitroaniline polymer was prepared to develop a sensitive and reliable solid phase extraction method (SPE) for extraction of PAH7 in tea samples, followed by high-performance liquid chromatography- fluorescence detection. The prepared adsorbent was validated in terms of linearity, the limit of detection, the limit of quantification, recovery (%), accuracy (%), and precision (%) for the determination of the PAH7 (benzo[a]pyrene, benzo[a]anthracene, benzo[b]fluoranthene, chrysene, benzo[b]fluoranthene, Dibenzo[a,h]anthracene and Benzo[g,h,i]perylene) in tea samples. The concentration was determined in two types of tea commercially available in Saudi Arabia, including black tea and green tea. The maximum mean of Σ7PAHs in black tea samples was 68.23 ± 0.02 ug kg-1 and 26.68 ± 0.01 ug kg-1 in green tea samples. The minimum mean of Σ7PAHs in black tea samples was 37.93 ± 0.01 ug kg-1 and 15.26 ± 0.01 ug kg-1 in green tea samples. The mean value of benzo[a]pyrene in black tea samples ranged from 6.85 to 12.17 ug kg-1, where two samples exceeded the standard level (10 ug kg-1) established by the European Union (UE), while in green tea ranged from 1.78 to 2.81 ug kg-1. Low levels of Σ7PAHs in green tea samples were detected in comparison with black tea samples.Keywords: polycyclic aromatic hydrocarbons, CS, PNA and GO, black/green tea, solid phase extraction, Saudi Arabia
Procedia PDF Downloads 962310 Chemical Study of Volatile Organic Compounds (VOCS) from Xylopia aromatica (LAM.) Mart (Annonaceae)
Authors: Vanessa G. P. Severino, JOÃO Gabriel M. Junqueira, Michelle N. G. do Nascimento, Francisco W. B. Aquino, João B. Fernandes, Ana P. Terezan
Abstract:
The scientific interest in analyzing VOCs represents a significant modern research field as a result of importance in most branches of the present life and industry. Therefore it is extremely important to investigate, identify and isolate volatile substances, since they can be used in different areas, such as food, medicine, cosmetics, perfumery, aromatherapy, pesticides, repellents and other household products through methods for extracting volatile constituents, such as solid phase microextraction (SPME), hydrodistillation (HD), solvent extraction (SE), Soxhlet extraction, supercritical fluid extraction (SFE), stream distillation (SD) and vacuum distillation (VD). The Chemometrics is an area of chemistry that uses statistical and mathematical tools for the planning and optimization of the experimental conditions, and to extract relevant chemical information multivariate chemical data. In this context, the focus of this work was the study of the chemical VOCs by SPME of the specie X. aromatica, in search of constituents that can be used in the industrial sector as well as in food, cosmetics and perfumery, since these areas industrial has a considerable role. In addition, by chemometric analysis, we sought to maximize the answers of this research, in order to search for the largest number of compounds. The investigation of flowers from X. aromatica in vitro and in alive mode proved consistent, but certain factors supposed influence the composition of metabolites, and the chemometric analysis strengthened the analysis. Thus, the study of the chemical composition of X. aromatica contributed to the VOCs knowledge of the species and a possible application.Keywords: chemometrics, flowers, HS-SPME, Xylopia aromatica
Procedia PDF Downloads 3622309 Hacking's 'Between Goffman and Foucault': A Theoretical Frame for Criminology
Authors: Tomás Speziale
Abstract:
This paper aims to analyse how Ian Hacking states the theoretical basis of his research on the classification of people. Although all his early philosophical education had been based in Foucault, it is also true that Erving Goffman’s perspective provided him with epistemological and methodological tools for understanding face-to-face relationships. Hence, all his works must be thought of as social science texts that combine the research on how the individuals are constituted ‘top-down’ (as in Foucault), with the inquiry into how people renegotiate ‘bottom-up’ the classifications about them. Thus, Hacking´s proposal constitutes a middle ground between the French Philosopher and the American Sociologist. Placing himself between both authors allows Hacking to build a frame that is expected to adjust to Social Sciences’ main particularity: the fact that they study interactive kinds. These are kinds of people, which imply that those who are classified can change in certain ways that prompt the need for changing previous classifications themselves. It is all about the interaction between the labelling of people and the people who are classified. Consequently, understanding the way in which Hacking uses Foucault’s and Goffman’s theories is essential to fully comprehend the social dynamic between individuals and concepts, what Bert Hansen had called dialectical realism. His theoretical proposal, therefore, is not only valuable because it combines diverse perspectives, but also because it constitutes an utterly original and relevant framework for Sociological theory and particularly for Criminology.Keywords: classification of people, Foucault's archaeology, Goffman's interpersonal sociology, interactive kinds
Procedia PDF Downloads 3432308 Proposing a New Design Method for Added Viscoelastic Damper’s Application in Steel Moment-Frame
Authors: Saeed Javaherzadeh, Babak Dindar Safa
Abstract:
Structure, given its ductility, can depreciate significant amount of seismic energy in the form of hysteresis behavior; the amount of energy depreciation depends on the structure ductility rate. So in seismic guidelines such as ASCE7-10 code, to reduce the number of design forces and using the seismic energy dissipation capacity of structure, when entering non-linear behavior range of the materials, the response modification factor is used. Various parameters such as ductility modification factor, overstrength factor and reliability factor, are effective in determining the value of this factor. Also, gradually, energy dissipation systems, especially added dampers, have become an inseparable part of the seismic design. In this paper, in addition to reviewing of previous studies, using the response modification factor caused by using more added viscoelastic dampers, a new design method has introduced for steel moment-frame with added dampers installed. To do this, in addition to using bilinear behavior models and quick ways such as using the equivalent lateral force method and capacity spectrum method for the proposed design methodology, the results has been controlled with non-linear time history analysis for a number of structural. The analysis is done by Opensees Software.Keywords: added viscoelastic damper, design base shear, response modification factor, non-linear time history
Procedia PDF Downloads 4422307 Evaluation of Arsenic Removal in Soils Contaminated by the Phytoremediation Technique
Authors: V. Ibujes, A. Guevara, P. Barreto
Abstract:
Concentration of arsenic represents a serious threat to human health. It is a bioaccumulable toxic element and is transferred through the food chain. In Ecuador, values of 0.0423 mg/kg As are registered in potatoes of the skirts of the Tungurahua volcano. The increase of arsenic contamination in Ecuador is mainly due to mining activity, since the process of gold extraction generates toxic tailings with mercury. In the Province of Azuay, due to the mining activity, the soil reaches concentrations of 2,500 to 6,420 mg/kg As whereas in the province of Tungurahua it can be found arsenic concentrations of 6.9 to 198.7 mg/kg due to volcanic eruptions. Since the contamination by arsenic, the present investigation is directed to the remediation of the soils in the provinces of Azuay and Tungurahua by phytoremediation technique and the definition of a methodology of extraction by means of analysis of arsenic in the system soil-plant. The methodology consists in selection of two types of plants that have the best arsenic removal capacity in synthetic solutions 60 μM As, a lower percentage of mortality and hydroponics resistance. The arsenic concentrations in each plant were obtained from taking 10 ml aliquots and the subsequent analysis of the ICP-OES (inductively coupled plasma-optical emission spectrometry) equipment. Soils were contaminated with synthetic solutions of arsenic with the capillarity method to achieve arsenic concentration of 13 and 15 mg/kg. Subsequently, two types of plants were evaluated to reduce the concentration of arsenic in soils for 7 weeks. The global variance for soil types was obtained with the InfoStat program. To measure the changes in arsenic concentration in the soil-plant system, the Rhizo and Wenzel arsenic extraction methodology was used and subsequently analyzed with the ICP-OES (optima 8000 Pekin Elmer). As a result, the selected plants were bluegrass and llanten, due to the high percentages of arsenic removal of 55% and 67% and low mortality rates of 9% and 8% respectively. In conclusion, Azuay soil with an initial concentration of 13 mg/kg As reached the concentrations of 11.49 and 11.04 mg/kg As for bluegrass and llanten respectively, and for the initial concentration of 15 mg/kg As reached 11.79 and 11.10 mg/kg As for blue grass and llanten after 7 weeks. For the Tungurahua soil with an initial concentration of 13 mg/kg As it reached the concentrations of 11.56 and 12.16 mg/kg As for the bluegrass and llanten respectively, and for the initial concentration of 15 mg/kg As reached 11.97 and 12.27 mg/kg Ace for bluegrass and llanten after 7 weeks. The best arsenic extraction methodology of soil-plant system is Wenzel.Keywords: blue grass, llanten, phytoremediation, soil of Azuay, soil of Tungurahua, synthetic arsenic solution
Procedia PDF Downloads 1042306 Nonlinear Analysis of Steel Fiber Reinforced Concrete Frames Considering Shear Behaviour of Members under Varying Axial Load
Authors: Habib Akbarzadeh Bengar, Mohammad Asadi Kiadehi, Ali Rameeh
Abstract:
The result of the past earthquakes has shown that insufficient amount of stirrups and brittle behavior of concrete lead to the shear and flexural failure in reinforced concrete (RC) members. In this paper, an analytical model proposed to predict the nonlinear behavior of RC and SFRC elements and frames. In this model, some important parameter such as shear effect, varying axial load, and longitudinal bar buckling are considered. The results of analytical model were verified with experimental tests. The results of verification have shown that the proposed analytical model can predict the nonlinear behavior of RC and SFRC members and also frames accurately. In addition, the results have shown that use of steel fibers increased bearing capacity and ductility of RC frame. Due to this enhancement in shear strength and ductility, insufficient amount of stirrups, which resulted in shear failure, can be offset with usage of the steel fibers. In addition to the steps taken, to analyze the effects of fibers percentages on the bearing capacity and ductility of frames parametric studies have been performed to investigate of these effects.Keywords: nonlinear analysis, SFRC frame, shear failure, varying an axial load
Procedia PDF Downloads 2192305 An Analysis of a Relational Frame Skills Training Intervention to Increase General Intelligence in Early Childhood
Authors: Ian M. Grey, Bryan Roche, Anna Dillon, Justin Thomas, Sarah Cassidy, Dylan Colbert, Ian Stewart
Abstract:
This paper presents findings from a study conducted in two schools in Abu Dhabi. The hypothesis is that teaching young children to derive various relations between stimuli leads to increases in full-scale IQ scores of typically developing children. In the experimental group, sixteen 6-7-year-old children were exposed over six weeks to an intensive training intervention designed specifically for their age group. This training intervention, presented on a tablet, aimed to improve their understanding of the relations Same, Opposite, Different, contextual control over the concept of Sameness and Difference, and purely arbitrary derived relational responding for Sameness and Difference. In the control group, sixteen 6-7-year-old children interacted with KIBO robotics over six weeks. KIBO purports to improve cognitive skills through engagement with STEAM activities. Increases in full-scale IQ were recorded for most children in the experimental group, while no increases in full-scale IQ were recorded for the control group. These findings support the hypothesis that relational skills underlie many aspects of general cognitive ability.Keywords: early childhood, derived relational responding, intelligence, relational frame theory, relational skills
Procedia PDF Downloads 1852304 Optimal and Best Timing for Capturing Satellite Thermal Images of Concrete Object
Authors: Toufic Abd El-Latif Sadek
Abstract:
The concrete object represents the concrete areas, like buildings. The best, easy, and efficient extraction of the concrete object from satellite thermal images occurred at specific times during the days of the year, by preventing the gaps in times which give the close and same brightness from different objects. Thus, to achieve the best original data which is the aim of the study and then better extraction of the concrete object and then better analysis. The study was done using seven sample objects, asphalt, concrete, metal, rock, dry soil, vegetation, and water, located at one place carefully investigated in a way that all the objects achieve the homogeneous in acquired data at the same time and same weather conditions. The samples of the objects were on the roof of building at position taking by global positioning system (GPS) which its geographical coordinates is: Latitude= 33 degrees 37 minutes, Longitude= 35 degrees 28 minutes, Height= 600 m. It has been found that the first choice and the best time in February is at 2:00 pm, in March at 4 pm, in April and may at 12 pm, in August at 5:00 pm, in October at 11:00 am. The best time in June and November is at 2:00 pm.Keywords: best timing, concrete areas, optimal, satellite thermal images
Procedia PDF Downloads 3542303 Dynamic High-Rise Moment Resisting Frame Dissipation Performances Adopting Glazed Curtain Walls with Superelastic Shape Memory Alloy Joints
Authors: Lorenzo Casagrande, Antonio Bonati, Ferdinando Auricchio, Antonio Occhiuzzi
Abstract:
This paper summarizes the results of a survey on smart non-structural element dynamic dissipation when installed in modern high-rise mega-frame prototypes. An innovative glazed curtain wall was designed using Shape Memory Alloy (SMA) joints in order to increase the energy dissipation and enhance the seismic/wind response of the structures. The studied buildings consisted of thirty- and sixty-storey planar frames, extracted from reference three-dimensional steel Moment Resisting Frame (MRF) with outriggers and belt trusses. The internal core was composed of a CBF system, whilst outriggers were placed every fifteen stories to limit second order effects and inter-storey drifts. These structural systems were designed in accordance with European rules and numerical FE models were developed with an open-source code, able to account for geometric and material nonlinearities. With regard to the characterization of non-structural building components, full-scale crescendo tests were performed on aluminium/glass curtain wall units at the laboratory of the Construction Technologies Institute (ITC) of the Italian National Research Council (CNR), deriving force-displacement curves. Three-dimensional brick-based inelastic FE models were calibrated according to experimental results, simulating the fac¸ade response. Since recent seismic events and extreme dynamic wind loads have generated the large occurrence of non-structural components failure, which causes sensitive economic losses and represents a hazard for pedestrians safety, a more dissipative glazed curtain wall was studied. Taking advantage of the mechanical properties of SMA, advanced smart joints were designed with the aim to enhance both the dynamic performance of the single non-structural unit and the global behavior. Thus, three-dimensional brick-based plastic FE models were produced, based on the innovated non-structural system, simulating the evolution of mechanical degradation in aluminium-to-glass and SMA-to-glass connections when high deformations occurred. Consequently, equivalent nonlinear links were calibrated to reproduce the behavior of both tested and smart designed units, and implemented on the thirty- and sixty-storey structural planar frame FE models. Nonlinear time history analyses (NLTHAs) were performed to quantify the potential of the new system, when considered in the lateral resisting frame system (LRFS) of modern high-rise MRFs. Sensitivity to the structure height was explored comparing the responses of the two prototypes. Trends in global and local performance were discussed to show that, if accurately designed, advanced materials in non-structural elements provide new sources of energy dissipation.Keywords: advanced technologies, glazed curtain walls, non-structural elements, seismic-action reduction, shape memory alloy
Procedia PDF Downloads 3292302 Experimental Study of Infill Walls with Joint Reinforcement Subjected to In-Plane Lateral Load
Authors: J. Martin Leal-Graciano, Juan J. Pérez-Gavilán, A. Reyes-Salazar, J. H. Castorena, J. L. Rivera-Salas
Abstract:
The experimental results about the global behavior of twelve 1:2 scaled reinforced concrete frames subject to in-plane lateral load are presented. The main objective was to generate experimental evidence about the use of steel bars within mortar bed joints as shear reinforcement in infill walls. Similar to the Canadian and New Zealand standards, the Mexican code includes specifications for this type of reinforcement. However, these specifications were obtained through experimental studies of load-bearing walls, mainly confined walls. Little information is found in the existing literature about the effects of joint reinforcement on the seismic behavior of infill masonry walls. Consequently, the Mexican code establishes the same equations to estimate the contribution of joint reinforcement for both confined walls and infill walls. Confined masonry construction and a reinforced concrete frame infilled with masonry walls have similar appearances. However, substantial differences exist between these two construction systems, which are mainly related to the sequence of construction and to how these structures support vertical and lateral loads. To achieve the objective established, ten reinforced concrete frames with masonry infill walls were built and tested in pairs, having both specimens in the pair identical characteristics except that one of them included joint reinforcement. The variables between pairs were the type of units, the size of the columns of the frame, and the aspect ratio of the wall. All cases included tie columns and tie beams on the perimeter of the wall to anchor the joint reinforcement. Also, two bare frames with identical characteristics to the infilled frames were tested. The purpose was to investigate the effects of the infill wall on the behavior of the system to in-plane lateral load. In addition, the experimental results were compared with the prediction of the Mexican code. All the specimens were tested in a cantilever under reversible cyclic lateral load. To simulate gravity load, constant vertical load was applied on the top of the columns. The results indicate that the contribution of the joint reinforcement to lateral strength depends on the size of the columns of the frame. Larger size columns produce a failure mode that is predominantly a sliding mode. Sliding inhibits the production of new inclined cracks, which are necessary to activate (deform) the joint reinforcement. Regarding the effects of joint reinforcement in the performance of confined masonry walls, many facts were confirmed for infill walls. This type of reinforcement increases the lateral strength of the wall, produces a more distributed cracking, and reduces the width of the cracks. Moreover, it reduces the ductility demand of the system at maximum strength. The prediction of the lateral strength provided by the Mexican code is a property in some cases; however, the effect of the size of the columns on the contribution of joint reinforcement needs to be better understood.Keywords: experimental study, infill wall, infilled frame, masonry wall
Procedia PDF Downloads 1752301 Influence of the Low Frequency Ultrasound on the Cadmium (II) Biosorption by an Ecofriendly Biocomposite (Extraction Solid Waste of Ammi visnaga / Calcium Alginate): Kinetic Modeling
Authors: L. Nouri Taiba, Y. Bouhamidi, F. Kaouah, Z. Bendjama, M. Trari
Abstract:
In the present study, an ecofriendly biocomposite namely calcium alginate immobilized Ammi Visnaga (Khella) extraction waste (SWAV/CA) was prepared by electrostatic extrusion method and used on the cadmium biosorption from aqueous phase with and without the assistance of ultrasound in batch conditions. The influence of low frequency ultrasound (37 and 80 KHz) on the cadmium biosorption kinetics was studied. The obtained results show that the ultrasonic irradiation significantly enhances and improves the efficiency of the cadmium removal. The Pseudo first order, Pseudo-second-order, Intraparticle diffusion, and Elovich models were evaluated using the non-linear curve fitting analysis method. Modeling of kinetic results shows that biosorption process is best described by the pseudo-second order and Elovich, in both the absence and presence of ultrasound.Keywords: biocomposite, biosorption, cadmium, non-linear analysis, ultrasound
Procedia PDF Downloads 2772300 The Influence of Shear Wall Position on Seismic Performance in Buildings
Authors: Akram Khelaifia, Nesreddine Djafar Henni
Abstract:
Reinforced concrete shear walls are essential components in protecting buildings from seismic forces by providing both strength and stiffness. This study focuses on optimizing the placement of shear walls in a high seismic zone. Through nonlinear analyses conducted on an eight-story building, various scenarios of shear wall positions are investigated to evaluate their impact on seismic performance. Employing a performance-based seismic design (PBSD) approach, the study aims to meet acceptance criteria related to inter-story drift ratio and damage levels. The findings emphasize the importance of concentrating shear walls in the central area of the building during the design phase. This strategic placement proves more effective compared to peripheral distributions, resulting in reduced inter-story drift and mitigated potential damage during seismic events. Additionally, the research explores the use of shear walls that completely infill the frame, forming compound shapes like Box configurations. It is discovered that incorporating such complete shear walls significantly enhances the structure's reliability concerning inter-story drift. Conversely, the absence of complete shear walls within the frame leads to reduced stiffness and the potential deterioration of short beams.Keywords: performance level, pushover analysis, shear wall, plastic hinge, nonlinear analyses
Procedia PDF Downloads 532299 Simulation of Multistage Extraction Process of Co-Ni Separation Using Ionic Liquids
Authors: Hongyan Chen, Megan Jobson, Andrew J. Masters, Maria Gonzalez-Miquel, Simon Halstead, Mayri Diaz de Rienzo
Abstract:
Ionic liquids offer excellent advantages over conventional solvents for industrial extraction of metals from aqueous solutions, where such extraction processes bring opportunities for recovery, reuse, and recycling of valuable resources and more sustainable production pathways. Recent research on the use of ionic liquids for extraction confirms their high selectivity and low volatility, but there is relatively little focus on how their properties can be best exploited in practice. This work addresses gaps in research on process modelling and simulation, to support development, design, and optimisation of these processes, focusing on the separation of the highly similar transition metals, cobalt, and nickel. The study exploits published experimental results, as well as new experimental results, relating to the separation of Co and Ni using trihexyl (tetradecyl) phosphonium chloride. This extraction agent is attractive because it is cheaper, more stable and less toxic than fluorinated hydrophobic ionic liquids. This process modelling work concerns selection and/or development of suitable models for the physical properties, distribution coefficients, for mass transfer phenomena, of the extractor unit and of the multi-stage extraction flowsheet. The distribution coefficient model for cobalt and HCl represents an anion exchange mechanism, supported by the literature and COSMO-RS calculations. Parameters of the distribution coefficient models are estimated by fitting the model to published experimental extraction equilibrium results. The mass transfer model applies Newman’s hard sphere model. Diffusion coefficients in the aqueous phase are obtained from the literature, while diffusion coefficients in the ionic liquid phase are fitted to dynamic experimental results. The mass transfer area is calculated from the surface to mean diameter of liquid droplets of the dispersed phase, estimated from the Weber number inside the extractor. New experiments measure the interfacial tension between the aqueous and ionic phases. The empirical models for predicting the density and viscosity of solutions under different metal loadings are also fitted to new experimental data. The extractor is modelled as a continuous stirred tank reactor with mass transfer between the two phases and perfect phase separation of the outlet flows. A multistage separation flowsheet simulation is set up to replicate a published experiment and compare model predictions with the experimental results. This simulation model is implemented in gPROMS software for dynamic process simulation. The results of single stage and multi-stage flowsheet simulations are shown to be in good agreement with the published experimental results. The estimated diffusion coefficient of cobalt in the ionic liquid phase is in reasonable agreement with published data for the diffusion coefficients of various metals in this ionic liquid. A sensitivity study with this simulation model demonstrates the usefulness of the models for process design. The simulation approach has potential to be extended to account for other metals, acids, and solvents for process development, design, and optimisation of extraction processes applying ionic liquids for metals separations, although a lack of experimental data is currently limiting the accuracy of models within the whole framework. Future work will focus on process development more generally and on extractive separation of rare earths using ionic liquids.Keywords: distribution coefficient, mass transfer, COSMO-RS, flowsheet simulation, phosphonium
Procedia PDF Downloads 190