Search results for: conjugate gradient method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 19031

Search results for: conjugate gradient method

18701 Comparative Study of Soliton Collisions in Uniform and Nonuniform Magnetized Plasma

Authors: Renu Tomar, Hitendra K. Malik, Raj P. Dahiya

Abstract:

Similar to the sound waves in air, plasmas support the propagation of ion waves, which evolve into the solitary structures when the effect of non linearity and dispersion are balanced. The ion acoustic solitary waves have been investigated in details in homogeneous plasmas, inhomogeneous plasmas, and magnetized plasmas. The ion acoustic solitary waves are also found to reflect from a density gradient or boundary present in the plasma after propagating. Another interesting feature of the solitary waves is their collision. In the present work, we carry out analytical calculations for the head-on collision of solitary waves in a magnetized plasma which has dust grains in addition to the ions and electrons. For this, we employ Poincar´e-Lighthill-Kuo (PLK) method. To lowest nonlinear order, the problem of colliding solitary waves leads to KdV (modified KdV) equations and also yields the phase shifts that occur in the interaction. These calculations are accomplished for the uniform and nonuniform plasmas, and the results on the soliton properties are discussed in detail.

Keywords: inhomogeneous magnetized plasma, dust charging, soliton collisions, magnetized plasma

Procedia PDF Downloads 451
18700 Electrokinetic Transport of Power Law Fluid through Hydrophobic Micro-Slits

Authors: Ainul Haque, Ameeye Kumar Nayak

Abstract:

Flow enhancement and species transport in a slit hydrophobic microchannel is studied for non-Newtonian fluids with the externally imposed electric field and pressure gradient. The incompressible Poisson-Nernst-Plank equations and the Navier-Stokes equations are approximated by lubrication theory to quantify the flow structure due to hydrophobic and hydrophilic surfaces. The analytical quantification of velocity and pressure of electroosmotic flow (EOF) is made with the numerical results due to the staggered grid based finite volume method for flow governing equations. The resistance force due to fluid friction and shear force along the surface are decreased by the hydrophobicity, enables the faster movement of fluid particles. The resulting flow enhancement factor Ef is increased with the low viscous fluid and provides maximum species transport. Also, the analytical comparison of EOF with pressure driven EOF justifies the flow enhancement due to hydrophobicity and shear impact on flow variation.

Keywords: electroosmotic flow, hydrophobic surface, power-law fluid, shear effect

Procedia PDF Downloads 348
18699 Multi-objective Rationality Optimisation for Robotic-fabrication-oriented Free-form Timber Structure Morphology Design

Authors: Yiping Meng, Yiming Sun

Abstract:

The traditional construction industry is unable to meet the requirements for novel fabrication and construction. Automated construction and digital design have emerged as industry development trends that compensate for this shortcoming under the backdrop of Industrial Revolution 4.0. Benefitting from more flexible working space and more various end-effector tools compared to CNC methods, robot fabrication and construction techniques have been used in irregular architectural design. However, there is a lack of a systematic and comprehensive design and optimisation workflow considering geometric form, material, and fabrication methods. This paper aims to propose a design optimisation workflow for improving the rationality of a free-form timber structure fabricated by the robotic arm. Firstly, the free-form surface is described by NURBS, while its structure is calculated using the finite element analysis method. Then, by considering the characteristics and limiting factors of robotic timber fabrication, strain energy and robustness are set as optimisation objectives to optimise structural morphology by gradient descent method. As a result, an optimised structure with axial force as the main force and uniform stress distribution is generated after the structure morphology optimisation process. With the decreased strain energy and the improved robustness, the generated structure's bearing capacity and mechanical properties have been enhanced. The results prove the feasibility and effectiveness of the proposed optimisation workflow for free-form timber structure morphology design.

Keywords: robotic fabrication, free-form timber structure, Multi-objective optimisation, Structural morphology, rational design

Procedia PDF Downloads 174
18698 Elvis Improved Method for Solving Simultaneous Equations in Two Variables with Some Applications

Authors: Elvis Adam Alhassan, Kaiyu Tian, Akos Konadu, Ernest Zamanah, Michael Jackson Adjabui, Ibrahim Justice Musah, Esther Agyeiwaa Owusu, Emmanuel K. A. Agyeman

Abstract:

In this paper, how to solve simultaneous equations using the Elvis improved method is shown. The Elvis improved method says; to make one variable in the first equation the subject; make the same variable in the second equation the subject; equate the results and simplify to obtain the value of the unknown variable; put the value of the variable found into one equation from the first or second steps and simplify for the remaining unknown variable. The difference between our Elvis improved method and the substitution method is that: with Elvis improved method, the same variable is made the subject in both equations, and the two resulting equations equated, unlike the substitution method where one variable is made the subject of only one equation and substituted into the other equation. After describing the Elvis improved method, findings from 100 secondary students and the views of 5 secondary tutors to demonstrate the effectiveness of the method are presented. The study's purpose is proved by hypothetical examples.

Keywords: simultaneous equations, substitution method, elimination method, graphical method, Elvis improved method

Procedia PDF Downloads 99
18697 Environmentally Sustainable Transparent Wood: A Fully Green Approach from Bleaching to Impregnation for Energy-Efficient Engineered Wood Components

Authors: Francesca Gullo, Paola Palmero, Massimo Messori

Abstract:

Transparent wood is considered a promising structural material for the development of environmentally friendly, energy-efficient engineered components. To obtain transparent wood from natural wood materials two approaches can be used: i) bottom-up and ii) top-down. Through the second method, the color of natural wood samples is lightened through a chemical bleaching process that acts on chromophore groups of lignin, such as the benzene ring, quinonoid, vinyl, phenolics, and carbonyl groups. These chromophoric units form complex conjugate systems responsible for the brown color of wood. There are two strategies to remove color and increase the whiteness of wood: i) lignin removal and ii) lignin bleaching. In the lignin removal strategy, strong chemicals containing chlorine (chlorine, hypochlorite, and chlorine dioxide) and oxidizers (oxygen, ozone, and peroxide) are used to completely destroy and dissolve the lignin. In lignin bleaching methods, a moderate reductive (hydrosulfite) or oxidative (hydrogen peroxide) is commonly used to alter or remove the groups and chromophore systems of lignin, selectively discoloring the lignin while keeping the macrostructure intact. It is, therefore, essential to manipulate nanostructured wood by precisely controlling the nanopores in the cell walls by monitoring both chemical treatments and process conditions, for instance, the treatment time, the concentration of chemical solutions, the pH value, and the temperature. The elimination of wood light scattering is the second step in the fabrication of transparent wood materials, which can be achieved through two-step approaches: i) the polymer impregnation method and ii) the densification method. For the polymer impregnation method, the wood scaffold is treated with polymers having a corresponding refractive index (e.g., PMMA and epoxy resins) under vacuum to obtain the transparent composite material, which can finally be pressed to align the cellulose fibers and reduce interfacial defects in order to have a finished product with high transmittance (>90%) and excellent light-guiding. However, both the solution-based bleaching and the impregnation processes used to produce transparent wood generally consume large amounts of energy and chemicals, including some toxic or pollutant agents, and are difficult to scale up industrially. Here, we report a method to produce optically transparent wood by modifying the lignin structure with a chemical reaction at room temperature using small amounts of hydrogen peroxide in an alkaline environment. This method preserves the lignin, which results only deconjugated and acts as a binder, providing both a strong wood scaffold and suitable porosity for infiltration of biobased polymers while reducing chemical consumption, the toxicity of the reagents used, polluting waste, petroleum by-products, energy and processing time. The resulting transparent wood demonstrates high transmittance and low thermal conductivity. Through the combination of process efficiency and scalability, the obtained materials are promising candidates for application in the field of construction for modern energy-efficient buildings.

Keywords: bleached wood, energy-efficient components, hydrogen peroxide, transparent wood, wood composites

Procedia PDF Downloads 26
18696 Customer Churn Prediction by Using Four Machine Learning Algorithms Integrating Features Selection and Normalization in the Telecom Sector

Authors: Alanoud Moraya Aldalan, Abdulaziz Almaleh

Abstract:

A crucial component of maintaining a customer-oriented business as in the telecom industry is understanding the reasons and factors that lead to customer churn. Competition between telecom companies has greatly increased in recent years. It has become more important to understand customers’ needs in this strong market of telecom industries, especially for those who are looking to turn over their service providers. So, predictive churn is now a mandatory requirement for retaining those customers. Machine learning can be utilized to accomplish this. Churn Prediction has become a very important topic in terms of machine learning classification in the telecommunications industry. Understanding the factors of customer churn and how they behave is very important to building an effective churn prediction model. This paper aims to predict churn and identify factors of customers’ churn based on their past service usage history. Aiming at this objective, the study makes use of feature selection, normalization, and feature engineering. Then, this study compared the performance of four different machine learning algorithms on the Orange dataset: Logistic Regression, Random Forest, Decision Tree, and Gradient Boosting. Evaluation of the performance was conducted by using the F1 score and ROC-AUC. Comparing the results of this study with existing models has proven to produce better results. The results showed the Gradients Boosting with feature selection technique outperformed in this study by achieving a 99% F1-score and 99% AUC, and all other experiments achieved good results as well.

Keywords: machine learning, gradient boosting, logistic regression, churn, random forest, decision tree, ROC, AUC, F1-score

Procedia PDF Downloads 113
18695 Classification on Statistical Distributions of a Complex N-Body System

Authors: David C. Ni

Abstract:

Contemporary models for N-body systems are based on temporal, two-body, and mass point representation of Newtonian mechanics. Other mainstream models include 2D and 3D Ising models based on local neighborhood the lattice structures. In Quantum mechanics, the theories of collective modes are for superconductivity and for the long-range quantum entanglement. However, these models are still mainly for the specific phenomena with a set of designated parameters. We are therefore motivated to develop a new construction directly from the complex-variable N-body systems based on the extended Blaschke functions (EBF), which represent a non-temporal and nonlinear extension of Lorentz transformation on the complex plane – the normalized momentum spaces. A point on the complex plane represents a normalized state of particle momentums observed from a reference frame in the theory of special relativity. There are only two key parameters, normalized momentum and nonlinearity for modelling. An algorithm similar to Jenkins-Traub method is adopted for solving EBF iteratively. Through iteration, the solution sets show a form of σ + i [-t, t], where σ and t are the real numbers, and the [-t, t] shows various distributions, such as 1-peak, 2-peak, and 3-peak etc. distributions and some of them are analog to the canonical distributions. The results of the numerical analysis demonstrate continuum-to-discreteness transitions, evolutional invariance of distributions, phase transitions with conjugate symmetry, etc., which manifest the construction as a potential candidate for the unification of statistics. We hereby classify the observed distributions on the finite convergent domains. Continuous and discrete distributions both exist and are predictable for given partitions in different regions of parameter-pair. We further compare these distributions with canonical distributions and address the impacts on the existing applications.

Keywords: blaschke, lorentz transformation, complex variables, continuous, discrete, canonical, classification

Procedia PDF Downloads 285
18694 Comparison of Different Machine Learning Algorithms for Solubility Prediction

Authors: Muhammet Baldan, Emel Timuçin

Abstract:

Molecular solubility prediction plays a crucial role in various fields, such as drug discovery, environmental science, and material science. In this study, we compare the performance of five machine learning algorithms—linear regression, support vector machines (SVM), random forests, gradient boosting machines (GBM), and neural networks—for predicting molecular solubility using the AqSolDB dataset. The dataset consists of 9981 data points with their corresponding solubility values. MACCS keys (166 bits), RDKit properties (20 properties), and structural properties(3) features are extracted for every smile representation in the dataset. A total of 189 features were used for training and testing for every molecule. Each algorithm is trained on a subset of the dataset and evaluated using metrics accuracy scores. Additionally, computational time for training and testing is recorded to assess the efficiency of each algorithm. Our results demonstrate that random forest model outperformed other algorithms in terms of predictive accuracy, achieving an 0.93 accuracy score. Gradient boosting machines and neural networks also exhibit strong performance, closely followed by support vector machines. Linear regression, while simpler in nature, demonstrates competitive performance but with slightly higher errors compared to ensemble methods. Overall, this study provides valuable insights into the performance of machine learning algorithms for molecular solubility prediction, highlighting the importance of algorithm selection in achieving accurate and efficient predictions in practical applications.

Keywords: random forest, machine learning, comparison, feature extraction

Procedia PDF Downloads 7
18693 Spin-Polarized Structural, Electronic, and Magnetic Properties of Co and Mn-Doped CdTe in Zinc-Blende Phase

Authors: A.Zitouni, S.Bentata, B.Bouadjemi, T.Lantri, W. Benstaali, Z.Aziz, S.Cherid, A. Sefir

Abstract:

Structural, electronic, and magnetic properties of Co and Mn-doped CdTe have been studied by employing the full potential linear augmented plane waves (FP-LAPW) method within the spin-polarized density functional theory (DFT). The electronic exchange-correlation energy is described by generalized gradient approximation (GGA) as exchange–correlation (XC) potential. We have calculated the lattice parameters, bulk modulii and the first pressure derivatives of the bulk modulii, spin-polarized band structures, and total and local densities of states. The value of calculated magnetic moment per Co and Mn impurity atoms is found to be 2.21 µB for CdCoTe and 3.20 µB for CdMnTe. The calculated densities of states presented in this study identify the half-metallic of Co and Mn-doped CdTe.

Keywords: electronic structure, density functional theory, band structures, half-metallic, magnetic moment

Procedia PDF Downloads 439
18692 Material Parameter Identification of Modified AbdelKarim-Ohno Model

Authors: Martin Cermak, Tomas Karasek, Jaroslav Rojicek

Abstract:

The key role in phenomenological modelling of cyclic plasticity is good understanding of stress-strain behaviour of given material. There are many models describing behaviour of materials using numerous parameters and constants. Combination of individual parameters in those material models significantly determines whether observed and predicted results are in compliance. Parameter identification techniques such as random gradient, genetic algorithm, and sensitivity analysis are used for identification of parameters using numerical modelling and simulation. In this paper genetic algorithm and sensitivity analysis are used to study effect of 4 parameters of modified AbdelKarim-Ohno cyclic plasticity model. Results predicted by Finite Element (FE) simulation are compared with experimental data from biaxial ratcheting test with semi-elliptical loading path.

Keywords: genetic algorithm, sensitivity analysis, inverse approach, finite element method, cyclic plasticity, ratcheting

Procedia PDF Downloads 429
18691 Different Methods of Fe3O4 Nano Particles Synthesis

Authors: Arezoo Hakimi, Afshin Farahbakhsh

Abstract:

Herein, we comparison synthesized Fe3O4 using, hydrothermal method, Mechanochemical processes and solvent thermal method. The Hydrothermal Technique has been the most popular one, gathering interest from scientists and technologists of different disciplines, particularly in the last fifteen years. In the hydrothermal method Fe3O4 microspheres, in which many nearly monodisperse spherical particles with diameters of about 400nm, in the mechanochemical method regular morphology indicates that the particles are well crystallized and in the solvent thermal method Fe3O4 nanoparticles have good properties of uniform size and good dispersion.

Keywords: Fe3O4 nanoparticles, hydrothermal method, mechanochemical processes, solvent thermal method

Procedia PDF Downloads 331
18690 An Evolutionary Approach for QAOA for Max-Cut

Authors: Francesca Schiavello

Abstract:

This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.

Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization

Procedia PDF Downloads 37
18689 A Comparison of Smoothing Spline Method and Penalized Spline Regression Method Based on Nonparametric Regression Model

Authors: Autcha Araveeporn

Abstract:

This paper presents a study about a nonparametric regression model consisting of a smoothing spline method and a penalized spline regression method. We also compare the techniques used for estimation and prediction of nonparametric regression model. We tried both methods with crude oil prices in dollars per barrel and the Stock Exchange of Thailand (SET) index. According to the results, it is concluded that smoothing spline method performs better than that of penalized spline regression method.

Keywords: nonparametric regression model, penalized spline regression method, smoothing spline method, Stock Exchange of Thailand (SET)

Procedia PDF Downloads 403
18688 Studying Relationship between Local Geometry of Decision Boundary with Network Complexity for Robustness Analysis with Adversarial Perturbations

Authors: Tushar K. Routh

Abstract:

If inputs are engineered in certain manners, they can influence deep neural networks’ (DNN) performances by facilitating misclassifications, a phenomenon well-known as adversarial attacks that question networks’ vulnerability. Recent studies have unfolded the relationship between vulnerability of such networks with their complexity. In this paper, the distinctive influence of additional convolutional layers at the decision boundaries of several DNN architectures was investigated. Here, to engineer inputs from widely known image datasets like MNIST, Fashion MNIST, and Cifar 10, we have exercised One Step Spectral Attack (OSSA) and Fast Gradient Method (FGM) techniques. The aftermaths of adding layers to the robustness of the architectures have been analyzed. For reasoning, separation width from linear class partitions and local geometry (curvature) near the decision boundary have been examined. The result reveals that model complexity has significant roles in adjusting relative distances from margins, as well as the local features of decision boundaries, which impact robustness.

Keywords: DNN robustness, decision boundary, local curvature, network complexity

Procedia PDF Downloads 48
18687 Temperature Distribution for Asphalt Concrete-Concrete Composite Pavement

Authors: Tetsya Sok, Seong Jae Hong, Young Kyu Kim, Seung Woo Lee

Abstract:

The temperature distribution for asphalt concrete (AC)-Concrete composite pavement is one of main influencing factor that affects to performance life of pavement. The temperature gradient in concrete slab underneath the AC layer results the critical curling stress and lead to causes de-bonding of AC-Concrete interface. These stresses, when enhanced by repetitive axial loadings, also contribute to the fatigue damage and eventual crack development within the slab. Moreover, the temperature change within concrete slab extremely causes the slab contracts and expands that significantly induces reflective cracking in AC layer. In this paper, the numerical prediction of pavement temperature was investigated using one-dimensional finite different method (FDM) in fully explicit scheme. The numerical predicted model provides a fundamental and clear understanding of heat energy balance including incoming and outgoing thermal energies in addition to dissipated heat in the system. By using the reliable meteorological data for daily air temperature, solar radiation, wind speech and variable pavement surface properties, the predicted pavement temperature profile was validated with the field measured data. Additionally, the effects of AC thickness and daily air temperature on the temperature profile in underlying concrete were also investigated. Based on obtained results, the numerical predicted temperature of AC-Concrete composite pavement using FDM provided a good accuracy compared to field measured data and thicker AC layer significantly insulates the temperature distribution in underlying concrete slab.

Keywords: asphalt concrete, finite different method (FDM), curling effect, heat transfer, solar radiation

Procedia PDF Downloads 251
18686 Genetic Diversity Analysis in Triticum Aestivum Using Microsatellite Markers

Authors: Prachi Sharma, Mukesh Kumar Rana

Abstract:

In the present study, the simple sequence repeat(SSR) markers have been used in analysis of genetic diversity of 37 genotypes of Triticum aestivum. The DNA was extracted using cTAB method. The DNA was quantified using the fluorimeter. The annealing temperatures for 27 primer pairs were standardized using gradient PCR, out of which 16 primers gave satisfactory amplification at temperature ranging from 50-62⁰ C. Out of 16 polymorphic SSR markers only 10 SSR primer pairs were used in the study generating 34 reproducible amplicons among 37 genotypes out of which 30 were polymorphic. Primer pairs Xgwm533, Xgwm 160, Xgwm 408, Xgwm 120, Xgwm 186, Xgwm 261 produced maximum percent of polymorphic bands (100%). The bands ranged on an average of 3.4 bands per primer. The genetic relationship was determined using Jaccard pair wise similarity co-efficient and UPGMA cluster analysis with NTSYS Pc.2 software. The values of similarity index range from 0-1. The similarity coefficient ranged from 0.13 to 0.97. A minimum genetic similarity (0.13) was observed between VL 804 and HPW 288, meaning they are only 13% similar. More number of available SSR markers can be useful for supporting the genetic diversity analysis in the above wheat genotypes.

Keywords: wheat, genetic diversity, microsatellite, polymorphism

Procedia PDF Downloads 586
18685 A Fast Multi-Scale Finite Element Method for Geophysical Resistivity Measurements

Authors: Mostafa Shahriari, Sergio Rojas, David Pardo, Angel Rodriguez- Rozas, Shaaban A. Bakr, Victor M. Calo, Ignacio Muga

Abstract:

Logging-While Drilling (LWD) is a technique to record down-hole logging measurements while drilling the well. Nowadays, LWD devices (e.g., nuclear, sonic, resistivity) are mostly used commercially for geo-steering applications. Modern borehole resistivity tools are able to measure all components of the magnetic field by incorporating tilted coils. The depth of investigation of LWD tools is limited compared to the thickness of the geological layers. Thus, it is a common practice to approximate the Earth’s subsurface with a sequence of 1D models. For a 1D model, we can reduce the dimensionality of the problem using a Hankel transform. We can solve the resulting system of ordinary differential equations (ODEs) either (a) analytically, which results in a so-called semi-analytic method after performing a numerical inverse Hankel transform, or (b) numerically. Semi-analytic methods are used by the industry due to their high performance. However, they have major limitations, namely: -The analytical solution of the aforementioned system of ODEs exists only for piecewise constant resistivity distributions. For arbitrary resistivity distributions, the solution of the system of ODEs is unknown by today’s knowledge. -In geo-steering, we need to solve inverse problems with respect to the inversion variables (e.g., the constant resistivity value of each layer and bed boundary positions) using a gradient-based inversion method. Thus, we need to compute the corresponding derivatives. However, the analytical derivatives of cross-bedded formation and the analytical derivatives with respect to the bed boundary positions have not been published to the best of our knowledge. The main contribution of this work is to overcome the aforementioned limitations of semi-analytic methods by solving each 1D model (associated with each Hankel mode) using an efficient multi-scale finite element method. The main idea is to divide our computations into two parts: (a) offline computations, which are independent of the tool positions and we precompute only once and use them for all logging positions, and (b) online computations, which depend upon the logging position. With the above method, (a) we can consider arbitrary resistivity distributions along the 1D model, and (b) we can easily and rapidly compute the derivatives with respect to any inversion variable at a negligible additional cost by using an adjoint state formulation. Although the proposed method is slower than semi-analytic methods, its computational efficiency is still high. In the presentation, we shall derive the mathematical variational formulation, describe the proposed multi-scale finite element method, and verify the accuracy and efficiency of our method by performing a wide range of numerical experiments and comparing the numerical solutions to semi-analytic ones when the latest are available.

Keywords: logging-While-Drilling, resistivity measurements, multi-scale finite elements, Hankel transform

Procedia PDF Downloads 365
18684 Development and Total Error Concept Validation of Common Analytical Method for Quantification of All Residual Solvents Present in Amino Acids by Gas Chromatography-Head Space

Authors: A. Ramachandra Reddy, V. Murugan, Prema Kumari

Abstract:

Residual solvents in Pharmaceutical samples are monitored using gas chromatography with headspace (GC-HS). Based on current regulatory and compendial requirements, measuring the residual solvents are mandatory for all release testing of active pharmaceutical ingredients (API). Generally, isopropyl alcohol is used as the residual solvent in proline and tryptophan; methanol in cysteine monohydrate hydrochloride, glycine, methionine and serine; ethanol in glycine and lysine monohydrate; acetic acid in methionine. In order to have a single method for determining these residual solvents (isopropyl alcohol, ethanol, methanol and acetic acid) in all these 7 amino acids a sensitive and simple method was developed by using gas chromatography headspace technique with flame ionization detection. During development, no reproducibility, retention time variation and bad peak shape of acetic acid peaks were identified due to the reaction of acetic acid with the stationary phase (cyanopropyl dimethyl polysiloxane phase) of column and dissociation of acetic acid with water (if diluent) while applying temperature gradient. Therefore, dimethyl sulfoxide was used as diluent to avoid these issues. But most the methods published for acetic acid quantification by GC-HS uses derivatisation technique to protect acetic acid. As per compendia, risk-based approach was selected as appropriate to determine the degree and extent of the validation process to assure the fitness of the procedure. Therefore, Total error concept was selected to validate the analytical procedure. An accuracy profile of ±40% was selected for lower level (quantitation limit level) and for other levels ±30% with 95% confidence interval (risk profile 5%). The method was developed using DB-Waxetr column manufactured by Agilent contains 530 µm internal diameter, thickness: 2.0 µm, and length: 30 m. A constant flow of 6.0 mL/min. with constant make up mode of Helium gas was selected as a carrier gas. The present method is simple, rapid, and accurate, which is suitable for rapid analysis of isopropyl alcohol, ethanol, methanol and acetic acid in amino acids. The range of the method for isopropyl alcohol is 50ppm to 200ppm, ethanol is 50ppm to 3000ppm, methanol is 50ppm to 400ppm and acetic acid 100ppm to 400ppm, which covers the specification limits provided in European pharmacopeia. The accuracy profile and risk profile generated as part of validation were found to be satisfactory. Therefore, this method can be used for testing of residual solvents in amino acids drug substances.

Keywords: amino acid, head space, gas chromatography, total error

Procedia PDF Downloads 124
18683 Influence of Optimization Method on Parameters Identification of Hyperelastic Models

Authors: Bale Baidi Blaise, Gilles Marckmann, Liman Kaoye, Talaka Dya, Moustapha Bachirou, Gambo Betchewe, Tibi Beda

Abstract:

This work highlights the capabilities of particles swarm optimization (PSO) method to identify parameters of hyperelastic models. The study compares this method with Genetic Algorithm (GA) method, Least Squares (LS) method, Pattern Search Algorithm (PSA) method, Beda-Chevalier (BC) method and the Levenberg-Marquardt (LM) method. Four classic hyperelastic models are used to test the different methods through parameters identification. Then, the study compares the ability of these models to reproduce experimental Treloar data in simple tension, biaxial tension and pure shear.

Keywords: particle swarm optimization, identification, hyperelastic, model

Procedia PDF Downloads 143
18682 C-Coordinated Chitosan Metal Complexes: Design, Synthesis and Antifungal Properties

Authors: Weixiang Liu, Yukun Qin, Song Liu, Pengcheng Li

Abstract:

Plant diseases can cause the death of crops with great economic losses. Particularly, those diseases are usually caused by pathogenic fungi. Metal fungicides are a type of pesticide that has advantages of a low-cost, broad antimicrobial spectrum and strong sterilization effect. However, the frequent and wide application of traditional metal fungicides has caused serious problems such as environmental pollution, the outbreak of mites and phytotoxicity. Therefore, it is critically necessary to discover new organic metal fungicides alternatives that have a low metal content, low toxicity, and little influence on mites. Chitosan, the second most abundant natural polysaccharide next to cellulose, was proved to have broad-spectrum antifungal activity against a variety of fungi. However, the use of chitosan was limited due to its poor solubility and weaker antifungal activity compared with commercial fungicide. Therefore, in order to improve the water solubility and antifungal activity, many researchers grafted the active groups onto chitosan. The present work was to combine free metal ions with chitosan, to prepare more potent antifungal chitosan derivatives, thus, based on condensation reaction, chitosan derivative bearing amino pyridine group was prepared and subsequently followed by coordination with cupric ions, zinc ions and nickel ions to synthesize chitosan metal complexes. The calculations by density functional theory (DFT) show that the copper ions and nickel ions underwent dsp2 hybridization, the zinc ions underwent sp3 hybridization, and all of them are coordinated by the carbon atom in the p-π conjugate group and the oxygen atoms in the acetate ion. The antifungal properties of chitosan metal complexes against Phytophthora capsici (P. capsici), Gibberella zeae (G. zeae), Fusarium oxysporum (F. oxysporum) and Botrytis cinerea (B. cinerea) were also assayed. In addition, a plant toxicity experiment was carried out. The experiments indicated that the derivatives have significantly enhanced antifungal activity after metal ions complexation compared with the original chitosan. It was shown that 0.20 mg/mL of O-CSPX-Cu can 100% inhibit the growth of P. capsici and 0.20 mg/mL of O-CSPX-Ni can 87.5% inhibit the growth of B. cinerea. In general, their activities are better than the positive control oligosaccharides. The combination of the pyridine formyl groups seems to favor biological activity. Additionally, the ligand fashion was precisely analyzed, and the results revealed that the copper ions and nickel ions underwent dsp2 hybridization, the zinc ions underwent sp3 hybridization, and the carbon atoms of the p-π conjugate group and the oxygen atoms of acetate ion are involved in the coordination of metal ions. The phytotoxicity assay of O-CSPX-M was also conducted, unlike the traditional metal fungicides, the metal complexes were not significantly toxic to the leaves of wheat. O-CSPX-Zn can even increase chlorophyll content in wheat leaves at 0.40 mg/mL. This is mainly because chitosan itself promotes plant growth and counteracts the phytotoxicity of metal ions. The chitosan derivative described here may lend themselves to future applicative studies in crop protection.

Keywords: coordination, chitosan, metal complex, antifungal properties

Procedia PDF Downloads 293
18681 First Principle Calculations of the Structural and Optoelectronic Properties of Cubic Perovskite CsSrF3

Authors: Meriem Harmel, Houari Khachai

Abstract:

We have investigated the structural, electronic and optical properties of a compound perovskite CsSrF3 using the full-potential linearized augmented plane wave (FP-LAPW) method within density functional theory (DFT). In this approach, both the local density approximation (LDA) and the generalized gradient approximation (GGA) were used for exchange-correlation potential calculation. The ground state properties such as lattice parameter, bulk modulus and its pressure derivative were calculated and the results are compared whit experimental and theoretical data. Electronic and bonding properties are discussed from the calculations of band structure, density of states and electron charge density, where the fundamental energy gap is direct under ambient conditions. The contribution of the different bands was analyzed from the total and partial density of states curves. The optical properties (namely: the real and the imaginary parts of the dielectric function ε(ω), the refractive index n(ω) and the extinction coefficient k(ω)) were calculated for radiation up to 35.0 eV. This is the first quantitative theoretical prediction of the optical properties for the investigated compound and still awaits experimental confirmations.

Keywords: DFT, fluoroperovskite, electronic structure, optical properties

Procedia PDF Downloads 438
18680 Mathematical Reconstruction of an Object Image Using X-Ray Interferometric Fourier Holography Method

Authors: M. K. Balyan

Abstract:

The main principles of X-ray Fourier interferometric holography method are discussed. The object image is reconstructed by the mathematical method of Fourier transformation. The three methods are presented – method of approximation, iteration method and step by step method. As an example the complex amplitude transmission coefficient reconstruction of a beryllium wire is considered. The results reconstructed by three presented methods are compared. The best results are obtained by means of step by step method.

Keywords: dynamical diffraction, hologram, object image, X-ray holography

Procedia PDF Downloads 371
18679 Modified Approximation Methods for Finding an Optimal Solution for the Transportation Problem

Authors: N. Guruprasad

Abstract:

This paper presents a modification of approximation method for transportation problems. The initial basic feasible solution can be computed using either Russel's or Vogel's approximation methods. Russell’s approximation method provides another excellent criterion that is still quick to implement on a computer (not manually) In most cases Russel's method yields a better initial solution, though it takes longer than Vogel's method (finding the next entering variable in Russel's method is in O(n1*n2), and in O(n1+n2) for Vogel's method). However, Russel's method normally has a lesser total running time because less pivots are required to reach the optimum for all but small problem sizes (n1+n2=~20). With this motivation behind we have incorporated a variation of the same – what we have proposed it has TMC (Total Modified Cost) to obtain fast and efficient solutions.

Keywords: computation, efficiency, modified cost, Russell’s approximation method, transportation, Vogel’s approximation method

Procedia PDF Downloads 507
18678 Probability Fuzzy Aggregation Operators in Vehicle Routing Problem

Authors: Anna Sikharulidze, Gia Sirbiladze

Abstract:

For the evaluation of unreliability levels of movement on the closed routes in the vehicle routing problem, the fuzzy operators family is constructed. The interactions between routing factors in extreme conditions on the roads are considered. A multi-criteria decision-making model (MCDM) is constructed. Constructed aggregations are based on the Choquet integral and the associated probability class of a fuzzy measure. Propositions on the correctness of the extension are proved. Connections between the operators and the compositions of dual triangular norms are described. The conjugate connections between the constructed operators are shown. Operators reflect interactions among all the combinations of the factors in the fuzzy MCDM process. Several variants of constructed operators are used in the decision-making problem regarding the assessment of unreliability and possibility levels of movement on closed routes.

Keywords: vehicle routing problem, associated probabilities of a fuzzy measure, choquet integral, fuzzy aggregation operator

Procedia PDF Downloads 291
18677 Vertical Distribution of the Monthly Average Values of the Air Temperature above the Territory of Kakheti in 2012-2017

Authors: Khatia Tavidashvili, Nino Jamrishvili, Valerian Omsarashvili

Abstract:

Studies of the vertical distribution of the air temperature in the atmosphere have great value for the solution of different problems of meteorology and climatology (meteorological forecast of showers, thunderstorms, and hail, weather modification, estimation of climate change, etc.). From the end of May 2015 in Kakheti after 25-year interruption, the work of anti-hail service was restored. Therefore, in connection with climate change, the need for the detailed study of the contemporary regime of the vertical distribution of the air temperature above this territory arose. In particular, the indicated information is necessary for the optimum selection of rocket means with the works on the weather modification (fight with the hail, the regulation of atmospheric precipitations, etc.). Construction of the detailed maps of the potential damage distribution of agricultural crops from the hail, etc. taking into account the dimensions of hailstones in the clouds according to the data of radar measurements and height of locality are the most important factors. For now, in Georgia, there is no aerological probing of atmosphere. To solve given problem we processed information about air temperature profiles above Telavi, at 27 km above earth's surface. Information was gathered during four observation time (4, 10, 16, 22 hours with local time. After research, we found vertical distribution of the average monthly values of the air temperature above Kakheti in ‎2012-2017 from January to December. Research was conducted from 0.543 to 27 km above sea level during four periods of research. In particular, it is obtained: -during January the monthly average air temperature linearly diminishes with 2.6 °C on the earth's surface to -57.1 °C at the height of 10 km, then little it changes up to the height of 26 km; the gradient of the air temperature in the layer of the atmosphere from 0.543 to 8 km - 6.3 °C/km; height of zero isotherm - is 1.33 km. -during July the air temperature linearly diminishes with 23.5 °C to -64.7 °C at the height of 17 km, then it grows to -47.5 °C at the height of 27 km; the gradient of the air temperature of - 6.1 °C/km; height of zero isotherm - is 4.39 km, which on 0.16 km is higher than in the sixties of past century.

Keywords: hail, Kakheti, meteorology, vertical distribution of the air temperature

Procedia PDF Downloads 151
18676 Investigation the Photocatalytic Properties of Fe3O4-TiO2 Nanocomposites Prepared by Sonochemical Method

Authors: Zh. Saffari, A. Naeimi, M. S. Ekrami-Kakhki, F. Hamidi

Abstract:

Fe3O4 is one of the important magnetic oxides with spinel structure; it has exhibited unique electric and magnetic properties based on the electron transfer between Fe2+ and Fe3+ in the octahedral sites. Fe3O4 has received considerable attention in various areas such as cancer therapy, drug targeting, enzyme immobilization catalysis, magnetic cell separation, magnetic refrigeration systems and super-paramagnetic materials Fe3O4–TiO2 nanostructures were synthesized by simple, effective and new co-precipitation method assisted by ultrasonic reaction at room temperatures with organic surfactant. The effect of various parameters such as temperature, time, and power on the size and morphology of the product was investigated. Alternating gradient force magnetometer shows that Fe3O4 nanoparticles exhibit super-paramagnetic behaviour at room temperature. For preparation of nanocomposite, 1 g of TiO2 nanostructures were dispersed in 100 mL of ethanol. 0.25 g of Fe(NO3)2 and 2 mL of octanoic acid was added to the solution as a surfactant. Then, NaOH solution (1.5 M) was slowly added into the solution until the pH of the mixture was 7–8. After complete precipitation, the solution placed under the ultrasonic irradiation for 30 min. The product was centrifuged, washed with distilled water and dried in an oven at 100 °C for 3 h. The resulting red powder was calcinated at 800 °C for 3 h to remove any organic residue. The photocatalytic behaviour of Fe3O4–TiO2 nanoparticles was evaluated using the degradation of a Methyl Violet (MV) aqueous solution under ultraviolet light irradiation. As time increased, more and more MV was adsorbed on the nanoparticles catalyst, until the absorption peak vanish. The MV concentration decreased rapidly with increasing UV-irradiation time

Keywords: magnetic, methyl violet, nanocomposite, photocatalytic

Procedia PDF Downloads 234
18675 Neural Network Approach for Solving Integral Equations

Authors: Bhavini Pandya

Abstract:

This paper considers Hη: T2 → T2 the Perturbed Cerbelli-Giona map. That is a family of 2-dimensional nonlinear area-preserving transformations on the torus T2=[0,1]×[0,1]= ℝ2/ ℤ2. A single parameter η varies between 0 and 1, taking the transformation from a hyperbolic toral automorphism to the “Cerbelli-Giona” map, a system known to exhibit multifractal properties. Here we study the multifractal properties of the family of maps. We apply a box-counting method by defining a grid of boxes Bi(δ), where i is the index and δ is the size of the boxes, to quantify the distribution of stable and unstable manifolds of the map. When the parameter is in the range 0.51< η <0.58 and 0.68< η <1 the map is ergodic; i.e., the unstable and stable manifolds eventually cover the whole torus, although not in a uniform distribution. For accurate numerical results we require correspondingly accurate construction of the stable and unstable manifolds. Here we use the piecewise linearity of the map to achieve this, by computing the endpoints of line segments which define the global stable and unstable manifolds. This allows the generalized fractal dimension Dq, and spectrum of dimensions f(α), to be computed with accuracy. Finally, the intersection of the unstable and stable manifold of the map will be investigated, and compared with the distribution of periodic points of the system.

Keywords: feed forward, gradient descent, neural network, integral equation

Procedia PDF Downloads 163
18674 Optimization Principles of Eddy Current Separator for Mixtures with Different Particle Sizes

Authors: Cao Bin, Yuan Yi, Wang Qiang, Amor Abdelkader, Ali Reza Kamali, Diogo Montalvão

Abstract:

The study of the electrodynamic behavior of non-ferrous particles in time-varying magnetic fields is a promising area of research with wide applications, including recycling of non-ferrous metals, mechanical transmission, and space debris. The key technology for recovering non-ferrous metals is eddy current separation (ECS), which utilizes the eddy current force and torque to separate non-ferrous metals. ECS has several advantages, such as low energy consumption, large processing capacity, and no secondary pollution, making it suitable for processing various mixtures like electronic scrap, auto shredder residue, aluminum scrap, and incineration bottom ash. Improving the separation efficiency of mixtures with different particle sizes in ECS can create significant social and economic benefits. Our previous study investigated the influence of particle size on separation efficiency by combining numerical simulations and separation experiments. Pearson correlation analysis found a strong correlation between the eddy current force in simulations and the repulsion distance in experiments, which confirmed the effectiveness of our simulation model. The interaction effects between particle size and material type, rotational speed, and magnetic pole arrangement were examined. It offer valuable insights for the design and optimization of eddy current separators. The underlying mechanism behind the effect of particle size on separation efficiency was discovered by analyzing eddy current and field gradient. The results showed that the magnitude and distribution heterogeneity of eddy current and magnetic field gradient increased with particle size in eddy current separation. Based on this, we further found that increasing the curvature of magnetic field lines within particles could also increase the eddy current force, providing a optimized method to improving the separation efficiency of fine particles. By combining the results of the studies, a more systematic and comprehensive set of optimization guidelines can be proposed for mixtures with different particle size ranges. The separation efficiency of fine particles could be improved by increasing the rotational speed, curvature of magnetic field lines, and electrical conductivity/density of materials, as well as utilizing the eddy current torque. When designing an ECS, the particle size range of the target mixture should be investigated in advance, and the suitable parameters for separating the mixture can be fixed accordingly. In summary, these results can guide the design and optimization of ECS, and also expand the application areas for ECS.

Keywords: eddy current separation, particle size, numerical simulation, metal recovery

Procedia PDF Downloads 56
18673 Stochastic Matrices and Lp Norms for Ill-Conditioned Linear Systems

Authors: Riadh Zorgati, Thomas Triboulet

Abstract:

In quite diverse application areas such as astronomy, medical imaging, geophysics or nondestructive evaluation, many problems related to calibration, fitting or estimation of a large number of input parameters of a model from a small amount of output noisy data, can be cast as inverse problems. Due to noisy data corruption, insufficient data and model errors, most inverse problems are ill-posed in a Hadamard sense, i.e. existence, uniqueness and stability of the solution are not guaranteed. A wide class of inverse problems in physics relates to the Fredholm equation of the first kind. The ill-posedness of such inverse problem results, after discretization, in a very ill-conditioned linear system of equations, the condition number of the associated matrix can typically range from 109 to 1018. This condition number plays the role of an amplifier of uncertainties on data during inversion and then, renders the inverse problem difficult to handle numerically. Similar problems appear in other areas such as numerical optimization when using interior points algorithms for solving linear programs leads to face ill-conditioned systems of linear equations. Devising efficient solution approaches for such system of equations is therefore of great practical interest. Efficient iterative algorithms are proposed for solving a system of linear equations. The approach is based on a preconditioning of the initial matrix of the system with an approximation of a generalized inverse leading to a stochastic preconditioned matrix. This approach, valid for non-negative matrices, is first extended to hermitian, semi-definite positive matrices and then generalized to any complex rectangular matrices. The main results obtained are as follows: 1) We are able to build a generalized inverse of any complex rectangular matrix which satisfies the convergence condition requested in iterative algorithms for solving a system of linear equations. This completes the (short) list of generalized inverse having this property, after Kaczmarz and Cimmino matrices. Theoretical results on both the characterization of the type of generalized inverse obtained and the convergence are derived. 2) Thanks to its properties, this matrix can be efficiently used in different solving schemes as Richardson-Tanabe or preconditioned conjugate gradients. 3) By using Lp norms, we propose generalized Kaczmarz’s type matrices. We also show how Cimmino's matrix can be considered as a particular case consisting in choosing the Euclidian norm in an asymmetrical structure. 4) Regarding numerical results obtained on some pathological well-known test-cases (Hilbert, Nakasaka, …), some of the proposed algorithms are empirically shown to be more efficient on ill-conditioned problems and more robust to error propagation than the known classical techniques we have tested (Gauss, Moore-Penrose inverse, minimum residue, conjugate gradients, Kaczmarz, Cimmino). We end on a very early prospective application of our approach based on stochastic matrices aiming at computing some parameters (such as the extreme values, the mean, the variance, …) of the solution of a linear system prior to its resolution. Such an approach, if it were to be efficient, would be a source of information on the solution of a system of linear equations.

Keywords: conditioning, generalized inverse, linear system, norms, stochastic matrix

Procedia PDF Downloads 113
18672 Production of Pre-Reduction of Iron Ore Nuggets with Lesser Sulphur Intake by Devolatisation of Boiler Grade Coal

Authors: Chanchal Biswas, Anrin Bhattacharyya, Gopes Chandra Das, Mahua Ghosh Chaudhuri, Rajib Dey

Abstract:

Boiler coals with low fixed carbon and higher ash content have always challenged the metallurgists to develop a suitable method for their utilization. In the present study, an attempt is made to establish an energy effective method for the reduction of iron ore fines in the form of nuggets by using ‘Syngas’. By devolatisation (expulsion of volatile matter by applying heat) of boiler coal, gaseous product (enriched with reducing agents like CO, CO2, H2, and CH4 gases) is generated. Iron ore nuggets are reduced by this syngas. For that reason, there is no direct contact between iron ore nuggets and coal ash. It helps to control the minimization of the sulphur intake of the reduced nuggets. A laboratory scale devolatisation furnace designed with reduction facility is evaluated after in-depth studies and exhaustive experimentations including thermo-gravimetric (TG-DTA) analysis to find out the volatile fraction present in boiler grade coal, gas chromatography (GC) to find out syngas composition in different temperature and furnace temperature gradient measurements to minimize the furnace cost by applying one heating coil. The nuggets are reduced in the devolatisation furnace at three different temperatures and three different times. The pre-reduced nuggets are subjected to analytical weight loss calculations to evaluate the extent of reduction. The phase and surface morphology analysis of pre-reduced samples are characterized using X-ray diffractometry (XRD), energy dispersive x-ray spectrometry (EDX), scanning electron microscopy (SEM), carbon sulphur analyzer and chemical analysis method. Degree of metallization of the reduced nuggets is 78.9% by using boiler grade coal. The pre-reduced nuggets with lesser sulphur content could be used in the blast furnace as raw materials or coolant which would reduce the high quality of coke rate of the furnace due to its pre-reduced character. These can be used in Basic Oxygen Furnace (BOF) as coolant also.

Keywords: alternative ironmaking, coal gasification, extent of reduction, nugget making, syngas based DRI, solid state reduction

Procedia PDF Downloads 246