Search results for: homogeneous linear systems
12230 Phase II Monitoring of First-Order Autocorrelated General Linear Profiles
Authors: Yihua Wang, Yunru Lai
Abstract:
Statistical process control has been successfully applied in a variety of industries. In some applications, the quality of a process or product is better characterized and summarized by a functional relationship between a response variable and one or more explanatory variables. A collection of this type of data is called a profile. Profile monitoring is used to understand and check the stability of this relationship or curve over time. The independent assumption for the error term is commonly used in the existing profile monitoring studies. However, in many applications, the profile data show correlations over time. Therefore, we focus on a general linear regression model with a first-order autocorrelation between profiles in this study. We propose an exponentially weighted moving average charting scheme to monitor this type of profile. The simulation study shows that our proposed methods outperform the existing schemes based on the average run length criterion.Keywords: autocorrelation, EWMA control chart, general linear regression model, profile monitoring
Procedia PDF Downloads 46012229 Comparative Study of Numerical and Analytical Buckling Analysis of a Steel Column with Various Slenderness Ratios
Authors: Lahlou Dahmani, Warda Mekiri, Ahmed Boudjemia
Abstract:
This scientific paper explores the comparison between the ultimate buckling load obtained through the Eurocode 3 methodology and the ultimate buckling load obtained through finite element simulations for steel columns under compression. The study aims to provide insights into the adequacy of the design rules proposed in Eurocode 3 for different slenderness ratios. The finite element simulations with the Ansys commercial program involve a geometrical and material non-linear analysis of the columns with imperfections. The loss of equilibrium is generally caused by the geometrically nonlinear effects where the column begins to buckle and lose its stability when the load reaches a certain critical value. The linear buckling analysis predicts the theoretical buckling strength of an elastic structure but the nonlinear one is more accurate with taking into account the initial imperfection.Keywords: Ansys, linear buckling, eigen value, nonlinear buckling, slenderness ratio, Eurocode 3
Procedia PDF Downloads 1912228 Propagation of Weak Non-Linear Waves in Non-Equilibrium Flow
Authors: J. Jena, Monica Saxena
Abstract:
In this paper, the propagation of weak nonlinear waves in non-equilibrium flow has been studied in detail using the perturbation method. The expansive action of receding piston undergoing infinite acceleration has been discussed. Central expansion fan, compression waves and shock fronts have been discussed and the solutions up to the first order in the characteristic plane and physical plane have been obtained.Keywords: Characteristic wave front, weak non-linear waves, central expansion fan, compression waves
Procedia PDF Downloads 36912227 Bulk Amounts of Linear and Cyclic Polypeptides on Our Hand within a Short Time
Abstract:
Polypeptides with defined peptide sequences illustrate the power of remarkable applications in drug delivery, tissue engineering, sensing and catalysis. Especially the cyclic polypeptides, the distinctive topological architecture imparts many characteristic properties comparing to linear polypeptides. Here, a facile and highly efficient strategy for the synthesis of linear and cyclic polypeptides is reported using N-heterocyclic carbenes (NHCs)-mediated ring-opening polymerization (ROP) of α-amino acid N-carboxyanhydrides (NCA) in the presence or absence of primary amine initiator. The polymerization proceeds rapidly in a quasi-living manner, allowing access to linear and cyclic polypeptides of well-defined chain length and narrow polydispersity, as evidenced by nuclear magnetic resonance spectrum (1H NMR and 13C NMR spectra) and size exclusion chromatography (SEC) analysis. The cyclic architecture of the polypeptides was further verified by matrix-assisted laser desorption ionization-time of flight (MALDI-TOF) mass spectra (MALDI-TOF MS) and electrospray ionization (ESI) mass spectra, as well as viscosity studies. This approach can also simplify workup procedures and make bulk scale synthesis possible, which thereby opens avenues for practical uses in diverse areas, opening up the new generation of polypeptide synthesis.Keywords: α-amino acid N-carboxyanhydrides, living polymerization, polypeptides, N-heterocyclic carbenes, ring-opening polymerization
Procedia PDF Downloads 16712226 Effectiveness of Catalysis in Ozonation for the Removal of Herbizide 2,4 Dichlorophenoxyacetic Acid from Contaminated Water
Authors: S. Shanthi
Abstract:
Catalyzed oxidation processes show extraordinary guarantee for application in numerous wastewater treatment ranges. Advanced oxidation processes are emerging innovation that might be utilized for particular objectives in wastewater treatment. This research work provides a solution for removal a refractory organic compound 2,4-dichlorophenoxyaceticacid a common water pollutant. All studies were done in batch mode in a constantly stirred reactor. Alternative ozonation processes catalysed by transition metals or granular activated carbon have been investigated for degradation of organics. Catalytic ozonation under study are homogeneous catalytic ozonation, which is based on ozone activation by transition metal ions present in aqueous solution, and secondly as heterogeneous catalytic ozonation in the presence of Granular Activated Carbon (GAC). The present studies reveal that heterogeneous catalytic ozonation using GAC favour the ozonation of 2,4-dichlorophenoxyaceticacid by increasing the rate of ozonation and a much higher degradation of substrates were obtained in a given time. Be that it may, Fe2+and Fe3+ ions decreased the rate of degradation of 2,4-dichlorophenoxyaceticacid indicating that it acts as a negative catalyst. In case of heterogeneous catalytic ozonation using GAC catalyst it was found that during the initial 5 minutes of contact solution concentration decreased significantly as the pollutants were adsorbed initially. Thereafter the substrate started getting oxidized and ozonation became a dominates the treatment process. The exhausted GAC was found to be regenerated in situ. The percentage reduction of the substrate was maximum achieved in minimum possible time when GAC catalyst is employed.Keywords: ozonation, homogeneous catalysis, heterogeneous catalysis, granular activated carbon
Procedia PDF Downloads 25012225 Displacement Solution for a Static Vertical Rigid Movement of an Interior Circular Disc in a Transversely Isotropic Tri-Material Full-Space
Authors: D. Mehdizadeh, M. Rahimian, M. Eskandari-Ghadi
Abstract:
This article is concerned with the determination of the static interaction of a vertically loaded rigid circular disc embedded at the interface of a horizontal layer sandwiched in between two different transversely isotropic half-spaces called as tri-material full-space. The axes of symmetry of different regions are assumed to be normal to the horizontal interfaces and parallel to the movement direction. With the use of a potential function method, and by implementing Hankel integral transforms in the radial direction, the government partial differential equation for the solely scalar potential function is transformed to an ordinary 4th order differential equation, and the mixed boundary conditions are transformed into a pair of integral equations called dual integral equations, which can be reduced to a Fredholm integral equation of the second kind, which is solved analytically. Then, the displacements and stresses are given in the form of improper line integrals, which is due to inverse Hankel integral transforms. It is shown that the present solutions are in exact agreement with the existing solutions for a homogeneous full-space with transversely isotropic material. To confirm the accuracy of the numerical evaluation of the integrals involved, the numerical results are compared with the solutions exists for the homogeneous full-space. Then, some different cases with different degrees of material anisotropy are compared to portray the effect of degree of anisotropy.Keywords: transversely isotropic, rigid disc, elasticity, dual integral equations, tri-material full-space
Procedia PDF Downloads 44012224 Fault Diagnosis and Fault-Tolerant Control of Bilinear-Systems: Application to Heating, Ventilation, and Air Conditioning Systems in Multi-Zone Buildings
Authors: Abderrhamane Jarou, Dominique Sauter, Christophe Aubrun
Abstract:
Over the past decade, the growing demand for energy efficiency in buildings has attracted the attention of the control community. Failures in HVAC (heating, ventilation and air conditioning) systems in buildings can have a significant impact on the desired and expected energy performance of buildings and on the user's comfort as well. FTC is a recent technology area that studies the adaptation of control algorithms to faulty operating conditions of a system. The application of Fault-Tolerant Control (FTC) in HVAC systems has gained attention in the last two decades. The objective is to maintain the variations in system performance due to faults within an acceptable range with respect to the desired nominal behavior. This paper considers the so-called active approach, which is based on fault and identification scheme combined with a control reconfiguration algorithm that consists in determining a new set of control parameters so that the reconfigured performance is "as close as possible, "in some sense, to the nominal performance. Thermal models of buildings and their HVAC systems are described by non-linear (usually bi-linear) equations. Most of the works carried out so far in FDI (fault diagnosis and isolation) or FTC consider a linearized model of the studied system. However, this model is only valid in a reduced range of variation. This study presents a new fault diagnosis (FD) algorithm based on a bilinear observer for the detection and accurate estimation of the magnitude of the HVAC system failure. The main contribution of the proposed FD algorithm is that instead of using specific linearized models, the algorithm inherits the structure of the actual bilinear model of the building thermal dynamics. As an immediate consequence, the algorithm is applicable to a wide range of unpredictable operating conditions, i.e., weather dynamics, outdoor air temperature, zone occupancy profile. A bilinear fault detection observer is proposed for a bilinear system with unknown inputs. The residual vector in the observer design is decoupled from the unknown inputs and, under certain conditions, is made sensitive to all faults. Sufficient conditions are given for the existence of the observer and results are given for the explicit computation of observer design matrices. Dedicated observer schemes (DOS) are considered for sensor FDI while unknown input bilinear observers are considered for actuator or system components FDI. The proposed strategy for FTC works as follows: At a first level, FDI algorithms are implemented, making it also possible to estimate the magnitude of the fault. Once the fault is detected, the fault estimation is then used to feed the second level and reconfigure the control low so that that expected performances are recovered. This paper is organized as follows. A general structure for fault-tolerant control of buildings is first presented and the building model under consideration is introduced. Then, the observer-based design for Fault Diagnosis of bilinear systems is studied. The FTC approach is developed in Section IV. Finally, a simulation example is given in Section V to illustrate the proposed method.Keywords: bilinear systems, fault diagnosis, fault-tolerant control, multi-zones building
Procedia PDF Downloads 17212223 Considerations for Effectively Using Probability of Failure as a Means of Slope Design Appraisal for Homogeneous and Heterogeneous Rock Masses
Authors: Neil Bar, Andrew Heweston
Abstract:
Probability of failure (PF) often appears alongside factor of safety (FS) in design acceptance criteria for rock slope, underground excavation and open pit mine designs. However, the design acceptance criteria generally provide no guidance relating to how PF should be calculated for homogeneous and heterogeneous rock masses, or what qualifies a ‘reasonable’ PF assessment for a given slope design. Observational and kinematic methods were widely used in the 1990s until advances in computing permitted the routine use of numerical modelling. In the 2000s and early 2010s, PF in numerical models was generally calculated using the point estimate method. More recently, some limit equilibrium analysis software offer statistical parameter inputs along with Monte-Carlo or Latin-Hypercube sampling methods to automatically calculate PF. Factors including rock type and density, weathering and alteration, intact rock strength, rock mass quality and shear strength, the location and orientation of geologic structure, shear strength of geologic structure and groundwater pore pressure influence the stability of rock slopes. Significant engineering and geological judgment, interpretation and data interpolation is usually applied in determining these factors and amalgamating them into a geotechnical model which can then be analysed. Most factors are estimated ‘approximately’ or with allowances for some variability rather than ‘exactly’. When it comes to numerical modelling, some of these factors are then treated deterministically (i.e. as exact values), while others have probabilistic inputs based on the user’s discretion and understanding of the problem being analysed. This paper discusses the importance of understanding the key aspects of slope design for homogeneous and heterogeneous rock masses and how they can be translated into reasonable PF assessments where the data permits. A case study from a large open pit gold mine in a complex geological setting in Western Australia is presented to illustrate how PF can be calculated using different methods and obtain markedly different results. Ultimately sound engineering judgement and logic is often required to decipher the true meaning and significance (if any) of some PF results.Keywords: probability of failure, point estimate method, Monte-Carlo simulations, sensitivity analysis, slope stability
Procedia PDF Downloads 20812222 Studying Second Language Development from a Complex Dynamic Systems Perspective
Authors: L. Freeborn
Abstract:
This paper discusses the application of complex dynamic system theory (DST) to the study of individual differences in second language development. This transdisciplinary framework allows researchers to view the trajectory of language development as a dynamic, non-linear process. A DST approach views language as multi-componential, consisting of multiple complex systems and nested layers. These multiple components and systems continuously interact and influence each other at both the macro- and micro-level. Dynamic systems theory aims to explain and describe the development of the language system, rather than make predictions about its trajectory. Such a holistic and ecological approach to second language development allows researchers to include various research methods from neurological, cognitive, and social perspectives. A DST perspective would involve in-depth analyses as well as mixed methods research. To illustrate, a neurobiological approach to second language development could include non-invasive neuroimaging techniques such as electroencephalography (EEG) and functional magnetic resonance imaging (fMRI) to investigate areas of brain activation during language-related tasks. A cognitive framework would further include behavioural research methods to assess the influence of intelligence and personality traits, as well as individual differences in foreign language aptitude, such as phonetic coding ability and working memory capacity. Exploring second language development from a DST approach would also benefit from including perspectives from the field of applied linguistics, regarding the teaching context, second language input, and the role of affective factors such as motivation. In this way, applying mixed research methods from neurobiological, cognitive, and social approaches would enable researchers to have a more holistic view of the dynamic and complex processes of second language development.Keywords: dynamic systems theory, mixed methods, research design, second language development
Procedia PDF Downloads 13512221 The Combination of the Mel Frequency Cepstral Coefficients (MFCC), Perceptual Linear Prediction (PLP), JITTER and SHIMMER Coefficients for the Improvement of Automatic Recognition System for Dysarthric Speech
Authors: Brahim-Fares Zaidi, Malika Boudraa, Sid-Ahmed Selouani
Abstract:
Our work aims to improve our Automatic Recognition System for Dysarthria Speech (ARSDS) based on the Hidden Models of Markov (HMM) and the Hidden Markov Model Toolkit (HTK) to help people who are sick. With pronunciation problems, we applied two techniques of speech parameterization based on Mel Frequency Cepstral Coefficients (MFCC's) and Perceptual Linear Prediction (PLP's) and concatenated them with JITTER and SHIMMER coefficients in order to increase the recognition rate of a dysarthria speech. For our tests, we used the NEMOURS database that represents speakers with dysarthria and normal speakers.Keywords: hidden Markov model toolkit (HTK), hidden models of Markov (HMM), Mel-frequency cepstral coefficients (MFCC), perceptual linear prediction (PLP’s)
Procedia PDF Downloads 16112220 Sum Capacity with Regularized Channel Inversion in Multi-Antenna Downlink Systems under Equal Power Constraint
Authors: Attaullah Khawaja, Amna Shabbir
Abstract:
Channel inversion is one of the simplest techniques for multiuser downlink systems with single-antenna users. In this paper regularized channel inversion under equal power constraint in the multiuser multiple input multiple output (MU-MIMO) broadcast channels has been considered. Sum capacity with plain channel inversion also known as Zero Forcing Beam Forming (ZFBF) and optimum sum capacity using Dirty Paper Coding (DPC) has also been investigated. Analysis and simulations show that regularization enhances the system performance and empower linear growth in Sum Capacity and specially work well at low signal to noise ratio (SNRs) regime.Keywords: broadcast channel, channel inversion, multiple antenna multiple-user wireless, multiple-input multiple-output (MIMO), regularization, dirty paper coding (DPC), sum capacity
Procedia PDF Downloads 52712219 A Case Comparative Study of Infant Mortality Rate in North-West Nigeria
Authors: G. I. Onwuka, A. Danbaba, S. U. Gulumbe
Abstract:
This study investigated of Infant Mortality Rate as observed at a general hospital in Kaduna-South, Kaduna State, North West Nigeria. The causes of infant Mortality were examined. The data used for this analysis were collected at the statistics unit of the Hospital. The analysis was carried out on the data using Multiple Linear regression Technique and this showed that there is linear relationship between the dependent variable (death) and the independent variables (malaria, measles, anaemia, and coronary heart disease). The resultant model also revealed that a unit increment in each of these diseases would result to a unit increment in death recorded, 98.7% of the total variation in mortality is explained by the given model. The highest number of mortality was recorded in July, 2005 and the lowest mortality recorded in October, 2009.Recommendations were however made based on the results of the study.Keywords: infant mortality rate, multiple linear regression, diseases, serial correlation
Procedia PDF Downloads 33112218 Improving Fault Tolerance and Load Balancing in Heterogeneous Grid Computing Using Fractal Transform
Authors: Saad M. Darwish, Adel A. El-Zoghabi, Moustafa F. Ashry
Abstract:
The popularity of the Internet and the availability of powerful computers and high-speed networks as low-cost commodity components are changing the way we use computers today. These technical opportunities have led to the possibility of using geographically distributed and multi-owner resources to solve large-scale problems in science, engineering, and commerce. Recent research on these topics has led to the emergence of a new paradigm known as Grid computing. To achieve the promising potentials of tremendous distributed resources, effective and efficient load balancing algorithms are fundamentally important. Unfortunately, load balancing algorithms in traditional parallel and distributed systems, which usually run on homogeneous and dedicated resources, cannot work well in the new circumstances. In this paper, the concept of a fast fractal transform in heterogeneous grid computing based on R-tree and the domain-range entropy is proposed to improve fault tolerance and load balancing algorithm by improve connectivity, communication delay, network bandwidth, resource availability, and resource unpredictability. A novel two-dimension figure of merit is suggested to describe the network effects on load balance and fault tolerance estimation. Fault tolerance is enhanced by adaptively decrease replication time and message cost while load balance is enhanced by adaptively decrease mean job response time. Experimental results show that the proposed method yields superior performance over other methods.Keywords: Grid computing, load balancing, fault tolerance, R-tree, heterogeneous systems
Procedia PDF Downloads 49012217 Optimal Tuning of Linear Quadratic Regulator Controller Using a Particle Swarm Optimization for Two-Rotor Aerodynamical System
Authors: Ayad Al-Mahturi, Herman Wahid
Abstract:
This paper presents an optimal state feedback controller based on Linear Quadratic Regulator (LQR) for a two-rotor aero-dynamical system (TRAS). TRAS is a highly nonlinear multi-input multi-output (MIMO) system with two degrees of freedom and cross coupling. There are two parameters that define the behavior of LQR controller: state weighting matrix and control weighting matrix. The two parameters influence the performance of LQR. Particle Swarm Optimization (PSO) is proposed to optimally tune weighting matrices of LQR. The major concern of using LQR controller is to stabilize the TRAS by making the beam move quickly and accurately for tracking a trajectory or to reach a desired altitude. The simulation results were carried out in MATLAB/Simulink. The system is decoupled into two single-input single-output (SISO) systems. Comparing the performance of the optimized proportional, integral and derivative (PID) controller provided by INTECO, results depict that LQR controller gives a better performance in terms of both transient and steady state responses when PSO is performed.Keywords: LQR controller, optimal control, particle swarm optimization (PSO), two rotor aero-dynamical system (TRAS)
Procedia PDF Downloads 32212216 Chassis Level Control Using Proportional Integrated Derivative Control, Fuzzy Logic and Deep Learning
Authors: Atakan Aral Ormancı, Tuğçe Arslantaş, Murat Özcü
Abstract:
This study presents the design and implementation of an experimental chassis-level system for various control applications. Specifically, the height level of the chassis is controlled using proportional integrated derivative, fuzzy logic, and deep learning control methods. Real-time data obtained from height and pressure sensors installed in a 6x2 truck chassis, in combination with pulse-width modulation signal values, are utilized during the tests. A prototype pneumatic system of a 6x2 truck is added to the setup, which enables the Smart Pneumatic Actuators to function as if they were in a real-world setting. To obtain real-time signal data from height sensors, an Arduino Nano is utilized, while a Raspberry Pi processes the data using Matlab/Simulink and provides the correct output signals to control the Smart Pneumatic Actuator in the truck chassis. The objective of this research is to optimize the time it takes for the chassis to level down and up under various loads. To achieve this, proportional integrated derivative control, fuzzy logic control, and deep learning techniques are applied to the system. The results show that the deep learning method is superior in optimizing time for a non-linear system. Fuzzy logic control with a triangular membership function as the rule base achieves better outcomes than proportional integrated derivative control. Traditional proportional integrated derivative control improves the time it takes to level the chassis down and up compared to an uncontrolled system. The findings highlight the superiority of deep learning techniques in optimizing the time for a non-linear system, and the potential of fuzzy logic control. The proposed approach and the experimental results provide a valuable contribution to the field of control, automation, and systems engineering.Keywords: automotive, chassis level control, control systems, pneumatic system control
Procedia PDF Downloads 8112215 Optimization of Structures with Mixed Integer Non-linear Programming (MINLP)
Authors: Stojan Kravanja, Andrej Ivanič, Tomaž Žula
Abstract:
This contribution focuses on structural optimization in civil engineering using mixed integer non-linear programming (MINLP). MINLP is characterized as a versatile method that can handle both continuous and discrete optimization variables simultaneously. Continuous variables are used to optimize parameters such as dimensions, stresses, masses, or costs, while discrete variables represent binary decisions to determine the presence or absence of structural elements within a structure while also calculating discrete materials and standard sections. The optimization process is divided into three main steps. First, a mechanical superstructure with a variety of different topology-, material- and dimensional alternatives. Next, a MINLP model is formulated to encapsulate the optimization problem. Finally, an optimal solution is searched in the direction of the defined objective function while respecting the structural constraints. The economic or mass objective function of the material and labor costs of a structure is subjected to the constraints known from structural analysis. These constraints include equations for the calculation of internal forces and deflections, as well as equations for the dimensioning of structural components (in accordance with the Eurocode standards). Given the complex, non-convex and highly non-linear nature of optimization problems in civil engineering, the Modified Outer-Approximation/Equality-Relaxation (OA/ER) algorithm is applied. This algorithm alternately solves subproblems of non-linear programming (NLP) and main problems of mixed-integer linear programming (MILP), in this way gradually refines the solution space up to the optimal solution. The NLP corresponds to the continuous optimization of parameters (with fixed topology, discrete materials and standard dimensions, all determined in the previous MILP), while the MILP involves a global approximation to the superstructure of alternatives, where a new topology, materials, standard dimensions are determined. The optimization of a convex problem is stopped when the MILP solution becomes better than the best NLP solution. Otherwise, it is terminated when the NLP solution can no longer be improved. While the OA/ER algorithm, like all other algorithms, does not guarantee global optimality due to the presence of non-convex functions, various modifications, including convexity tests, are implemented in OA/ER to mitigate these difficulties. The effectiveness of the proposed MINLP approach is demonstrated by its application to various structural optimization tasks, such as mass optimization of steel buildings, cost optimization of timber halls, composite floor systems, etc. Special optimization models have been developed for the optimization of these structures. The MINLP optimizations, facilitated by the user-friendly software package MIPSYN, provide insights into a mass or cost-optimal solutions, optimal structural topologies, optimal material and standard cross-section choices, confirming MINLP as a valuable method for the optimization of structures in civil engineering.Keywords: MINLP, mixed-integer non-linear programming, optimization, structures
Procedia PDF Downloads 4612214 Bioinspired Green Synthesis of Magnetite Nanoparticles Using Room-Temperature Co-Precipitation: A Study of the Effect of Amine Additives on Particle Morphology in Fluidic Systems
Authors: Laura Norfolk, Georgina Zimbitas, Jan Sefcik, Sarah Staniland
Abstract:
Magnetite nanoparticles (MNP) have been an area of increasing research interest due to their extensive applications in industry, such as in carbon capture, water purification, and crucially, the biomedical industry. The use of MNP in the biomedical industry is rising, with studies on their effect as Magnetic resonance imaging contrast agents, drug delivery systems, and as hyperthermic cancer treatments becoming prevalent in the nanomaterial research community. Particles used for biomedical purposes must meet stringent criteria; the particles must have consistent shape and size between particles. Variation between particle morphology can drastically alter the effective surface area of the material, making it difficult to correctly dose particles that are not homogeneous. Particles of defined shape such as octahedral and cubic have been shown to outperform irregular shaped particles in some applications, leading to the need to synthesize particles of defined shape. In nature, highly homogeneous MNP are found within magnetotactic bacteria, a unique bacteria capable of producing magnetite nanoparticles internally under ambient conditions. Biomineralisation proteins control the properties of the MNPs, enhancing their homogeneity. One of these proteins, Mms6, has been successfully isolated and used in vitro as an additive in room-temperature co-precipitation reactions (RTCP) to produce particles of defined mono-dispersed size & morphology. When considering future industrial scale-up it is crucial to consider the costs and feasibility of an additive, as an additive that is not readily available or easily synthesized at a competitive price will not be sustainable. As such, additives selected for this research are inspired by the functional groups of biomineralisation proteins, but cost-effective, environmentally friendly, and compatible with scale-up. Diethylenetriamine (DETA), triethylenetetramine (TETA), tetraethylenepentamine (TEPA), and pentaethylenehexamine (PEHA) have been successfully used in RTCP to modulate the properties of particles synthesized, leading to the formation of octahedral nanoparticles with no use of organic solvents, heating, or toxic precursors. By extending this principle to a fluidic system, ongoing research will reveal whether the amine additives can also exert morphological control in an environment which is suited toward higher particle yield. Two fluidic systems have been employed; a peristaltic turbulent flow mixing system suitable for the rapid production of MNP, and a macrofluidic system for the synthesis of tailored nanomaterials under a laminar flow regime. The presence of the amine additives in the turbulent flow system in initial results appears to offer similar morphological control as observed under RTCP conditions, with higher proportions of octahedral particles formed. This is a proof of concept which may pave the way to green synthesis of tailored MNP on an industrial scale. Mms6 and amine additives have been used in the macrofluidic system, with Mms6 allowing magnetite to be synthesized at unfavourable ferric ratios, but no longer influencing particle size. This suggests this synthetic technique while still benefiting from the addition of additives, may not allow additives to fully influence the particles formed due to the faster timescale of reaction. The amine additives have been tested at various concentrations, the results of which will be discussed in this paper.Keywords: bioinspired, green synthesis, fluidic, magnetite, morphological control, scale-up
Procedia PDF Downloads 11312213 Method of Successive Approximations for Modeling of Distributed Systems
Authors: A. Torokhti
Abstract:
A new method of mathematical modeling of the distributed nonlinear system is developed. The system is represented by a combination of the set of spatially distributed sensors and the fusion center. Its mathematical model is obtained from the iterative procedure that converges to the model which is optimal in the sense of minimizing an associated cost function.Keywords: mathematical modeling, non-linear system, spatially distributed sensors, fusion center
Procedia PDF Downloads 38112212 Productivity and Structural Design of Manufacturing Systems
Authors: Ryspek Usubamatov, Tan San Chin, Sarken Kapaeva
Abstract:
Productivity of the manufacturing systems depends on technological processes, a technical data of machines and a structure of systems. Technology is presented by the machining mode and data, a technical data presents reliability parameters and auxiliary time for discrete production processes. The term structure of manufacturing systems includes the number of serial and parallel production machines and links between them. Structures of manufacturing systems depend on the complexity of technological processes. Mathematical models of productivity rate for manufacturing systems are important attributes that enable to define best structure by criterion of a productivity rate. These models are important tool in evaluation of the economical efficiency for production systems.Keywords: productivity, structure, manufacturing systems, structural design
Procedia PDF Downloads 58412211 Modified Newton's Iterative Method for Solving System of Nonlinear Equations in Two Variables
Authors: Sara Mahesar, Saleem M. Chandio, Hira Soomro
Abstract:
Nonlinear system of equations in two variables is a system which contains variables of degree greater or equal to two or that comprises of the transcendental functions. Mathematical modeling of numerous physical problems occurs as a system of nonlinear equations. In applied and pure mathematics it is the main dispute to solve a system of nonlinear equations. Numerical techniques mainly used for finding the solution to problems where analytical methods are failed, which leads to the inexact solutions. To find the exact roots or solutions in case of the system of non-linear equations there does not exist any analytical technique. Various methods have been proposed to solve such systems with an improved rate of convergence and accuracy. In this paper, a new scheme is developed for solving system of non-linear equation in two variables. The iterative scheme proposed here is modified form of the conventional Newton’s Method (CN) whose order of convergence is two whereas the order of convergence of the devised technique is three. Furthermore, the detailed error and convergence analysis of the proposed method is also examined. Additionally, various numerical test problems are compared with the results of its counterpart conventional Newton’s Method (CN) which confirms the theoretic consequences of the proposed method.Keywords: conventional Newton’s method, modified Newton’s method, order of convergence, system of nonlinear equations
Procedia PDF Downloads 25712210 Seizure Effects of FP Bearings on the Seismic Reliability of Base-Isolated Systems
Authors: Paolo Castaldo, Bruno Palazzo, Laura Lodato
Abstract:
This study deals with the seizure effects of friction pendulum (FP) bearings on the seismic reliability of a 3D base-isolated nonlinear structural system, designed according to Italian seismic code (NTC08). The isolated system consists in a 3D reinforced concrete superstructure, a r.c. substructure and the FP devices, described by employing a velocity dependent model. The seismic input uncertainty is considered as a random variable relevant to the problem, by employing a set of natural seismic records selected in compliance with L’Aquila (Italy) seismic hazard as provided from NTC08. Several non-linear dynamic analyses considering the three components of each ground motion have been performed with the aim to evaluate the seismic reliability of the superstructure, substructure, and isolation level, also taking into account the seizure event of the isolation devices. Finally, a design solution aimed at increasing the seismic robustness of the base-isolated systems with FPS is analyzed.Keywords: FP devices, seismic reliability, seismic robustness, seizure
Procedia PDF Downloads 41312209 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation
Authors: H. Khanfari, M. Johari Fard
Abstract:
Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.Keywords: carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L)
Procedia PDF Downloads 22012208 The Integration of Geographical Information Systems and Capacitated Vehicle Routing Problem with Simulated Demand for Humanitarian Logistics in Tsunami-Prone Area: A Case Study of Phuket, Thailand
Authors: Kiatkulchai Jitt-Aer, Graham Wall, Dylan Jones
Abstract:
As a result of the Indian Ocean tsunami in 2004, logistics applied to disaster relief operations has received great attention in the humanitarian sector. As learned from such disaster, preparing and responding to the aspect of delivering essential items from distribution centres to affected locations are of the importance for relief operations as the nature of disasters is uncertain especially in suffering figures, which are normally proportional to quantity of supplies. Thus, this study proposes a spatial decision support system (SDSS) for humanitarian logistics by integrating Geographical Information Systems (GIS) and the capacitated vehicle routing problem (CVRP). The GIS is utilised for acquiring demands simulated from the tsunami flooding model of the affected area in the first stage, and visualising the simulation solutions in the last stage. While CVRP in this study encompasses designing the relief routes of a set of homogeneous vehicles from a relief centre to a set of geographically distributed evacuation points in which their demands are estimated by using both simulation and randomisation techniques. The CVRP is modeled as a multi-objective optimization problem where both total travelling distance and total transport resources used are minimized, while demand-cost efficiency of each route is maximized in order to determine route priority. As the model is a NP-hard combinatorial optimization problem, the Clarke and Wright Saving heuristics is proposed to solve the problem for the near-optimal solutions. The real-case instances in the coastal area of Phuket, Thailand are studied to perform the SDSS that allows a decision maker to visually analyse the simulation scenarios through different decision factors.Keywords: demand simulation, humanitarian logistics, geographical information systems, relief operations, capacitated vehicle routing problem
Procedia PDF Downloads 24812207 Classification of ECG Signal Based on Mixture of Linear and Non-Linear Features
Authors: Mohammad Karimi Moridani, Mohammad Abdi Zadeh, Zahra Shahiazar Mazraeh
Abstract:
In recent years, the use of intelligent systems in biomedical engineering has increased dramatically, especially in the diagnosis of various diseases. Also, due to the relatively simple recording of the electrocardiogram signal (ECG), this signal is a good tool to show the function of the heart and diseases associated with it. The aim of this paper is to design an intelligent system for automatically detecting a normal electrocardiogram signal from abnormal one. Using this diagnostic system, it is possible to identify a person's heart condition in a very short time and with high accuracy. The data used in this article are from the Physionet database, available in 2016 for use by researchers to provide the best method for detecting normal signals from abnormalities. Data is of both genders and the data recording time varies between several seconds to several minutes. All data is also labeled normal or abnormal. Due to the low positional accuracy and ECG signal time limit and the similarity of the signal in some diseases with the normal signal, the heart rate variability (HRV) signal was used. Measuring and analyzing the heart rate variability with time to evaluate the activity of the heart and differentiating different types of heart failure from one another is of interest to the experts. In the preprocessing stage, after noise cancelation by the adaptive Kalman filter and extracting the R wave by the Pan and Tampkinz algorithm, R-R intervals were extracted and the HRV signal was generated. In the process of processing this paper, a new idea was presented that, in addition to using the statistical characteristics of the signal to create a return map and extraction of nonlinear characteristics of the HRV signal due to the nonlinear nature of the signal. Finally, the artificial neural networks widely used in the field of ECG signal processing as well as distinctive features were used to classify the normal signals from abnormal ones. To evaluate the efficiency of proposed classifiers in this paper, the area under curve ROC was used. The results of the simulation in the MATLAB environment showed that the AUC of the MLP and SVM neural network was 0.893 and 0.947, respectively. As well as, the results of the proposed algorithm in this paper indicated that the more use of nonlinear characteristics in normal signal classification of the patient showed better performance. Today, research is aimed at quantitatively analyzing the linear and non-linear or descriptive and random nature of the heart rate variability signal, because it has been shown that the amount of these properties can be used to indicate the health status of the individual's heart. The study of nonlinear behavior and dynamics of the heart's neural control system in the short and long-term provides new information on how the cardiovascular system functions, and has led to the development of research in this field. Given that the ECG signal contains important information and is one of the common tools used by physicians to diagnose heart disease, but due to the limited accuracy of time and the fact that some information about this signal is hidden from the viewpoint of physicians, the design of the intelligent system proposed in this paper can help physicians with greater speed and accuracy in the diagnosis of normal and patient individuals and can be used as a complementary system in the treatment centers.Keywords: neart rate variability, signal processing, linear and non-linear features, classification methods, ROC Curve
Procedia PDF Downloads 26212206 Reconstruction and Rejection of External Disturbances in a Dynamical System
Authors: Iftikhar Ahmad, A. Benallegue, A. El Hadri
Abstract:
In this paper, we have proposed an observer for the reconstruction and a control law for the rejection application of unknown bounded external disturbance in a dynamical system. The strategy of both the observer and the controller is designed like a second order sliding mode with a proportional-integral (PI) term. Lyapunov theory is used to prove the exponential convergence and stability. Simulations results are given to show the performance of this method.Keywords: non-linear systems, sliding mode observer, disturbance rejection, nonlinear control
Procedia PDF Downloads 33412205 Targeting Mineral Resources of the Upper Benue trough, Northeastern Nigeria Using Linear Spectral Unmixing
Authors: Bello Yusuf Idi
Abstract:
The Gongola arm of the Upper Banue Trough, Northeastern Nigeria is predominantly covered by the outcrops of Limestone-bearing rocks in form of Sandstone with intercalation of carbonate clay, shale, basaltic, felsphatic and migmatide rocks at subpixel dimension. In this work, subpixel classification algorithm was used to classify the data acquired from landsat 7 Enhance Thematic Mapper (ETM+) satellite system with the aim of producing fractional distribution image for three most economically important solid minerals of the area: Limestone, Basalt and Migmatide. Linear Spectral Unmixing (LSU) algorithm was used to produce fractional distribution image of abundance of the three mineral resources within a 100Km2 portion of the area. The results show that the minerals occur at different proportion all over the area. The fractional map could therefore serve as a guide to the ongoing reconnaissance for the economic potentiality of the formation.Keywords: linear spectral un-mixing, upper benue trough, gongola arm, geological engineering
Procedia PDF Downloads 37412204 Inverse Problem Method for Microwave Intrabody Medical Imaging
Authors: J. Chamorro-Servent, S. Tassani, M. A. Gonzalez-Ballester, L. J. Roca, J. Romeu, O. Camara
Abstract:
Electromagnetic and microwave imaging (MWI) have been used in medical imaging in the last years, being the most common applications of breast cancer and stroke detection or monitoring. In those applications, the subject or zone to observe is surrounded by a number of antennas, and the Nyquist criterium can be satisfied. Additionally, the space between the antennas (transmitting and receiving the electromagnetic fields) and the zone to study can be prepared in a homogeneous scenario. However, this may differ in other cases as could be intracardiac catheters, stomach monitoring devices, pelvic organ systems, liver ablation monitoring devices, or uterine fibroids’ ablation systems. In this work, we analyzed different MWI algorithms to find the most suitable method for dealing with an intrabody scenario. Due to the space limitations usually confronted on those applications, the device would have a cylindrical configuration of a maximum of eight transmitters and eight receiver antennas. This together with the positioning of the supposed device inside a body tract impose additional constraints in order to choose a reconstruction method; for instance, it inhabitants the use of well-known algorithms such as filtered backpropagation for diffraction tomography (due to the unusual configuration with probes enclosed by the imaging region). Finally, the difficulty of simulating a realistic non-homogeneous background inside the body (due to the incomplete knowledge of the dielectric properties of other tissues between the antennas’ position and the zone to observe), also prevents the use of Born and Rytov algorithms due to their limitations with a heterogeneous background. Instead, we decided to use a time-reversed algorithm (mostly used in geophysics) due to its characteristics of ignoring heterogeneities in the background medium, and of focusing its generated field onto the scatters. Therefore, a 2D time-reversed finite difference time domain was developed based on the time-reversed approach for microwave breast cancer detection. Simultaneously an in-silico testbed was also developed to compare ground-truth dielectric properties with corresponding microwave imaging reconstruction. Forward and inverse problems were computed varying: the frequency used related to a small zone to observe (7, 7.5 and 8 GHz); a small polyp diameter (5, 7 and 10 mm); two polyp positions with respect to the closest antenna (aligned or disaligned); and the (transmitters-to-receivers) antenna combination used for the reconstruction (1-1, 8-1, 8-8 or 8-3). Results indicate that when using the existent time-reversed method for breast cancer here for the different combinations of transmitters and receivers, we found false positives due to the high degrees of freedom and unusual configuration (and the possible violation of Nyquist criterium). Those false positives founded in 8-1 and 8-8 combinations, highly reduced with the 1-1 and 8-3 combination, being the 8-3 configuration de most suitable (three neighboring receivers at each time). The 8-3 configuration creates a region-of-interest reduced problem, decreasing the ill-posedness of the inverse problem. To conclude, the proposed algorithm solves the main limitations of the described intrabody application, successfully detecting the angular position of targets inside the body tract.Keywords: FDTD, time-reversed, medical imaging, microwave imaging
Procedia PDF Downloads 12712203 Cooperation of Unmanned Vehicles for Accomplishing Missions
Authors: Ahmet Ozcan, Onder Alparslan, Anil Sezgin, Omer Cetin
Abstract:
The use of unmanned systems for different purposes has become very popular over the past decade. Expectations from these systems have also shown an incredible increase in this parallel. But meeting the demands of the tasks are often not possible with the usage of a single unmanned vehicle in a mission, so it is necessary to use multiple autonomous vehicles with different abilities together in coordination. Therefore the usage of the same type of vehicles together as a swarm is helped especially to satisfy the time constraints of the missions effectively. In other words, it allows sharing the workload by the various numbers of homogenous platforms together. Besides, it is possible to say there are many kinds of problems that require the usage of the different capabilities of the heterogeneous platforms together cooperatively to achieve successful results. In this case, cooperative working brings additional problems beyond the homogeneous clusters. In the scenario presented as an example problem, it is expected that an autonomous ground vehicle, which is lack of its position information, manage to perform point-to-point navigation without losing its way in a previously unknown labyrinth. Furthermore, the ground vehicle is equipped with very limited sensors such as ultrasonic sensors that can detect obstacles. It is very hard to plan or complete the mission for the ground vehicle by self without lost its way in the unknown labyrinth. Thus, in order to assist the ground vehicle, the autonomous air drone is also used to solve the problem cooperatively. The autonomous drone also has limited sensors like downward looking camera and IMU, and it also lacks computing its global position. In this context, it is aimed to solve the problem effectively without taking additional support or input from the outside, just benefiting capabilities of two autonomous vehicles. To manage the point-to-point navigation in a previously unknown labyrinth, the platforms have to work together coordinated. In this paper, cooperative work of heterogeneous unmanned systems is handled in an applied sample scenario, and it is mentioned that how to work together with an autonomous ground vehicle and the autonomous flying platform together in a harmony to take advantage of different platform-specific capabilities. The difficulties of using heterogeneous multiple autonomous platforms in a mission are put forward, and the successful solutions are defined and implemented against the problems like spatially distributed tasks planning, simultaneous coordinated motion, effective communication, and sensor fusion.Keywords: unmanned systems, heterogeneous autonomous vehicles, coordination, task planning
Procedia PDF Downloads 12812202 Optimal Hybrid Linear and Nonlinear Control for a Quadcopter Drone
Authors: Xinhuang Wu, Yousef Sardahi
Abstract:
A hybrid and optimal multi-loop control structure combining linear and nonlinear control algorithms are introduced in this paper to regulate the position of a quadcopter unmanned aerial vehicle (UAV) driven by four brushless DC motors. To this end, a nonlinear mathematical model of the UAV is derived and then linearized around one of its operating points. Using the nonlinear version of the model, a sliding mode control is used to derive the control laws of the motor thrust forces required to drive the UAV to a certain position. The linear model is used to design two controllers, XG-controller and YG-controller, responsible for calculating the required roll and pitch to maneuver the vehicle to the desired X and Y position. Three attitude controllers are designed to calculate the desired angular rates of rotors, assuming that the Euler angles are minimal. After that, a many-objective optimization problem involving 20 design parameters and ten objective functions is formulated and solved by HypE (Hypervolume estimation algorithm), one of the widely used many-objective optimization algorithms approaches. Both stability and performance constraints are imposed on the optimization problem. The optimization results in terms of Pareto sets and fronts are obtained and show that some of the design objectives are competing. That is, when one objective goes down, the other goes up. Also, Numerical simulations conducted on the nonlinear UAV model show that the proposed optimization method is quite effective.Keywords: optimal control, many-objective optimization, sliding mode control, linear control, cascade controllers, UAV, drones
Procedia PDF Downloads 7312201 Capacity of Cold-Formed Steel Warping-Restrained Members Subjected to Combined Axial Compressive Load and Bending
Authors: Maryam Hasanali, Syed Mohammad Mojtabaei, Iman Hajirasouliha, G. Charles Clifton, James B. P. Lim
Abstract:
Cold-formed steel (CFS) elements are increasingly being used as main load-bearing components in the modern construction industry, including low- to mid-rise buildings. In typical multi-storey buildings, CFS structural members act as beam-column elements since they are exposed to combined axial compression and bending actions, both in moment-resisting frames and stud wall systems. Current design specifications, including the American Iron and Steel Institute (AISI S100) and the Australian/New Zealand Standard (AS/NZS 4600), neglect the beneficial effects of warping-restrained boundary conditions in the design of beam-column elements. Furthermore, while a non-linear relationship governs the interaction of axial compression and bending, the combined effect of these actions is taken into account through a simplified linear expression combining pure axial and flexural strengths. This paper aims to evaluate the reliability of the well-known Direct Strength Method (DSM) as well as design proposals found in the literature to provide a better understanding of the efficiency of the code-prescribed linear interaction equation in the strength predictions of CFS beam columns and the effects of warping-restrained boundary conditions on their behavior. To this end, the experimentally validated finite element (FE) models of CFS elements under compression and bending were developed in ABAQUS software, which accounts for both non-linear material properties and geometric imperfections. The validated models were then used for a comprehensive parametric study containing 270 FE models, covering a wide range of key design parameters, such as length (i.e., 0.5, 1.5, and 3 m), thickness (i.e., 1, 2, and 4 mm) and cross-sectional dimensions under ten different load eccentricity levels. The results of this parametric study demonstrated that using the DSM led to the most conservative strength predictions for beam-column members by up to 55%, depending on the element’s length and thickness. This can be sourced by the errors associated with (i) the absence of warping-restrained boundary condition effects, (ii) equations for the calculations of buckling loads, and (iii) the linear interaction equation. While the influence of warping restraint is generally less than 6%, the code suggested interaction equation led to an average error of 4% to 22%, based on the element lengths. This paper highlights the need to provide more reliable design solutions for CFS beam-column elements for practical design purposes.Keywords: beam-columns, cold-formed steel, finite element model, interaction equation, warping-restrained boundary conditions
Procedia PDF Downloads 104