Search results for: GNSS time series
6939 Vector Control Using Series Iron Loss Model of Induction, Motors and Power Loss Minimization
Authors: Kheldoun Aissa, Khodja Djalal Eddine
Abstract:
The iron loss is a source of detuning in vector controlled induction motor drives if the classical rotor vector controller is used for decoupling. In fact, the field orientation will not be satisfied and the output torque will not truck the reference torque mostly used by Loss Model Controllers (LMCs). In addition, this component of loss, among others, may be excessive if the vector controlled induction motor is driving light loads. In this paper, the series iron loss model is used to develop a vector controller immune to iron loss effect and then an LMC to minimize the total power loss using the torque generated by the speed controller.Keywords: Field Oriented Controller, Induction Motor, Loss ModelController, Series Iron Loss.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27036938 A Ground Observation Based Climatology of Winter Fog: Study over the Indo-Gangetic Plains, India
Authors: Sanjay Kumar Srivastava, Anu Rani Sharma, Kamna Sachdeva
Abstract:
Every year, fog formation over the Indo-Gangetic Plains (IGPs) of Indian region during the winter months of December and January is believed to create numerous hazards, inconvenience, and economic loss to the inhabitants of this densely populated region of Indian subcontinent. The aim of the paper is to analyze the spatial and temporal variability of winter fog over IGPs. Long term ground observations of visibility and other meteorological parameters (1971-2010) have been analyzed to understand the formation of fog phenomena and its relevance during the peak winter months of January and December over IGP of India. In order to examine the temporal variability, time series and trend analysis were carried out by using the Mann-Kendall Statistical test. Trend analysis performed by using the Mann-Kendall test, accepts the alternate hypothesis with 95% confidence level indicating that there exists a trend. Kendall tau’s statistics showed that there exists a positive correlation between time series and fog frequency. Further, the Theil and Sen’s median slope estimate showed that the magnitude of trend is positive. Magnitude is higher during January compared to December for the entire IGP except in December when it is high over the western IGP. Decade wise time series analysis revealed that there has been continuous increase in fog days. The net overall increase of 99 % was observed over IGP in last four decades. Diurnal variability and average daily persistence were computed by using descriptive statistical techniques. Geo-statistical analysis of fog was carried out to understand the spatial variability of fog. Geo-statistical analysis of fog revealed that IGP is a high fog prone zone with fog occurrence frequency of more than 66% days during the study period. Diurnal variability indicates the peak occurrence of fog is between 06:00 and 10:00 local time and average daily fog persistence extends to 5 to 7 hours during the peak winter season. The results would offer a new perspective to take proactive measures in reducing the irreparable damage that could be caused due to changing trends of fog.
Keywords: Fog, climatology, Mann-Kendall test, trend analysis, spatial variability, temporal variability, visibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17556937 Approximated Solutions of Two-Point Nonlinear Boundary Problem by a Combination of Taylor Series Expansion and Newton Raphson Method
Authors: Chinwendu. B. Eleje, Udechukwu P. Egbuhuzor
Abstract:
One of the difficulties encountered in solving nonlinear Boundary Value Problems (BVP) by many researchers is finding approximated solutions with minimum deviations from the exact solutions without so much rigor and complications. In this paper, we propose an approach to solve a two point BVP which involves a combination of Taylor series expansion method and Newton Raphson method. Furthermore, the fourth and sixth order approximated solutions are obtained and we compare their relative error and rate of convergence to the exact solution. Finally, some numerical simulations are presented to show the behavior of the solution and its derivatives.
Keywords: Newton Raphson method, non-linear boundary value problem, Taylor series approximation, Michaelis-Menten equation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3286936 A Quantitative Study of the Evolution of Open Source Software Communities
Authors: M. R. Martinez-Torres, S. L. Toral, M. Olmedilla
Abstract:
Typically, virtual communities exhibit the well-known phenomenon of participation inequality, which means that only a small percentage of users is responsible of the majority of contributions. However, the sustainability of the community requires that the group of active users must be continuously nurtured with new users that gain expertise through a participation process. This paper analyzes the time evolution of Open Source Software (OSS) communities, considering users that join/abandon the community over time and several topological properties of the network when modeled as a social network. More specifically, the paper analyzes the role of those users rejoining the community and their influence in the global characteristics of the network.
Keywords: Open source communities, social network analysis, time series, virtual communities.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20026935 Measurement of I-V Characteristics of a PtSi/p-Si Schottky Barrier Diode at low Temperatures
Authors: Somayeh Gholami, Meysam Khakbaz
Abstract:
The current-voltage characteristics of a PtSi/p-Si Schottky barrier diode was measured at the temperature of 85 K and from the forward bias region of the I-V curve, the electrical parameters of the diode were measured by three methods. The results obtained from the two methods which considered the series resistance were in close agreement with each other and from them barrier height (), ideality factor (n) and series resistance () were found to be 0.2045 eV, 2.877 and 14.556 K respectively. By measuring the I-V characteristics in the temperature range of 85-136 K the electrical parameters were observed to have strong dependency on temperature. The increase of barrier height and decrease of ideality factor with increasing temperature is attributed to the existence of barrier height inhomogeneities in the silicide-semiconductor structure.Keywords: Schottky diode, barrier height, series resistance, I-V, barrier height inhomogeneities.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 83866934 Investigation of SSR Characteristics of SSSC With GA Based Voltage Controller
Authors: R. Thirumalaivasan, M.Janaki, Nagesh Prabhu
Abstract:
In this paper, investigation of subsynchronous resonance (SSR) characteristics of a hybrid series compensated system and the design of voltage controller for three level 24-pulse Voltage Source Converter based Static Synchronous Series Compensator (SSSC) is presented. Hybrid compensation consists of series fixed capacitor and SSSC which is a active series FACTS controller. The design of voltage controller for SSSC is based on damping torque analysis, and Genetic Algorithm (GA) is adopted for tuning the controller parameters. The SSR Characteristics of SSSC with constant reactive voltage control modes has been investigated. The results show that the constant reactive voltage control of SSSC has the effect of reducing the electrical resonance frequency, which detunes the SSR.The analysis of SSR with SSSC is carried out based on frequency domain method, eigenvalue analysis and transient simulation. While the eigenvalue and damping torque analysis are based on D-Q model of SSSC, the transient simulation considers both D-Q and detailed three phase nonlinear system model using switching functions.Keywords: FACTS, SSR, SSSC, damping torque, GA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17346933 Solar Radiation Time Series Prediction
Authors: Cameron Hamilton, Walter Potter, Gerrit Hoogenboom, Ronald McClendon, Will Hobbs
Abstract:
A model was constructed to predict the amount of solar radiation that will make contact with the surface of the earth in a given location an hour into the future. This project was supported by the Southern Company to determine at what specific times during a given day of the year solar panels could be relied upon to produce energy in sufficient quantities. Due to their ability as universal function approximators, an artificial neural network was used to estimate the nonlinear pattern of solar radiation, which utilized measurements of weather conditions collected at the Griffin, Georgia weather station as inputs. A number of network configurations and training strategies were utilized, though a multilayer perceptron with a variety of hidden nodes trained with the resilient propagation algorithm consistently yielded the most accurate predictions. In addition, a modeled direct normal irradiance field and adjacent weather station data were used to bolster prediction accuracy. In later trials, the solar radiation field was preprocessed with a discrete wavelet transform with the aim of removing noise from the measurements. The current model provides predictions of solar radiation with a mean square error of 0.0042, though ongoing efforts are being made to further improve the model’s accuracy.
Keywords: Artificial Neural Networks, Resilient Propagation, Solar Radiation, Time Series Forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27636932 Prediction on Housing Price Based on Deep Learning
Authors: Li Yu, Chenlu Jiao, Hongrun Xin, Yan Wang, Kaiyang Wang
Abstract:
In order to study the impact of various factors on the housing price, we propose to build different prediction models based on deep learning to determine the existing data of the real estate in order to more accurately predict the housing price or its changing trend in the future. Considering that the factors which affect the housing price vary widely, the proposed prediction models include two categories. The first one is based on multiple characteristic factors of the real estate. We built Convolution Neural Network (CNN) prediction model and Long Short-Term Memory (LSTM) neural network prediction model based on deep learning, and logical regression model was implemented to make a comparison between these three models. Another prediction model is time series model. Based on deep learning, we proposed an LSTM-1 model purely regard to time series, then implementing and comparing the LSTM model and the Auto-Regressive and Moving Average (ARMA) model. In this paper, comprehensive study of the second-hand housing price in Beijing has been conducted from three aspects: crawling and analyzing, housing price predicting, and the result comparing. Ultimately the best model program was produced, which is of great significance to evaluation and prediction of the housing price in the real estate industry.
Keywords: Deep learning, convolutional neural network, LSTM, housing prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49906931 Estimation of Time -Varying Linear Regression with Unknown Time -Volatility via Continuous Generalization of the Akaike Information Criterion
Authors: Elena Ezhova, Vadim Mottl, Olga Krasotkina
Abstract:
The problem of estimating time-varying regression is inevitably concerned with the necessity to choose the appropriate level of model volatility - ranging from the full stationarity of instant regression models to their absolute independence of each other. In the stationary case the number of regression coefficients to be estimated equals that of regressors, whereas the absence of any smoothness assumptions augments the dimension of the unknown vector by the factor of the time-series length. The Akaike Information Criterion is a commonly adopted means of adjusting a model to the given data set within a succession of nested parametric model classes, but its crucial restriction is that the classes are rigidly defined by the growing integer-valued dimension of the unknown vector. To make the Kullback information maximization principle underlying the classical AIC applicable to the problem of time-varying regression estimation, we extend it onto a wider class of data models in which the dimension of the parameter is fixed, but the freedom of its values is softly constrained by a family of continuously nested a priori probability distributions.Keywords: Time varying regression, time-volatility of regression coefficients, Akaike Information Criterion (AIC), Kullback information maximization principle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15346930 Modelling Silica Optical Fibre Reliability: A Software Application
Authors: I. Severin, M. Caramihai, R. El Abdi, M. Poulain, A. Avadanii
Abstract:
In order to assess optical fiber reliability in different environmental and stress conditions series of testing are performed simulating overlapping of chemical and mechanical controlled varying factors. Each series of testing may be compared using statistical processing: i.e. Weibull plots. Due to the numerous data to treat, a software application has appeared useful to interpret selected series of experiments in function of envisaged factors. The current paper presents a software application used in the storage, modelling and interpretation of experimental data gathered from optical fibre testing. The present paper strictly deals with the software part of the project (regarding the modelling, storage and processing of user supplied data).
Keywords: Optical fibres, computer aided analysis, data models, data processing, graphical user interfaces.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18236929 Lagrange-s Inversion Theorem and Infiltration
Authors: Pushpa N. Rathie, Prabhata K. Swamee, André L. B. Cavalcante, Luan Carlos de S. M. Ozelim
Abstract:
Implicit equations play a crucial role in Engineering. Based on this importance, several techniques have been applied to solve this particular class of equations. When it comes to practical applications, in general, iterative procedures are taken into account. On the other hand, with the improvement of computers, other numerical methods have been developed to provide a more straightforward methodology of solution. Analytical exact approaches seem to have been continuously neglected due to the difficulty inherent in their application; notwithstanding, they are indispensable to validate numerical routines. Lagrange-s Inversion Theorem is a simple mathematical tool which has proved to be widely applicable to engineering problems. In short, it provides the solution to implicit equations by means of an infinite series. To show the validity of this method, the tree-parameter infiltration equation is, for the first time, analytically and exactly solved. After manipulating these series, closed-form solutions are presented as H-functions.Keywords: Green-Ampt Equation, Lagrange's Inversion Theorem, Talsma-Parlange Equation, Three-Parameter Infiltration Equation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18886928 Accurate Fault Classification and Section Identification Scheme in TCSC Compensated Transmission Line using SVM
Authors: Pushkar Tripathi, Abhishek Sharma, G. N. Pillai, Indira Gupta
Abstract:
This paper presents a new approach for the protection of Thyristor-Controlled Series Compensator (TCSC) line using Support Vector Machine (SVM). One SVM is trained for fault classification and another for section identification. This method use three phase current measurement that results in better speed and accuracy than other SVM based methods which used single phase current measurement. This makes it suitable for real-time protection. The method was tested on 10,000 data instances with a very wide variation in system conditions such as compensation level, source impedance, location of fault, fault inception angle, load angle at source bus and fault resistance. The proposed method requires only local current measurement.Keywords: Fault Classification, Section Identification, Feature Selection, Support Vector Machine (SVM), Thyristor-Controlled Series Compensator (TCSC)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25286927 Numerical Analysis of Fractured Process in Locomotive Steel Wheels
Authors: J. Alizadeh K., R. S. Ashofteh, A. Asadi Lari
Abstract:
Railway vehicle wheels are designed to operate in harsh environments and to withstand high hydrostatic contact pressures. This situation may result in critical circumstances, in particular wheel breakage. This paper presents a time history of a series of broken wheels during a time interval [2007-2008] belongs to locomotive fleet on Iranian Railways. Such fractures in locomotive wheels never reported before. Due to the importance of this issue, a research study has been launched to find the potential reasons of this problem. The authors introduce a FEM model to indicate how and where the wheels could have been affected during their operation. Then, the modeling results are presented and discussed in detail.
Keywords: Crack, fatigue, FE analysis, wheel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23586926 Evaluating the Nexus between Energy Demand and Economic Growth Using the VECM Approach: Case Study of Nigeria, China, and the United States
Authors: Rita U. Onolemhemhen, Saheed L. Bello, Akin P. Iwayemi
Abstract:
The effectiveness of energy demand policy depends on identifying the key drivers of energy demand both in the short-run and the long-run. This paper examines the influence of regional differences on the link between energy demand and other explanatory variables for Nigeria, China and USA using the Vector Error Correction Model (VECM) approach. This study employed annual time series data on energy consumption (ED), real gross domestic product (GDP) per capita (RGDP), real energy prices (P) and urbanization (N) for a thirty-six-year sample period. The utilized time-series data are sourced from World Bank’s World Development Indicators (WDI, 2016) and US Energy Information Administration (EIA). Results from the study, shows that all the independent variables (income, urbanization, and price) substantially affect the long-run energy consumption in Nigeria, USA and China, whereas, income has no significant effect on short-run energy demand in USA and Nigeria. In addition, the long-run effect of urbanization is relatively stronger in China. Urbanization is a key factor in energy demand, it therefore recommended that more attention should be given to the development of rural communities to reduce the inflow of migrants into urban communities which causes the increase in energy demand and energy excesses should be penalized while energy management should be incentivized.Keywords: Economic growth, energy demand, income, real GDP, urbanization, VECM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9896925 Generalized Differential Quadrature Nonlinear Consolidation Analysis of Clay Layer with Time-Varied Drainage Conditions
Authors: A. Bahmanikashkouli, O.R. Bahadori Nezhad
Abstract:
In this article, the phenomenon of nonlinear consolidation in saturated and homogeneous clay layer is studied. Considering time-varied drainage model, the excess pore water pressure in the layer depth is calculated. The Generalized Differential Quadrature (GDQ) method is used for the modeling and numerical analysis. For the purpose of analysis, first the domain of independent variables (i.e., time and clay layer depth) is discretized by the Chebyshev-Gauss-Lobatto series and then the nonlinear system of equations obtained from the GDQ method is solved by means of the Newton-Raphson approach. The obtained results indicate that the Generalized Differential Quadrature method, in addition to being simple to apply, enjoys a very high accuracy in the calculation of excess pore water pressure.Keywords: Generalized Differential Quadrature method, Nonlinear consolidation, Nonlinear system of equations, Time-varied drainage
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20296924 Performance Evaluation and Cost Analysis of Standby Systems
Authors: M. A. Hajeeh
Abstract:
Pumping systems are an integral part of water desalination plants, their effective functioning is vital for the operation of a plant. In this research work, the reliability and availability of pressurized pumps in a reverse osmosis desalination plant are studied with the objective of finding configurations that provides optimal performance. Six configurations of a series system with different number of warm and cold standby components were examined. Closed form expressions for the mean time to failure (MTTF) and the long run availability are derived and compared under the assumption that the time between failures and repair times of the primary and standby components are exponentially distributed. Moreover, a cost/ benefit analysis is conducted in order to identify a configuration with the best performance and least cost. It is concluded that configurations with cold standby components are preferable especially when the pumps are of the size.
Keywords: Availability, Cost/ benefit, Mean time to failure, Pumps.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17266923 Brand Placement Strategies in Turkey: The Case of “Yalan Dünya”
Authors: Burçe Boyraz
Abstract:
This study examines appearances of brand placement as an alternative communication strategy in television series by focusing on Yalan Dünya which is one of the most popular television series in Turkey. Consequently, this study has a descriptive research design and quantitative content analysis method is used in order to analyze frequency and time data of brand placement appearances in first 3 seasons of Yalan Dünya with 16 episodes. Analysis of brand placement practices in Yalan Dünya is dealt in three categories: episode-based analysis, season-based analysis and comparative analysis. At the end, brand placement practices in Yalan Dünya are evaluated in terms of type, form, duration and legal arrangements. As a result of this study, it is seen that brand placement plays a determinant role in Yalan Dünya content. Also, current legal arrangements make brand placement closer to other traditional communication strategies instead of differing brand placement from them distinctly.
Keywords: Advertising, Alternative communication strategy, Brand placement, Yalan Dünya.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43716922 High-Resolution 12-Bit Segmented Capacitor DAC in Successive Approximation ADC
Authors: Wee Leong Son, Hasmayadi Abdul Majid, Rohana Musa
Abstract:
This paper study the segmented split capacitor Digital-to-Analog Converter (DAC) implemented in a differentialtype 12-bit Successive Approximation Analog-to-Digital Converter (SA-ADC). The series capacitance split array method employed as it reduced the total area of the capacitors required for high resolution DACs. A 12-bit regular binary array structure requires 2049 unit capacitors (Cs) while the split array needs 127 unit Cs. These results in the reduction of the total capacitance and power consumption of the series split array architectures as to regular binary-weighted structures. The paper will show the 12-bit DAC series split capacitor with 4-bit thermometer coded DAC architectures as well as the simulation and measured results.Keywords: Successive Approximation Register Analog-to- Digital Converter, SAR ADC, Low voltage ADC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 95626921 Dynamic Time Warping in Gait Classificationof Motion Capture Data
Authors: Adam Świtoński, Agnieszka Michalczuk, Henryk Josiński, Andrzej Polański, KonradWojciechowski
Abstract:
The method of gait identification based on the nearest neighbor classification technique with motion similarity assessment by the dynamic time warping is proposed. The model based kinematic motion data, represented by the joints rotations coded by Euler angles and unit quaternions is used. The different pose distance functions in Euler angles and quaternion spaces are considered. To evaluate individual features of the subsequent joints movements during gait cycle, joint selection is carried out. To examine proposed approach database containing 353 gaits of 25 humans collected in motion capture laboratory is used. The obtained results are promising. The classifications, which takes into consideration all joints has accuracy over 91%. Only analysis of movements of hip joints allows to correctly identify gaits with almost 80% precision.
Keywords: Biometrics, dynamic time warping, gait identification, motion capture, time series classification, quaternion distance functions, attribute ranking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26116920 Performance Evaluation of a Limited Round-Robin System
Authors: Yoshiaki Shikata
Abstract:
Performance of a limited Round-Robin (RR) rule is studied in order to clarify the characteristics of a realistic sharing model of a processor. Under the limited RR rule, the processor allocates to each request a fixed amount of time, called a quantum, in a fixed order. The sum of the requests being allocated these quanta is kept below a fixed value. Arriving requests that cannot be allocated quanta because of such a restriction are queued or rejected. Practical performance measures, such as the relationship between the mean sojourn time, the mean number of requests, or the loss probability and the quantum size are evaluated via simulation. In the evaluation, the requested service time of an arriving request is converted into a quantum number. One of these quanta is included in an RR cycle, which means a series of quanta allocated to each request in a fixed order. The service time of the arriving request can be evaluated using the number of RR cycles required to complete the service, the number of requests receiving service, and the quantum size. Then an increase or decrease in the number of quanta that are necessary before service is completed is reevaluated at the arrival or departure of other requests. Tracking these events and calculations enables us to analyze the performance of our limited RR rule. In particular, we obtain the most suitable quantum size, which minimizes the mean sojourn time, for the case in which the switching time for each quantum is considered.Keywords: Limited RR rule, quantum, processor sharing, sojourn time, performance measures, simulation, loss probability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12466919 Lineup Optimization Model of Basketball Players Based on the Prediction of Recursive Neural Networks
Authors: Wang Yichen, Haruka Yamashita
Abstract:
In recent years, in the field of sports, decision making such as member in the game and strategy of the game based on then analysis of the accumulated sports data are widely attempted. In fact, in the NBA basketball league where the world's highest level players gather, to win the games, teams analyze the data using various statistical techniques. However, it is difficult to analyze the game data for each play such as the ball tracking or motion of the players in the game, because the situation of the game changes rapidly, and the structure of the data should be complicated. Therefore, it is considered that the analysis method for real time game play data is proposed. In this research, we propose an analytical model for "determining the optimal lineup composition" using the real time play data, which is considered to be difficult for all coaches. In this study, because replacing the entire lineup is too complicated, and the actual question for the replacement of players is "whether or not the lineup should be changed", and “whether or not Small Ball lineup is adopted”. Therefore, we propose an analytical model for the optimal player selection problem based on Small Ball lineups. In basketball, we can accumulate scoring data for each play, which indicates a player's contribution to the game, and the scoring data can be considered as a time series data. In order to compare the importance of players in different situations and lineups, we combine RNN (Recurrent Neural Network) model, which can analyze time series data, and NN (Neural Network) model, which can analyze the situation on the field, to build the prediction model of score. This model is capable to identify the current optimal lineup for different situations. In this research, we collected all the data of accumulated data of NBA from 2019-2020. Then we apply the method to the actual basketball play data to verify the reliability of the proposed model.Keywords: Recurrent Neural Network, players lineup, basketball data, decision making model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8346918 Stochastic Comparisons of Heterogeneous Samples with Homogeneous Exponential Samples
Authors: Nitin Gupta, Rakesh Kumar Bajaj
Abstract:
In the present communication, stochastic comparison of a series (parallel) system having heterogeneous components with random lifetimes and series (parallel) system having homogeneous exponential components with random lifetimes has been studied. Further, conditions under which such a comparison is possible has been established.Keywords: Exponential distribution, Order statistics, Star ordering, Stochastic ordering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15656917 In vitro Biological Activity of Some Synthesized Monoazo Heterocycles Based On Thiophene and Thiazolyl-Thiophene Analogues
Authors: M. E. Khalifa, A. A. Gobouri
Abstract:
Potential synthesis of a series of 3-amino-4-arylazothiophene derivatives from reaction of 2-cyano-2-phenylthiocarbamoyl acetamide and the appropriate α-halogenated reagents, followed by coupling with different aryl diazonium salts (Japp-Klingemann reaction), and another series of 5-arylazo-thiazol-2-ylcarbamoyl-thiophene derivatives from base-catalyzed intramolecular condensation of 5-arylazo-2-(N-chloroacetyl)amino-thiazole with selected b-keto compounds (Thorpe-Ziegler reaction) was performed. The biological activity of the two series was studied in vitro. Their versatility for pharmaceutical purposes was reported, where they displayed remarkable activities against selected pathogenic microorganisms; Bacillus subtilis, Staphylococcus aureus (Gram positive bacteria), Escherichia coli, Pseudomonas aeruginosa (Gram negative bacteria), and Aspergillus flavus, Candida albicans (fungi) with various degrees related to their chemical structures.
Keywords: 2-Aminothiazole, antimicrobial, monoazo compounds, thiophene, pathogenic microorganisms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22296916 Numerical Inverse Laplace Transform Using Chebyshev Polynomial
Authors: Vinod Mishra, Dimple Rani
Abstract:
In this paper, numerical approximate Laplace transform inversion algorithm based on Chebyshev polynomial of second kind is developed using odd cosine series. The technique has been tested for three different functions to work efficiently. The illustrations show that the new developed numerical inverse Laplace transform is very much close to the classical analytic inverse Laplace transform.
Keywords: Chebyshev polynomial, Numerical inverse Laplace transform, Odd cosine series.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14036915 Sampling of Variables in Discrete-Event Simulation using the Example of Inventory Evolutions in Job-Shop-Systems Based on Deterministic and Non-Deterministic Data
Authors: Bernd Scholz-Reiter, Christian Toonen, Jan Topi Tervo, Dennis Lappe
Abstract:
Time series analysis often requires data that represents the evolution of an observed variable in equidistant time steps. In order to collect this data sampling is applied. While continuous signals may be sampled, analyzed and reconstructed applying Shannon-s sampling theorem, time-discrete signals have to be dealt with differently. In this article we consider the discrete-event simulation (DES) of job-shop-systems and study the effects of different sampling rates on data quality regarding completeness and accuracy of reconstructed inventory evolutions. At this we discuss deterministic as well as non-deterministic behavior of system variables. Error curves are deployed to illustrate and discuss the sampling rate-s impact and to derive recommendations for its wellfounded choice.Keywords: discrete-event simulation, job-shop-system, sampling rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18286914 Retail Strategy to Reduce Waste Keeping High Profit Utilizing Taylor's Law in Point-of-Sales Data
Authors: Gen Sakoda, Hideki Takayasu, Misako Takayasu
Abstract:
Waste reduction is a fundamental problem for sustainability. Methods for waste reduction with point-of-sales (POS) data are proposed, utilizing the knowledge of a recent econophysics study on a statistical property of POS data. Concretely, the non-stationary time series analysis method based on the Particle Filter is developed, which considers abnormal fluctuation scaling known as Taylor's law. This method is extended for handling incomplete sales data because of stock-outs by introducing maximum likelihood estimation for censored data. The way for optimal stock determination with pricing the cost of waste reduction is also proposed. This study focuses on the examination of the methods for large sales numbers where Taylor's law is obvious. Numerical analysis using aggregated POS data shows the effectiveness of the methods to reduce food waste maintaining a high profit for large sales numbers. Moreover, the way of pricing the cost of waste reduction reveals that a small profit loss realizes substantial waste reduction, especially in the case that the proportionality constant of Taylor’s law is small. Specifically, around 1% profit loss realizes half disposal at =0.12, which is the actual value of processed food items used in this research. The methods provide practical and effective solutions for waste reduction keeping a high profit, especially with large sales numbers.
Keywords: Food waste reduction, particle filter, point of sales, sustainable development goals, Taylor's Law, time series analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8716913 Two-dimensional Differential Transform Method for Solving Linear and Non-linear Goursat Problem
Authors: H. Taghvafard, G. H. Erjaee
Abstract:
A method for solving linear and non-linear Goursat problem is given by using the two-dimensional differential transform method. The approximate solution of this problem is calculated in the form of a series with easily computable terms and also the exact solutions can be achieved by the known forms of the series solutions. The method can easily be applied to many linear and non-linear problems and is capable of reducing the size of computational work. Several examples are given to demonstrate the reliability and the performance of the presented method.Keywords: Quadrature, Spline interpolation, Trapezoidal rule, Numericalintegration, Error analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23856912 Prediction of Scour Profile Caused by Submerged Three-Dimensional Wall Jets
Authors: Abdullah Al Faruque, Ram Balachandar
Abstract:
Series of laboratory tests were carried out to study the extent of scour caused by a three-dimensional wall jets exiting from a square cross-section nozzle and into a non-cohesive sand beds. Previous observations have indicated that the effect of the tail water depth was significant for densimetric Froude number greater than ten. However, the present results indicate that the cut off value could be lower depending on the value of grain size-to-nozzle width ratio. Numbers of equations are drawn out for a better scaling of numerous scour parameters. Also suggested the empirical prediction of scour to predict the scour centre line profile and plan view of scour profile at any particular time.
Keywords: Densimetric Froude Number, Jets, Nozzle, Sand, Scour, Tailwater, Time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19806911 Haar wavelet Method for Solving Initial and Boundary Value Problems of Bratu-type
Authors: S.G.Venkatesh, S.K.Ayyaswamy, G.Hariharan
Abstract:
In this paper, we present a framework to determine Haar solutions of Bratu-type equations that are widely applicable in fuel ignition of the combustion theory and heat transfer. The method is proposed by applying Haar series for the highest derivatives and integrate the series. Several examples are given to confirm the efficiency and the accuracy of the proposed algorithm. The results show that the proposed way is quite reasonable when compared to exact solution.
Keywords: Haar wavelet method, Bratu's problem, boundary value problems, initial value problems, adomain decomposition method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29666910 Numerical Simulation of Different Configurations for a Combined Gasification/Carbonization Reactors
Authors: Mahmoud Amer, Ibrahim El-Sharkawy, Shinichi Ookawara, Ahmed Elwardany
Abstract:
Gasification and carbonization are two of the most common ways for biomass utilization. Both processes are using part of the waste to be accomplished, either by incomplete combustion or for heating for both gasification and carbonization, respectively. The focus of this paper is to minimize the part of the waste that is used for heating biomass for gasification and carbonization. This will occur by combining both gasifiers and carbonization reactors in a single unit to utilize the heat in the product biogas to heating up the wastes in the carbonization reactors. Three different designs are proposed for the combined gasification/carbonization (CGC) reactor. These include a parallel combination of two gasifiers and carbonized syngas, carbonizer and combustion chamber, and one gasifier, carbonizer, and combustion chamber. They are tested numerically using ANSYS Fluent Computational Fluid Dynamics to ensure homogeneity of temperature distribution inside the carbonization part of the CGC reactor. 2D simulations are performed for the three cases after performing both mesh-size and time-step independent solutions. The carbonization part is common among the three different cases, and the difference among them is how this carbonization reactor is heated. The simulation results showed that the first design could provide only partial homogeneous temperature distribution, not across the whole reactor. This means that the produced carbonized biomass will be reduced as it will only fill a specified height of the reactor. To keep the carbonized product production high, a series combination is proposed. This series configuration resulted in a uniform temperature distribution across the whole reactor as it has only one source for heat with no temperature distribution on any surface of the carbonization section. The simulations provided a satisfactory result that either the first parallel combination of gasifier and carbonization reactor could be used with a reduced carbonized amount or a series configuration to keep the production rate high.
Keywords: Numerical simulation, carbonization, gasification, reactor, biomass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 727