Search results for: time prediction algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20468

Search results for: time prediction algorithms

16208 Localized Detection of ᴅ-Serine by Using an Enzymatic Amperometric Biosensor and Scanning Electrochemical Microscopy

Authors: David Polcari, Samuel C. Perry, Loredano Pollegioni, Matthias Geissler, Janine Mauzeroll

Abstract:

ᴅ-serine acts as an endogenous co-agonist for N-methyl-ᴅ-aspartate receptors in neuronal synapses. This makes it a key component in the development and function of a healthy brain, especially given its role in several neurodegenerative diseases such as Alzheimer’s disease and dementia. Despite such clear research motivations, the primary site and mechanism of ᴅ-serine release is still currently unclear. For this reason, we are developing a biosensor for the detection of ᴅ-serine utilizing a microelectrode in combination with a ᴅ-amino acid oxidase enzyme, which produces stoichiometric quantities of hydrogen peroxide in response to ᴅ-serine. For the fabrication of a biosensor with good selectivity, we use a permselective poly(meta-phenylenediamine) film to ensure only the target molecule is reacted, according to the size exclusion principle. In this work, we investigated the effect of the electrodeposition conditions used on the biosensor’s response time and selectivity. Careful optimization of the fabrication process allowed for enhanced biosensor response time. This allowed for the real time sensing of ᴅ-serine in a bulk solution, and also provided in means to map the efflux of ᴅ-serine in real time. This was done using scanning electrochemical microscopy (SECM) with the optimized biosensor to measure localized release of ᴅ-serine from an agar filled glass capillary sealed in an epoxy puck, which acted as a model system. The SECM area scan simultaneously provided information regarding the rate of ᴅ-serine flux from the model substrate, as well as the size of the substrate itself. This SECM methodology, which provides high spatial and temporal resolution, could be useful to investigate the primary site and mechanism of ᴅ-serine release in other biological samples.

Keywords: ᴅ-serine, enzymatic biosensor, microelectrode, scanning electrochemical microscopy

Procedia PDF Downloads 215
16207 Performance Prediction of a SANDIA 17-m Vertical Axis Wind Turbine Using Improved Double Multiple Streamtube

Authors: Abolfazl Hosseinkhani, Sepehr Sanaye

Abstract:

Different approaches have been used to predict the performance of the vertical axis wind turbines (VAWT), such as experimental, computational fluid dynamics (CFD), and analytical methods. Analytical methods, such as momentum models that use streamtubes, have low computational cost and sufficient accuracy. The double multiple streamtube (DMST) is one of the most commonly used of momentum models, which divide the rotor plane of VAWT into upwind and downwind. In fact, results from the DMST method have shown some discrepancy compared with experiment results; that is because the Darrieus turbine is a complex and aerodynamically unsteady configuration. In this study, analytical-experimental-based corrections, including dynamic stall, streamtube expansion, and finite blade length correction are used to improve the DMST method. Results indicated that using these corrections for a SANDIA 17-m VAWT will lead to improving the results of DMST.

Keywords: vertical axis wind turbine, analytical, double multiple streamtube, streamtube expansion model, dynamic stall model, finite blade length correction

Procedia PDF Downloads 124
16206 An Analysis of Non-Elliptic Curve Based Primality Tests

Authors: William Wong, Zakaria Alomari, Hon Ching Lai, Zhida Li

Abstract:

Modern-day information security depends on implementing Diffie-Hellman, which requires the generation of prime numbers. Because the number of primes is infinite, it is impractical to store prime numbers for use, and therefore, primality tests are indispensable in modern-day information security. A primality test is a test to determine whether a number is prime or composite. There are two types of primality tests, which are deterministic tests and probabilistic tests. Deterministic tests are adopting algorithms that provide a definite answer whether a given number is prime or composite. While in probabilistic tests, a probabilistic result would be provided, there is a degree of uncertainty. In this paper, we review three probabilistic tests: the Fermat Primality Test, the Miller-Rabin Test, and the Baillie-PSW Test, as well as one deterministic test, the Agrawal-Kayal-Saxena (AKS) Test. Furthermore, we do an analysis of these tests. All of the reviews discussed are not based on the Elliptic Curve. The analysis demonstrates that, in the majority of real-world scenarios, the Baillie- PSW test’s favorability stems from its typical operational complexity of O(log 3n) and its capacity to deliver accurate results for numbers below 2^64.

Keywords: primality tests, Fermat’s primality test, Miller-Rabin primality test, Baillie-PSW primality test, AKS primality test

Procedia PDF Downloads 72
16205 An Adaptive Neuro-Fuzzy Inference System (ANFIS) Modelling of Bleeding

Authors: Seyed Abbas Tabatabaei, Fereydoon Moghadas Nejad, Mohammad Saed

Abstract:

The bleeding prediction of the asphalt is one of the most complex subjects in the pavement engineering. In this paper, an Adaptive Neuro Fuzzy Inference System (ANFIS) is used for modeling the effect of important parameters on bleeding is trained and tested with the experimental results. bleeding index based on the asphalt film thickness differential as target parameter,asphalt content, temperature depth of two centemeter, heavy traffic, dust to effective binder, Marshall strength, passing 3/4 sieves, passing 3/8 sieves,passing 3/16 sieves, passing NO8, passing NO50, passing NO100, passing NO200 as input parameters. Then, we randomly divided empirical data into train and test sections in order to accomplish modeling. We instructed ANFIS network by 72 percent of empirical data. 28 percent of primary data which had been considered for testing the approprativity of the modeling were entered into ANFIS model. Results were compared by two statistical criterions (R2, RMSE) with empirical ones. Considering the results, it is obvious that our proposed modeling by ANFIS is efficient and valid and it can also be promoted to more general states.

Keywords: bleeding, asphalt film thickness differential, Anfis Modeling

Procedia PDF Downloads 259
16204 Investing in Shares of Innovative Companies: The Risk and the Return, Evidence from Polish Capital Market

Authors: Tomasz L. Nawrocki

Abstract:

Due to the growing global interest of investment society in innovative enterprises, as the objective of this research was adopted to examine the investment efficiency in shares of companies with innovative characteristics in the risk-return layout. The research was carried out for companies listed on the Warsaw Stock Exchange taking into various consideration time ranges of investment. Obtained results show that in shorter periods of time, investors buy expectations connected with innovative companies and therefore the efficiency of investment in their shares is relatively high, but in the longer term expectations are revised by companies financial results, which in turn negatively affects the efficiency of investment in their shares.

Keywords: capital market, innovative company, investment strategies, risk and return analysis

Procedia PDF Downloads 333
16203 Validation of Modern Work Modules and Their Impact on Sustainable Human Resource Management in the Construction Industry

Authors: Robin Becker, Nane Roetmann, Manfred Helmus

Abstract:

The construction industry faces a significant challenge due to a shortage of skilled work-ers, especially in construction management, despite an increase in graduates. This is main-ly because the job is associated with high stress, long hours, and poor work-life balance. A survey revealed that the profession is unattractive to students, who prioritize personal growth, flexibility, and digitalization in their careers. To address this issue, companies can consider implementing various work modules like "working time documentation," "home office," "job sharing," and "time off." These modules can improve control, work-life bal-ance, and efficiency if tailored to the company's framework. They offer a way to make the field more appealing to future employees while benefiting existing staff, provided that both employers and employees are flexible and considerate of project-specific conditions and teams. The feasibility of these models depends on the company's overall framework, with potential for cost-neutral implementation and positive effects on efficiency and men-tal health. However, their success also relies on the specific context of the company, and more data is needed to assess their full impact.

Keywords: modern construction management, construction industry, work modules, shortage of junior staff, sustainable personnel management, making construction management more attractive, working time model

Procedia PDF Downloads 7
16202 Contagion and Stock Interdependence in the BRIC+M Block

Authors: Christian Bucio Pacheco, Miriam Magnolia Sosa Castro, María Alejandra Cabello Rosales

Abstract:

This paper aims to analyze the contagion effect among the stock markets of the BRIC+M block (Brazil, Russia, India, China plus Mexico). The contagion effect is proved through increasing on dependence parameters during crisis periods. The dependence parameters are estimated through copula approach in a period of time from July 1997 to December 2015. During this period there are instability and calm episodes, allowing to analyze changes in the relations of dependence. Empirical results show strong evidence of time-varying dependence among the BRIC+M markets and an increasing dependence relation during global financial crisis period.

Keywords: BRIC+M Block, Contagion effect, Copula, dependence

Procedia PDF Downloads 332
16201 A Computational Investigation of Knocking Tendency in a Hydrogen-Fueled SI Engine

Authors: Hammam Aljabri, Hong G. Im

Abstract:

Hydrogen is a promising future fuel to support the transition of the energy sector toward carbon neutrality. The direct utilization of H2 in Internal Combustion Engines (ICEs) is possible, and this technology faces mainly two challenges; high NOx emissions and severe knocking at mid to high loads. In this study, we numerically investigated the potential of H2 combustion in a truck-size engine operated in SI mode. To mitigate the knocking nature of H2 combustion, we have focused on studying the effects of three primary parameters; the compression ratio (CR), the air-fuel ratio, and the spark time. The baseline case was set using a CR of 16.5 and an equivalence ratio of 0.35. In simulations, the auto-ignition tendency was evaluated based on the maximum pressure rise rate and the local pressure fluctuations at the monitoring points set along the wall of the combustion chamber. To mitigate the auto-ignition tendency while enabling a wider range of engine operation, the effect of lowering the compression ratio was assessed. The results indicate that by lowering the compression ratio from 16.5:1 to 12.5:1, an indicated thermal efficiency of 47.5% can be achieved. Aiming to restrain the auto-ignition while maintaining good efficiency, a reduction in the equivalence ratio was examined under different compression ratios. The result indicates that higher compression ratios will require lower equivalence ratios, and due to practical limitations, a lower equivalence ratio of 0.25 was set as the limit. Using a compression ratio of 13.5 combined with an equivalence ratio of 0.3 resulted in an indicated thermal efficiency of 48.6%, that is, at a fixed spark time. It is found that under such lean conditions, the incomplete combustion losses and exhaust losses were high. Thus, advancing the spark time was assessed as a possible solution. The results demonstrated the advantages of advancing the spark time, where an indicated thermal efficiency exceeding 50% was achieved using a compression ratio of 14.5:1 and an equivalence ratio of 0.25.

Keywords: hydrogen, combustion, engine knock, SI engine

Procedia PDF Downloads 115
16200 Workflow Based Inspection of Geometrical Adaptability from 3D CAD Models Considering Production Requirements

Authors: Tobias Huwer, Thomas Bobek, Gunter Spöcker

Abstract:

Driving forces for enhancements in production are trends like digitalization and individualized production. Currently, such developments are restricted to assembly parts. Thus, complex freeform surfaces are not addressed in this context. The need for efficient use of resources and near-net-shape production will require individualized production of complex shaped workpieces. Due to variations between nominal model and actual geometry, this can lead to changes in operations in Computer-aided process planning (CAPP) to make CAPP manageable for an adaptive serial production. In this context, 3D CAD data can be a key to realizing that objective. Along with developments in the geometrical adaptation, a preceding inspection method based on CAD data is required to support the process planner by finding objective criteria to make decisions about the adaptive manufacturability of workpieces. Nowadays, this kind of decisions is depending on the experience-based knowledge of humans (e.g. process planners) and results in subjective decisions – leading to a variability of workpiece quality and potential failure in production. In this paper, we present an automatic part inspection method, based on design and measurement data, which evaluates actual geometries of single workpiece preforms. The aim is to automatically determine the suitability of the current shape for further machining, and to provide a basis for an objective decision about subsequent adaptive manufacturability. The proposed method is realized by a workflow-based approach, keeping in mind the requirements of industrial applications. Workflows are a well-known design method of standardized processes. Especially in applications like aerospace industry standardization and certification of processes are an important aspect. Function blocks, providing a standardized, event-driven abstraction to algorithms and data exchange, will be used for modeling and execution of inspection workflows. Each analysis step of the inspection, such as positioning of measurement data or checking of geometrical criteria, will be carried out by function blocks. One advantage of this approach is its flexibility to design workflows and to adapt algorithms specific to the application domain. In general, within the specified tolerance range it will be checked if a geometrical adaption is possible. The development of particular function blocks is predicated on workpiece specific information e.g. design data. Furthermore, for different product lifecycle phases, appropriate logics and decision criteria have to be considered. For example, tolerances for geometric deviations are different in type and size for new-part production compared to repair processes. In addition to function blocks, appropriate referencing systems are important. They need to support exact determination of position and orientation of the actual geometries to provide a basis for precise analysis. The presented approach provides an inspection methodology for adaptive and part-individual process chains. The analysis of each workpiece results in an inspection protocol and an objective decision about further manufacturability. A representative application domain is the product lifecycle of turbine blades containing a new-part production and a maintenance process. In both cases, a geometrical adaptation is required to calculate individual production data. In contrast to existing approaches, the proposed initial inspection method provides information to decide between different potential adaptive machining processes.

Keywords: adaptive, CAx, function blocks, turbomachinery

Procedia PDF Downloads 286
16199 Co-Seismic Surface Deformation Induced By 24 September 2019 Mirpur, Pakistan Earthquake Along an Active Blind Fault Estimated Using Sentinel-1 TOPS Interferometry

Authors: Muhammad Ali, Zeeshan Afzal, Giampaolo Ferraioli, Gilda Schirinzi, Muhammad Saleem Mughal, Vito Pascazio

Abstract:

On 24 September 2019, an earthquake with 5.6 Mw and 10 km depth stroke in Mirpur. The Mirpur area was highly affected by this earthquake, with the death of 34 people. This study aims to estimate the surface deformation associated with this earthquake. The interferometric synthetic aperture radar (InSAR) technique is applied to study earthquake induced surface motion. InSAR data using 9 Sentinel-1A SAR images from 11 August 2019 to 22 October 2019 is used to investigate the pre, co-, and post-seismic deformation trends. Time series investigation reveals that there was not such deformation in pre-seismic time period. In the co-seismic time period, strong displacement was observed, and in post-seismic results, small displacement is seen due to aftershocks. Our results show the existence of a previously unpublished blind fault in Mirpur and help to locate the fault line. Previous this fault line was triggered during the 2005 earthquake, and now it’s activated on 24 September 2019. Study area is already facing many problems due to natural hazards where additional surface deformations, particularly because of an earthquake with an activated blind fault, have increased its vulnerability.

Keywords: surface deformation, InSAR, earthquake, sentinel-1, mirpur

Procedia PDF Downloads 114
16198 Quality Assurance in Cardiac Disorder Detection Images

Authors: Anam Naveed, Asma Andleeb, Mehreen Sirshar

Abstract:

In the article, Image processing techniques have been applied on cardiac images for enhancing the image quality. Two types of methodologies considers for survey, invasive techniques and non-invasive techniques. Different image processes for improvement of cardiac image quality and reduce the amount of radiation exposure for invasive techniques are explored. Different image processing algorithms for enhancing the noninvasive cardiac image qualities are described. Beside these two methodologies, third methodology has applied on live streaming of heart rate on ECG window for extracting necessary information, removing noise and enhancing quality. Sensitivity analyses have been carried out to investigate the impacts of cardiac images for diagnosis of cardiac arteries disease and how the enhancement on images will help the cardiologist to diagnoses disease. The paper evaluates strengths and weaknesses of different techniques applied for improved the image quality and draw a conclusion. Some specific limitations must be considered for whole survey, like the patient heart beat must be 70-75 beats/minute while doing the angiography, similarly patient weight and exposure radiation amount has some limitation.

Keywords: cardiac images, CT angiography, critical analysis, exposure radiation, invasive techniques, invasive techniques, non-invasive techniques

Procedia PDF Downloads 328
16197 Comparison between Pushover Analysis Techniques and Validation of the Simplified Modal Pushover Analysis

Authors: N. F. Hanna, A. M. Haridy

Abstract:

One of the main drawbacks of the Modal Pushover Analysis (MPA) is the need to perform nonlinear time-history analysis, which complicates the analysis method and time. A simplified version of the MPA has been proposed based on the concept of the inelastic deformation ratio. Furthermore, the effect of the higher modes of vibration is considered by assuming linearly-elastic responses, which enables the use of standard elastic response spectrum analysis. In this thesis, the simplified MPA (SMPA) method is applied to determine the target global drift and the inter-story drifts of steel frame building. The effect of the higher vibration modes is considered within the framework of the SMPA. A comprehensive survey about the inelastic deformation ratio is presented. After that, a suitable expression from literature is selected for the inelastic deformation ratio and then implemented in the SMPA. The estimated seismic demands using the SMPA, such as target drift, base shear, and the inter-story drifts, are compared with the seismic responses determined by applying the standard MPA. The accuracy of the estimated seismic demands is validated by comparing with the results obtained by the nonlinear time-history analysis using real earthquake records.

Keywords: modal analysis, pushover analysis, seismic performance, target displacement

Procedia PDF Downloads 348
16196 Database Management System for Orphanages to Help Track of Orphans

Authors: Srivatsav Sanjay Sridhar, Asvitha Raja, Prathit Kalra, Soni Gupta

Abstract:

Database management is a system that keeps track of details about a person in an organisation. Not a lot of orphanages these days are shifting to a computer and program-based system, but unfortunately, most have only pen and paper-based records, which not only consumes space but it is also not eco-friendly. It comes as a hassle when one has to view a record of a person as they have to search through multiple records, and it will consume time. This program will organise all the data and can pull out any information about anyone whose data is entered. This is also a safe way of storage as physical data gets degraded over time or, worse, destroyed due to natural disasters. In this developing world, it is only smart enough to shift all data to an electronic-based storage system. The program comes with all features, including creating, inserting, searching, and deleting the data, as well as printing them.

Keywords: database, orphans, programming, C⁺⁺

Procedia PDF Downloads 129
16195 Offset Dependent Uniform Delay Mathematical Optimization Model for Signalized Traffic Network Using Differential Evolution Algorithm

Authors: Tahseen Saad, Halim Ceylan, Jonathan Weaver, Osman Nuri Çelik, Onur Gungor Sahin

Abstract:

A new concept of uniform delay offset dependent mathematical optimization problem is derived as the main objective for this study using a differential evolution algorithm. To control the coordination problem, which depends on offset selection and to estimate uniform delay based on the offset choice in a traffic signal network. The assumption is the periodic sinusoidal function for arrival and departure patterns. The cycle time is optimized at the entry links and the optimized value is used in the non-entry links as a common cycle time. The offset optimization algorithm is used to calculate the uniform delay at each link. The results are illustrated by using a case study and are compared with the canonical uniform delay model derived by Webster and the highway capacity manual’s model. The findings show new model minimizes the total uniform delay to almost half compared to conventional models. The mathematical objective function is robust. The algorithm convergence time is fast.

Keywords: area traffic control, traffic flow, differential evolution, sinusoidal periodic function, uniform delay, offset variable

Procedia PDF Downloads 263
16194 Implementation of Model Reference Adaptive Control in Tuning of Controller Gains for Following-Vehicle System with Fixed Time Headway

Authors: Fatemeh Behbahani, Rubiyah Yusof

Abstract:

To avoid collision between following vehicles and vehicles in front, it is vital to keep appropriate, safe spacing between both vehicles over all speeds. Therefore, the following vehicle needs to have exact information regarding the speed and spacing between vehicles. This project is conducted to simulate the tuning of controller gain for a vehicle-following system through the selected control strategy, spacing control policy and fixed-time headway policy. In addition, the paper simulates and designs an adaptive gain controller for a road-vehicle-following system which uses information on the spacing, velocity and also acceleration of a preceding vehicle in the proposed one-vehicle look-ahead strategy. The mathematical model is implemented using Kirchhoff and Newton’s Laws, and stability simulated. The trial-error method was used to obtain a suitable value of controller gain. However, the adaptive-based controller system was able to optimize the gain value automatically. Model Reference Adaptive Control (MRAC) is designed and utilized and based on firstly the Gradient and secondly the Lyapunov approach. The Lyapunov approach considers stability. The Gradient approach was found to improve the best value of gain in the controller system with fixed-time headway.

Keywords: one-vehicle look-ahead, model reference adaptive, stability, tuning gain controller, MRAC

Procedia PDF Downloads 222
16193 E-Government Continuance Intention of Media Psychology: Some Insights from Psychographic Characteristics

Authors: Azlina Binti Abu Bakar, Fahmi Zaidi Bin Abdul Razak, Wan Salihin Wong Abdullah

Abstract:

Psychographic is a psychological study of values, attitudes, interests and it is used mostly in prediction, opinion research and social research. This study predicts the influence of performance expectancy, effort expectancy, social influence and facilitating condition on e-government acceptance among Malaysian citizens. The survey responses of 543 e-government users have been validated and analyzed by means of covariance-based Structural Equation Modeling. The findings indicate that e-government acceptance among Malaysian citizens are mainly influenced by performance expectancy (β = 0.66, t = 11.53, p < 0.01) and social influence (β = 0.20, t = 4.23, p < 0.01). Surprisingly, there is no significant effect of facilitating condition and effort expectancy on e-government continuance intention (β = 0.01, t = 0.27, p > 0.05; β = -0.01, t = -0.40, p > 0.05). This study offers government and vendors a frame of reference to analyze citizen’s situation before initiating new innovations. In case of Malaysian e-government technology, adoption strategies should be built around fostering level of citizens’ technological expectation and social influence on e-government usage.

Keywords: continuance intention, Malaysian citizen, media psychology, structural equation modeling

Procedia PDF Downloads 309
16192 Transient Heat Transfer: Experimental Investigation near the Critical Point

Authors: Andreas Kohlhepp, Gerrit Schatte, Wieland Christoph, Spliethoff Hartmut

Abstract:

In recent years the research of heat transfer phenomena of water and other working fluids near the critical point experiences a growing interest for power engineering applications. To match the highly volatile characteristics of renewable energies, conventional power plants need to shift towards flexible operation. This requires speeding up the load change dynamics of steam generators and their heating surfaces near the critical point. In dynamic load transients, both a high heat flux with an unfavorable ratio to the mass flux and a high difference in fluid and wall temperatures, may cause problems. It may lead to deteriorated heat transfer (at supercritical pressures), dry-out or departure from nucleate boiling (at subcritical pressures), all cases leading to an extensive rise of temperatures. For relevant technical applications, the heat transfer coefficients need to be predicted correctly in case of transient scenarios to prevent damage to the heated surfaces (membrane walls, tube bundles or fuel rods). In transient processes, the state of the art method of calculating the heat transfer coefficients is using a multitude of different steady-state correlations for the momentarily existing local parameters for each time step. This approach does not necessarily reflect the different cases that may lead to a significant variation of the heat transfer coefficients and shows gaps in the individual ranges of validity. An algorithm was implemented to calculate the transient behavior of steam generators during load changes. It is used to assess existing correlations for transient heat transfer calculations. It is also desirable to validate the calculation using experimental data. By the use of a new full-scale supercritical thermo-hydraulic test rig, experimental data is obtained to describe the transient phenomena under dynamic boundary conditions as mentioned above and to serve for validation of transient steam generator calculations. Aiming to improve correlations for the prediction of the onset of deteriorated heat transfer in both, stationary and transient cases the test rig was specially designed for this task. It is a closed loop design with a directly electrically heated evaporation tube, the total heating power of the evaporator tube and the preheater is 1MW. To allow a big range of parameters, including supercritical pressures, the maximum pressure rating is 380 bar. The measurements contain the most important extrinsic thermo-hydraulic parameters. Moreover, a high geometric resolution allows to accurately predict the local heat transfer coefficients and fluid enthalpies.

Keywords: departure from nucleate boiling, deteriorated heat transfer, dryout, supercritical working fluid, transient operation of steam generators

Procedia PDF Downloads 210
16191 Numerical Study of Steel Structures Responses to External Explosions

Authors: Mohammad Abdallah

Abstract:

Due to the constant increase in terrorist attacks, the research and engineering communities have given significant attention to building performance under explosions. This paper presents a methodology for studying and simulating the dynamic responses of steel structures during external detonations, particularly for accurately investigating the impact of incrementing charge weight on the members total behavior, resistance and failure. Prediction damage method was introduced to evaluate the damage level of the steel members based on five scenarios of explosions. Johnson–Cook strength and failure model have been used as well as ABAQUS finite element code to simulate the explicit dynamic analysis, and antecedent field tests were used to verify the acceptance and accuracy of the proposed material strength and failure model. Based on the structural response, evaluation criteria such as deflection, vertical displacement, drift index, and damage level; the obtained results show the vulnerability of steel columns and un-braced steel frames which are designed and optimized to carry dead and live load to resist and endure blast loading.

Keywords: steel structure, blast load, terrorist attacks, charge weight, damage level

Procedia PDF Downloads 352
16190 Study on The Model of Microscopic Contact Parameters for Grinding M300 Using Elastic Abrasive Tool

Authors: Wu Xiaojun, Liu Ruiping, Yu Xingzhan, Wu Qian

Abstract:

In precision grinding, utilizing the elastic matrix ball has higher processing efficiency and better superficial quality than traditional grinding. The diversity of characteristics which elastic abrasive tool contact with bend surface results in irregular wear abrasion,and abrasive tool machining status get complicated. There is no theoretical interpretation that parameters affect the grinding accuracy.Aiming at corrosion resistance, wear resistance and other characteristics of M 300 material, it is often used as a material on aerospace precision components. The paper carried out grinding and polishing experiments by using material of M 300,to theoretically show the relationship between stress magnitude and grinding efficiency,and predict the optimal combination of grinding parameter for effective grinding, just for the high abrasion resistance features of M 300, analyzing the micro-contact of elastic ball abrasive tool (Whetstone), using mathematical methods deduce the functional relationship between residual peak removal rate and the main parameters which impact the grinding accuracy on the plane case.Thus laying the foundation for the study of elastic abrasive prediction and compensation.

Keywords: flexible abrasive tool, polishing parameters, Hertz theory, removal rate

Procedia PDF Downloads 527
16189 Digital Control Algorithm Based on Delta-Operator for High-Frequency DC-DC Switching Converters

Authors: Renkai Wang, Tingcun Wei

Abstract:

In this paper, a digital control algorithm based on delta-operator is presented for high-frequency digitally-controlled DC-DC switching converters. The stability and the controlling accuracy of the DC-DC switching converters are improved by using the digital control algorithm based on delta-operator without increasing the hardware circuit scale. The design method of voltage compensator in delta-domain using PID (Proportion-Integration- Differentiation) control is given in this paper, and the simulation results based on Simulink platform are provided, which have verified the theoretical analysis results very well. It can be concluded that, the presented control algorithm based on delta-operator has better stability and controlling accuracy, and easier hardware implementation than the existed control algorithms based on z-operator, therefore it can be used for the voltage compensator design in high-frequency digitally- controlled DC-DC switching converters.

Keywords: digitally-controlled DC-DC switching converter, digital voltage compensator, delta-operator, finite word length, stability

Procedia PDF Downloads 394
16188 Modelling Consistency and Change of Social Attitudes in 7 Years of Longitudinal Data

Authors: Paul Campbell, Nicholas Biddle

Abstract:

There is a complex, endogenous relationship between individual circumstances, attitudes, and behaviour. This study uses longitudinal panel data to assess changes in social and political attitudes over a 7-year period. Attitudes are captured with the question 'what is the most important issue facing Australia today', collected at multiple time points in a longitudinal survey of 2200 Australians. Consistency of attitudes, and factors predicting change over time, are assessed. The consistency of responses has methodological implications for data collection, specifically how often such questions ought to be asked of a population. When change in attitude is observed, this study assesses the extent to which individual demographic characteristics, personality traits, and broader societal events predict change.

Keywords: attitudes, longitudinal survey analysis, personality, social values

Procedia PDF Downloads 115
16187 Equalization Algorithm for the Optical OFDM System Based on the Fractional Fourier Transform

Authors: A. Cherifi, B. Bouazza, A. O. Dahmane, B. Yagoubi

Abstract:

Transmission over Optical channels will introduce inter-symbol interference (ISI) as well as inter-channel (or inter-carrier) interference (ICI). To decrease the effects of ICI, this paper proposes equalizer for the Optical OFDM system based on the fractional Fourier transform (FrFFT). In this FrFT-OFDM system, traditional Fourier transform is replaced by fractional Fourier transform to modulate and demodulate the data symbols. The equalizer proposed consists of sampling the received signal in the different time per time symbol. Theoretical analysis and numerical simulation are discussed.

Keywords: OFDM, (FrFT) fractional fourier transform, optical OFDM, equalization algorithm

Procedia PDF Downloads 418
16186 Non-Parametric Regression over Its Parametric Couterparts with Large Sample Size

Authors: Jude Opara, Esemokumo Perewarebo Akpos

Abstract:

This paper is on non-parametric linear regression over its parametric counterparts with large sample size. Data set on anthropometric measurement of primary school pupils was taken for the analysis. The study used 50 randomly selected pupils for the study. The set of data was subjected to normality test, and it was discovered that the residuals are not normally distributed (i.e. they do not follow a Gaussian distribution) for the commonly used least squares regression method for fitting an equation into a set of (x,y)-data points using the Anderson-Darling technique. The algorithms for the nonparametric Theil’s regression are stated in this paper as well as its parametric OLS counterpart. The use of a programming language software known as “R Development” was used in this paper. From the analysis, the result showed that there exists a significant relationship between the response and the explanatory variable for both the parametric and non-parametric regression. To know the efficiency of one method over the other, the Akaike Information Criterion (AIC), Bayesian Information Criterion (BIC) are used, and it is discovered that the nonparametric regression performs better than its parametric regression counterparts due to their lower values in both the AIC and BIC. The study however recommends that future researchers should study a similar work by examining the presence of outliers in the data set, and probably expunge it if detected and re-analyze to compare results.

Keywords: Theil’s regression, Bayesian information criterion, Akaike information criterion, OLS

Procedia PDF Downloads 290
16185 Clustering of Association Rules of ISIS & Al-Qaeda Based on Similarity Measures

Authors: Tamanna Goyal, Divya Bansal, Sanjeev Sofat

Abstract:

In world-threatening terrorist attacks, where early detection, distinction, and prediction are effective diagnosis techniques and for functionally accurate and precise analysis of terrorism data, there are so many data mining & statistical approaches to assure accuracy. The computational extraction of derived patterns is a non-trivial task which comprises specific domain discovery by means of sophisticated algorithm design and analysis. This paper proposes an approach for similarity extraction by obtaining the useful attributes from the available datasets of terrorist attacks and then applying feature selection technique based on the statistical impurity measures followed by clustering techniques on the basis of similarity measures. On the basis of degree of participation of attributes in the rules, the associative dependencies between the attacks are analyzed. Consequently, to compute the similarity among the discovered rules, we applied a weighted similarity measure. Finally, the rules are grouped by applying using hierarchical clustering. We have applied it to an open source dataset to determine the usability and efficiency of our technique, and a literature search is also accomplished to support the efficiency and accuracy of our results.

Keywords: association rules, clustering, similarity measure, statistical approaches

Procedia PDF Downloads 305
16184 The Relationship between Conceptual Organizational Culture and the Level of Tolerance in Employees

Authors: M. Sadoughi, R. Ehsani

Abstract:

The aim of the present study is examining the relationship between conceptual organizational culture and the level of tolerance in employees of Islamic Azad University of Shahre Ghods. This research is a correlational and analytic-descriptive one. The samples included 144 individuals. A 24-item standard questionnaire of organizational culture by Cameron and Queen was used in this study. This questionnaire has six criteria and each criterion includes four items that each item indicates one cultural dimension. Reliability coefficient of this questionnaire was normed using Cronbach's alpha of 0.91. Also, the 25-item questionnaire of tolerance by Conor and Davidson was used. This questionnaire is in a five-degree Likert scale form. It has seven criteria and is designed to measure the power of coping with pressure and threat. It has the needed content reliability and its reliability coefficient is normed using Cronbach's alpha of 0.87. Data were analyzed using Pearson correlation coefficient and multivariable regression. The results showed among various dimensions of organizational culture, there is a positive significant relationship between three dimensions (family, adhocracy, bureaucracy) and tolerance, there is a negative significant relationship between dimension of market and tolerance and components of organizational culture have the power of prediction and explaining the tolerance. In this explanation, the component of family is the most effective and the best predictor of tolerance.

Keywords: adhocracy, bureaucracy, organizational culture, tolerance

Procedia PDF Downloads 438
16183 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem

Authors: Newroz Nooralddin Abdulrazaq, Thuraya Mahmood Qaradaghi

Abstract:

The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.

Keywords: McEliece cryptosystem, Goppa code, separable, irreducible

Procedia PDF Downloads 256
16182 Anti-Bubble Painting Booth for Wood Coating Resins

Authors: Abasali Masoumi, Amir Gholamian Bozorgi

Abstract:

To have the best quality in wood products such as tabletops and inlay-woods, applying two principles are required: aesthetic and protection against the destructive agent. Artists spent a lot of time creating a masterwork project and also for better demonstrating beautiful appearance and preserving it for hundred years. So they need good material and appropriate method to finish it. As usual, wood painters use polyester or epoxy resins. These finishes need a special skill to use and then give a fantastic paint film and clearness. If we let resins dry in exposure to environmental agents such as unstable temperature, dust and etc., no doubt it becomes cloudy, crack, blister and much wood dust and air bubbles in it. We have designed a special wood coating booth (IR-Patent No: 70429) for wood-coating resins (polyester and epoxy), and this booth provides an adjustable space to control factors that is necessary to have a good finish in the end. Anti-bubble painting booth has the ability to remove bubbles from resin, precludes the cracking process and causes the resin to be the best. With this booth drying time of resin is reduced from 24 hours to 6 hours by fixing the optimum temperature, and it is very good for saving time. This booth is environment-friendly and never lets the poisonous vapors and other VOC (Volatile organic components) enter to workplace atmosphere because they are very harmful to humans.

Keywords: wood coating, epoxy resin, polyester resin, wood finishes

Procedia PDF Downloads 207
16181 Comparative Analysis of Edge Detection Techniques for Extracting Characters

Authors: Rana Gill, Chandandeep Kaur

Abstract:

Segmentation of images can be implemented using different fundamental algorithms like edge detection (discontinuity based segmentation), region growing (similarity based segmentation), iterative thresholding method. A comprehensive literature review relevant to the study gives description of different techniques for vehicle number plate detection and edge detection techniques widely used on different types of images. This research work is based on edge detection techniques and calculating threshold on the basis of five edge operators. Five operators used are Prewitt, Roberts, Sobel, LoG and Canny. Segmentation of characters present in different type of images like vehicle number plate, name plate of house and characters on different sign boards are selected as a case study in this work. The proposed methodology has seven stages. The proposed system has been implemented using MATLAB R2010a. Comparison of all the five operators has been done on the basis of their performance. From the results it is found that Canny operators produce best results among the used operators and performance of different edge operators in decreasing order is: Canny>Log>Sobel>Prewitt>Roberts.

Keywords: segmentation, edge detection, text, extracting characters

Procedia PDF Downloads 414
16180 MICA-TM Peptide Selectively Binds to HLAs Associated with Behçet's Disease

Authors: Sirilak Kongkaew, Pathumwadee Yodmanee, Nopporn Kaiyawet, Arthitaya Meeprasert, Thanyada Rungrotmongkol, Toshikatsu Kaburaki, Hiroshi Noguchi, Fujio Takeuch, Nawee Kungwan, Supot Hannongbua

Abstract:

Behçet’s disease (BD) is a genetic autoimmune expressed by multisystemic inflammatory disorder mostly occurred at the skin, joints, gastrointestinal tract, and genitalia, including ocular, oral, genital, and central nervous systems. Most BD patients in Japan and Korea were strongly indicated by the genetic factor namely HLA-B*51 (especially, HLA-B*51:01) marker in HMC class I, while HLA-A*26:01 allele has been detected from the BD patients in Greek, Japan, and Taiwan. To understand the selective binding of the MICA-TM peptide towards the HLAs associated with BD, the molecular dynamics simulations were applied on the four HLA alleles (B*51:01, B*35:01, A*26:01, and A*11:01) in complex with such peptide. As a result, the key residues in the binding groove of HLA protein which play an important role in the MICA-TM peptide binding and stabilization were revealed. The Van der Waals force was found to be the main protein-protein interaction. Based on the binding free energy prediction by MM/PBSA method, the MICA-TM peptide interacted stronger to the HLA alleles associated to BD in the identical class by 7-12 kcal/mol. The obtained results from the present study could help to differentiate the HLA alleles and explain a source of Behçet’s disease.

Keywords: Behçet’s disease, MD simulations, HMC class I, autoimmune

Procedia PDF Downloads 383
16179 Water Detection in Aerial Images Using Fuzzy Sets

Authors: Caio Marcelo Nunes, Anderson da Silva Soares, Gustavo Teodoro Laureano, Clarimar Jose Coelho

Abstract:

This paper presents a methodology to pixel recognition in aerial images using fuzzy $c$-means algorithm. This algorithm is a alternative to recognize areas considering uncertainties and inaccuracies. Traditional clustering technics are used in recognizing of multispectral images of earth's surface. This technics recognize well-defined borders that can be easily discretized. However, in the real world there are many areas with uncertainties and inaccuracies which can be mapped by clustering algorithms that use fuzzy sets. The methodology presents in this work is applied to multispectral images obtained from Landsat-5/TM satellite. The pixels are joined using the $c$-means algorithm. After, a classification process identify the types of surface according the patterns obtained from spectral response of image surface. The classes considered are, exposed soil, moist soil, vegetation, turbid water and clean water. The results obtained shows that the fuzzy clustering identify the real type of the earth's surface.

Keywords: aerial images, fuzzy clustering, image processing, pattern recognition

Procedia PDF Downloads 457