Search results for: Regularization Parameter Search.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1846

Search results for: Regularization Parameter Search.

376 Application and Assessment of Artificial Neural Networks for Biodiesel Iodine Value Prediction

Authors: Raquel M. de Sousa, Sofiane Labidi, Allan Kardec D. Barros, Alex O. Barradas Filho, Aldalea L. B. Marques

Abstract:

Several parameters are established in order to measure biodiesel quality. One of them is the iodine value, which is an important parameter that measures the total unsaturation within a mixture of fatty acids. Limitation of unsaturated fatty acids is necessary since warming of higher quantity of these ones ends in either formation of deposits inside the motor or damage of lubricant. Determination of iodine value by official procedure tends to be very laborious, with high costs and toxicity of the reagents, this study uses artificial neural network (ANN) in order to predict the iodine value property as an alternative to these problems. The methodology of development of networks used 13 esters of fatty acids in the input with convergence algorithms of back propagation of back propagation type were optimized in order to get an architecture of prediction of iodine value. This study allowed us to demonstrate the neural networks’ ability to learn the correlation between biodiesel quality properties, in this caseiodine value, and the molecular structures that make it up. The model developed in the study reached a correlation coefficient (R) of 0.99 for both network validation and network simulation, with Levenberg-Maquardt algorithm.

Keywords: Artificial Neural Networks, Biodiesel, Iodine Value, Prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2351
375 Robust FACTS Controller Design Employing Modern Heuristic Optimization Techniques

Authors: A.K.Balirsingh, S.C.Swain, S. Panda

Abstract:

Recently, Genetic Algorithms (GA) and Differential Evolution (DE) algorithm technique have attracted considerable attention among various modern heuristic optimization techniques. Since the two approaches are supposed to find a solution to a given objective function but employ different strategies and computational effort, it is appropriate to compare their performance. This paper presents the application and performance comparison of DE and GA optimization techniques, for flexible ac transmission system (FACTS)-based controller design. The design objective is to enhance the power system stability. The design problem of the FACTS-based controller is formulated as an optimization problem and both the PSO and GA optimization techniques are employed to search for optimal controller parameters. The performance of both optimization techniques has been compared. Further, the optimized controllers are tested on a weekly connected power system subjected to different disturbances, and their performance is compared with the conventional power system stabilizer (CPSS). The eigenvalue analysis and non-linear simulation results are presented and compared to show the effectiveness of both the techniques in designing a FACTS-based controller, to enhance power system stability.

Keywords: Differential Evolution, Flexible AC TransmissionSystems (FACTS), Genetic Algorithm, Low Frequency Oscillations, Single-machine Infinite Bus Power System.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1764
374 Dynamic Traffic Simulation for Traffic Congestion Problem Using an Enhanced Algorithm

Authors: Wong Poh Lee, Mohd. Azam Osman, Abdullah Zawawi Talib, Ahmad Izani Md. Ismail

Abstract:

Traffic congestion has become a major problem in many countries. One of the main causes of traffic congestion is due to road merges. Vehicles tend to move slower when they reach the merging point. In this paper, an enhanced algorithm for traffic simulation based on the fluid-dynamic algorithm and kinematic wave theory is proposed. The enhanced algorithm is used to study traffic congestion at a road merge. This paper also describes the development of a dynamic traffic simulation tool which is used as a scenario planning and to forecast traffic congestion level in a certain time based on defined parameter values. The tool incorporates the enhanced algorithm as well as the two original algorithms. Output from the three above mentioned algorithms are measured in terms of traffic queue length, travel time and the total number of vehicles passing through the merging point. This paper also suggests an efficient way of reducing traffic congestion at a road merge by analyzing the traffic queue length and travel time.

Keywords: Dynamic, fluid-dynamic, kinematic wave theory, simulation, traffic congestion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3118
373 Estimation of Vertical Handover Probability in an Integrated UMTS and WLAN Networks

Authors: Diganta Kumar Pathak, Manashjyoti Bhuyan, Vaskar Deka

Abstract:

Vertical Handover(VHO) among different communication technologies ensuring uninterruption and service continuity is one of the most important performance parameter in Heterogenous networks environment. In an integrated Universal Mobile Telecommunicatin System(UMTS) and Wireless Local Area Network(WLAN), WLAN is given an inherent priority over UMTS because of its high data rates with low cost. Therefore mobile users want to be associated with WLAN maximum of the time while roaming, to enjoy best possible services with low cost. That encourages reduction of number of VHO. In this work the reduction of number of VHO with respect to varying number of WLAN Access Points(APs) in an integrated UMTS and WLAN network is investigated through simulation to provide best possible cost effective service to the users. The simulation has been carried out for an area (7800 × 9006)m2 where COST-231 Hata model and 3GPP (TR 101 112 V 3.1.0) specified models are used for WLAN and UMTS path loss models respectively. The handover decision is triggered based on the received signal level as compared to the fade margin. Fade margin gives a probabilistic measure of the reliability of the communication link. A relationship between number of WLAN APs and the number of VHO is also established in this work.

Keywords: VHO, UMTS, WLAN, MT, AP, BS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2013
372 Computational Analysis of the Scaling Effects on the Performance of an Axial Compressor

Authors: Junting Xiang, Jörg Uwe Schlüter, Fei Duan

Abstract:

The miniaturization of gas turbines promises many advantages. Miniature gas turbines can be used for local power generation or the propulsion of small aircraft, such as UAV and MAV. However, experience shows that the miniaturization of conventional gas turbines, which are optimized at their current large size, leads to a substantial loss of efficiency and performance at smaller scales. This may be due to a number of factors, such as the Reynolds-number effect, the increased heat transfer, and manufacturing tolerances. In the present work, we focus on computational investigations of the Reynolds number effect and the wall heat transfer on the performance of axial compressor during its size change. The NASA stage 35 compressor is selected as the configuration in this study and computational fluid dynamics (CFD) is used to carry out the miniaturization process and simulations. We perform parameter studies on the effect of Reynolds number and wall thermal conditions. Our results indicate a decrease of efficiency, if the compressor is miniaturized based on its original geometry due to the increase of viscous effects. The increased heat transfer through wall has only a small effect and will actually benefit compressor performance based on our study.

Keywords: Axial compressor, CFD, heat transfer, miniature gas turbines, Reynolds number.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3181
371 Load Frequency Control of Nonlinear Interconnected Hydro-Thermal System Using Differential Evolution Technique

Authors: Banaja Mohanty, Prakash Kumar Hota

Abstract:

This paper presents a differential evolution algorithm to design a robust PI and PID controllers for Load Frequency Control (LFC) of nonlinear interconnected power systems considering the boiler dynamics, Governor Dead Band (GDB), Generation Rate Constraint (GRC). Differential evolution algorithm is employed to search for the optimal controller parameters. The proposed method easily copes of with nonlinear constraints. Further the proposed controller is simple, effective and can ensure the desirable overall system performance. The superiority of the proposed approach has been shown by comparing the results with published fuzzy logic controller for the same power systems. The comparison is done using various performance measures like overshoot, settling time and standard error criteria of frequency and tie-line power deviation following a 1% step load perturbation in hydro area. It is noticed that, the dynamic performance of proposed controller is better than fuzzy logic controller. Furthermore, it is also seen that the proposed system is robust and is not affected by change in the system parameters.

Keywords: Automatic Generation control (AGC), Generation Rate Constraint (GRC), Governor Dead Band (GDB), Differential Evolution (DE)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3342
370 An Efficient Approach to Mining Frequent Itemsets on Data Streams

Authors: Sara Ansari, Mohammad Hadi Sadreddini

Abstract:

The increasing importance of data stream arising in a wide range of advanced applications has led to the extensive study of mining frequent patterns. Mining data streams poses many new challenges amongst which are the one-scan nature, the unbounded memory requirement and the high arrival rate of data streams. In this paper, we propose a new approach for mining itemsets on data stream. Our approach SFIDS has been developed based on FIDS algorithm. The main attempts were to keep some advantages of the previous approach and resolve some of its drawbacks, and consequently to improve run time and memory consumption. Our approach has the following advantages: using a data structure similar to lattice for keeping frequent itemsets, separating regions from each other with deleting common nodes that results in a decrease in search space, memory consumption and run time; and Finally, considering CPU constraint, with increasing arrival rate of data that result in overloading system, SFIDS automatically detect this situation and discard some of unprocessing data. We guarantee that error of results is bounded to user pre-specified threshold, based on a probability technique. Final results show that SFIDS algorithm could attain about 50% run time improvement than FIDS approach.

Keywords: Data stream, frequent itemset, stream mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1392
369 Financial Information and Collective Bargaining: Conflicting or Complementing?

Authors: Humayun Murshed, Shibly Abdullah

Abstract:

The research conducted in early seventies apparently assumed the existence of a universal decision model for union negotiators and furthermore tended to regard financial information as a ‘neutral’ input into a rational decision making process. However, research in the eighties began to question the neutrality of financial information as an input in collective bargaining rather viewing it as a potentially effective means for controlling the labour force. Furthermore, this later research also started challenging the simplistic assumptions relating particularly to union objectives which have underpinned the earlier search for universal union decision models. Despite the above developments there seems to be a dearth of studies in developing countries concerning the use of financial information in collective bargaining. This paper seeks to begin to remedy this deficiency. Utilising a case study approach based on two enterprises, one in the public sector and the other a multinational, the universal decision model is rejected and it is argued that the decision whether or not to use financial information is a contingent one and such a contingency is largely defined by the context and environment in which both union and management negotiators work. An attempt is also made to identify the factors constraining as well as promoting the use of financial information in collective bargaining, these being regarded as unique to the organisations within which the case studies are conducted.

Keywords: Collective Bargaining, Developing Countries, Disclosures, Financial Information.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1546
368 An Optimal Algorithm for Finding (r, Q) Policy in a Price-Dependent Order Quantity Inventory System with Soft Budget Constraint

Authors: S. Hamid Mirmohammadi, Shahrazad Tamjidzad

Abstract:

This paper is concerned with the single-item continuous review inventory system in which demand is stochastic and discrete. The budget consumed for purchasing the ordered items is not restricted but it incurs extra cost when exceeding specific value. The unit purchasing price depends on the quantity ordered under the all-units discounts cost structure. In many actual systems, the budget as a resource which is occupied by the purchased items is limited and the system is able to confront the resource shortage by charging more costs. Thus, considering the resource shortage costs as a part of system costs, especially when the amount of resource occupied by the purchased item is influenced by quantity discounts, is well motivated by practical concerns. In this paper, an optimization problem is formulated for finding the optimal (r, Q) policy, when the system is influenced by the budget limitation and a discount pricing simultaneously. Properties of the cost function are investigated and then an algorithm based on a one-dimensional search procedure is proposed for finding an optimal (r, Q) policy which minimizes the expected system costs.

Keywords: (r, Q) policy, Stochastic demand, backorders, limited resource, quantity discounts.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1837
367 Study of Natural Convection in a Triangular Cavity Filled with Water: Application of the Lattice Boltzmann Method

Authors: Imen Mejri, Ahmed Mahmoudi, Mohamed A. Abbassi, Ahmed Omri

Abstract:

The Lattice Boltzmann Method (LBM) with double populations is applied to solve the steady-state laminar natural convective heat transfer in a triangular cavity filled with water. The bottom wall is heated, the vertical wall is cooled, and the inclined wall is kept adiabatic. The buoyancy effect was modeled by applying the Boussinesq approximation to the momentum equation. The fluid velocity is determined by D2Q9 LBM and the energy equation is discritized by D2Q4 LBM to compute the temperature field. Comparisons with previously published work are performed and found to be in excellent agreement. Numerical results are obtained for a wide range of parameters: the Rayleigh number from  to  and the inclination angle from 0° to 360°. Flow and thermal fields were exhibited by means of streamlines and isotherms. It is observed that inclination angle can be used as a relevant parameter to control heat transfer in right-angled triangular enclosures.

 

Keywords: Heat transfer, inclination angle, Lattice Boltzmann Method, Nusselt number, Natural convection, Rayleigh number.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2726
366 Stability Analysis of Three-Dimensional Flow and Heat Transfer over a Permeable Shrinking Surface in a Cu-Water Nanofluid

Authors: Roslinda Nazar, Amin Noor, Khamisah Jafar, Ioan Pop

Abstract:

In this paper, the steady laminar three-dimensional boundary layer flow and heat transfer of a copper (Cu)-water nanofluid in the vicinity of a permeable shrinking flat surface in an otherwise quiescent fluid is studied. The nanofluid mathematical model in which the effect of the nanoparticle volume fraction is taken into account is considered. The governing nonlinear partial differential equations are transformed into a system of nonlinear ordinary differential equations using a similarity transformation which is then solved numerically using the function bvp4c from Matlab. Dual solutions (upper and lower branch solutions) are found for the similarity boundary layer equations for a certain range of the suction parameter. A stability analysis has been performed to show which branch solutions are stable and physically realizable. The numerical results for the skin friction coefficient and the local Nusselt number as well as the velocity and temperature profiles are obtained, presented and discussed in detail for a range of various governing parameters.

Keywords: Heat Transfer, Nanofluid, Shrinking Surface, Stability Analysis, Three-Dimensional Flow.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2170
365 Research on Transformer Condition-based Maintenance System using the Method of Fuzzy Comprehensive Evaluation

Authors: Po-Chun Lin, Jyh-Cherng Gu

Abstract:

This study adopted previous fault patterns, results of detection analysis, historical records and data, and experts- experiences to establish fuzzy principles and estimate the failure probability index of components of a power transformer. Considering that actual parameters and limiting conditions of parameters may differ, this study used the standard data of IEC, IEEE, and CIGRE as condition parameters. According to the characteristics of each condition parameter, relative degradation was introduced to reflect the degree of influence of the factors on the transformer condition. The method of fuzzy mathematics was adopted to determine the subordinate function of the transformer condition. The calculation used the Matlab Fuzzy Tool Box to select the condition parameters of coil winding, iron core, bushing, OLTC, insulating oil and other auxiliary components and factors (e.g., load records, performance history, and maintenance records) of the transformer to establish the fuzzy principles. Examples were presented to support the rationality and effectiveness of the evaluation method of power transformer performance conditions, as based on fuzzy comprehensive evaluation.

Keywords: Fuzzy, relative degradation degree, condition-basedmaintenance, power transformer

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2438
364 An Angioplasty Intervention Simulator with a Specific Virtual Environment

Authors: G. Aloisio, L. T. De Paolis, A. De Mauro, A. Mongelli

Abstract:

One of the essential requirements of a realistic surgical simulator is to reproduce haptic sensations due to the interactions in the virtual environment. However, the interaction need to be performed in real-time, since a delay between the user action and the system reaction reduces the immersion sensation. In this paper, a prototype of a coronary stent implant simulator is present; this system allows real-time interactions with an artery by means of a specific haptic device. To improve the realism of the simulation, the building of the virtual environment is based on real patients- images and a Web Portal is used to search in the geographically remote medical centres a virtual environment with specific features in terms of pathology or anatomy. The functional architecture of the system defines several Medical Centres in which virtual environments built from the real patients- images and related metadata with specific features in terms of pathology or anatomy are stored. The searched data are downloaded from the Medical Centre to the Training Centre provided with a specific haptic device and with the software necessary both to manage the interaction in the virtual environment. After the integration of the virtual environment in the simulation system it is possible to perform training on the specific surgical procedure.

Keywords: Medical Simulation, Web Portal, Virtual Reality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1774
363 Bubble Point Pressures of CO2+Ethyl Palmitate by a Cubic Equation of State and the Wong-Sandler Mixing Rule

Authors: M. A. Sedghamiz, S. Raeissi

Abstract:

This study presents three different approaches to estimate bubble point pressures for the binary system of CO2 and ethyl palmitate fatty acid ethyl ester. The first method involves the Peng-Robinson (PR) Equation of State (EoS) with the conventional mixing rule of Van der Waals. The second approach involves the PR EOS together with the Wong Sandler (WS) mixing rule, coupled with the UNIQUAC GE model. In order to model the bubble point pressures with this approach, the volume and area parameter for ethyl palmitate were estimated by the Hansen group contribution method. The last method involved the Peng-Robinson, combined with the Wong-Sandler method, but using NRTL as the GE model. Results using the Van der Waals mixing rule clearly indicated that this method has the largest errors among all three methods, with errors in the range of 3.96-6.22%. The PR-WS-UNIQUAC method exhibited small errors, with average absolute deviations between 0.95 to 1.97 percent. The PR-WS-NRTL method led to the least errors, where average absolute deviations ranged between 0.65-1.7%.

Keywords: Bubble pressure, Gibbs excess energy model, mixing rule, CO2 solubility, ethyl palmitate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1824
362 Bayesian Network Model for Students- Laboratory Work Performance Assessment: An Empirical Investigation of the Optimal Construction Approach

Authors: Ifeyinwa E. Achumba, Djamel Azzi, Rinat Khusainov

Abstract:

There are three approaches to complete Bayesian Network (BN) model construction: total expert-centred, total datacentred, and semi data-centred. These three approaches constitute the basis of the empirical investigation undertaken and reported in this paper. The objective is to determine, amongst these three approaches, which is the optimal approach for the construction of a BN-based model for the performance assessment of students- laboratory work in a virtual electronic laboratory environment. BN models were constructed using all three approaches, with respect to the focus domain, and compared using a set of optimality criteria. In addition, the impact of the size and source of the training, on the performance of total data-centred and semi data-centred models was investigated. The results of the investigation provide additional insight for BN model constructors and contribute to literature providing supportive evidence for the conceptual feasibility and efficiency of structure and parameter learning from data. In addition, the results highlight other interesting themes.

Keywords: Bayesian networks, model construction, parameterlearning, structure learning, performance index, model comparison.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1701
361 DSLEP (Data Structure Learning Platform to Aid in Higher Education IT Courses)

Authors: Estevan B. Costa, Armando M. Toda, Marcell A. A. Mesquita, Jacques D. Brancher

Abstract:

The advances in technology in the last five years allowed an improvement in the educational area, as the increasing in the development of educational software. One of the techniques that emerged in this lapse is called Gamification, which is the utilization of video game mechanics outside its bounds. Recent studies involving this technique provided positive results in the application of these concepts in many areas as marketing, health and education. In the last area there are studies that covers from elementary to higher education, with many variations to adequate to the educators methodologies. Among higher education, focusing on IT courses, data structures are an important subject taught in many of these courses, as they are base for many systems. Based on the exposed this paper exposes the development of an interactive web learning environment, called DSLEP (Data Structure Learning Platform), to aid students in higher education IT courses. The system includes basic concepts seen on this subject such as stacks, queues, lists, arrays, trees and was implemented to ease the insertion of new structures. It was also implemented with gamification concepts, such as points, levels, and leader boards, to engage students in the search for knowledge and stimulate self-learning.

Keywords: Gamification, Interactive learning environment, Data structures, e-learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2373
360 Extracting Single Trial Visual Evoked Potentials using Selective Eigen-Rate Principal Components

Authors: Samraj Andrews, Ramaswamy Palaniappan, Nidal Kamel

Abstract:

In single trial analysis, when using Principal Component Analysis (PCA) to extract Visual Evoked Potential (VEP) signals, the selection of principal components (PCs) is an important issue. We propose a new method here that selects only the appropriate PCs. We denote the method as selective eigen-rate (SER). In the method, the VEP is reconstructed based on the rate of the eigen-values of the PCs. When this technique is applied on emulated VEP signals added with background electroencephalogram (EEG), with a focus on extracting the evoked P3 parameter, it is found to be feasible. The improvement in signal to noise ratio (SNR) is superior to two other existing methods of PC selection: Kaiser (KSR) and Residual Power (RP). Though another PC selection method, Spectral Power Ratio (SPR) gives a comparable SNR with high noise factors (i.e. EEGs), SER give more impressive results in such cases. Next, we applied SER method to real VEP signals to analyse the P3 responses for matched and non-matched stimuli. The P3 parameters extracted through our proposed SER method showed higher P3 response for matched stimulus, which confirms to the existing neuroscience knowledge. Single trial PCA using KSR and RP methods failed to indicate any difference for the stimuli.

Keywords: Electroencephalogram, P3, Single trial VEP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1613
359 A High-Speed and Low-Energy Ternary Content Addressable Memory Design Using Feedback in Match-Line Sense Amplifier

Authors: Syed Iftekhar Ali, M. S. Islam

Abstract:

In this paper we present an energy efficient match-line (ML) sensing scheme for high-speed ternary content-addressable memory (TCAM). The proposed scheme isolates the sensing unit of the sense amplifier from the large and variable ML capacitance. It employs feedback in the sense amplifier to successfully detect a match while keeping the ML voltage swing low. This reduced voltage swing results in large energy saving. Simulation performed using 130nm 1.2V CMOS logic shows at least 30% total energy saving in our scheme compared to popular current race (CR) scheme for similar search speed. In terms of speed, dynamic energy, peak power consumption and transistor count our scheme also shows better performance than mismatch-dependant (MD) power allocation technique which also employs feedback in the sense amplifier. Additionally, the implementation of our scheme is simpler than CR or MD scheme because of absence of analog control voltage and programmable delay circuit as have been used in those schemes.

Keywords: content-addressable memory, energy consumption, feedback, peak power, sensing scheme, sense amplifier, ternary.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1791
358 Effect of Bentonite on the Rheological Behavior of Cement Grout in Presence of Superplasticizer

Authors: K. Benyounes, A. Benmounah

Abstract:

Cement-based grouts has been used successfully to repair cracks in many concrete structures such as bridges, tunnels, buildings and to consolidate soils or rock foundations. In the present study the rheological characterization of cement grout with water/binder ratio (W/B) is fixed at 0.5. The effect of the replacement of cement by bentonite (2 to 10% wt) in presence of superplasticizer (0.5% wt) was investigated. Several rheological tests were carried out by using controlled-stress rheometer equipped with vane geometry in temperature of 20°C. To highlight the influence of bentonite and superplasticizer on the rheological behavior of grout cement, various flow tests in a range of shear rate from 0 to 200 s-1 were observed. Cement grout showed a non-Newtonian viscosity behavior at all concentrations of bentonite. Three parameter model Herschel- Bulkley was chosen for fitting of experimental data. Based on the values of correlation coefficients of the estimated parameters, The Herschel-Bulkley law model well described the rheological behavior of the grouts. Test results showed that the dosage of bentonite increases the viscosity and yield stress of the system and introduces more thixotropy. While the addition of both bentonite and superplasticizer with cement grout improve significantly the fluidity and reduced the yield stress due to the action of dispersion of SP.

Keywords: Cement grout, bentonite, superplasticizer, viscosity, yield stress.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3535
357 Spike Sorting Method Using Exponential Autoregressive Modeling of Action Potentials

Authors: Sajjad Farashi

Abstract:

Neurons in the nervous system communicate with each other by producing electrical signals called spikes. To investigate the physiological function of nervous system it is essential to study the activity of neurons by detecting and sorting spikes in the recorded signal. In this paper a method is proposed for considering the spike sorting problem which is based on the nonlinear modeling of spikes using exponential autoregressive model. The genetic algorithm is utilized for model parameter estimation. In this regard some selected model coefficients are used as features for sorting purposes. For optimal selection of model coefficients, self-organizing feature map is used. The results show that modeling of spikes with nonlinear autoregressive model outperforms its linear counterpart. Also the extracted features based on the coefficients of exponential autoregressive model are better than wavelet based extracted features and get more compact and well-separated clusters. In the case of spikes different in small-scale structures where principal component analysis fails to get separated clouds in the feature space, the proposed method can obtain well-separated cluster which removes the necessity of applying complex classifiers.

Keywords: Exponential autoregressive model, Neural data, spike sorting, time series modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1745
356 Impact of Process Parameters on Tensile Strength of Fused Deposition Modeling Printed Crisscross Poylactic Acid

Authors: Shilpesh R. Rajpurohit, Harshit K. Dave

Abstract:

Additive manufacturing gains the popularity in recent times, due to its capability to create prototype as well functional as end use product directly from CAD data without any specific requirement of tooling. Fused deposition modeling (FDM) is one of the widely used additive manufacturing techniques that are used to create functional end use part of polymer that is comparable with the injection-molded parts. FDM printed part has an application in various fields such as automobile, aerospace, medical, electronic, etc. However, application of FDM part is greatly affected by poor mechanical properties. Proper selection of the process parameter could enhance the mechanical performance of the printed part. In the present study, experimental investigation has been carried out to study the behavior of the mechanical performance of the printed part with respect to process variables. Three process variables viz. raster angle, raster width and layer height have been varied to understand its effect on tensile strength. Further, effect of process variables on fractured surface has been also investigated.

Keywords: 3D printing, fused deposition modeling, layer height, raster angle, raster width, tensile strength.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1632
355 In Search of an SVD and QRcp Based Optimization Technique of ANN for Automatic Classification of Abnormal Heart Sounds

Authors: Samit Ari, Goutam Saha

Abstract:

Artificial Neural Network (ANN) has been extensively used for classification of heart sounds for its discriminative training ability and easy implementation. However, it suffers from overparameterization if the number of nodes is not chosen properly. In such cases, when the dataset has redundancy within it, ANN is trained along with this redundant information that results in poor validation. Also a larger network means more computational expense resulting more hardware and time related cost. Therefore, an optimum design of neural network is needed towards real-time detection of pathological patterns, if any from heart sound signal. The aims of this work are to (i) select a set of input features that are effective for identification of heart sound signals and (ii) make certain optimum selection of nodes in the hidden layer for a more effective ANN structure. Here, we present an optimization technique that involves Singular Value Decomposition (SVD) and QR factorization with column pivoting (QRcp) methodology to optimize empirically chosen over-parameterized ANN structure. Input nodes present in ANN structure is optimized by SVD followed by QRcp while only SVD is required to prune undesirable hidden nodes. The result is presented for classifying 12 common pathological cases and normal heart sound.

Keywords: ANN, Classification of heart diseases, murmurs, optimization, Phonocardiogram, QRcp, SVD.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2039
354 The Use Management of the Knowledge Management and the Information Technologies in the Competitive Strategy of a Self-Propelling Industry

Authors: Guerrero Ramírez Sandra, Ramos Salinas Norma Maricela, Muriel Amezcua Vanesa

Abstract:

This article presents the beginning of a wider study that intends to demonstrate how within organizations of the automotive industry from the city of Querétaro. Knowledge management and technological management are required, as well as people’s initiative and the interaction embedded at the interior of it, with the appropriate environment that facilitates information conversion with wide information technologies management (ITM) range. A company was identified for the pilot study of this research, where descriptive and inferential research information was obtained. The results of the pilot suggest that some respondents did noted entity the knowledge management topic, even if staffs have access to information technology (IT) that serve to enhance access to knowledge (through internet, email, databases, external and internal company personnel, suppliers, customers and competitors) data, this implicates that there are Knowledge Management (KM) problems. The data shows that academically well-prepared organizations normally do not recognize the importance of knowledge in the business, nor in the implementation of it, which at the end is a great influence on how to manage it, so that it should guide the company to greater in sight towards a competitive strategy search, given that the company has an excellent technological infrastructure and KM was not exploited. Cultural diversity is another factor that was observed by the staff.

Keywords: Knowledge management, technological knowledge management, technology information management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 904
353 An Evaluation of Average Run Length of MaxEWMA and MaxGWMA Control Charts

Authors: S. Phanyaem

Abstract:

Exponentially weighted moving average control chart (EWMA) is a popular chart used for detecting shift in the mean of parameter of distributions in quality control. The objective of this paper is to compare the efficiency of control chart to detect an increases in the mean of a process. In particular, we compared the Maximum Exponentially Weighted Moving Average (MaxEWMA) and Maximum Generally Weighted Moving Average (MaxGWMA) control charts when the observations are Exponential distribution. The criteria for evaluate the performance of control chart is called, the Average Run Length (ARL). The result of comparison show that in the case of process is small sample size, the MaxEWMA control chart is more efficiency to detect shift in the process mean than MaxGWMA control chart. For the case of large sample size, the MaxEWMA control chart is more sensitive to detect small shift in the process mean than MaxGWMA control chart, and when the process is a large shift in mean, the MaxGWMA control chart is more sensitive to detect mean shift than MaxEWMA control chart.

Keywords: Maximum Exponentially Weighted Moving Average, Maximum General Weighted Moving Average, Average Run Length.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2143
352 On the Analysis and a Few Optimization Issues of a New iCIM 3000 System at an Academic-Research Oriented Institution

Authors: D. R. Delgado Sobrino, R. Holubek, R. Ružarovský

Abstract:

In the past years, the world has witnessed significant work in the field of Manufacturing. Special efforts have been made in the implementation of new technologies, management and control systems, among many others which have all evolved the field. Closely following all this, due to the scope of new projects and the need of turning the existing flexible ideas into more autonomous and intelligent ones, i.e.: moving toward a more intelligent manufacturing, the present paper emerges with the main aim of contributing to the analysis and a few customization issues of a new iCIM 3000 system at the IPSAM. In this process, special emphasis in made on the material flow problem. For this, besides offering a description and analysis of the system and its main parts, also some tips on how to define other possible alternative material flow scenarios and a partial analysis of the combinatorial nature of the problem are offered as well. All this is done with the intentions of relating it with the use of simulation tools, for which these have been briefly addressed with a special focus on the Witness simulation package. For a better comprehension, the previous elements are supported by a few figures and expressions which would help obtaining necessary data. Such data and others will be used in the future, when simulating the scenarios in the search of the best material flow configurations.

Keywords: Flexible/Intelligent assembly/disassembly cell (F/IA/DC), Flexible/Intelligent Manufacturing Systems/Cell (F/IMS/C), Material Flow Optimization/Combinations/Design (MFO/C/D).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048
351 Effect of Size of the Step in the Response Surface Methodology using Nonlinear Test Functions

Authors: Jesús Everardo Olguín Tiznado, Rafael García Martínez, Claudia Camargo Wilson, Juan Andrés López Barreras, Everardo Inzunza González, Javier Ordorica Villalvazo

Abstract:

The response surface methodology (RSM) is a collection of mathematical and statistical techniques useful in the modeling and analysis of problems in which the dependent variable receives the influence of several independent variables, in order to determine which are the conditions under which should operate these variables to optimize a production process. The RSM estimated a regression model of first order, and sets the search direction using the method of maximum / minimum slope up / down MMS U/D. However, this method selects the step size intuitively, which can affect the efficiency of the RSM. This paper assesses how the step size affects the efficiency of this methodology. The numerical examples are carried out through Monte Carlo experiments, evaluating three response variables: efficiency gain function, the optimum distance and the number of iterations. The results in the simulation experiments showed that in response variables efficiency and gain function at the optimum distance were not affected by the step size, while the number of iterations is found that the efficiency if it is affected by the size of the step and function type of test used.

Keywords: RSM, dependent variable, independent variables, efficiency, simulation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1965
350 Magnetohydrodynamic Maxwell Nanofluids Flow over a Stretching Surface through a Porous Medium: Effects of Non-Linear Thermal Radiation, Convective Boundary Conditions and Heat Generation/Absorption

Authors: Sameh E. Ahmed, Ramadan A. Mohamed, Abd Elraheem M. Aly, Mahmoud S. Soliman

Abstract:

In this paper, an enhancement of the heat transfer using non-Newtonian nanofluids by magnetohydrodynamic (MHD) mixed convection along stretching sheets embedded in an isotropic porous medium is investigated. Case of the Maxwell nanofluids is studied using the two phase mathematical model of nanofluids and the Darcy model is applied for the porous medium. Important effects are taken into account, namely, non-linear thermal radiation, convective boundary conditions, electromagnetic force and presence of the heat source/sink. Suitable similarity transformations are used to convert the governing equations to a system of ordinary differential equations then it is solved numerically using a fourth order Runge-Kutta method with shooting technique. The main results of the study revealed that the velocity profiles are decreasing functions of the Darcy number, the Deborah number and the magnetic field parameter. Also, the increase in the non-linear radiation parameters causes an enhancement in the local Nusselt number.

Keywords: MHD, nanofluids, stretching surface, non-linear thermal radiation, convective condition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 911
349 Mathematical Model of Depletion of Forestry Resource: Effect of Synthetic Based Industries

Authors: Manisha Chaudhary, Joydip Dhar, Govind Prasad Sahu

Abstract:

A mathematical model is proposed considering the forest biomass density B(t), density of wood based industries W(t) and density of synthetic industries S(t). It is assumed that the forest biomass grows logistically in the absence of wood based industries, but depletion of forestry biomass is due to presence of wood based industries. The growth of wood based industries depends on B(t), while S(t) grows at a constant rate, independent of B(t). Further there is a competition between W(t) and S(t) according to market demand. The proposed model has four ecologically feasible steady states, namely, E1: forest biomass free and wood industries free equilibrium; E2: wood industries free equilibrium and two coexisting equilibria E∗1 , E∗2 . Behavior of the system near all feasible equilibria is analyzed using the stability theory of differential equations. In the proposed model, the natural depletion rate h1 is a crucial parameter and system exhibits Hopf-bifurcation about the non-trivial equilibrium with respect to h1. The analytical results are verified using numerical simulation.

Keywords: A mathematical model, Competition between wood based and synthetic industries, Hopf-bifurcation, Stability analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3456
348 Toxicity Depletion Rates of Water Lettuce (Pistia stratoites) in an Aquaculture Effluent Hydroponic System

Authors: E. A. Kiridi, A. O. Ogunlela

Abstract:

The control of ammonia build-up and its by-product is a limiting factor for a successful commercial aquaculture in a developing country like Nigeria. The technology for an advanced treatment of fish tank effluent is uneconomical to local fish farmers which have led to indiscriminate disposal of aquaculture wastewater, thereby increasing the concentrations of these nitrogenous compound and other contaminants in surface and groundwater above the permissible level. Phytoremediation using water lettuce could offer cheaper and sustainable alternative. On the first day of experimentation, approximately 100 g of water lettuce were replicated in four hydroponic units containing aquaculture effluents. The water quality parameters measured were concentration of ammonium–nitrogen (NH4+-N), nitrite-nitrogen (NO2--N), nitrate-nitrogen (NO3--N), and phosphate–phosphorus (PO43--P). Others were total suspended solids (TSS), pH, electrical conductivity (EC), and biomass value. At phytoremediation intervals of 7, 14, 21 and 28 days, the biomass recorded were 361.2 g, 498.7 g, 561.2 g, and 623.7 g. Water lettuce was able to reduce the pollutant concentration of all the selected parameter. The percentage reduction of pH ranged from 3.9% to 14.4%, EC from 49.8% to 96.2%, TDS from 50.4% to 96.2%, TSS from 38.3% to 81.7%, NH4+-N from 38.9% to 90.7%, NO2--N from 0% to 74.9%, NO3--N from 63.2% to 95.9% and PO43--P from 0% to 76.3%. At 95% confidence level, the analysis of variance shows that F(critical) is less than F(cal) and p < 0.05; therefore, it can be concluded statistically that the inequality between the pre-treatment and post-treatment values are significant. This suggests the potency of water lettuce for remediation of aquaculture effluent.

Keywords: Aquaculture effluent, nitrification, phytoremediation, water lettuce.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1102
347 Optimization of Surface Roughness in Turning Process Utilizing Live Tooling via Taguchi Methodology

Authors: Weinian Wang, Joseph C. Chen

Abstract:

The objective of this research is to optimize the process of cutting cylindrical workpieces utilizing live tooling on a HAAS ST-20 lathe. Surface roughness (Ra) has been investigated as the indicator of quality characteristics for machining process. Aluminum alloy was used to conduct experiments due to its wide range usages in engineering structures and components where light weight or corrosion resistance is required. In this study, Taguchi methodology is utilized to determine the effects that each of the parameters has on surface roughness (Ra). A total of 18 experiments of each process were designed according to Taguchi’s L9 orthogonal array (OA) with four control factors at three levels of each and signal-to-noise ratios (S/N) were computed with Smaller the better equation for minimizing the system. The optimal parameters identified for the surface roughness of the turning operation utilizing live tooling were a feed rate of 3 inches/min(A3); a spindle speed of 1300 rpm(B3); a 2-flute titanium nitrite coated 3/8” endmill (C1); and a depth of cut of 0.025 inches (D2). The mean surface roughness of the confirmation runs in turning operation was 8.22 micro inches. The final results demonstrate that Taguchi methodology is a sufficient way of process improvement in turning process on surface roughness.

Keywords: Live tooling, surface roughness, Taguchi Parameter Design, CNC turning operation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 766