Search results for: real time traffic reporting
4271 Variation of Spot Price and Profits of Andhra Pradesh State Grid in Deregulated Environment
Authors: Chava Sunil Kumar, P.S. Subrahmanyan, J. Amarnath
Abstract:
In this paper variation of spot price and total profits of the generating companies- through wholesale electricity trading are discussed with and without Central Generating Stations (CGS) share and seasonal variations are also considered. It demonstrates how proper analysis of generators- efficiencies and capabilities, types of generators owned, fuel costs, transmission losses and settling price variation using the solutions of Optimal Power Flow (OPF), can allow companies to maximize overall revenue. It illustrates how solutions of OPF can be used to maximize companies- revenue under different scenarios. And is also extended to computation of Available Transfer Capability (ATC) is very important to the transmission system security and market forecasting. From these results it is observed that how crucial it is for companies to plan their daily operations and is certainly useful in an online environment of deregulated power system. In this paper above tasks are demonstrated on 124 bus real-life Indian utility power system of Andhra Pradesh State Grid and results have been presented and analyzed.Keywords: OPF, ATC, Electricity Market, Bid, Spot Price
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18184270 Micromechanics of Stress Transfer across the Interface Fiber-Matrix Bonding
Authors: Fatiha Teklal, Bachir Kacimi, Arezki Djebbar
Abstract:
The study and application of composite materials are a truly interdisciplinary endeavor that has been enriched by contributions from chemistry, physics, materials science, mechanics and manufacturing engineering. The understanding of the interface (or interphase) in composites is the central point of this interdisciplinary effort. From the early development of composite materials of various nature, the optimization of the interface has been of major importance. Even more important, the ideas linking the properties of composites to the interface structure are still emerging. In our study, we need a direct characterization of the interface; the micromechanical tests we are addressing seem to meet this objective and we chose to use two complementary tests simultaneously. The microindentation test that can be applied to real composites and the drop test, preferred to the pull-out because of the theoretical possibility of studying systems with high adhesion (which is a priori the case with our systems). These two tests are complementary because of the principle of the model specimen used for both the first "compression indentation" and the second whose fiber is subjected to tensile stress called the drop test. Comparing the results obtained by the two methods can therefore be rewarding.Keywords: Interface, micromechanics, pull-out, composite, fiber, matrix.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6344269 Approximating Maximum Speed on Road from Curvature Information of Bezier Curve
Authors: M. Y. Misro, A. Ramli, J. M. Ali
Abstract:
Bezier curves have useful properties for path generation problem, for instance, it can generate the reference trajectory for vehicles to satisfy the path constraints. Both algorithms join cubic Bezier curve segment smoothly to generate the path. Some of the useful properties of Bezier are curvature. In mathematics, curvature is the amount by which a geometric object deviates from being flat, or straight in the case of a line. Another extrinsic example of curvature is a circle, where the curvature is equal to the reciprocal of its radius at any point on the circle. The smaller the radius, the higher the curvature thus the vehicle needs to bend sharply. In this study, we use Bezier curve to fit highway-like curve. We use different approach to find the best approximation for the curve so that it will resembles highway-like curve. We compute curvature value by analytical differentiation of the Bezier Curve. We will then compute the maximum speed for driving using the curvature information obtained. Our research works on some assumptions; first, the Bezier curve estimates the real shape of the curve which can be verified visually. Even though, fitting process of Bezier curve does not interpolate exactly on the curve of interest, we believe that the estimation of speed are acceptable. We verified our result with the manual calculation of the curvature from the map.Keywords: Speed estimation, path constraints, reference trajectory, Bezier curve.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40604268 An Efficient Algorithm for Delay Delay-variation Bounded Least Cost Multicast Routing
Authors: Manas Ranjan Kabat, Manoj Kumar Patel, Chita Ranjan Tripathy
Abstract:
Many multimedia communication applications require a source to transmit messages to multiple destinations subject to quality of service (QoS) delay constraint. To support delay constrained multicast communications, computer networks need to guarantee an upper bound end-to-end delay from the source node to each of the destination nodes. This is known as multicast delay problem. On the other hand, if the same message fails to arrive at each destination node at the same time, there may arise inconsistency and unfairness problem among users. This is related to multicast delayvariation problem. The problem to find a minimum cost multicast tree with delay and delay-variation constraints has been proven to be NP-Complete. In this paper, we propose an efficient heuristic algorithm, namely, Economic Delay and Delay-Variation Bounded Multicast (EDVBM) algorithm, based on a novel heuristic function, to construct an economic delay and delay-variation bounded multicast tree. A noteworthy feature of this algorithm is that it has very high probability of finding the optimal solution in polynomial time with low computational complexity.Keywords: EDVBM, Heuristic algorithm, Multicast tree, QoS routing, Shortest path.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16474267 Some Morphological Characteristics of Perennial Ryegrass Genotypes and Correlations among Their Characteristics
Abstract:
The present study involved analysis of certain characteristics of the perennial ryegrass (Lolium perenne L.) genotypes collected from the natural flora of Ankara, and explores a correlation among them. In order to evaluate the plants for breeding purpose as per Turkey's environmental conditions, the perennial ryegrass plants were collected from natural pasture of Ankara in 2004 and were utilized for the study. Seeds of the collected plants were sown in pots and seedlings were prepared in a greenhouse. In 2005, the seedlings were transplanted at 50 × 50 cm2 intervals in Randomized Complete Blocks Design in an experimental field. In 2007 and 2008, data were recorded from the observations and measurements of 568 perennial ryegrasses. The plant characteristics, which were investigated, included re-growth time in spring, color, density, growth habit, tendency to form inflorescence, time of inflorescence, plant height, length of upper internode, spike length, leaf length, leaf width, leaf area, leaf shape, number of spikelets per spike, seed yield per spike and 1000 grain weight and the correlation analyses were made using this data. Correlation coefficients were estimated between all paired combinations of the studied traits. The yield components exhibited varying trends of association among themselves. Seed yield per spike showed significant and positive association with the number of spikelets per spike, 1000 grain weight, plant height, length of upper internode, spike length, leaf length, leaf width, leaf area and color, but significant and negative association with the growth habit and re-growth time in spring.
Keywords: Correlation, morphological traits, Lolium perenne.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19484266 Blind Channel Estimation for Frequency Hopping System Using Subspace Based Method
Authors: M. M. Qasaymeh, M. A. Khodeir
Abstract:
Subspace channel estimation methods have been studied widely, where the subspace of the covariance matrix is decomposed to separate the signal subspace from noise subspace. The decomposition is normally done by using either the eigenvalue decomposition (EVD) or the singular value decomposition (SVD) of the auto-correlation matrix (ACM). However, the subspace decomposition process is computationally expensive. This paper considers the estimation of the multipath slow frequency hopping (FH) channel using noise space based method. In particular, an efficient method is proposed to estimate the multipath time delays by applying multiple signal classification (MUSIC) algorithm which is based on the null space extracted by the rank revealing LU (RRLU) factorization. As a result, precise information is provided by the RRLU about the numerical null space and the rank, (i.e., important tool in linear algebra). The simulation results demonstrate the effectiveness of the proposed novel method by approximately decreasing the computational complexity to the half as compared with RRQR methods keeping the same performance.
Keywords: Time Delay Estimation, RRLU, RRQR, MUSIC, LS-ESPRIT, LS-ESPRIT, Frequency Hopping.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20504265 Impact of Safety and Quality Considerations of Housing Clients on the Construction Firms’ Intention to Adopt Quality Function Deployment: A Case of Construction Sector
Authors: Saif Ul Haq
Abstract:
The current study intends to examine the safety and quality considerations of clients of housing projects and their impact on the adoption of Quality Function Deployment (QFD) by the construction firm. Mixed method research technique has been used to collect and analyze the data wherein a survey was conducted to collect the data from 220 clients of housing projects in Saudi Arabia. Then, the telephonic and Skype interviews were conducted to collect data of 15 professionals working in the top ten real estate companies of Saudi Arabia. Data were analyzed by using partial least square (PLS) and thematic analysis techniques. Findings reveal that today’s customer prioritizes the safety and quality requirements of their houses and as a result, construction firms adopt QFD to address the needs of customers. The findings are of great importance for the clients of housing projects as well as for the construction firms as they could apply QFD in housing projects to address the safety and quality concerns of their clients.Keywords: Construction industry, quality considerations, quality function deployment, safety considerations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9084264 Effect of Ripening Conditions and Storage Time on Oxidative and Sensory Stability of Petrovská Klobása Sausage
Authors: Branislav V. Šojić, Ljiljana S. Petrović, Vladimir M. Tomović, Natalija R. Džinić, Anamarija I. Mandić, Snežana B. Škaljac, Marija R. Jokanović, Predrag M. Ikonić, Tatjana A. Tasić, Ivana J. Sedej
Abstract:
The influence of ripening conditions (traditional and industrial) on oxidative and sensory stability of dry fermented sausage (Petrovská klobása), during 7 months of storage, was investigated. During the storage period the content of free fatty acids was significantly higher (P<0.05), while the content of malondialdehyde was significantly lower in the sausage subjected to traditional conditions of drying. At the end of the storage period, content of hexanal in the sausage subjected to traditional conditions of ripening (1.67μg/g) was significantly lower (P<0.05) in comparison with this content in the sausage subjected to industrial conditions of ripening (4.94µg/g). Traditional conditions of ripening at lower temperatures have led to better sensory properties of odor and taste of traditional dry fermented sausage, Petrovská klobása after 2 and 7 months of storage.
Keywords: Lipid oxidation, Petrovská klobása, sensory stability, storage time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39484263 BugCatcher.Net: Detecting Bugs and Proposing Corrective Solutions
Authors: Sheetal Chavan, P. J. Kulkarni, Vivek Shanbhag
Abstract:
Although achieving zero-defect software release is practically impossible, software industries should take maximum care to detect defects/bugs well ahead in time allowing only bare minimums to creep into released version. This is a clear indicator of time playing an important role in the bug detection. In addition to this, software quality is the major factor in software engineering process. Moreover, early detection can be achieved only through static code analysis as opposed to conventional testing. BugCatcher.Net is a static analysis tool, which detects bugs in .NET® languages through MSIL (Microsoft Intermediate Language) inspection. The tool utilizes a Parser based on Finite State Automata to carry out bug detection. After being detected, bugs need to be corrected immediately. BugCatcher.Net facilitates correction, by proposing a corrective solution for reported warnings/bugs to end users with minimum side effects. Moreover, the tool is also capable of analyzing the bug trend of a program under inspection.Keywords: Dependence, Early solution, Finite State Automata, Grammar, Late solution, Parser State Transition Diagram, StaticProgram Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15164262 Augmented Reality in Advertising and Brand Communication: An Experimental Study
Authors: O. Mauroner, L. Le, S. Best
Abstract:
Digital technologies offer many opportunities in the design and implementation of brand communication and advertising. Augmented reality (AR) is an innovative technology in marketing communication that focuses on the fact that virtual interaction with a product ad offers additional value to consumers. AR enables consumers to obtain (almost) real product experiences by the way of virtual information even before the purchase of a certain product. Aim of AR applications in relation with advertising is in-depth examination of product characteristics to enhance product knowledge as well as brand knowledge. Interactive design of advertising provides observers with an intense examination of a specific advertising message and therefore leads to better brand knowledge. The elaboration likelihood model and the central route to persuasion strongly support this argumentation. Nevertheless, AR in brand communication is still in an initial stage and therefore scientific findings about the impact of AR on information processing and brand attitude are rare. The aim of this paper is to empirically investigate the potential of AR applications in combination with traditional print advertising. To that effect an experimental design with different levels of interactivity is built to measure the impact of interactivity of an ad on different variables o advertising effectiveness.Keywords: Advertising effectiveness, augmented reality, brand communication, brand recall, interactivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 49304261 Flocculation on the Treatment of Olive Oil Mill Wastewater: Pretreatment
Authors: G. Hodaifa, J. A. Páez, C. Agabo, E. Ramos, J. C. Gutiérrez, A. Rosal
Abstract:
Currently, continuous two-phase decanter process used for olive oil production is the more internationally widespread. The wastewaters generated from this industry (OMW) are a real environmental problem because of its high organic load. Among proposed treatments for these wastewaters, advanced oxidation technologies (Fenton, ozone, photoFenton, etc.) are the most favourable. The direct application of these processes is somewhat expensive. Therefore, the application of a previous stage based on a flocculation-sedimentation operation is of high importance. In this research five commercial flocculants (three cationic, and two anionic) have been used to achieve the separation of phases (liquid clarifiedsludge). For each flocculant, different concentrations (0-1000 mg/L) have been studied. In these experiments, sludge volume formed and the final water quality were determined. The final removal percentages of total phenols (11.3-25.1%), COD (5.6-20.4%), total carbon (2.3-26.5%), total organic carbon (1.50-23.8%), total nitrogen (1.45-24.8%), and turbidity (27.9-61.4%) were determined. The variation on electric conductivity reduction percentage (1-8%) was also determined. Finally, the best flocculants with highest removal percentages have been determined (QG2001 and Flocudex CS49).Keywords: Flocculants, flocculation, olive oil mill wastewater, water quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25584260 Adversarial Disentanglement Using Latent Classifier for Pose-Independent Representation
Authors: Hamed Alqahtani, Manolya Kavakli-Thorne
Abstract:
The large pose discrepancy is one of the critical challenges in face recognition during video surveillance. Due to the entanglement of pose attributes with identity information, the conventional approaches for pose-independent representation lack in providing quality results in recognizing largely posed faces. In this paper, we propose a practical approach to disentangle the pose attribute from the identity information followed by synthesis of a face using a classifier network in latent space. The proposed approach employs a modified generative adversarial network framework consisting of an encoder-decoder structure embedded with a classifier in manifold space for carrying out factorization on the latent encoding. It can be further generalized to other face and non-face attributes for real-life video frames containing faces with significant attribute variations. Experimental results and comparison with state of the art in the field prove that the learned representation of the proposed approach synthesizes more compelling perceptual images through a combination of adversarial and classification losses.Keywords: Video surveillance, disentanglement, face detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6144259 Iraqi Short Term Electrical Load Forecasting Based On Interval Type-2 Fuzzy Logic
Authors: Firas M. Tuaimah, Huda M. Abdul Abbas
Abstract:
Accurate Short Term Load Forecasting (STLF) is essential for a variety of decision making processes. However, forecasting accuracy can drop due to the presence of uncertainty in the operation of energy systems or unexpected behavior of exogenous variables. Interval Type 2 Fuzzy Logic System (IT2 FLS), with additional degrees of freedom, gives an excellent tool for handling uncertainties and it improved the prediction accuracy. The training data used in this study covers the period from January 1, 2012 to February 1, 2012 for winter season and the period from July 1, 2012 to August 1, 2012 for summer season. The actual load forecasting period starts from January 22, till 28, 2012 for winter model and from July 22 till 28, 2012 for summer model. The real data for Iraqi power system which belongs to the Ministry of Electricity.
Keywords: Short term load forecasting, prediction interval, type 2 fuzzy logic systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18934258 Agent/Group/Role Organizational Model to Simulate an Industrial Control System
Authors: Noureddine Seddari, Mohamed Belaoued, Salah Bougueroua
Abstract:
The modeling of complex systems is generally based on the decomposition of their components into sub-systems easier to handle. This division has to be made in a methodical way. In this paper, we introduce an industrial control system modeling and simulation based on the Multi-Agent System (MAS) methodology AALAADIN and more particularly the underlying conceptual model Agent/Group/Role (AGR). Indeed, in this division using AGR model, the overall system is decomposed into sub-systems in order to improve the understanding of regulation and control systems, and to simplify the implementation of the obtained agents and their groups, which are implemented using the Multi-Agents Development KIT (MAD-KIT) platform. This approach appears to us to be the most appropriate for modeling of this type of systems because, due to the use of MAS, it is possible to model real systems in which very complex behaviors emerge from relatively simple and local interactions between many different individuals, therefore a MAS is well adapted to describe a system from the standpoint of the activity of its components, that is to say when the behavior of the individuals is complex (difficult to describe with equations). The main aim of this approach is the take advantage of the performance, the scalability and the robustness that are intuitively provided by MAS.
Keywords: Complex systems, modeling and simulation, industrial control system, MAS, AALAADIN, AGR, MAD-KIT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11944257 Adaptive Kernel Principal Analysis for Online Feature Extraction
Authors: Mingtao Ding, Zheng Tian, Haixia Xu
Abstract:
The batch nature limits the standard kernel principal component analysis (KPCA) methods in numerous applications, especially for dynamic or large-scale data. In this paper, an efficient adaptive approach is presented for online extraction of the kernel principal components (KPC). The contribution of this paper may be divided into two parts. First, kernel covariance matrix is correctly updated to adapt to the changing characteristics of data. Second, KPC are recursively formulated to overcome the batch nature of standard KPCA.This formulation is derived from the recursive eigen-decomposition of kernel covariance matrix and indicates the KPC variation caused by the new data. The proposed method not only alleviates sub-optimality of the KPCA method for non-stationary data, but also maintains constant update speed and memory usage as the data-size increases. Experiments for simulation data and real applications demonstrate that our approach yields improvements in terms of both computational speed and approximation accuracy.
Keywords: adaptive method, kernel principal component analysis, online extraction, recursive algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15574256 Automated Heart Sound Classification from Unsegmented Phonocardiogram Signals Using Time Frequency Features
Authors: Nadia Masood Khan, Muhammad Salman Khan, Gul Muhammad Khan
Abstract:
Cardiologists perform cardiac auscultation to detect abnormalities in heart sounds. Since accurate auscultation is a crucial first step in screening patients with heart diseases, there is a need to develop computer-aided detection/diagnosis (CAD) systems to assist cardiologists in interpreting heart sounds and provide second opinions. In this paper different algorithms are implemented for automated heart sound classification using unsegmented phonocardiogram (PCG) signals. Support vector machine (SVM), artificial neural network (ANN) and cartesian genetic programming evolved artificial neural network (CGPANN) without the application of any segmentation algorithm has been explored in this study. The signals are first pre-processed to remove any unwanted frequencies. Both time and frequency domain features are then extracted for training the different models. The different algorithms are tested in multiple scenarios and their strengths and weaknesses are discussed. Results indicate that SVM outperforms the rest with an accuracy of 73.64%.Keywords: Pattern recognition, machine learning, computer aided diagnosis, heart sound classification, and feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12944255 Survey of Curriculum Quality of Postgraduate Studies of Insurance Management Field Case: University of Allameh Taba Tabaee
Authors: F. Havas Beigi, E. Mohammadi, M.Vafaee Yeganeh
Abstract:
Curriculum is one of the most important inputs in higher education system and for knowing the strong and weak spots of it we need evaluation. The main purpose of this study was to survey of the curriculum quality of Insurance Management field. Case: University of Allameh Taba Tabaee(according to view point of students,alumni,employer and faculty members).Descriptive statistics (mean, tables, percentages, frequency distribution) and inferential statistics (CHI SQUARE) were used to analyze the data. Six criterions considered for the Quality of curriculum: objectives, content, teaching and learning methods, space and facilities, Time, assessment of learning. objectives, teaching and learning methods criterions was desirable level, content criteria was undesirable level, space and facilities, time and assessment of learning were rather desirable level. The quality of curriculum of insurance management field was relatively desirable level.
Keywords: Quality, curriculum, insurance management, higher education.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13354254 MLOps Scaling Machine Learning Lifecycle in an Industrial Setting
Authors: Yizhen Zhao, Adam S. Z. Belloum, Gonc¸alo Maia da Costa, Zhiming Zhao
Abstract:
Machine learning has evolved from an area of academic research to a real-world applied field. This change comes with challenges, gaps and differences exist between common practices in academic environments and the ones in production environments. Following continuous integration, development and delivery practices in software engineering, similar trends have happened in machine learning (ML) systems, called MLOps. In this paper we propose a framework that helps to streamline and introduce best practices that facilitate the ML lifecycle in an industrial setting. This framework can be used as a template that can be customized to implement various machine learning experiments. The proposed framework is modular and can be recomposed to be adapted to various use cases (e.g. data versioning, remote training on Cloud). The framework inherits practices from DevOps and introduces other practices that are unique to the machine learning system (e.g.data versioning). Our MLOps practices automate the entire machine learning lifecycle, bridge the gap between development and operation.
Keywords: Cloud computing, continuous development, data versioning, DevOps, industrial setting, MLOps, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10934253 Fenton’s Oxidation as Post-Treatment of a Mature Municipal Landfill Leachate
Authors: Susana Cortez, Pilar Teixeira, Rosário Oliveira, Manuel Mota
Abstract:
Mature landfill leachates contain some macromolecular organic substances that are resistant to biological degradation. Recently, Fenton-s oxidation has been investigated for chemical treatment or pre-treatment of mature landfill leachates. The aim of this study was to reduce the recalcitrant organic load still remaining after the complete treatment of a mature landfill leachate by Fenton-s oxidation post-treatment. The effect of various parameters such as H2O2 to Fe2+ molar ratio, dosage of Fe2+ reagent, initial pH, reaction time and initial chemical oxygen demand (COD) strength, that have an important role on the oxidation, was analysed. A molar ratio H2O2/Fe2+ = 3, a Fe2+ dosage of 4 mmol·L-1, pH 3, and a reaction time of 40 min were found to achieve better oxidation performances. At these favorable conditions, COD removal efficiency was 60.9% and 31.1% for initial COD of 93 and 743 mg·L-1 respectively (diluted and non diluted leachate). Fenton-s oxidation also presented good results for color removal. In spite of being extremely difficult to treat this leachate, the above results seem rather encouraging on the application of Fenton-s oxidation.
Keywords: Fenton's oxidation, mature landfill leachate, recalcitrant organic matter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17214252 Hybrid Neuro Fuzzy Approach for Automatic Generation Control of Two -Area Interconnected Power System
Authors: Gayadhar Panda, Sidhartha Panda, C. Ardil
Abstract:
The main objective of Automatic Generation Control (AGC) is to balance the total system generation against system load losses so that the desired frequency and power interchange with neighboring systems is maintained. Any mismatch between generation and demand causes the system frequency to deviate from its nominal value. Thus high frequency deviation may lead to system collapse. This necessitates a very fast and accurate controller to maintain the nominal system frequency. This paper deals with a novel approach of artificial intelligence (AI) technique called Hybrid Neuro-Fuzzy (HNF) approach for an (AGC). The advantage of this controller is that it can handle the non-linearities at the same time it is faster than other conventional controllers. The effectiveness of the proposed controller in increasing the damping of local and inter area modes of oscillation is demonstrated in a two area interconnected power system. The result shows that intelligent controller is having improved dynamic response and at the same time faster than conventional controller.
Keywords: Automatic Generation Control (AGC), Dynamic Model, Two-area Power System, Fuzzy Logic Controller, Neural Network, Hybrid Neuro-Fuzzy(HNF).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24674251 The Application of FSI Techniques in Modeling of Realist Pulmonary Systems
Authors: Abdurrahim Bolukbasi, Hassan Athari, Dogan Ciloglu
Abstract:
The modeling lung respiratory system that has complex anatomy and biophysics presents several challenges including tissue-driven flow patterns and wall motion. Also, the pulmonary lung system because of that they stretch and recoil with each breath, has not static walls and structures. The direct relationship between air flow and tissue motion in the lung structures naturally prefers an FSI simulation technique. Therefore, in order to toward the realistic simulation of pulmonary breathing mechanics the development of a coupled FSI computational model is an important step. A simple but physiologically relevant three-dimensional deep long geometry is designed and fluid-structure interaction (FSI) coupling technique is utilized for simulating the deformation of the lung parenchyma tissue that produces airflow fields. The real understanding of respiratory tissue system as a complex phenomenon have been investigated with respect to respiratory patterns, fluid dynamics and tissue viscoelasticity and tidal breathing period.
Keywords: Lung deformation and mechanics, tissue mechanics, viscoelasticity, fluid-structure interactions, ANSYS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23304250 Neural Network Tuned Fuzzy Controller for MIMO System
Authors: Seema Chopra, R. Mitra, Vijay Kumar
Abstract:
In this paper, a neural network tuned fuzzy controller is proposed for controlling Multi-Input Multi-Output (MIMO) systems. For the convenience of analysis, the structure of MIMO fuzzy controller is divided into single input single-output (SISO) controllers for controlling each degree of freedom. Secondly, according to the characteristics of the system-s dynamics coupling, an appropriate coupling fuzzy controller is incorporated to improve the performance. The simulation analysis on a two-level mass–spring MIMO vibration system is carried out and results show the effectiveness of the proposed fuzzy controller. The performance though improved, the computational time and memory used is comparatively higher, because it has four fuzzy reasoning blocks and number may increase in case of other MIMO system. Then a fuzzy neural network is designed from a set of input-output training data to reduce the computing burden during implementation. This control strategy can not only simplify the implementation problem of fuzzy control, but also reduce computational time and consume less memory.Keywords: Fuzzy Control, Neural Network, MIMO System, Optimization of Membership functions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32174249 Statistical Analysis of Stresses in Rigid Pavement
Authors: Aleš Florian, Lenka Ševelová, Rudolf Hela
Abstract:
Complex statistical analysis of stresses in concrete slab of the real type of rigid pavement is performed. The computational model of the pavement is designed as a spatial (3D) model, is based on a nonlinear variant of the finite element method that respects the structural nonlinearity, enables to model different arrangement of joints, and the entire model can be loaded by the thermal load. Interaction of adjacent slabs in joints and contact of the slab and the subsequent layer are modeled with help of special contact elements. Four concrete slabs separated by transverse and longitudinal joints and the additional subgrade layers and soil to the depth of about 3m are modeled. The thickness of individual layers, physical and mechanical properties of materials, characteristics of joints, and the temperature of the upper and lower surface of slabs are supposed to be random variables. The modern simulation technique Updated Latin Hypercube Sampling with 20 simulations is used for statistical analysis. As results, the estimates of basic statistics of the principal stresses s1 and s3 in 53 points on the upper and lower surface of the slabs are obtained.Keywords: concrete, FEM, pavement, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15814248 Performance Evaluation of Data Mining Techniques for Predicting Software Reliability
Authors: Pradeep Kumar, Abdul Wahid
Abstract:
Accurate software reliability prediction not only enables developers to improve the quality of software but also provides useful information to help them for planning valuable resources. This paper examines the performance of three well-known data mining techniques (CART, TreeNet and Random Forest) for predicting software reliability. We evaluate and compare the performance of proposed models with Cascade Correlation Neural Network (CCNN) using sixteen empirical databases from the Data and Analysis Center for Software. The goal of our study is to help project managers to concentrate their testing efforts to minimize the software failures in order to improve the reliability of the software systems. Two performance measures, Normalized Root Mean Squared Error (NRMSE) and Mean Absolute Errors (MAE), illustrate that CART model is accurate than the models predicted using Random Forest, TreeNet and CCNN in all datasets used in our study. Finally, we conclude that such methods can help in reliability prediction using real-life failure datasets.
Keywords: Classification, Cascade Correlation Neural Network, Random Forest, Software reliability, TreeNet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18434247 Examining Foreign Student Visual Perceptions of Online Marketing Tools at a Hungarian University
Authors: Anita Kéri
Abstract:
Higher education marketing has been a widely researched field in recent years. Due to the increasing competition among higher education institutions worldwide, it has become crucial to target foreign students with effective marketing tools. Online marketing tools became central to attracting, retaining, and satisfying the needs of foreign students. Therefore, the aim of the current study is to reveal how the online marketing tools of a Hungarian university are perceived visually by its first-year foreign students, with special emphasis on the university webpage content. Eye-camera tracking and retrospective think aloud interviews were used to measure visual perceptions. Results show that freshmen students remember those online marketing content more that have familiar content on them. Pictures of real-life students and their experiences attract students’ attention more, and they also remember information on these webpage elements more, compared to designs with stock photos. This research uses eye camera tracking in the field of higher education marketing, thereby providing insight into the perception of online higher education marketing for foreign students.
Keywords: Higher education, marketing, eye-camera, visual perception.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1794246 Customer Relations and Use of Online Shopping Sites
Authors: Bahar Urhan Torun, Havva Nur Tarakcı
Abstract:
At the present time, online marketing has become the common target of small and full-scale organizations. Today’s humanbeing who has to spend most of their time in front of the computer because of his job, prefers to socialize by internet due to the easy access to technology. So online marketing area expands day by day. All business organizations from the smallest to the biggest are in a race in order to get a cut from the virtual market share in an extreme competitive environment. However these organizations which use the internet to reach more consumers cannot determine their target group accurately, so this is the biggest handicap of online marketing sales nowadays. The aim of this study is to determine some significant elements about need for communicating efficiently with the consumer on the internet on online marketing. The strategies that can be used in order to increase sales and the limitations of virtual environment where cannot be communicated with the consumer face to face are argued in this study’s scope. As a consequence it is thought that to study on this subject because of lacking and also being limited efficiency of researches and outputs. Within this scope suggesting some proposals about how to communicate efficiently with the consumer and also offering the consumers’ demands efficiently is the essential objective of this study.
Keywords: Communication, competition, consumer, online marketing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6294245 A Hybrid Nature Inspired Algorithm for Generating Optimal Query Plan
Authors: R. Gomathi, D. Sharmila
Abstract:
The emergence of the Semantic Web technology increases day by day due to the rapid growth of multiple web pages. Many standard formats are available to store the semantic web data. The most popular format is the Resource Description Framework (RDF). Querying large RDF graphs becomes a tedious procedure with a vast increase in the amount of data. The problem of query optimization becomes an issue in querying large RDF graphs. Choosing the best query plan reduces the amount of query execution time. To address this problem, nature inspired algorithms can be used as an alternative to the traditional query optimization techniques. In this research, the optimal query plan is generated by the proposed SAPSO algorithm which is a hybrid of Simulated Annealing (SA) and Particle Swarm Optimization (PSO) algorithms. The proposed SAPSO algorithm has the ability to find the local optimistic result and it avoids the problem of local minimum. Experiments were performed on different datasets by changing the number of predicates and the amount of data. The proposed algorithm gives improved results compared to existing algorithms in terms of query execution time.
Keywords: Semantic web, RDF, Query optimization, Nature inspired algorithms, PSO, SA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22444244 Convergence Analysis of Training Two-Hidden-Layer Partially Over-Parameterized ReLU Networks via Gradient Descent
Authors: Zhifeng Kong
Abstract:
Over-parameterized neural networks have attracted a great deal of attention in recent deep learning theory research, as they challenge the classic perspective of over-fitting when the model has excessive parameters and have gained empirical success in various settings. While a number of theoretical works have been presented to demystify properties of such models, the convergence properties of such models are still far from being thoroughly understood. In this work, we study the convergence properties of training two-hidden-layer partially over-parameterized fully connected networks with the Rectified Linear Unit activation via gradient descent. To our knowledge, this is the first theoretical work to understand convergence properties of deep over-parameterized networks without the equally-wide-hidden-layer assumption and other unrealistic assumptions. We provide a probabilistic lower bound of the widths of hidden layers and proved linear convergence rate of gradient descent. We also conducted experiments on synthetic and real-world datasets to validate our theory.Keywords: Over-parameterization, Rectified Linear Units (ReLU), convergence, gradient descent, neural networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9084243 Fuzzy Risk-Based Life Cycle Assessment for Estimating Environmental Aspects in EMS
Authors: Kevin Fong-Rey Liu, Ken Yeh, Cheng-Wu Chen, Han-Hsi Liang
Abstract:
Environmental aspects plays a central role in environmental management system (EMS) because it is the basis for the identification of an organization-s environmental targets. The existing methods for the assessment of environmental aspects are grouped into three categories: risk assessment-based (RA-based), LCA-based and criterion-based methods. To combine the benefits of these three categories of research, this study proposes an integrated framework, combining RA-, LCA- and criterion-based methods. The integrated framework incorporates LCA techniques for the identification of the causal linkage for aspect, pathway, receptor and impact, uses fuzzy logic to assess aspects, considers fuzzy conditions, in likelihood assessment, and employs a new multi-criteria decision analysis method - multi-criteria and multi-connection comprehensive assessment (MMCA) - to estimate significant aspects in EMS. The proposed model is verified, using a real case study and the results show that this method successfully prioritizes the environmental aspects.Keywords: Environmental management system, environmental aspect, risk assessment, life cycle assessment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22234242 A Novel In-Place Sorting Algorithm with O(n log z) Comparisons and O(n log z) Moves
Authors: Hanan Ahmed-Hosni Mahmoud, Nadia Al-Ghreimil
Abstract:
In-place sorting algorithms play an important role in many fields such as very large database systems, data warehouses, data mining, etc. Such algorithms maximize the size of data that can be processed in main memory without input/output operations. In this paper, a novel in-place sorting algorithm is presented. The algorithm comprises two phases; rearranging the input unsorted array in place, resulting segments that are ordered relative to each other but whose elements are yet to be sorted. The first phase requires linear time, while, in the second phase, elements of each segment are sorted inplace in the order of z log (z), where z is the size of the segment, and O(1) auxiliary storage. The algorithm performs, in the worst case, for an array of size n, an O(n log z) element comparisons and O(n log z) element moves. Further, no auxiliary arithmetic operations with indices are required. Besides these theoretical achievements of this algorithm, it is of practical interest, because of its simplicity. Experimental results also show that it outperforms other in-place sorting algorithms. Finally, the analysis of time and space complexity, and required number of moves are presented, along with the auxiliary storage requirements of the proposed algorithm.
Keywords: Auxiliary storage sorting, in-place sorting, sorting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1915