Search results for: loading problem
537 Evolutionary Multi-objective Optimization for Positioning of Residential Houses
Authors: Ayman El Ansary, Mohamed Shalaby
Abstract:
The current study describes a multi-objective optimization technique for positioning of houses in a residential neighborhood. The main task is the placement of residential houses in a favorable configuration satisfying a number of objectives. Solving the house layout problem is a challenging task. It requires an iterative approach to satisfy design requirements (e.g. energy efficiency, skyview, daylight, roads network, visual privacy, and clear access to favorite views). These design requirements vary from one project to another based on location and client preferences. In the Gulf region, the most important socio-cultural factor is the visual privacy in indoor space. Hence, most of the residential houses in this region are surrounded by high fences to provide privacy, which has a direct impact on other requirements (e.g. daylight and direction to favorite views). This investigation introduces a novel technique to optimally locate and orient residential buildings to satisfy a set of design requirements. The developed technique explores the search space for possible solutions. This study considers two dimensional house planning problems. However, it can be extended to solve three dimensional cases.
Keywords: Evolutionary optimization, Houses planning, Urban modeling, Daylight, Visual Privacy, Residential compounds.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1545536 Active Intra-ONU Scheduling with Cooperative Prediction Mechanism in EPONs
Authors: Chuan-Ching Sue, Shi-Zhou Chen, Ting-Yu Huang
Abstract:
Dynamic bandwidth allocation in EPONs can be generally separated into inter-ONU scheduling and intra-ONU scheduling. In our previous work, the active intra-ONU scheduling (AS) utilizes multiple queue reports (QRs) in each report message to cooperate with the inter-ONU scheduling and makes the granted bandwidth fully utilized without leaving unused slot remainder (USR). This scheme successfully solves the USR problem originating from the inseparability of Ethernet frame. However, without proper setting of threshold value in AS, the number of QRs constrained by the IEEE 802.3ah standard is not enough, especially in the unbalanced traffic environment. This limitation may be solved by enlarging the threshold value. The large threshold implies the large gap between the adjacent QRs, thus resulting in the large difference between the best granted bandwidth and the real granted bandwidth. In this paper, we integrate AS with a cooperative prediction mechanism and distribute multiple QRs to reduce the penalty brought by the prediction error. Furthermore, to improve the QoS and save the usage of queue reports, the highest priority (EF) traffic which comes during the waiting time is granted automatically by OLT and is not considered in the requested bandwidth of ONU. The simulation results show that the proposed scheme has better performance metrics in terms of bandwidth utilization and average delay for different classes of packets.Keywords: EPON, Inter-ONU and Intra-ONU scheduling, Prediction, Unused slot remainder
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1595535 MaxMin Share Based Medium Access for Attaining Fairness and Channel Utilization in Mobile Adhoc Networks
Authors: P. Priakanth, P. Thangaraj
Abstract:
Due to the complex network architecture, the mobile adhoc network-s multihop feature gives additional problems to the users. When the traffic load at each node gets increased, the additional contention due its traffic pattern might cause the nodes which are close to destination to starve the nodes more away from the destination and also the capacity of network is unable to satisfy the total user-s demand which results in an unfairness problem. In this paper, we propose to create an algorithm to compute the optimal MAC-layer bandwidth assigned to each flow in the network. The bottleneck links contention area determines the fair time share which is necessary to calculate the maximum allowed transmission rate used by each flow. To completely utilize the network resources, we compute two optimal rates namely, the maximum fair share and minimum fair share. We use the maximum fair share achieved in order to limit the input rate of those flows which crosses the bottleneck links contention area when the flows that are not allocated to the optimal transmission rate and calculate the following highest fair share. Through simulation results, we show that the proposed protocol achieves improved fair share and throughput with reduced delay.Keywords: MAC-layer, MANETs, Multihop, optimal rate, Transmission.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1509534 An Inter-banking Auditing Security Solution for Detecting Unauthorised Financial Transactions entered by Authorised Insiders
Authors: C. A. Corzo, N. Zhang, F. Corzo
Abstract:
Insider abuse has recently been reported as one of the more frequently occurring security incidents, suggesting that more security is required for detecting and preventing unauthorised financial transactions entered by authorised users. To address the problem, and based on the observation that all authorised interbanking financial transactions trigger or are triggered by other transactions in a workflow, we have developed a security solution based on a redefined understanding of an audit workflow. One audit workflow where there is a log file containing the complete workflow activity of financial transactions directly related to one financial transaction (an electronic deal recorded at an e-trading system). The new security solution contemplates any two parties interacting on the basis of financial transactions recorded by their users in related but distinct automated financial systems. In the new definition interorganizational and intra-organization interactions can be described in one unique audit trail. This concept expands the current ideas of audit trails by adapting them to actual e-trading workflow activity, i.e. intra-organizational and inter-organizational activity. With the above, a security auditing service is designed to detect integrity drifts with and between organizations in order to detect unauthorised financial transactions entered by authorised users.Keywords: Intrusion Detection and Prevention, Authentica-transtionand Identification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1541533 Data Mining to Capture User-Experience: A Case Study in Notebook Product Appearance Design
Authors: Rhoann Kerh, Chen-Fu Chien, Kuo-Yi Lin
Abstract:
In the era of rapidly increasing notebook market, consumer electronics manufacturers are facing a highly dynamic and competitive environment. In particular, the product appearance is the first part for user to distinguish the product from the product of other brands. Notebook product should differ in its appearance to engage users and contribute to the user experience (UX). The UX evaluates various product concepts to find the design for user needs; in addition, help the designer to further understand the product appearance preference of different market segment. However, few studies have been done for exploring the relationship between consumer background and the reaction of product appearance. This study aims to propose a data mining framework to capture the user’s information and the important relation between product appearance factors. The proposed framework consists of problem definition and structuring, data preparation, rules generation, and results evaluation and interpretation. An empirical study has been done in Taiwan that recruited 168 subjects from different background to experience the appearance performance of 11 different portable computers. The results assist the designers to develop product strategies based on the characteristics of consumers and the product concept that related to the UX, which help to launch the products to the right customers and increase the market shares. The results have shown the practical feasibility of the proposed framework.
Keywords: Consumers Decision Making, Product Design, Rough Set Theory, User Experience.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3512532 Hydrogeological Risk and Mining Tunnels: the Fontane-Rodoretto Mine Turin (Italy)
Authors: Paola Gattinoni, Laura Scesi, Elena Cerino Adbin, Daniele Cremonesi
Abstract:
The interaction of tunneling or mining with groundwater has become a very relevant problem not only due to the need to guarantee the safety of workers and to assure the efficiency of the tunnel drainage systems, but also to safeguard water resources from impoverishment and pollution risk. Therefore it is very important to forecast the drainage processes (i.e., the evaluation of drained discharge and drawdown caused by the excavation). The aim of this study was to know better the system and to quantify the flow drained from the Fontane mines, located in Val Germanasca (Turin, Italy). This allowed to understand the hydrogeological local changes in time. The work has therefore been structured as follows: the reconstruction of the conceptual model with the geological, hydrogeological and geological-structural study; the calculation of the tunnel inflows (through the use of structural methods) and the comparison with the measured flow rates; the water balance at the basin scale. In this way it was possible to understand what are the relationships between rainfall, groundwater level variations and the effect of the presence of tunnels as a means of draining water. Subsequently, it the effects produced by the excavation of the mining tunnels was quantified, through numerical modeling. In particular, the modeling made it possible to observe the drawdown variation as a function of number, excavation depth and different mines linings.Keywords: Groundwater, Italy, numerical model, tunneling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1928531 Proposal of Commutation Protocol in Hybrid Sensors and Vehicular Networks for Intelligent Transport Systems
Authors: Taha Bensiradj, Samira Moussaoui
Abstract:
Hybrid Sensors and Vehicular Networks (HSVN), represent a hybrid network, which uses several generations of Ad-Hoc networks. It is used especially in Intelligent Transport Systems (ITS). The HSVN allows making collaboration between the Wireless Sensors Network (WSN) deployed on the border of the road and the Vehicular Network (VANET). This collaboration is defined by messages exchanged between the two networks for the purpose to inform the drivers about the state of the road, provide road safety information and more information about traffic on the road. Moreover, this collaboration created by HSVN, also allows the use of a network and the advantage of improving another network. For example, the dissemination of information between the sensors quickly decreases its energy, and therefore, we can use vehicles that do not have energy constraint to disseminate the information between sensors. On the other hand, to solve the disconnection problem in VANET, the sensors can be used as gateways that allow sending the messages received by one vehicle to another. However, because of the short communication range of the sensor and its low capacity of storage and processing of data, it is difficult to ensure the exchange of road messages between it and the vehicle, which can be moving at high speed at the time of exchange. This represents the time where the vehicle is in communication range with the sensor. This work is the proposition of a communication protocol between the sensors and the vehicle used in HSVN. The latter has as the purpose to ensure the exchange of road messages in the available time of exchange.
Keywords: HSVN, ITS, VANET, WSN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1233530 Conceptual Model for Massive Open Online Blended Courses Based on Disciplines’ Concepts Capitalization and Obstacles’ Detection
Authors: N. Hammid, F. Bouarab-Dahmani, T. Berkane
Abstract:
Since its appearance, the MOOC (massive open online course) is gaining more and more intention of the educational communities over the world. Apart from the current MOOCs design and purposes, the creators of MOOC focused on the importance of the connection and knowledge exchange between individuals in learning. In this paper, we present a conceptual model for massive open online blended courses where teachers over the world can collaborate and exchange their experience to get a common efficient content designed as a MOOC opened to their students to live a better learning experience. This model is based on disciplines’ concepts capitalization and the detection of the obstacles met by their students when faced with problem situations (exercises, projects, case studies, etc.). This detection is possible by analyzing the frequently of semantic errors committed by the students. The participation of teachers in the design of the course and the attendance by their students can guarantee an efficient and extensive participation (an important number of participants) in the course, the learners’ motivation and the evaluation issues, in the way that the teachers designing the course assess their students. Thus, the teachers review, together with their knowledge, offer a better assessment and efficient connections to their students.
Keywords: MOOC, Massive Open Online Courses, Online learning, E-learning, Blended learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 942529 Optimal Manufacturing Scheduling for Dependent Details Processing
Authors: Ivan C. Mustakerov, Daniela I. Borissova
Abstract:
The increasing competitiveness in manufacturing industry is forcing manufacturers to seek effective processing schedules. The paper presents an optimization manufacture scheduling approach for dependent details processing with given processing sequences and times on multiple machines. By defining decision variables as start and end moments of details processing it is possible to use straightforward variables restrictions to satisfy different technological requirements and to formulate easy to understand and solve optimization tasks for multiple numbers of details and machines. A case study example is solved for seven base moldings for CNC metalworking machines processed on five different machines with given processing order among details and machines and known processing time-s duration. As a result of linear optimization task solution the optimal manufacturing schedule minimizing the overall processing time is obtained. The manufacturing schedule defines the moments of moldings delivery thus minimizing storage costs and provides mounting due-time satisfaction. The proposed optimization approach is based on real manufacturing plant problem. Different processing schedules variants for different technological restrictions were defined and implemented in the practice of Bulgarian company RAIS Ltd. The proposed approach could be generalized for other job shop scheduling problems for different applications.Keywords: Optimal manufacturing scheduling, linear programming, metalworking machines production, dependant details processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1487528 Increase of Organization in Complex Systems
Authors: Georgi Yordanov Georgiev, Michael Daly, Erin Gombos, Amrit Vinod, Gajinder Hoonjan
Abstract:
Measures of complexity and entropy have not converged to a single quantitative description of levels of organization of complex systems. The need for such a measure is increasingly necessary in all disciplines studying complex systems. To address this problem, starting from the most fundamental principle in Physics, here a new measure for quantity of organization and rate of self-organization in complex systems based on the principle of least (stationary) action is applied to a model system - the central processing unit (CPU) of computers. The quantity of organization for several generations of CPUs shows a double exponential rate of change of organization with time. The exact functional dependence has a fine, S-shaped structure, revealing some of the mechanisms of self-organization. The principle of least action helps to explain the mechanism of increase of organization through quantity accumulation and constraint and curvature minimization with an attractor, the least average sum of actions of all elements and for all motions. This approach can help describe, quantify, measure, manage, design and predict future behavior of complex systems to achieve the highest rates of self organization to improve their quality. It can be applied to other complex systems from Physics, Chemistry, Biology, Ecology, Economics, Cities, network theory and others where complex systems are present.
Keywords: Organization, self-organization, complex system, complexification, quantitative measure, principle of least action, principle of stationary action, attractor, progressive development, acceleration, stochastic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1641527 Power System Damping Using Hierarchical Fuzzy Multi- Input Power System Stabilizer and Static VAR Compensator
Authors: Mohammad Hasan Raouf, Ebrahim Rasooli Anarmarzi, Hamid Lesani, Javad Olamaei
Abstract:
This paper proposes the application of a hierarchical fuzzy system (HFS) based on multi-input power system stabilizer (MPSS) and also Static Var Compensator (SVC) in multi-machine environment.The number of rules grows exponentially with the number of variables in a conventional fuzzy logic system. The proposed HFS method is developed to solve this problem. To reduce the number of rules the HFS consists of a number of low-dimensional fuzzy systems in a hierarchical structure. In fact, by using HFS the total number of involved rules increases only linearly with the number of input variables. In the MPSS, to have better efficiency an auxiliary signal of reactive power deviation (ΔQ) is added with ΔP+ Δω input type Power system stabilizer (PSS). Phasor model of SVC is described and used in this paper. The performances of MPSS, Conventional power system stabilizer (CPSS), hierarchical Fuzzy Multi-input Power System Stabilizer (HFMPSS) and the proposed method in damping inter-area mode of oscillation are examined in response to disturbances. By using digital simulations the comparative study is illustrated. It can be seen that the proposed PSS is performing satisfactorily within the whole range of disturbances.
Keywords: Power system stabilizer (PSS), hierarchical fuzzysystem (HFS), Static VAR compensator (SVC)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1526526 Performance Evaluation of Neural Network Prediction for Data Prefetching in Embedded Applications
Authors: Sofien Chtourou, Mohamed Chtourou, Omar Hammami
Abstract:
Embedded systems need to respect stringent real time constraints. Various hardware components included in such systems such as cache memories exhibit variability and therefore affect execution time. Indeed, a cache memory access from an embedded microprocessor might result in a cache hit where the data is available or a cache miss and the data need to be fetched with an additional delay from an external memory. It is therefore highly desirable to predict future memory accesses during execution in order to appropriately prefetch data without incurring delays. In this paper, we evaluate the potential of several artificial neural networks for the prediction of instruction memory addresses. Neural network have the potential to tackle the nonlinear behavior observed in memory accesses during program execution and their demonstrated numerous hardware implementation emphasize this choice over traditional forecasting techniques for their inclusion in embedded systems. However, embedded applications execute millions of instructions and therefore millions of addresses to be predicted. This very challenging problem of neural network based prediction of large time series is approached in this paper by evaluating various neural network architectures based on the recurrent neural network paradigm with pre-processing based on the Self Organizing Map (SOM) classification technique.Keywords: Address, data set, memory, prediction, recurrentneural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1675525 The Relations between Seismic Results and Groundwater near the Gokpinar Damp Area, Denizli, Turkey
Authors: Mahmud Gungor, Ali Aydin, Erdal Akyol, Suat Tasdelen
Abstract:
The understanding of geotechnical characteristics of near-surface material and the effects of the groundwater is very important problem in such as site studies. For showing the relations between seismic data and groundwater, we selected about 25 km2 as the study area. It has been presented which is a detailed work of seismic data and groundwater depths of Gokpinar Damp area. Seismic waves velocity (Vp and Vs) are very important parameters showing the soil properties. The seismic records were used the method of the multichannel analysis of surface waves near area of Gokpinar Damp area. Sixty sites in this area have been investigated with survey lines about 60 m in length. MASW (Multichannel analysis of surface wave) method has been used to generate onedimensional shear wave velocity profile at locations. These shear wave velocities are used to estimate equivalent shear wave velocity in the study area at every 2 and 5 m intervals up to a depth of 45 m. Levels of equivalent shear wave velocity of soil are used the classified of the study area. After the results of the study, it must be considered as components of urban planning and building design of Gokpinar Damp area, Denizli and the application and use of these results should be required and enforced by municipal authorities.
Keywords: Seismic data, Gokpinar Damp, urban planning, Denizli.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2358524 Optimal Placement of Capacitors for Achieve the Best Total Generation Cost by Genetic Algorithm
Authors: Mohammad Reza Tabatabaei, Mohammad Bagher Haddadi, Mojtaba Saeedimoghadam, Ali Vaseghi Ardekani
Abstract:
Economic Dispatch (ED) is one of the most challenging problems of power system since it is difficult to determine the optimum generation scheduling to meet the particular load demand with the minimum fuel costs while all constraints are satisfied. The objective of the Economic Dispatch Problems (EDPs) of electric power generation is to schedule the committed generating units outputs so as to meet the required load demand at minimum operating cost while satisfying all units and system equality and inequality constraints. In this paper, an efficient and practical steady-state genetic algorithm (SSGAs) has been proposed for solving the economic dispatch problem. The objective is to minimize the total generation fuel cost and keep the power flows within the security limits. To achieve that, the present work is developed to determine the optimal location and size of capacitors in transmission power system where, the Participation Factor Algorithm and the Steady State Genetic Algorithm are proposed to select the best locations for the capacitors and determine the optimal size for them.
Keywords: Economic Dispatch, Lagrange, Capacitors Placement, Losses Reduction, Genetic Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1672523 A Bio-Ecological Perspective on Risk Awareness and Factors Associated with Substance Use during Pregnancy in Communities of the Western Cape Province, South Africa
Authors: Mutshinye Manguvhewa, Maria Florence, Mansoo Yu
Abstract:
Substance use among pregnant women is a perennial problem in the Western Cape Province of South Africa. There are many influential elements related with substance use among women of childbearing-age. Factors associated with substance use during pregnancy were explored using qualitative research approach and bio-ecological theoretical framework was utilised to guide the study. Participants were selected using purposive sampling. Participants accessed from the Department of Social Development who met the inclusion criteria of the study were interviewed using semi structured interviews. Participants were referred for psychological intervention during the interview if deemed necessary. Braun and Clarke’s six phases of thematic analysis were used to analyse the data. The study adhered to ethical measures for the participants’ protection. Participants had been knowledgeable about the study earlier than the initiation of the interviews and the important points of their voluntary participation had been explained. The key findings from this study illustrate that social factors, individual area and romantic relationship are the major contributing factors to substance use among pregnant ladies in this sample. Recommendations arising from the study encompass that the stakeholders, rehabilitation centers, Department of Health and future researchers ought to act proactively against substance use all through pregnancy.
Keywords: Bio-ecological factors, pregnancy risk awareness, antenatal care, substance use.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 455522 Experiment Study on the Influence of Tool Materials on the Drilling of Thick Stacked Plate of 2219 Aluminum Alloy
Authors: G. H. Li, M. Liu, H. J. Qi, Q. Zhu, W. Z. He
Abstract:
The drilling and riveting processes are widely used in the assembly of carrier rocket, which makes the efficiency and quality of drilling become the important factor affecting the assembly process. According to the problem existing in the drilling of thick stacked plate (thickness larger than 10mm) of carrier rocket, such as drill break, large noise and burr etc., experimental study of the influence of tool material on the drilling was carried out. The cutting force was measured by a piezoelectric dynamometer, the aperture was measured with an outline projector, and the burr is observed and measured by a digital stereo microscope. Through the measurement, the effects of tool material on the drilling were analyzed from the aspects of drilling force, diameter, and burr. The results show that, compared with carbide drill and coated carbide one, the drilling force of high speed steel is larger. But, the application of high speed steel also has some advantages, e.g. a higher number of hole can be obtained, the height of burr is small, the exit is smooth and the slim burr is less, and the tool experiences wear but not fracture. Therefore, the high speed steel tool is suitable for the drilling of thick stacked plate of 2219 Aluminum alloy.
Keywords: 2219 aluminum alloy, thick stacked plate, drilling, tool material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1283521 A Renovated Cook's Distance Based On The Buckley-James Estimate In Censored Regression
Authors: Nazrina Aziz, Dong Q. Wang
Abstract:
There have been various methods created based on the regression ideas to resolve the problem of data set containing censored observations, i.e. the Buckley-James method, Miller-s method, Cox method, and Koul-Susarla-Van Ryzin estimators. Even though comparison studies show the Buckley-James method performs better than some other methods, it is still rarely used by researchers mainly because of the limited diagnostics analysis developed for the Buckley-James method thus far. Therefore, a diagnostic tool for the Buckley-James method is proposed in this paper. It is called the renovated Cook-s Distance, (RD* i ) and has been developed based on the Cook-s idea. The renovated Cook-s Distance (RD* i ) has advantages (depending on the analyst demand) over (i) the change in the fitted value for a single case, DFIT* i as it measures the influence of case i on all n fitted values Yˆ∗ (not just the fitted value for case i as DFIT* i) (ii) the change in the estimate of the coefficient when the ith case is deleted, DBETA* i since DBETA* i corresponds to the number of variables p so it is usually easier to look at a diagnostic measure such as RD* i since information from p variables can be considered simultaneously. Finally, an example using Stanford Heart Transplant data is provided to illustrate the proposed diagnostic tool.
Keywords: Buckley-James estimators, censored regression, censored data, diagnostic analysis, product-limit estimator, renovated Cook's Distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1438520 Early Warning System of Financial Distress Based On Credit Cycle Index
Authors: Bi-Huei Tsai
Abstract:
Previous studies on financial distress prediction choose the conventional failing and non-failing dichotomy; however, the distressed extent differs substantially among different financial distress events. To solve the problem, “non-distressed”, “slightlydistressed” and “reorganization and bankruptcy” are used in our article to approximate the continuum of corporate financial health. This paper explains different financial distress events using the two-stage method. First, this investigation adopts firm-specific financial ratios, corporate governance and market factors to measure the probability of various financial distress events based on multinomial logit models. Specifically, the bootstrapping simulation is performed to examine the difference of estimated misclassifying cost (EMC). Second, this work further applies macroeconomic factors to establish the credit cycle index and determines the distressed cut-off indicator of the two-stage models using such index. Two different models, one-stage and two-stage prediction models are developed to forecast financial distress, and the results acquired from different models are compared with each other, and with the collected data. The findings show that the one-stage model has the lower misclassification error rate than the two-stage model. The one-stage model is more accurate than the two-stage model.
Keywords: Multinomial logit model, corporate governance, company failure, reorganization, bankruptcy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2680519 Continuous Feature Adaptation for Non-Native Speech Recognition
Authors: Y. Deng, X. Li, C. Kwan, B. Raj, R. Stern
Abstract:
The current speech interfaces in many military applications may be adequate for native speakers. However, the recognition rate drops quite a lot for non-native speakers (people with foreign accents). This is mainly because the nonnative speakers have large temporal and intra-phoneme variations when they pronounce the same words. This problem is also complicated by the presence of large environmental noise such as tank noise, helicopter noise, etc. In this paper, we proposed a novel continuous acoustic feature adaptation algorithm for on-line accent and environmental adaptation. Implemented by incremental singular value decomposition (SVD), the algorithm captures local acoustic variation and runs in real-time. This feature-based adaptation method is then integrated with conventional model-based maximum likelihood linear regression (MLLR) algorithm. Extensive experiments have been performed on the NATO non-native speech corpus with baseline acoustic model trained on native American English. The proposed feature-based adaptation algorithm improved the average recognition accuracy by 15%, while the MLLR model based adaptation achieved 11% improvement. The corresponding word error rate (WER) reduction was 25.8% and 2.73%, as compared to that without adaptation. The combined adaptation achieved overall recognition accuracy improvement of 29.5%, and WER reduction of 31.8%, as compared to that without adaptation. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3217518 Principle Components Updates via Matrix Perturbations
Authors: Aiman Elragig, Hanan Dreiwi, Dung Ly, Idriss Elmabrook
Abstract:
This paper highlights a new approach to look at online principle components analysis (OPCA). Given a data matrix X ∈ R,^m x n we characterise the online updates of its covariance as a matrix perturbation problem. Up to the principle components, it turns out that online updates of the batch PCA can be captured by symmetric matrix perturbation of the batch covariance matrix. We have shown that as n→ n0 >> 1, the batch covariance and its update become almost similar. Finally, utilize our new setup of online updates to find a bound on the angle distance of the principle components of X and its update.Keywords: Online data updates, covariance matrix, online principle component analysis (OPCA), matrix perturbation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1038517 Performance Optimization of Data Mining Application Using Radial Basis Function Classifier
Authors: M. Govindarajan, R. M.Chandrasekaran
Abstract:
Text data mining is a process of exploratory data analysis. Classification maps data into predefined groups or classes. It is often referred to as supervised learning because the classes are determined before examining the data. This paper describes proposed radial basis function Classifier that performs comparative crossvalidation for existing radial basis function Classifier. The feasibility and the benefits of the proposed approach are demonstrated by means of data mining problem: direct Marketing. Direct marketing has become an important application field of data mining. Comparative Cross-validation involves estimation of accuracy by either stratified k-fold cross-validation or equivalent repeated random subsampling. While the proposed method may have high bias; its performance (accuracy estimation in our case) may be poor due to high variance. Thus the accuracy with proposed radial basis function Classifier was less than with the existing radial basis function Classifier. However there is smaller the improvement in runtime and larger improvement in precision and recall. In the proposed method Classification accuracy and prediction accuracy are determined where the prediction accuracy is comparatively high.Keywords: Text Data Mining, Comparative Cross-validation, Radial Basis Function, runtime, accuracy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1555516 Indicators as Early Warning Signal Performance to Solve Underlying Safety Problem before They Emerge as Accident Risks
Authors: Benson Chizubem
Abstract:
Because of the severe hazards that substantially impact workers' lives and assets lost, the oil and gas industry has established a goal of establishing zero occurrences or accidents in operations. Using leading indicators to measure and assess an organization's safety performance is a proactive approach to safety management. Also, it will provide early warning signals to solve inherent safety issues before they lead to an accident in the study industry. The analysis of these indicators' performance was based on a questionnaire-based methodology. A total number of 1000 questionnaires were disseminated to the workers, of which 327 were returned to the researcher team. The data collected were analysed to evaluate their safety perceptions on indicators performance. Data analysis identified safety training, safety system, safety supervision, safety rules and procedures, safety auditing, strategies and policies, management commitment, safety meeting and safety behaviour, as potential leading indicators that are capable of measuring organizational safety performance and as capable of providing early warning signals of weak safety area in an operational environment. The findings of this study have provided safety researchers and industrial safety practitioners with helpful information on the improvement of the existing safety monitoring process in the oil and gas industry, both locally and globally, as proactive actions.
Keywords: Early warning, safety, accident risks, oil and gas industry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 373515 The Development of Online-Class Scheduling Management System Conducted by the Case Study of Department of Social Science: Faculty of Humanities and Social Sciences Suan Sunandha Rajabhat University
Authors: Wipada Chaiwchan, Patcharee Klinhom
Abstract:
This research is aimed to develop the online-class scheduling management system and improve as a complex problem solution, this must take into consideration in various conditions and factors. In addition to the number of courses, the number of students and a timetable to study, the physical characteristics of each class room and regulations used in the class scheduling must also be taken into consideration. This system is developed to assist management in the class scheduling for convenience and efficiency. It can provide several instructors to schedule simultaneously. Both lecturers and students can check and publish a timetable and other documents associated with the system online immediately. It is developed in a web-based application. PHP is used as a developing tool. The database management system was MySQL. The tool that is used for efficiency testing of the system is questionnaire. The system was evaluated by using a Black-Box testing. The sample was composed of 2 groups: 5 experts and 100 general users. The average and the standard deviation of results from the experts were 3.50 and 0.67. The average and the standard deviation of results from the general users were 3.54 and 0.54. In summary, the results from the research indicated that the satisfaction of users were in a good level. Therefore, this system could be implemented in an actual workplace and satisfy the users’ requirement effectively.
Keywords: Timetable, schedule, management system, online.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6404514 Artificial Neural Networks Technique for Seismic Hazard Prediction Using Seismic Bumps
Authors: Belkacem Selma, Boumediene Selma, Samira Chouraqui, Hanifi Missoum, Tourkia Guerzou
Abstract:
Natural disasters have occurred and will continue to cause human and material damage. Therefore, the idea of "preventing" natural disasters will never be possible. However, their prediction is possible with the advancement of technology. Even if natural disasters are effectively inevitable, their consequences may be partly controlled. The rapid growth and progress of artificial intelligence (AI) had a major impact on the prediction of natural disasters and risk assessment which are necessary for effective disaster reduction. Earthquake prediction to prevent the loss of human lives and even property damage is an important factor; that, is why it is crucial to develop techniques for predicting this natural disaster. This study aims to analyze the ability of artificial neural networks (ANNs) to predict earthquakes that occur in a given area. The used data describe the problem of high energy (higher than 104 J) seismic bumps forecasting in a coal mine using two long walls as an example. For this purpose, seismic bumps data obtained from mines have been analyzed. The results obtained show that the ANN is able to predict earthquake parameters with high accuracy; the classification accuracy through neural networks is more than 94%, and the models developed are efficient and robust and depend only weakly on the initial database.
Keywords: Earthquake prediction, artificial intelligence, AI, Artificial Neural Network, ANN, seismic bumps.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1187513 Evaluation of Heavy Metal Concentrations of Stem and Seed of Juncus acutus for Grazing Animals and Birds in Kızılırmak Delta
Authors: N. Cetinkaya, F. Erdem
Abstract:
Juncus acutus (Juncaceae) is a perennial wetland plant and it is commonly known as spiny rush or sharp rush. It is the most abundant plant in Kizilirmak grassland, Samsun, Turkey. Heavy metals are significant environmental contaminants in delta and their toxicity is an increasing problem for animals whose natural habitat is delta. The objective of this study was to evaluate heavy metal concentrations mainly As, Cd, Sb, Ba, Pb and Hg in stem and seed of Juncus acutus for grazing animals and birds in delta. The Juncus acutus stem and seed samples were collected from Kizilirmak Delta in July, August and September. Heavy metal concentrations of collected samples were analyzed by Inductively Coupled Plasma – Mass Spectrometer (ICP-MS). The obtained mean values of three months for As, Cd, Sb, Ba, Pb and Hg of stem and seed samples of Juncus acutus were 0.11 and 0.23 mg/kg; 0.07 and 0.11 mg/kg; 0.02 and 0.02 mg/kg; 5.26 and 1.75 mg/kg; 0.05 and not detectable in July respectively. Hg was not detected in both stem and seed of Juncus acutus, Pb concentration was determined only in stem of Juncus acutus but not in seed. There were no significant differences between the values of three months for As, Cd, Sb, Ba, Pb and Hg of stem and seed samples of Juncus acutus. The obtained As, Cd, Sb, Ba, Pb and Hg results of stem and seed of Juncus acutus show that seed and stem of Juncus acutus may be safely consumed for grazing animals and birds regarding to heavy metals contamination in Kizilirmak Delta.
Keywords: Heavy metals, Juncus acutus, Kizilirmak Delta, wetland.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 721512 Investigating Iraqi EFL Undergraduates' Performance in the Production of Number Forms in English
Authors: Adnan Z. Mkhelif
Abstract:
The production of number forms in English tends to be problematic for Iraqi learners of English as a foreign language (EFL), even at the undergraduate level. To help better understand and consequently address this problem, it is important to identify its sources. This study aims at: (1) statistically analysing Iraqi EFL undergraduates' performance in the production of number forms in English; (2) classifying learners' errors in terms of their possible major causes; and (3) outlining some pedagogical recommendations relevant to the teaching of number forms in English. It is hypothesized in this study that (1) Iraqi EFL undergraduates still face problems in the production of number forms in English and (2) errors pertaining to the context of learning are more numerous than those attributable to the other possible causes. After reviewing the literature available on the topic, a written test comprising 50 items has been constructed and administered to a randomly chosen sample of 50 second-year college students from the Department of English, College of Education, Wasit University. The findings of the study showed that Iraqi EFL undergraduates still face problems in the production of number forms in English and that the possible major sources of learners’ errors can be arranged hierarchically in terms of the percentages of errors to which they can be ascribed as follows: (1) context of learning (50%), (2) intralingual transfer (37%), and (3) interlingual transfer (13%). It is hoped that the implications of the study findings will be beneficial to researchers, syllabus designers, as well as teachers of English as a foreign/second language.
Keywords: L2 morphology, L2 number forms, L2 vocabulary learning, productive knowledge.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 655511 Addressing Scalability Issues of Named Entity Recognition Using Multi-Class Support Vector Machines
Authors: Mona Soliman Habib
Abstract:
This paper explores the scalability issues associated with solving the Named Entity Recognition (NER) problem using Support Vector Machines (SVM) and high-dimensional features. The performance results of a set of experiments conducted using binary and multi-class SVM with increasing training data sizes are examined. The NER domain chosen for these experiments is the biomedical publications domain, especially selected due to its importance and inherent challenges. A simple machine learning approach is used that eliminates prior language knowledge such as part-of-speech or noun phrase tagging thereby allowing for its applicability across languages. No domain-specific knowledge is included. The accuracy measures achieved are comparable to those obtained using more complex approaches, which constitutes a motivation to investigate ways to improve the scalability of multiclass SVM in order to make the solution more practical and useable. Improving training time of multi-class SVM would make support vector machines a more viable and practical machine learning solution for real-world problems with large datasets. An initial prototype results in great improvement of the training time at the expense of memory requirements.Keywords: Named entity recognition, support vector machines, language independence, bioinformatics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1690510 PSS with Multiple FACTS Controllers Coordinated Design and Real-Time Implementation Using Advanced Adaptive PSO
Authors: Rajendraprasad Narne, P. C. Panda
Abstract:
In this article, coordinated tuning of power system stabilizer (PSS) with static var compensator (SVC) and thyristor controlled series capacitor (TCSC) in multi-machine power system is proposed. The design of proposed coordinated damping controller is formulated as an optimization problem and the controller gains are optimized instantaneously using advanced adaptive particle swarm optimization (AAPSO). The objective function is framed with the inter-area speed deviations of the generators and it is minimized using AAPSO to improve the dynamic stability of power system under severe disturbance. The proposed coordinated controller performance is evaluated under a wide range of system operating conditions with three-phase fault disturbance. Using time domain simulations the damping characteristics of proposed controller is compared with individually tuned PSS, SVC and TCSC controllers. Finally, the real-time simulations are carried out in Opal-RT hardware simulator to synchronize the proposed controller performance in the real world.
Keywords: Advanced adaptive particle swarm optimization, Coordinated design, Power system stabilizer, Real-time implementation, static var compensator, Thyristor controlled series capacitor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2591509 Corporate Environmentalism: A Case Study in the Czech Republic
Authors: Pavel Adámek
Abstract:
This study examines perception of environmental approach in small and medium-sized enterprises (SMEs) – the process by which firms integrate environmental concern into business. Based on a review of the literature, the paper synthesizes focus on environmental issues with the reflection in a case study in the Czech Republic. Two themes of corporate environmentalism are discussed – corporate environmental orientation and corporate stances toward environmental concerns. It provides theoretical material on greening organizational culture that is helpful in understanding the response of contemporary business to environmental problems. We integrate theoretical predictions with empirical findings confronted with reality. Scales to measure these themes are tested in a survey of managers in 229 Czech firms. We used the process of in-depth questioning. The research question was derived and answered in the context of the corresponding literature and conducted research. A case study showed us that environmental approach is variety different (depending on the size of the firm) in SMEs sector. The results of the empirical mapping demonstrate Czech company’s approach to environment and define the problem areas and pinpoint the main limitation in the expansion of environmental aspects. We contribute to the debate for recognition of the particular role of environmental issues in business reality.
Keywords: Corporate environmentalism, Czech Republic, empirical mapping, environmental performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2085508 Comparison of Router Intelligent and Cooperative Host Intelligent Algorithms in a Continuous Model of Fixed Telecommunication Networks
Authors: Dávid Csercsik, Sándor Imre
Abstract:
The performance of state of the art worldwide telecommunication networks strongly depends on the efficiency of the applied routing mechanism. Game theoretical approaches to this problem offer new solutions. In this paper a new continuous network routing model is defined to describe data transfer in fixed telecommunication networks of multiple hosts. The nodes of the network correspond to routers whose latency is assumed to be traffic dependent. We propose that the whole traffic of the network can be decomposed to a finite number of tasks, which belong to various hosts. To describe the different latency-sensitivity, utility functions are defined for each task. The model is used to compare router and host intelligent types of routing methods, corresponding to various data transfer protocols. We analyze host intelligent routing as a transferable utility cooperative game with externalities. The main aim of the paper is to provide a framework in which the efficiency of various routing algorithms can be compared and the transferable utility game arising in the cooperative case can be analyzed.
Keywords: Routing, Telecommunication networks, Performance evaluation, Cooperative game theory, Partition function form games
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1852