Search results for: distribution process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6988

Search results for: distribution process

6298 Concentrations and History of Heavy Metals in Sediment Cores: Geochemistry and Geochronology Using 210Pb

Authors: F. Fernandes, C. Poleto

Abstract:

This paper aims at assessing the concentrations of heavy metals and the isotopic composition of lead 210Pb in different fractions of sediment produced in the watershed that makes up the Mãe d'água dam and thus characterizing the distribution of metals along the sedimentary column and inferencing in the urbanization of the same process. Sample collection was carried out in June 2014; eight sediment cores were sampled in the lake of the dam. For extraction of the sediments core, a core sampler “Piston Core” was used. The trace metal concentrations were determined by conventional atomic absorption spectrophotometric methods. The samples were subjected to radiochemical analysis of 210Po. 210Pb activity was obtained by measuring 210Po activity. The chronology was calculated using the constant rate of supply (CRS). 210Pb is used to estimate the sedimentation rate.

Keywords: 210Pb dating method, heavy metal, lakes urban, pollution history.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1090
6297 Experimental Investigation of On-Body Channel Modelling at 2.45 GHz

Authors: Hasliza A. Rahim, Fareq Malek, Nur A. M. Affendi, Azuwa Ali, Norshafinash Saudin, Latifah Mohamed

Abstract:

This paper presents the experimental investigation of on-body channel fading at 2.45 GHz considering two effects of the user body movement; stationary and mobile. A pair of body-worn antennas was utilized in this measurement campaign. A statistical analysis was performed by comparing the measured on-body path loss to five well-known distributions; lognormal, normal, Nakagami, Weibull and Rayleigh. The results showed that the average path loss of moving arm varied higher than the path loss in sitting position for upper-arm-to-left-chest link, up to 3.5 dB. The analysis also concluded that the Nakagami distribution provided the best fit for most of on-body static link path loss in standing still and sitting position, while the arm movement can be best described by log-normal distribution.

Keywords: On-Body channel communications, fading characteristics, statistical model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1526
6296 A Study of Dose Distribution and Image Quality under an Automatic Tube Current Modulation (ATCM) System for a Toshiba Aquilion 64 CT Scanner Using a New Design of Phantom

Authors: S. Sookpeng, C. J. Martin, D. J. Gentle

Abstract:

Automatic tube current modulation (ATCM) systems are available for all CT manufacturers and are used for the majority of patients. Understanding how the systems work and their influence on patient dose and image quality is important for CT users, in order to gain the most effective use of the systems. In the present study, a new phantom was used for evaluating dose distribution and image quality under the ATCM operation for the Toshiba Aquilion 64 CT scanner using different ATCM options and a fixed mAs technique. A routine chest, abdomen and pelvis (CAP) protocol was selected for study and Gafchromic film was used to measure entrance surface dose (ESD), peripheral dose and central axis dose in the phantom. The results show the dose reductions achievable with various ATCM options, in relation with the target noise. The doses and image noise distribution were more uniform when the ATCM system was implemented compared with the fixed mAs technique. The lower limit set for the tube current will affect the modulations especially for the lower dose option. This limit prevented the tube current being reduced further and therefore the lower dose ATCM setting resembled a fixed mAs technique. Selection of a lower tube current limit is likely to reduce doses for smaller patients in scans of chest and neck regions.

Keywords: Computed Tomography (CT), Automatic Tube Current Modulation (ATCM), Automatic Exposure Control (AEC).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2612
6295 Heavy Metal Contamination of the Landscape at the ─¢ubietová Deposit (Slovakia)

Authors: Peter Andráš, Adam Lichý, Jana Rusková, Lenka Matúšková

Abstract:

The heavy metal contamination of the technogenous sediments and soils at the investigated dump-field show irregular planar distribution. Also the heavy metal content in the surface water, drainage water and in the groundwater was studied both in the dry as well as during the rainy periods. The cementation process causes substitution of iron by copper. Natural installation and development of plant species was observed at the old mine waste dumps, specific to the local chemical conditions such as low content of essential nutrients and high content of heavy metals. The individual parts of the plant tissues (roots, branches/stems, leaves/needles, flowers/ fruits) are contaminated by heavy metals and tissues are damaged differently, respectively.

Keywords: Contamination, dump-field, heavy metals, plants, sediment, water.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3060
6294 A Hybrid Approach to Fault Detection and Diagnosis in a Diesel Fuel Hydrotreatment Process

Authors: Salvatore L., Pires B., Campos M. C. M., De Souza Jr M. B.

Abstract:

It is estimated that the total cost of abnormal conditions to US process industries is around $20 billion dollars in annual losses. The hydrotreatment (HDT) of diesel fuel in petroleum refineries is a conversion process that leads to high profitable economical returns. However, this is a difficult process to control because it is operated continuously, with high hydrogen pressures and it is also subject to disturbances in feed properties and catalyst performance. So, the automatic detection of fault and diagnosis plays an important role in this context. In this work, a hybrid approach based on neural networks together with a pos-processing classification algorithm is used to detect faults in a simulated HDT unit. Nine classes (8 faults and the normal operation) were correctly classified using the proposed approach in a maximum time of 5 minutes, based on on-line data process measurements.

Keywords: Fault detection, hydrotreatment, hybrid systems, neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1627
6293 The System Identification and PID Lead-lag Control for Two Poles Unstable SOPDT Process by Improved Relay Method

Authors: V. K. Singh, P. K. Padhy

Abstract:

This paper describes identification of the two poles unstable SOPDT process, especially with large time delay. A new modified relay feedback identification method for two poles unstable SOPDT process is proposed. Furthermore, for the two poles unstable SOPDT process, an additional Derivative controller is incorporated parallel with relay to relax the constraint on the ratio of delay to the unstable time constant, so that the exact model parameters of unstable processes can be identified. To cope with measurement noise in practice, a low pass filter is suggested to get denoised output signal toimprove the exactness of model parameter of unstable process. PID Lead-lag tuning formulas are derived for two poles unstable (SOPDT) processes based on IMC principle. Simulation example illustrates the effectiveness and the simplicity of the proposed identification and control method.

Keywords: IMC structure, PID Lead-lag controller, Relayfeedback, SOPDT

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2085
6292 Identifying Significant Factors of Brick Laying Process through Design of Experiment and Computer Simulation: A Case Study

Authors: M. H. Zarei, A. Nikakhtar, A. H. Roudsari, N. Madadi, K. Y. Wong

Abstract:

Improving performance measures in the construction processes has been a major concern for managers and decision makers in the industry. They seek for ways to recognize the key factors which have the largest effect on the process. Identifying such factors can guide them to focus on the right parts of the process in order to gain the best possible result. In the present study design of experiment (DOE) has been applied to a computer simulation model of brick laying process to determine significant factors while productivity has been chosen as the response of the experiment. To this end, four controllable factors and their interaction have been experimented and the best factor level has been calculated for each one. The results indicate that three factors, namely, labor of brick, labor of mortar and inter arrival time of mortar along with interaction of labor of brick and labor of mortar are significant.

Keywords: Brick laying process, computer simulation, design of experiment, significant factors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2079
6291 Research on Weakly Hard Real-Time Constraints and Their Boolean Combination to Support Adaptive QoS

Authors: Xiangbin Zhu

Abstract:

Advances in computing applications in recent years have prompted the demand for more flexible scheduling models for QoS demand. Moreover, in practical applications, partly violated temporal constraints can be tolerated if the violation meets certain distribution. So we need extend the traditional Liu and Lanland model to adapt to these circumstances. There are two extensions, which are the (m, k)-firm model and Window-Constrained model. This paper researches on weakly hard real-time constraints and their combination to support QoS. The fact that a practical application can tolerate some violations of temporal constraint under certain distribution is employed to support adaptive QoS on the open real-time system. The experiment results show these approaches are effective compared to traditional scheduling algorithms.

Keywords: Weakly Hard Real-Time, Real-Time, Scheduling, Quality of Service.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567
6290 Automation of Heat Exchanger using Neural Network

Authors: Sudhir Agashe, Ashok Ghatol, Sujata Agashe

Abstract:

In this paper the development of a heat exchanger as a pilot plant for educational purpose is discussed and the use of neural network for controlling the process is being presented. The aim of the study is to highlight the need of a specific Pseudo Random Binary Sequence (PRBS) to excite a process under control. As the neural network is a data driven technique, the method for data generation plays an important role. In light of this a careful experimentation procedure for data generation was crucial task. Heat exchange is a complex process, which has a capacity and a time lag as process elements. The proposed system is a typical pipe-in- pipe type heat exchanger. The complexity of the system demands careful selection, proper installation and commissioning. The temperature, flow, and pressure sensors play a vital role in the control performance. The final control element used is a pneumatically operated control valve. While carrying out the experimentation on heat exchanger a welldrafted procedure is followed giving utmost attention towards safety of the system. The results obtained are encouraging and revealing the fact that if the process details are known completely as far as process parameters are concerned and utilities are well stabilized then feedback systems are suitable, whereas neural network control paradigm is useful for the processes with nonlinearity and less knowledge about process. The implementation of NN control reinforces the concepts of process control and NN control paradigm. The result also underlined the importance of excitation signal typically for that process. Data acquisition, processing, and presentation in a typical format are the most important parameters while validating the results.

Keywords: Process identification, neural network, heat exchanger.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1556
6289 Flow Characteristics of Pulp Liquid in Straight Ducts

Authors: M. Sumida

Abstract:

An experimental investigation was performed on pulp liquid flow in straight ducts with a square cross section. Fully developed steady flow was visualized and the fiber concentration was obtained using a light-section method developed by the author et al. The obtained results reveal quantitatively, in a definite form, the distribution of the fiber concentration. From the results and measurements of pressure loss, it is found that the flow characteristics of pulp liquid in ducts can be classified into five patterns. The relationships among the distributions of mean and fluctuation of fiber concentration, the pressure loss and the flow velocity are discussed, and then the features for each pattern are extracted. The degree of nonuniformity of the fiber concentration, which is indicated by the standard deviation of its distribution, is decreased from 0.3 to 0.05 with an increase in the velocity of the tested pulp liquid from 0.4 to 0.8%.

Keywords: Fiber Concentration, Flow Characteristic, Pulp Liquid, Straight Duct.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1564
6288 Understanding and Predicting Foam in Anaerobic Digester

Authors: I. R. Kanu, T. J. Aspray, A. J. Adeloye

Abstract:

As a result of the ambiguity and complexity surrounding anaerobic digester foaming, efforts have been made by various researchers to understand the process of anaerobic digester foaming so as to proffer a solution that can be universally applied rather than site specific. All attempts ranging from experimental analysis to comparative review of other process has not fully explained the conditions and process of foaming in anaerobic digester. Studying the current available knowledge on foam formation and relating it to anaerobic digester process and operating condition, this piece of work presents a succinct and enhanced understanding of foaming in anaerobic digesters as well as introducing a simple method to identify the onset of anaerobic digester foaming based on analysis of historical data from a field scale system.

Keywords: Anaerobic digester, foam, biogas, surfactants, wastewater sludge.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2948
6287 Neutral to Earth Voltage Analysis in Harmonic Polluted Distribution Networks with Multi- Grounded Neutrals

Authors: G. Ahmadi, S.M. Shahrtash

Abstract:

A multiphase harmonic load flow algorithm is developed based on backward/forward sweep to examine the effects of various factors on the neutral to earth voltage (NEV), including unsymmetrical system configuration, load unbalance and harmonic injection. The proposed algorithm composes fundamental frequency and harmonic frequencies power flows. The algorithm and the associated models are tested on IEEE 13 bus system. The magnitude of NEV is investigated under various conditions of the number of grounding rods per feeder lengths, the grounding rods resistance and the grounding resistance of the in feeding source. Additionally, the harmonic injection of nonlinear loads has been considered and its influences on NEV under different conditions are shown.

Keywords: NEV, Distribution System, Multi-grounded, Backward/Forward Sweep, Harmonic Analysis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048
6286 Detergent Removal from Rinsing Water by Peroxi Electrocoagulation Process

Authors: A. Benhadji, M. Taleb Ahmed

Abstract:

Among the various methods of treatment, advanced oxidation processes (AOP) are the most promising ones. In this study, Peroxi Electrocoagulation Process (PEP) was investigated for the treatment of detergent wastewater. The process was compared with electrooxidation treatment. The results showed that chemical oxygen demand (COD) was high 7584 mgO2.L-1, while the biochemical oxygen demand was low (250 mgO2.L-1). This wastewater was hardly biodegradable. Electrochemical process was carried out for the removal of detergent using a glass reactor with a volume of 1 L and fitted with three electrodes. A direct current (DC) supply was used. Samples were taken at various current density (0.0227 A/cm2 to 0.0378 A/cm2) and reaction time (1-2-3-4 and 5 hour). Finally, the COD was determined. The results indicated that COD removal efficiency of PEP was observed to increase with current intensity and reached to 77% after 5 h. The highest removal efficiency was observed after 5 h of treatment.

Keywords: Advanced oxidation processes, chemical oxygen demand, COD, detergent, peroxi electrocoagulation process, PEP, wastewater

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 890
6285 The Effects of Asymmetric Bracing on Steel Structures under Seismic Loads

Authors: Mahmoud Miri, Soleiman Maramaee

Abstract:

Because of architectural condition and structure application, sometimes mass source and stiffness source are not coincidence, and the structure is irregular. The structure is also might be asymmetric as an asymmetric bracing in plan which leads to unbalance distribution of stiffness or because of unbalance distribution of the mass. Both condition lead to eccentricity and torsion in the structure. The deficiency of ordinary code to evaluate the performance of steel structures against earthquake has been caused designing based on performance level or capacity spectrum be used. By using the mentioned methods it is possible to design a structure that its behavior against different earthquakes be predictive. In this article 5- story buildings with different percentage of asymmetric which is because of stiffness changes have been designed. The static and dynamic nonlinear analysis under three acceleration recording has been done. Finally performance level of the structure has been evaluated.

Keywords: Seismic analysis, torsion, asymmetric, irregular building, stiffness source.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2156
6284 Weight Comparison of Oil and Dry Type Distribution Transformers

Authors: Murat Toren, Mehmet Çelebi

Abstract:

Reducing the weight of transformers while providing good performance, cost reduction and increased efficiency is important. Weight is one of the most significant factors in all electrical machines, and as such, many transformer design parameters are related to weight calculations. This study presents a comparison of the weight of oil type transformers and dry type transformer weight. Oil type transformers are mainly used in industry; however, dry type transformers are becoming more widespread in recent years. MATLAB is typically used for designing transformers and design parameters (rated voltages, core loss, etc.) along with design in ANSYS Maxwell. Similar to other studies, this study presented that the dry type transformer option is limited. Moreover, the commonly-used 50 kVA distribution transformers in the industry are oil type and dry type transformers are designed and considered in terms of weight. Currently, the preference for low-cost oil-type transformers would change if costs for dry-type transformer were more competitive. The aim of this study was to compare the weight of transformers, which is a substantial cost factor, and to provide an evaluation about increasing the use of dry type transformers.

Keywords: Weight, oil-type transformers, dry-type transformers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3291
6283 Dynamic Economic Dispatch Constrained by Wind Power Weibull Distribution: A Here-and-Now Strategy

Authors: Mostafa A. Elshahed, Magdy M. Elmarsfawy, Hussain M. Zain Eldain

Abstract:

In this paper, a Dynamic Economic Dispatch (DED) model is developed for the system consisting of both thermal generators and wind turbines. The inclusion of a significant amount of wind energy into power systems has resulted in additional constraints on DED to accommodate the intermittent nature of the output. The probability of stochastic wind power based on the Weibull probability density function is included in the model as a constraint; A Here-and-Now Approach. The Environmental Protection Agency-s hourly emission target, which gives the maximum emission during the day, is used as a constraint to reduce the atmospheric pollution. A 69-bus test system with non-smooth cost function is used to illustrate the effectiveness of the proposed model compared with static economic dispatch model with including the wind power.

Keywords: Dynamic Economic Dispatch, StochasticOptimization, Weibull Distribution, Wind Power

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2943
6282 Modeling And Analysis of Simple Open Cycle Gas Turbine Using Graph Networks

Authors: Naresh Yadav, I.A. Khan, Sandeep Grover

Abstract:

This paper presents a unified approach based graph theory and system theory postulates for the modeling and analysis of Simple open cycle Gas turbine system. In the present paper, the simple open cycle gas turbine system has been modeled up to its subsystem level and system variables have been identified to develop the process subgraphs. The theorems and algorithms of the graph theory have been used to represent behavioural properties of the system like rate of heat and work transfers rates, pressure drops and temperature drops in the involved processes of the system. The processes have been represented as edges of the process subgraphs and their limits as the vertices of the process subgraphs. The system across variables and through variables has been used to develop terminal equations of the process subgraphs of the system. The set of equations developed for vertices and edges of network graph are used to solve the system for its process variables.

Keywords: Simple open cycle gas turbine, Graph theoretic approach, process subgraphs, gas turbines system modeling, systemtheory

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2612
6281 Design of the Large Dimension Cold Shield Cooled by G-M Cryocooler

Authors: Gong Jie, Yu Qianxu, Liu Min, Shan Weiwei

Abstract:

The design of methods of the 20 K large dimension cold shield used for infrared radiation demarcating in space environment simulation test were introduced in this paper. The cold shield were cooled by five G-M cryocoolers , and the dimension of the cold shield is the largest in our country.Cold shield installation and distribution and compensator for contraction on cooling were introduced detailedly. The temperature distribution and cool-down time of cold shield surface were also calculated and analysed in this paper. The design of cold shield resolves the difficulty of compensator for contraction on cooling successfully. Test results show that the actual technical performance indicators of cold shield met and exceeded the design requirements.

Keywords: cold shield, G-M cryocooler,infrared radiometer demarcating, satellite, space environment simulation equipments

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1671
6280 Change Management in Business Process Modeling Based on Object Oriented Petri Net

Authors: Bassam Atieh Rajabi, Sai Peck Lee

Abstract:

Business Process Modeling (BPM) is the first and most important step in business process management lifecycle. Graph based formalism and rule based formalism are the two most predominant formalisms on which process modeling languages are developed. BPM technology continues to face challenges in coping with dynamic business environments where requirements and goals are constantly changing at the execution time. Graph based formalisms incur problems to react to dynamic changes in Business Process (BP) at the runtime instances. In this research, an adaptive and flexible framework based on the integration between Object Oriented diagramming technique and Petri Net modeling language is proposed in order to support change management techniques for BPM and increase the representation capability for Object Oriented modeling for the dynamic changes in the runtime instances. The proposed framework is applied in a higher education environment to achieve flexible, updatable and dynamic BP.

Keywords: Business Process Modeling, Change Management, Graph Based Modeling, Rule Based Modeling, Object Oriented PetriNet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2017
6279 Stress Distribution in Axisymmetric Indentation of an Elastic Layer-Substrate Body

Authors: Kotaro Miura, Makoto Sakamoto, Yuji Tanabe

Abstract:

We focus on internal stress and displacement of an elastic axisymmetric contact problem for indentation of a layer-substrate body. An elastic layer is assumed to be perfectly bonded to an elastic semi-infinite substrate. The elastic layer is smoothly indented with a flat-ended cylindrical indenter. The analytical and exact solutions were obtained by solving an infinite system of simultaneous equations using the method to express a normal contact stress at the upper surface of the elastic layer as an appropriate series. This paper presented the numerical results of internal stress and displacement distributions for hard-coating system with constant values of Poisson’s ratio and the thickness of elastic layer.

Keywords: Indentation, contact problem, stress distribution, coating materials, layer-substrate body.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 800
6278 Developing New Processes and Optimizing Performance Using Response Surface Methodology

Authors: S. Raissi

Abstract:

Response surface methodology (RSM) is a very efficient tool to provide a good practical insight into developing new process and optimizing them. This methodology could help engineers to raise a mathematical model to represent the behavior of system as a convincing function of process parameters. Through this paper the sequential nature of the RSM surveyed for process engineers and its relationship to design of experiments (DOE), regression analysis and robust design reviewed. The proposed four-step procedure in two different phases could help system analyst to resolve the parameter design problem involving responses. In order to check accuracy of the designed model, residual analysis and prediction error sum of squares (PRESS) described. It is believed that the proposed procedure in this study can resolve a complex parameter design problem with one or more responses. It can be applied to those areas where there are large data sets and a number of responses are to be optimized simultaneously. In addition, the proposed procedure is relatively simple and can be implemented easily by using ready-made standard statistical packages.

Keywords: Response Surface Methodology (RSM), Design of Experiments (DOE), Process modeling, Process setting, Process optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1822
6277 A Statistical Approach for Predicting and Optimizing Depth of Cut in AWJ Machining for 6063-T6 Al Alloy

Authors: Farhad Kolahan, A. Hamid Khajavi

Abstract:

In this paper, a set of experimental data has been used to assess the influence of abrasive water jet (AWJ) process parameters in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. The effects of these input parameters are studied on depth of cut (h); one of most important characteristics of AWJ. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. In the next stage, the proposed model is embedded into a Simulated Annealing (SA) algorithm to optimize the AWJ process parameters. The objective is to determine a suitable set of process parameters that can produce a desired depth of cut, considering the ranges of the process parameters. Computational results prove the effectiveness of the proposed model and optimization procedure.

Keywords: AWJ machining, Mathematical modeling, Simulated Annealing, Optimization

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1761
6276 Distributed Automation System Based Remote Monitoring of Power Quality Disturbance on LV Network

Authors: Emmanuel D. Buedi, K. O. Boateng, Griffith S. Klogo

Abstract:

Electrical distribution networks are prone to power quality disturbances originating from the complexity of the distribution network, mode of distribution (overhead or underground) and types of loads used by customers. Data on the types of disturbances present and frequency of occurrence is needed for economic evaluation and hence finding solution to the problem. Utility companies have resorted to using secondary power quality devices such as smart meters to help gather the required data. Even though this approach is easier to adopt, data gathered from these devices may not serve the required purpose, since the installation of these devices in the electrical network usually does not conform to available PQM placement methods. This paper presents a design of a PQM that is capable of integrating into an existing DAS infrastructure to take advantage of available placement methodologies. The monitoring component of the design is implemented and installed to monitor an existing LV network. Data from the monitor is analyzed and presented. A portion of the LV network of the Electricity Company of Ghana is modeled in MATLAB-Simulink and analyzed under various earth fault conditions. The results presented show the ability of the PQM to detect and analyze PQ disturbance such as voltage sag and overvoltage. By adopting a placement methodology and installing these nodes, utilities are assured of accurate and reliable information with respect to the quality of power delivered to consumers.

Keywords: Power quality, remote monitoring, distributed automation system, economic evaluation, LV network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1114
6275 Clustering of Variables Based On a Probabilistic Approach Defined on the Hypersphere

Authors: Paulo Gomes, Adelaide Figueiredo

Abstract:

We consider n individuals described by p standardized variables, represented by points of the surface of the unit hypersphere Sn-1. For a previous choice of n individuals we suppose that the set of observables variables comes from a mixture of bipolar Watson distribution defined on the hypersphere. EM and Dynamic Clusters algorithms are used for identification of such mixture. We obtain estimates of parameters for each Watson component and then a partition of the set of variables into homogeneous groups of variables. Additionally we will present a factor analysis model where unobservable factors are just the maximum likelihood estimators of Watson directional parameters, exactly the first principal component of data matrix associated to each group previously identified. Such alternative model it will yield us to directly interpretable solutions (simple structure), avoiding factors rotations.

Keywords: Dynamic Clusters algorithm, EM algorithm, Factor analysis model, Hierarchical Clustering, Watson distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1608
6274 Electroencephalography Based Brain-Computer Interface for Cerebellum Impaired Patients

Authors: Young-Seok Choi

Abstract:

In healthy humans, the cortical brain rhythm shows specific mu (~6-14 Hz) and beta (~18-24 Hz) band patterns in the cases of both real and imaginary motor movements. As cerebellar ataxia is associated with impairment of precise motor movement control as well as motor imagery, ataxia is an ideal model system in which to study the role of the cerebellocortical circuit in rhythm control. We hypothesize that the EEG characteristics of ataxic patients differ from those of controls during the performance of a Brain-Computer Interface (BCI) task. Ataxia and control subjects showed a similar distribution of mu power during cued relaxation. During cued motor imagery, however, the ataxia group showed significant spatial distribution of the response, while the control group showed the expected decrease in mu-band power (localized to the motor cortex).

Keywords: Brain-computer interface, EEG, modulation, ataxia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1952
6273 A New Time-Frequency Speech Analysis Approach Based On Adaptive Fourier Decomposition

Authors: Liming Zhang

Abstract:

In this paper, a new adaptive Fourier decomposition (AFD) based time-frequency speech analysis approach is proposed. Given the fact that the fundamental frequency of speech signals often undergo fluctuation, the classical short-time Fourier transform (STFT) based spectrogram analysis suffers from the difficulty of window size selection. AFD is a newly developed signal decomposition theory. It is designed to deal with time-varying non-stationary signals. Its outstanding characteristic is to provide instantaneous frequency for each decomposed component, so the time-frequency analysis becomes easier. Experiments are conducted based on the sample sentence in TIMIT Acoustic-Phonetic Continuous Speech Corpus. The results show that the AFD based time-frequency distribution outperforms the STFT based one.

Keywords: Adaptive fourier decomposition, instantaneous frequency, speech analysis, time-frequency distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1707
6272 Reliable Capacitated Facility Location Problem Considering Maximal Covering

Authors: Mehdi Seifbarghy, Sajjad Jalali, Seyed Habib A. Rahmati

Abstract:

This paper provides a framework in order to incorporate reliability issue as a sign of disruption in distribution systems and partial covering theory as a response to limitation in coverage radios and economical preferences, simultaneously into the traditional literatures of capacitated facility location problems. As a result we develop a bi-objective model based on the discrete scenarios for expected cost minimization and demands coverage maximization through a three echelon supply chain network by facilitating multi-capacity levels for provider side layers and imposing gradual coverage function for distribution centers (DCs). Additionally, in spite of objectives aggregation for solving the model through LINGO software, a branch of LP-Metric method called Min- Max approach is proposed and different aspects of corresponds model will be explored.

Keywords: Reliability Cost, Partial Covering, LP-Metric

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1883
6271 Manufacturing Process of S-Glass Fiber Reinforced PEKK Prepregs

Authors: Nassier A. Nassir, Robert Birch, Zhongwei Guan

Abstract:

The aim of this study is to investigate the fundamental science/technology related to novel S-glass fiber reinforced polyether- ketone-ketone (GF/PEKK) composites and to gain insight into bonding strength and failure mechanisms. Different manufacturing techniques to make this high-temperature pre-impregnated composite (prepreg) were conducted i.e. mechanical deposition, electrostatic powder deposition, and dry powder prepregging techniques. Generally, the results of this investigation showed that it was difficult to control the distribution of the resin powder evenly on the both sides of the fibers within a specific percentage. Most successful approach was by using a dry powder prepregging where the fibers were coated evenly with an adhesive that served as a temporary binder to hold the resin powder in place onto the glass fiber fabric.

Keywords: Dry powder technique, PEKK, S-glass, thermoplastic prepreg.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1047
6270 Asymmetric and Kind of Bracing Effects on Steel Frames Under Earthquake Loads

Authors: Mahmoud Miri, Soliman Maramaee

Abstract:

Because of architectural condition and structure application, sometimes mass source and stiffness source are not coincidence, and the structure is irregular. The structure is also might be asymmetric as an asymmetric bracing in plan which leads to unbalance distribution of stiffness or because of unbalance distribution of the mass. Both condition lead to eccentricity and torsion in the structure. The deficiency of ordinary code to evaluate the performance of steel structures against earthquake has been caused designing based on performance level or capacity spectrum be used. By using the mentioned methods it is possible to design a structure that its behavior against different earthquakes be predictive. In this article 5- story buildings with different percentage of asymmetric which is because of stiffness changes and kind of bracing (x and chevron bracing) have been designed. The static and dynamic nonlinear analysis under three acceleration recording has been done. Finally performance level of the structure has been evaluated.

Keywords: Asymmetric, irregular, seismic analysis, torsion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1623
6269 Strengthening the HCI Approaches in the Software Development Process

Authors: Rogayah A. Majid, Nor Laila Md. Noor, Wan Adilah Wan Adnan

Abstract:

User-Centered Design (UCD), Usability Engineering (UE) and Participatory Design (PD) are the common Human- Computer Interaction (HCI) approaches that are practiced in the software development process, focusing towards issues and matters concerning user involvement. It overlooks the organizational perspective of HCI integration within the software development organization. The Management Information Systems (MIS) perspective of HCI takes a managerial and organizational context to view the effectiveness of integrating HCI in the software development process. The Human-Centered Design (HCD) which encompasses all of the human aspects including aesthetic and ergonomic, is claimed as to provide a better approach in strengthening the HCI approaches to strengthen the software development process. In determining the effectiveness of HCD in the software development process, this paper presents the findings of a content analysis of HCI approaches by viewing those approaches as a technology which integrates user requirements, ranging from the top management to other stake holder in the software development process. The findings obtained show that HCD approach is a technology that emphasizes on human, tools and knowledge in strengthening the HCI approaches to strengthen the software development process in the quest to produce a sustainable, usable and useful software product.

Keywords: Human-Centered Design (HCD), Management Information Systems (MIS), Participatory Design (PD), User- Centered Design (UCD), Usability Engineering (UE)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2228