Search results for: distribution process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6988

Search results for: distribution process

4888 A Dirty Page Migration Method in Process of Memory Migration Based on Pre-copy Technology

Authors: Kang Zijian, Zhang Tingyu, Burra Venkata Durga Kumar

Abstract:

This article investigates the challenges in memory migration during the live migration of virtual machines. We found three challenges probably existing in pre-copy technology. One of the main challenges is the challenge of downtime migration. Decreasing the downtime could promise the normal work for a virtual machine. Although pre-copy technology is greatly decreasing the downtime, we still need to shut down the machine in order to finish the last round of data transfer. This paper provides an optimization scheme for the problems existing in pro-copy technology, mainly the optimization of the dirty page migration mechanism. The typical pre-copy technology copies n-1th’s dirty pages in nth turn. However, our idea is to create a double iteration method to solve this problem.

Keywords: Virtual machine, pre-copy technology, memory migration process, downtime, dirty pages migration method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 126
4887 Confidence Interval for the Inverse of a Normal Mean with a Known Coefficient of Variation

Authors: Arunee Wongkha, Suparat Niwitpong, Sa-aat Niwitpong

Abstract:

In this paper, we propose two new confidence intervals for the inverse of a normal mean with a known coefficient of variation. One of new confidence intervals for the inverse of a normal mean with a known coefficient of variation is constructed based on the pivotal statistic Z where Z is a standard normal distribution and another confidence interval is constructed based on the generalized confidence interval, presented by Weerahandi. We examine the performance of these confidence intervals in terms of coverage probabilities and average lengths via Monte Carlo simulation.

Keywords: The inverse of a normal mean, confidence interval, generalized confidence intervals, known coefficient of variation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2571
4886 Kazakhstani Koreans- Conflict of Linguistic Identity: In–between the Sovietized and Kazakhstani Citizens

Authors: Soon-ok Myong, Byong-soon Chun

Abstract:

This paper intends to identify the ethnic Kazakhstani Koreans- political process of identity formation by exploring their narrative and practice about the state language represented in the course of their becoming the new citizens of a new independent state. The Russophone Kazakhstani Koreans- inability to speak the official language of their affiliated state is considered there as dissatisfying the basic requirement of citizens of the independent state, so that they are becoming marginalized from the public sphere. Their contradictory attitude that at once demonstrates nominal reception and practical rejection of the obligatory state language unveils a high barrier inside between their self-language and other-language. In this paper, the ethnic Korean group-s conflicting linguistic identity is not seen as a free and simple choice, but as a dynamic struggle and political process in which the subject-s past experiences and memories intersect with the external elements of pressure.

Keywords: Ethnic Kazakhstani Koreans, Soviet Korean's Russification, Linguistic Identity, Russian-Kazakh Dichotomy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1571
4885 How Team Efficacy Beliefs Impact Project Performance: An Empirical Investigation of Team Potency in Capital Projects in the Process Industries

Authors: C. Scott-Young, D. Samson

Abstract:

Team efficacy beliefs show promise in enhancing team performance. Using a model-based quantitative research design, we investigated the antecedents and performance consequences of generalized team efficacy (potency) in a sample of 56 capital projects executed by 15 Fortune 500 companies in the process industries. Empirical analysis of our field survey identified that generalized team efficacy beliefs were positively associated with an objective measure of project cost performance. Regression analysis revealed that team competence, empowering leadership, and performance feedback all predicted generalized team efficacy beliefs. Tests of mediation revealed that generalized team efficacy fully mediated between these three inputs and project cost performance.

Keywords: Team efficacy, Potency, Leadership, Feedback, Project cost.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2145
4884 GA based Optimal Sizing and Placement of Distributed Generation for Loss Minimization

Authors: Deependra Singh, Devender Singh, K. S. Verma

Abstract:

This paper addresses a novel technique for placement of distributed generation (DG) in electric power systems. A GA based approach for sizing and placement of DG keeping in view of system power loss minimization in different loading conditions is explained. Minimal system power loss is obtained under voltage and line loading constraints. Proposed strategy is applied to power distribution systems and its effectiveness is verified through simulation results on 16, 37-bus and 75-bus test systems.

Keywords: Distributed generation (DG), Genetic algorithms (GA), optimal sizing and placement, Power loss.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3444
4883 Recycling of Sclareolide in the Crystallization Mother Liquid of Sclareolide by Adsorption and Chromatography

Authors: Xiang Li, Kui Chen, Bin Wu, Min Zhou

Abstract:

Sclareolide is made from sclareol by oxidiative synthesis and subsequent crystallization, while the crystallization mother liquor still contains 15%~30%wt of sclareolide to be reclaimed. With the reaction material of sclareol is provided as plant extract, many sorts of complex impurities exist in the mother liquor. Due to the difficulty in recycling sclareolide after solvent recovery, it is common practice for the factories to discard the mother liquor, which not only results in loss of sclareolide, but also contributes extra environmental burden. In this paper, a process based on adsorption and elution has been presented for recycling of sclareolide from mother liquor. After pretreatment of the crystallization mother liquor by HZ-845 resin to remove parts of impurities, sclareolide is adsorbed by HZ-816 resin. The HZ-816 resin loaded with sclareolide is then eluted by elution solvent. Finally, the eluent containing sclareolide is concentrated and fed into the crystallization step in the process. By adoption of the recycle from mother liquor, total yield of sclareolide increases from 86% to 90% with a stable purity of the final sclareolide products maintained.

Keywords: Sclareolide, resin, adsorption, chromatography.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1808
4882 Development and Optimization of Automated Dry-Wafer Separation

Authors: Tim Giesen, Christian Fischmann, Fabian Böttinger, Alexander Ehm, Alexander Verl

Abstract:

In a state-of-the-art industrial production line of photovoltaic products the handling and automation processes are of particular importance and implication. While processing a fully functional crystalline solar cell an as-cut photovoltaic wafer is subject to numerous repeated handling steps. With respect to stronger requirements in productivity and decreasing rejections due to defects the mechanical stress on the thin wafers has to be reduced to a minimum as the fragility increases by decreasing wafer thicknesses. In relation to the increasing wafer fragility, researches at the Fraunhofer Institutes IPA and CSP showed a negative correlation between multiple handling processes and the wafer integrity. Recent work therefore focused on the analysis and optimization of the dry wafer stack separation process with compressed air. The achievement of a wafer sensitive process capability and a high production throughput rate is the basic motivation in this research.

Keywords: Automation, Photovoltaic Manufacturing, Thin Wafer, Material Handling

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1658
4881 Cost Valuation Method for Development Concurrent Phase Appropriate Requirement Valuation Using the Example of Load Carrier Development in the Lithium-Ion-Battery Production

Authors: Achim Kampker, Christoph Deutskens, Heiner Hans Heimes, Mathias Ordung, Felix Optehostert

Abstract:

In the past years electric mobility became part of a public discussion. The trend to fully electrified vehicles instead of vehicles fueled with fossil energy has notably gained momentum. Today nearly every big car manufacturer produces and sells fully electrified vehicles, but electrified vehicles are still not as competitive as conventional powered vehicles. As the traction battery states the largest cost driver, lowering its price is a crucial objective. In addition to improvements in product and production processes a nonnegligible, but widely underestimated cost driver of production can be found in logistics, since the production technology is not continuous yet and neither are the logistics systems. This paper presents an approach to evaluate cost factors on different designs of load carrier systems. Due to numerous interdependencies, the combination of costs factors for a particular scenario is not transparent. This is effecting actions for cost reduction negatively, but still cost reduction is one of the major goals for simultaneous engineering processes. Therefore a concurrent and phase appropriate cost valuation method is necessary to serve cost transparency. In this paper the four phases of this cost valuation method are defined and explained, which based upon a new approach integrating the logistics development process in to the integrated product and process development.

Keywords: Research and development, technology and Innovation, lithium-ion-battery production, load carrier development process, cost valuation method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2259
4880 Geochemical Assessment of Heavy Metals Concentration in Surface Sediment of West Port, Malaysia

Authors: B.Tavakoly Sany, A. Salleh, A.H .Sulaiman, A. Mehdinia, GH. Monazami

Abstract:

One year (November 2009-October 2010) sediment monitoring was used to evaluate pollution status, concentration and distribution of heavy metals (As, Cu, Cd, Cr, Hg, Ni, Pb and Zn) in West Port of Malaysia. Sediment sample were collected from nine stations every four months. Geo-accumulation factor and Pollution Load Index (PLI) were estimated to better understand the pollution level in study area. The heavy metal concentration (Mg/g dry weight) were ranged from 20.2 to 162 for As, 7.4 to 27.6 for Cu, 0.244 to 3.53 for Cd, 11.5 to 61.5 for Cr, 0.11 to 0.409 for Hg, 7.2 to 22.2 for Ni, 22.3 to 80 for Pb and 23 to 98.3 for Zn. In general, concentration some metals (As,Cd, Hg and Pb) was higher than background values that are considered as serious concern for aquatic life and the human health.

Keywords: Heavy metals, Sediment Quality, geo-accumulationindex, Pollution Load Index

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2516
4879 A Review of Quality Relationship between IT Processes, IT Products and IT Services

Authors: Whee Yen Wong, Chan Wai Lee, Kim Yeow Tshai

Abstract:

Producing IT products/services required carefully designed. IT development process is intangible and labour intensive. Making optimal use of available resources, both soft (knowledge, skill-set etc.) and hard (computer system, ancillary equipment etc.), is vital if IT development is to achieve sensible economical advantages. Apart from the norm of Project Life Cycle and System Development Life Cycle (SDLC), there is an urgent need to establish a general yet widely acceptable guideline on the most effective and efficient way to precede an IT project in the broader view of Product Life Cycle. The current paper proposes such a framework with two major areas of concern: (1) an integration of IT Products and IT Services within an existing IT Process architecture and; (2) how IT Product and IT Services are built into the framework of Product Life Cycle, Project Life Cycle and SDLC.

Keywords: Mapping of Quality Relationship, IT Processes/IT Products/IT Services, Product Life Cycle, System Development Life Cycle.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2149
4878 A Hybrid Radial-Based Neuro-GA Multiobjective Design of Laminated Composite Plates under Moisture and Thermal Actions

Authors: Mohammad Reza Ghasemi, Ali Ehsani

Abstract:

In this paper, the optimum weight and cost of a laminated composite plate is seeked, while it undergoes the heaviest load prior to a complete failure. Various failure criteria are defined for such structures in the literature. In this work, the Tsai-Hill theory is used as the failure criterion. The theory of analysis was based on the Classical Lamination Theory (CLT). A newly type of Genetic Algorithm (GA) as an optimization technique with a direct use of real variables was employed. Yet, since the optimization via GAs is a long process, and the major time is consumed through the analysis, Radial Basis Function Neural Networks (RBFNN) was employed in predicting the output from the analysis. Thus, the process of optimization will be carried out through a hybrid neuro-GA environment, and the procedure will be carried out until a predicted optimum solution is achieved.

Keywords: Composite Laminates, GA, Multi-objectiveOptimization, Neural Networks, RBFNN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1454
4877 Properties of the Research Teaching Organization of Military Masters

Authors: A. Bulatbayeva, A. Kusainov, P. Makhanova, A. Bashchikulov

Abstract:

In the article there have been revealed the properties of designing the research teaching the military masters and in the context it has been offered the program of mastering by the masters military men the methodology of research work, in the course of practical teaching activity there has been considered the developed and approbated model of organization of the process of mastering by the masters the methodology of research work. As a whole, the research direction of master preparation leaves its sign to the content of education, forms of organization of educational process, scientific work of masters. In this connection the offered in the article properties of organization of research teaching and a model of organization of mastering by the masters military men the methodology of research work can be taken into account when designing the content of master preparation.

Keywords: Masters militaries, Methodology of research work, Military knowledge, Research teaching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1223
4876 A Discretizing Method for Reliability Computation in Complex Stress-strength Models

Authors: Alessandro Barbiero

Abstract:

This paper proposes, implements and evaluates an original discretization method for continuous random variables, in order to estimate the reliability of systems for which stress and strength are defined as complex functions, and whose reliability is not derivable through analytic techniques. This method is compared to other two discretizing approaches appeared in literature, also through a comparative study involving four engineering applications. The results show that the proposal is very efficient in terms of closeness of the estimates to the true (simulated) reliability. In the study we analyzed both a normal and a non-normal distribution for the random variables: this method is theoretically suitable for each parametric family.

Keywords: Approximation, asymmetry, experimental design, interference theory, Monte Carlo simulations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1752
4875 A New Extended Group Mutual Exclusion Algorithm with Low Message Complexity in Distributed Systems

Authors: S. Dehghan, A.M. Rahmani

Abstract:

The group mutual exclusion (GME) problem is an interesting generalization of the mutual exclusion problem. In the group mutual exclusion, multiple processes can enter a critical section simultaneously if they belong to the same group. In the extended group mutual exclusion, each process is a member of multiple groups at the same time. As a result, after the process by selecting a group enter critical section, other processes can select the same group with its belonging group and can enter critical section at the moment, so that it avoids their unnecessary blocking. This paper presents a quorum-based distributed algorithm for the extended group mutual exclusion problem. The message complexity of our algorithm is O(4Q ) in the best case and O(5Q) in the worst case, where Q is a quorum size.

Keywords: Group Mutual Exclusion (GME), Extended GME, Distributed systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1509
4874 Image Ranking to Assist Object Labeling for Training Detection Models

Authors: Tonislav Ivanov, Oleksii Nedashkivskyi, Denis Babeshko, Vadim Pinskiy, Matthew Putman

Abstract:

Training a machine learning model for object detection that generalizes well is known to benefit from a training dataset with diverse examples. However, training datasets usually contain many repeats of common examples of a class and lack rarely seen examples. This is due to the process commonly used during human annotation where a person would proceed sequentially through a list of images labeling a sufficiently high total number of examples. Instead, the method presented involves an active process where, after the initial labeling of several images is completed, the next subset of images for labeling is selected by an algorithm. This process of algorithmic image selection and manual labeling continues in an iterative fashion. The algorithm used for the image selection is a deep learning algorithm, based on the U-shaped architecture, which quantifies the presence of unseen data in each image in order to find images that contain the most novel examples. Moreover, the location of the unseen data in each image is highlighted, aiding the labeler in spotting these examples. Experiments performed using semiconductor wafer data show that labeling a subset of the data, curated by this algorithm, resulted in a model with a better performance than a model produced from sequentially labeling the same amount of data. Also, similar performance is achieved compared to a model trained on exhaustive labeling of the whole dataset. Overall, the proposed approach results in a dataset that has a diverse set of examples per class as well as more balanced classes, which proves beneficial when training a deep learning model.

Keywords: Computer vision, deep learning, object detection, semiconductor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 797
4873 Detection Characteristics of the Random and Deterministic Signals in Antenna Arrays

Authors: Olesya Bolkhovskaya, Alexey Davydov, Alexander Maltsev

Abstract:

In this paper, approach to incoherent signal detection in multi-element antenna array are researched and modeled. Two types of useful signals with unknown wavefront were considered: first one, deterministic (Barker code), and second one, random (Gaussian distribution). The derivation of the sufficient statistics took into account the linearity of the antenna array. The performance characteristics and detecting curves are modeled and compared for different useful signals parameters and for different number of elements of the antenna array. Results of researches in case of some additional conditions can be applied to a digital communications systems.

Keywords: Antenna array, detection curves, performance characteristics, quadrature processing, signal detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1760
4872 Hexavalent Chromium Pollution Abatement by use of Scrap Iron

Authors: Marius Gheju, Laura Cocheci

Abstract:

In this study, the reduction of Cr(VI) by use of scrap iron, a cheap and locally available industrial waste, was investigated in continuous system. The greater scrap iron efficiency observed for the first two sections of the column filling indicate that most of the reduction process was carried out in the bottom half of the column filling. This was ascribed to a constant decrease of Cr(VI) concentration inside the filling, as the water front passes from the bottom to the top end of the column. While the bottom section of the column filling was heavily passivated with secondary mineral phases, the top section was less affected by the passivation process; therefore the column filling would likely ensure the reduction of Cr(VI) for time periods longer than 216 hours. The experimental results indicate that fixed beds columns packed with scrap iron could be successfully used for the first step of Cr(VI) polluted wastewater treatment. However, the mass of scrap iron filling should be carefully estimated since it significantly affects the Cr(VI) reduction efficiency.

Keywords: hexavalent chromium, heavy metals, scrap iron, reduction capacity, wastewater treatment

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1825
4871 A Wavelet-Based Watermarking Method Exploiting the Contrast Sensitivity Function

Authors: John N. Ellinas, Panagiotis Kenterlis

Abstract:

The efficiency of an image watermarking technique depends on the preservation of visually significant information. This is attained by embedding the watermark transparently with the maximum possible strength. The current paper presents an approach for still image digital watermarking in which the watermark embedding process employs the wavelet transform and incorporates Human Visual System (HVS) characteristics. The sensitivity of a human observer to contrast with respect to spatial frequency is described by the Contrast Sensitivity Function (CSF). The strength of the watermark within the decomposition subbands, which occupy an interval on the spatial frequencies, is adjusted according to this sensitivity. Moreover, the watermark embedding process is carried over the subband coefficients that lie on edges where distortions are less noticeable. The experimental evaluation of the proposed method shows very good results in terms of robustness and transparency.

Keywords: Image watermarking, wavelet transform, human visual system, contrast sensitivity function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2081
4870 Analysis of the Shielding Effectiveness of Several Magnetic Shields

Authors: Diako Azizi, Hosein Heydari, Ahmad Gholami

Abstract:

Today with the rapid growth of telecommunications equipment, electronic and developing more and more networks of power, influence of electromagnetic waves on one another has become hot topic discussions. So in this article, this issue and appropriate mechanisms for EMC operations have been presented. First, a source of alternating current (50 Hz) and a clear victim in a certain distance from the source is placed. With this simple model, the effects of electromagnetic radiation from the source to the victim will be investigated and several methods to reduce these effects have been presented. Therefore passive and active shields have been used. In some steps, shielding effectiveness of proposed shields will be compared. . It should be noted that simulations have been done by the finite element method (FEM).

Keywords: Electrical field, field distribution, finite element method, shielding effectiveness

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1732
4869 Modeling of Statistically Multiplexed Non Uniform Activity VBR Video

Authors: J. P. Dubois

Abstract:

This paper reports the feasibility of the ARMA model to describe a bursty video source transmitting over a AAL5 ATM link (VBR traffic). The traffic represents the activity of the action movie "Lethal Weapon 3" transmitted over the ATM network using the Fore System AVA-200 ATM video codec with a peak rate of 100 Mbps and a frame rate of 25. The model parameters were estimated for a single video source and independently multiplexed video sources. It was found that the model ARMA (2, 4) is well-suited for the real data in terms of average rate traffic profile, probability density function, autocorrelation function, burstiness measure, and the pole-zero distribution of the filter model.

Keywords: ARMA, ATM networks, burstiness, multimediatraffic, VBR video.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1357
4868 Informative, Inclusive and Transparent Planning Methods for Sustainable Heritage Management

Authors: Mathilde Kirkegaard

Abstract:

The paper will focus on management of heritage that integrates the local community, and argue towards an obligation to integrate this social aspect in heritage management. By broadening the understanding of heritage, a sustainable heritage management takes its departure in more than a continual conservation of the physicality of heritage. The social aspect, or the local community, is in many govern heritage management situations being overlooked and it is not managed through community based urban planning methods, e.g.: citizen-inclusion, a transparent process, informative and inviting initiatives, etc. Historical sites are often being described by embracing terms such as “ours” and “us”: “our history” and “a history that is part of us”. Heritage is not something static, it is a link between the life that has been lived in the historical frames, and the life that is defining it today. This view on heritage is rooted in the strive to ensure that heritage sites, besides securing the national historical interest, have a value for those people who are affected by it: living in it or visiting it. Antigua Guatemala is a UNESCO-defined heritage site and this site is being ‘threatened’ by tourism, habitation and recreation. In other words: ‘the use’ of the site is considered a threat of the preservation of the heritage. Contradictory the same types of use (tourism and habitation) can also be considered development ability, and perhaps even a sustainable management solution. ‘The use’ of heritage is interlinked with the perspective that heritage sites ought to have a value for people today. In other words, the heritage sites should be comprised of a contemporary substance. Heritage is entwined in its context of physical structures and the social layer. A synergy between the use of heritage and the knowledge about the heritage can generate a sustainable preservation solution. The paper will exemplify this symbiosis with different examples of a heritage management that is centred around a local community inclusion. The inclusive method is not new in architectural planning and it refers to a top-down and bottom-up balance in decision making. It can be endeavoured through designs of an inclusive nature. Catalyst architecture is a planning method that strives to move the process of design solutions into the public space. Through process-orientated designs, or catalyst designs, the community can gain an insight into the process or be invited to participate in the process. A balance between bottom-up and top-down in the development process of a heritage site can, in relation to management measures, be understood to generate a socially sustainable solution. The ownership and engagement that can be created among the local community, along with the use that ultimately can gain an economic benefit, can delegate the maintenance and preservation. Informative, inclusive and transparent planning methods can generate a heritage management that is long-term due to the collective understanding and effort. This method handles sustainable management on two levels: the current preservation necessities and the long-term management, while ensuring a value for people today.

Keywords: Community, intangible, inclusion, planning, heritage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 743
4867 Visual Construction of Youth in Czechoslovak Press Photographs: 1959-1989

Authors: Jana Teplá

Abstract:

This text focuses on the visual construction of youth in press photographs in socialist Czechoslovakia. It deals with photographs in a magazine for young readers, Mladý svět, published by the Socialist Union of Youth of Czechoslovakia. The aim of this study was to develop a methodological tool for uncovering the values and the ideological messages in the strategies used in the visual construction of reality in the socialist press. Two methods of visual analysis were applied to the photographs, a quantitative content analysis and a social semiotic analysis. The social semiotic analysis focused on images representing youth in their free time. The study shows that the meaning of a socialist press photograph is a result of a struggle for ideological power between formal and informal ideologies. This struggle takes place within the process of production of the photograph and also within the process of interpretation of the photograph.

Keywords: Ideology, press photography, socialist regime, social semiotics, youth.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 874
4866 Modeling and Simulation of Dynamic Voltage Restorer for Mitigation of Voltage Sags

Authors: S. Ganesh, L. Raguraman, E. Anushya, J. krishnasree

Abstract:

Voltage sags are the most common power quality disturbance in the distribution system. It occurs due to the fault in the electrical network or by the starting of a large induction motor and this can be solved by using the custom power devices such as Dynamic Voltage Restorer (DVR). In this paper DVR is proposed to compensate voltage sags on critical loads dynamically. The DVR consists of VSC, injection transformers, passive filters and energy storage (lead acid battery). By injecting an appropriate voltage, the DVR restores a voltage waveform and ensures constant load voltage. The simulation and experimental results of a DVR using MATLAB software shows clearly the performance of the DVR in mitigating voltage sags.

Keywords: Dynamic voltage restorer, Voltage sags, Power quality, Injection methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4266
4865 A Comparison of Single of Decision Tree, Decision Tree Forest and Group Method of Data Handling to Evaluate the Surface Roughness in Machining Process

Authors: S. Ghorbani, N. I. Polushin

Abstract:

The machinability of workpieces (AISI 1045 Steel, AA2024 aluminum alloy, A48-class30 gray cast iron) in turning operation has been carried out using different types of cutting tool (conventional, cutting tool with holes in toolholder and cutting tool filled up with composite material) under dry conditions on a turning machine at different stages of spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev), depth of cut (0.05-0.15 mm) and tool overhang (41-65 mm). Experimentation was performed as per Taguchi’s orthogonal array. To evaluate the relative importance of factors affecting surface roughness the single decision tree (SDT), Decision tree forest (DTF) and Group method of data handling (GMDH) were applied.

Keywords: Decision Tree Forest, GMDH, surface roughness, taguchi method, turning process.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 936
4864 Biological Soil Conservation Planning by Spatial Multi-Criteria Evaluation Techniques (Case Study: Bonkuh Watershed in Iran)

Authors: Ali Akbar Jamali

Abstract:

This paper discusses site selection process for biological soil conservation planning. It was supported by a valuefocused approach and spatial multi-criteria evaluation techniques. A first set of spatial criteria was used to design a number of potential sites. Next, a new set of spatial and non-spatial criteria was employed, including the natural factors and the financial costs, together with the degree of suitability for the Bonkuh watershed to biological soil conservation planning and to recommend the most acceptable program. The whole process was facilitated by a new software tool that supports spatial multiple criteria evaluation, or SMCE in GIS software (ILWIS). The application of this tool, combined with a continual feedback by the public attentions, has provided an effective methodology to solve complex decisional problem in biological soil conservation planning.

Keywords: GIS, Biological soil conservation planning, Spatial multi-criteria evaluation, Iran

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1707
4863 The DAQ Debugger for iFDAQ of the COMPASS Experiment

Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius

Abstract:

In general, state-of-the-art Data Acquisition Systems (DAQ) in high energy physics experiments must satisfy high requirements in terms of reliability, efficiency and data rate capability. This paper presents the development and deployment of a debugging tool named DAQ Debugger for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. Utilizing a hardware event builder, the iFDAQ is designed to be able to readout data at the average maximum rate of 1.5 GB/s of the experiment. In complex softwares, such as the iFDAQ, having thousands of lines of code, the debugging process is absolutely essential to reveal all software issues. Unfortunately, conventional debugging of the iFDAQ is not possible during the real data taking. The DAQ Debugger is a tool for identifying a problem, isolating the source of the problem, and then either correcting the problem or determining a way to work around it. It provides the layer for an easy integration to any process and has no impact on the process performance. Based on handling of system signals, the DAQ Debugger represents an alternative to conventional debuggers provided by most integrated development environments. Whenever problem occurs, it generates reports containing all necessary information important for a deeper investigation and analysis. The DAQ Debugger was fully incorporated to all processes in the iFDAQ during the run 2016. It helped to reveal remaining software issues and improved significantly the stability of the system in comparison with the previous run. In the paper, we present the DAQ Debugger from several insights and discuss it in a detailed way.

Keywords: DAQ debugger, data acquisition system, FPGA, system signals, Qt framework.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 879
4862 Estimating the Population Mean by Using Stratified Double Extreme Ranked Set Sample

Authors: Mahmoud I. Syam, Kamarulzaman Ibrahim, Amer I. Al-Omari

Abstract:

Stratified double extreme ranked set sampling (SDERSS) method is introduced and considered for estimating the population mean. The SDERSS is compared with the simple random sampling (SRS), stratified ranked set sampling (SRSS) and stratified simple set sampling (SSRS). It is shown that the SDERSS estimator is an unbiased of the population mean and more efficient than the estimators using SRS, SRSS and SSRS when the underlying distribution of the variable of interest is symmetric or asymmetric.

Keywords: Double extreme ranked set sampling, Extreme ranked set sampling, Ranked set sampling, Stratified double extreme ranked set sampling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2331
4861 Revisiting the Concept of Risk Analysis within the Context of Geospatial Database Design: A Collaborative Framework

Authors: J. Grira, Y. Bédard, S. Roche

Abstract:

The aim of this research is to design a collaborative framework that integrates risk analysis activities into the geospatial database design (GDD) process. Risk analysis is rarely undertaken iteratively as part of the present GDD methods in conformance to requirement engineering (RE) guidelines and risk standards. Accordingly, when risk analysis is performed during the GDD, some foreseeable risks may be overlooked and not reach the output specifications especially when user intentions are not systematically collected. This may lead to ill-defined requirements and ultimately in higher risks of geospatial data misuse. The adopted approach consists of 1) reviewing risk analysis process within the scope of RE and GDD, 2) analyzing the challenges of risk analysis within the context of GDD, and 3) presenting the components of a risk-based collaborative framework that improves the collection of the intended/forbidden usages of the data and helps geo-IT experts to discover implicit requirements and risks.

Keywords: Collaborative risk analysis, intention of use, Geospatial database design, Geospatial data misuse.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1648
4860 Industrial Development, Environment And Occupational Problems: The Case Of Iran

Authors: Ghaffari, H., Changi Ashtiani, A., Younessi, A.

Abstract:

There are three distinct stages in the evolution of economic thought, namely: 1. in the first stage, the major concern was to accelerate economic growth with increased availability of material goods, especially in developing economies with very low living standards, because poverty eradication meant faster economic growth. 2. in the second stage, economists made distinction between growth and development. Development was seen as going beyond economic growth, and bringing certain changes in the structure of the economy with more equitable distribution of the benefits of growth, with the growth coming automatic and sustained. 3. the third stage is now reached. Our concern is now with “sustainable development", that is, development not only for the present but also of the future. Thus the focus changed from “sustained growth" to “sustained development". Sustained development brings to the fore the long term relationship between the ecology and economic development. Since the creation of UNEP in 1972 it has worked for development without destruction for environmentally sound and sustained development. It was realised that the environment cannot be viewed in a vaccum, it is not separate from development, nor is it competing. It suggested for the integration of the environment with development whereby ecological factors enter development planning, socio-economic policies, cost-benefit analysis, trade, technology transfer, waste management, educational and other specific areas. Industrialisation has contributed to the growth of economy of several countries. It has improved the standards of living of its people and provided benefits to the society. It has also created in the process great environmental problems like climate change, forest destruction and denudation, soil erosion and desertification etc. On the other hand, industry has provided jobs and improved the prospects of wealth for the industrialists. The working class communities had to simply put up with the high levels of pollution in order to keep up their jobs and also to save their income. There are many roots of the environmental problem. They may be political, economic, cultural and technological conditions of the modern society. The experts concede that industrial growth lies somewhere close to the heart of the matter. Therefore, the objective of this paper is not to document all roots of an environmental crisis but rather to discuss the effects of industrial growth and development. We have come to the conclusion that although public intervention is often unnecessary to ensure that perfectly competitive markets will function in society-s best interests, such intervention is necessary when firms or consumers pollute.

Keywords: Development, Environment, Industrial Development, Iran, Occupational problems, Pollution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1642
4859 Mathematical Models for Overall Gas Transfer Coefficient Using Different Theories and Evaluating Their Measurement Accuracy

Authors: Shashank.B. Thakre, Lalit.B. Bhuyar, Samir.J. Deshmukh

Abstract:

Oxygen transfer, the process by which oxygen is transferred from the gaseous to liquid phase, is a vital part of the waste water treatment process. Because of low solubility of oxygen and consequent low rate of oxygen transfer, sufficient oxygen to meet the requirement of aerobic waste does not enter through normal surface air water interface. Many theories have come up in explaining the mechanism of gas transfer and absorption of non-reacting gases in a liquid, of out of which, Two film theory is important. An exiting mathematical model determines approximate value of Overall Gas Transfer coefficient. The Overall Gas Transfer coefficient, in case of Penetration theory, is 1.13 time more than that obtained in case of Two film theory. The difference is due to the difference in assumptions in the two theories. The paper aims at development of mathematical model which determines the value of Overall Gas Transfer coefficient with greater accuracy than the existing model.

Keywords: Theories, Dissolved oxygen, Mathematical model, Gas Transfer coefficient, Accuracy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1538