Search results for: standard methodology.
2932 Assessing and Improving Ramp-Up Capability
Authors: Sebastian Tschöpe, Konja Knüppel, Peter Nyhuis
Abstract:
In times when product life cycles are decreasing, while market demands are increasing, manufacturing enterprises are confronted with the challenge of more frequent and more complex ramp-ups. Thus it becomes obvious that ramp-up management is going to be a topic enterprises have to focus on in the future. Since each ramp-up is unique concerning the product, the process, the technology, the circumstances and the coaction of these four factors, the knowledge of the ramp-up situation and the current ramp-up capability of the enterprise are fundamental requirements for the subsequent improvement of the ramp-up capability of the production system.
In this article a methodology is going to be presented which can be used to define typical production ramp-up situations, to identify the current ramp-up capability of a production system and to improve it with respect to a specific situation. Additionally there will be a description of the functionality of a software-tool developed based on this methodology.
Keywords: Assessment methodology, ramp-up, ramp-up capability, software-tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20092931 A Methodology for Reducing the BGP Convergence Time
Authors: Eatedal A. Alabdulkreem, Hamed S. Al-Raweshidy, Maysam F. Abbod
Abstract:
Border Gateway Protocol (BGP) is the standard routing protocol between various autonomous systems (AS) in the internet. In the event of failure, a considerable delay in the BGP convergence has been shown by empirical measurements. During the convergence time the BGP will repeatedly advertise new routes to some destination and withdraw old ones until it reach a stable state. It has been found that the KEEPALIVE message timer and the HOLD time are tow parameters affecting the convergence speed. This paper aims to find the optimum value for the KEEPALIVE timer and the HOLD time that maximally reduces the convergence time without increasing the traffic. The KEEPALIVE message timer optimal value founded by this paper is 30 second instead of 60 seconds, and the optimal value for the HOLD time is 90 seconds instead of 180 seconds.
Keywords: BGP, Convergence Time, HOLD time, Keep alive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20412930 The Determination of Heavy Metal in Herb Used in Dusit Community to Develop a Sustainable Quality of Life
Authors: Chinnawat Satsananan
Abstract:
This research aimed to find amount of heavy metal in herb used in Dusit community and compare of heavy metal in each part by quantity in herb and standard determination in Thai herb books to develop a sustainable quality of life, the result of study in 14 herbs do not find sample of heavy metal., by quantity of heavy contamination of 4 kinds: Cd, Co, Fe and Pb have lower than standard of 2 organizations: Thai herb standard, and World Health Organization, from the test 14 herbs have Fe in every part of herbs and all 14 kinds has Fe that is necessary for our health.
Keywords: Herbs Plants, Heavy Metal, Dusit District
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18362929 Factors That Affect the Effectiveness of Enterprise Architecture Implementation Methodology
Authors: Babak Darvish Rouhani, Mohd Naz’ri Mahrin, Fatemeh Nikpay, Pourya Nikfard, Maryam Khanian Najafabadi
Abstract:
Enterprise Architecture (EA) is a strategy that is employed by enterprises in order to align their business and Information Technology (IT). EA is managed, developed, and maintained through Enterprise Architecture Implementation Methodology (EAIM). Effectiveness of EA implementation is the degree in which EA helps to achieve the collective goals of the organization. This paper analyzes the results of a survey that aims to explore the factors that affect the effectiveness of EAIM and specifically the relationship between factors and effectiveness of the output and functionality of EA project. The exploratory factor analysis highlights a specific set of five factors: alignment, adaptiveness, support, binding, and innovation. The regression analysis shows that there is a statistically significant and positive relationship between each of the five factors and the effectiveness of EAIM. Consistent with theory and practice, the most prominent factor for developing an effective EAIM is innovation. The findings contribute to the measuring the effectiveness of EA implementation project by providing an indication of the measurement implementation approaches which is used by the Enterprise Architects, and developing an effective EAIM.
Keywords: Enterprise Architecture, Enterprise Architecture Implementation Methodology, EA, Effectiveness, Factors, Implementation Methodology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32692928 Standard Deviation of Mean and Variance of Rows and Columns of Images for CBIR
Authors: H. B. Kekre, Kavita Patil
Abstract:
This paper describes a novel and effective approach to content-based image retrieval (CBIR) that represents each image in the database by a vector of feature values called “Standard deviation of mean vectors of color distribution of rows and columns of images for CBIR". In many areas of commerce, government, academia, and hospitals, large collections of digital images are being created. This paper describes the approach that uses contents as feature vector for retrieval of similar images. There are several classes of features that are used to specify queries: colour, texture, shape, spatial layout. Colour features are often easily obtained directly from the pixel intensities. In this paper feature extraction is done for the texture descriptor that is 'variance' and 'Variance of Variances'. First standard deviation of each row and column mean is calculated for R, G, and B planes. These six values are obtained for one image which acts as a feature vector. Secondly we calculate variance of the row and column of R, G and B planes of an image. Then six standard deviations of these variance sequences are calculated to form a feature vector of dimension six. We applied our approach to a database of 300 BMP images. We have determined the capability of automatic indexing by analyzing image content: color and texture as features and by applying a similarity measure Euclidean distance.
Keywords: Standard deviation Image retrieval, color distribution, Variance, Variance of Variance, Euclidean distance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37452927 Optimization of Extraction of Phenolic Compounds from Avicennia marina (Forssk.)Vierh using Response Surface Methodology
Authors: V.Bharathi, Jamila Patterson, R.Rajendiran
Abstract:
Optimization of extraction of phenolic compounds from Avicennia marina using response surface methodology was carried out during the present study. Five levels, three factors rotatable design (CCRD) was utilized to examine the optimum combination of extraction variables based on the TPC of Avicennia marina leaves. The best combination of response function was 78.41 °C, drying temperature; 26.18°C; extraction temperature and 36.53 minutes of extraction time. However, the procedure can be promptly extended to the study of several others pharmaceutical processes like purification of bioactive substances, drying of extracts and development of the pharmaceutical dosage forms for the benefit of consumers.Keywords: Avicennia marina, Central Composite RotatableDesign (CCRD), Response Surface Methodology, Total Phenoliccontents (TPC)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20652926 A Methodology for Creating a Conceptual Model Under Uncertainty
Authors: Bogdan Walek, Jiri Bartos, Cyril Klimes
Abstract:
This article deals with the conceptual modeling under uncertainty. First, the division of information systems with their definition will be described, focusing on those where the construction of a conceptual model is suitable for the design of future information system database. Furthermore, the disadvantages of the traditional approach in creating a conceptual model and database design will be analyzed. A comprehensive methodology for the creation of a conceptual model based on analysis of client requirements and the selection of a suitable domain model is proposed here. This article presents the expert system used for the construction of a conceptual model and is a suitable tool for database designers to create a conceptual model.
Keywords: Conceptual model, conceptual modeling, database, methodology, uncertainty, information system, entity, attribute, relationship, conceptual domain model, fuzzy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15842925 An Ontology Model for Systems Engineering Derived from ISO/IEC/IEEE 15288: 2015: Systems and Software Engineering - System Life Cycle Processes
Authors: Lan Yang, Kathryn Cormican, Ming Yu
Abstract:
ISO/IEC/IEEE 15288: 2015, Systems and Software Engineering - System Life Cycle Processes is an international standard that provides generic top-level process descriptions to support systems engineering (SE). However, the processes defined in the standard needs improvement to lift integrity and consistency. The goal of this research is to explore the way by building an ontology model for the SE standard to manage the knowledge of SE. The ontology model gives a whole picture of the SE knowledge domain by building connections between SE concepts. Moreover, it creates a hierarchical classification of the concepts to fulfil different requirements of displaying and analysing SE knowledge.Keywords: Knowledge management, model-based systems engineering, ontology modelling, systems engineering ontology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18642924 Fuzzy Multiple Criteria Decision Making for Unmanned Combat Aircraft Selection Using Proximity Measure Method
Authors: C. Ardil
Abstract:
Intuitionistic fuzzy sets (IFS), Pythagorean fuzzy sets (PyFS), Picture fuzzy sets (PFS), q-rung orthopair fuzzy sets (q-ROF), Spherical fuzzy sets (SFS), T-spherical FS, and Neutrosophic sets (NS) are reviewed as multidimensional extensions of fuzzy sets in order to more explicitly and informatively describe the opinions of decision-making experts under uncertainty. To handle operations with standard fuzzy sets (SFS), the necessary operators; weighted arithmetic mean (WAM), weighted geometric mean (WGM), and Minkowski distance function are defined. The algorithm of the proposed proximity measure method (PMM) is provided with a multiple criteria group decision making method (MCDM) for use in a standard fuzzy set environment. To demonstrate the feasibility of the proposed method, the problem of selecting the best drone for an Air Force procurement request is used. The proximity measure method (PMM) based multidimensional standard fuzzy sets (SFS) is introduced to demonstrate its use with an issue involving unmanned combat aircraft selection.
Keywords: standard fuzzy sets (SFS), unmanned combat aircraft selection, multiple criteria decision making (MCDM), proximity measure method (PMM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3592923 Investigation of Flow Characteristics on Upstream and Downstream of Orifice Using Computational Fluid Dynamics
Authors: War War Min Swe, Aung Myat Thu, Khin Cho Thet, Zaw Moe Htet, Thuzar Mon
Abstract:
The main parameter of the orifice hole diameter was designed according to the range of throttle diameter ratio which gave the required discharge coefficient. The discharge coefficient is determined by difference diameter ratios. The value of discharge coefficient is 0.958 occurred at throttle diameter ratio 0.5. The throttle hole diameter is 80 mm. The flow analysis is done numerically using ANSYS 17.0, computational fluid dynamics. The flow velocity was analyzed in the upstream and downstream of the orifice meter. The downstream velocity of non-standard orifice meter is 2.5% greater than that of standard orifice meter. The differential pressure is 515.379 Pa in standard orifice.
Keywords: CFD-CFX, discharge coefficients, flow characteristics, inclined.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5742922 Towards Modeling for Crashes A Low-Cost Adaptive Methodology for Karachi
Authors: Mohammad Ahmed Rehmatullah
Abstract:
The aim of this paper is to discuss a low-cost methodology that can predict traffic flow conflicts and quantitatively rank crash expectancies (based on relative probability) for various traffic facilities. This paper focuses on the application of statistical distributions to model traffic flow and Monte Carlo techniques to simulate traffic and discusses how to create a tool in order to predict the possibility of a traffic crash. A low-cost data collection methodology has been discussed for the heterogeneous traffic flow that exists and a GIS platform has been proposed to thematically represent traffic flow from simulations and the probability of a crash. Furthermore, discussions have been made to reflect the dynamism of the model in reference to its adaptability, adequacy, economy, and efficiency to ensure adoption.
Keywords: Heterogeneous traffic data collection, Monte CarloSimulation, Traffic Flow Modeling, GIS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14332921 eLearning Tools Evaluation based on Quality Concept Distance Computing. A Case Study
Authors: Mihai Caramihai, Irina Severin
Abstract:
Despite the extensive use of eLearning systems, there is no consensus on a standard framework for evaluating this kind of quality system. Hence, there is only a minimum set of tools that can supervise this judgment and gives information about the course content value. This paper presents two kinds of quality set evaluation indicators for eLearning courses based on the computational process of three known metrics, the Euclidian, Hamming and Levenshtein distances. The “distance" calculus is applied to standard evaluation templates (i.e. the European Commission Programme procedures vs. the AFNOR Z 76-001 Standard), determining a reference point in the evaluation of the e-learning course quality vs. the optimal concept(s). The case study, based on the results of project(s) developed in the framework of the European Programme “Leonardo da Vinci", with Romanian contractors, try to put into evidence the benefits of such a method.Keywords: eLearning, European programme, metrics, quality evaluation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15182920 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved
Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben
Abstract:
Model transformation, as a pivotal aspect of Modeldriven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: crossdomain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.Keywords: Automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21912919 The Methodology of Out-Migration in Georgia
Authors: Shorena Tsiklauri
Abstract:
Out-migration is an important issue for Georgia as well as since independence has loosed due to emigration one fifth of its population. During Soviet time out-migration from USSR was almost impossible and one of the most important instruments in regulating population movement within the Soviet Union was the system of compulsory residential registrations, so-called “propiska”. Since independent here was not any regulation for migration from Georgia. The majorities of Georgian migrants go abroad by tourist visa and then overstay, becoming the irregular labor migrants. The official statistics on migration published for this period was based on the administrative system of population registration, were insignificant in terms of numbers and did not represent the real scope of these migration movements. This paper discusses the data quality and methodology of migration statistics in Georgia and we are going to answer the questions: what is the real reason of increasing immigration flows according to the official numbers since 2000s?Keywords: Data quality, Georgia, methodology, out-migration, policy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19112918 Sorption of Nickel by Hypnea Valentiae: Application of Response Surface Methodology
Authors: M. Rajasimman, K. Murugaiyan
Abstract:
In this work, sorption of nickel from aqueous solution on hypnea valentiae, red macro algae, was investigated. Batch experiments have been carried out to find the effect of various parameters such as pH, temperature, sorbent dosage, metal concentration and contact time on the sorption of nickel using hypnea valentiae. Response surface methodology (RSM) is employed to optimize the process parameters. Based on the central composite design, quadratic model was developed to correlate the process variables to the response. The most influential factor on each experimental design response was identified from the analysis of variance (ANOVA). The optimum conditions for the sorption of nickel were found to be: pH – 5.1, temperature – 36.8oC, sorbent dosage – 5.1 g/L, metal concentration – 100 mg/L and contact time – 30 min. At these optimized conditions the maximum removal of nickel was found to be 91.97%. A coefficient of determination R2 value 0.9548 shows the fitness of response surface methodology in this work.
Keywords: Optimization, metal, Hypnea valentia, response surface methodology, red algae.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16612917 Suicide Wrongful Death: Standard of Care Problems Involving the Inaccurate Discernment of Lethal Risk When Focusing on the Elicitation of Suicide Ideation
Authors: Bill D. Geis, Frederick Newman
Abstract:
Suicide and wrongful death forensic cases are the fastest rising tort in mental health law. Most suicide-related personal injury claims fall into the legal category of “wrongful death.” Though mental health experts may be called on to address a range of forensic questions in wrongful death cases, the central consultation that most experts provide is about the negligence element—specifically, the issue of whether the clinician met the clinical standard of care in assessing, treating, and managing the deceased person’s mental health care. Standards of care, varying from US state to state, are broad and address what a reasonable clinician might do in a similar circumstance. This fact leaves the issue of the suicide standard of care, in each case, up to forensic experts to put forth a reasoned estimate of what the standard of care should have been in the specific case under litigation. Because the general state guidelines for standard of care are broad, forensic experts are readily retained to provide scientific and clinical opinions about whether or not a clinician met the standard of care in their suicide assessment, treatment, and management of the case. In the past and in much of current practice, the assessment of suicide has centered on the elicitation of verbalized suicide ideation. But suicide ideation, in the matter of suicide risk determination, may be a necessary but insufficient target of lethal suicide risk assessment. Assessment of near-term suicide risk—assessment that goes beyond verbalized suicide ideation and relates to acute crisis variables—is likely needed. Specifically, such other or additional suicide risk variable assessment may be required in the context of lethal suicide risk situations, as opposed to the discernment of general, nonlethal suicide behavior as a standard of practice (whether a patient is having suicidal thoughts or exhibiting an ambivalent suicide attempt potential). In the current study, verbalized suicide ideation information was unhelpful in the assessment of lethal risk. The Lethal Suicide Risk Assessment, Acute Model, and other dynamic, near-term risk models (such as the Acute Suicide Affective Disorder Model and the Suicide Crisis Syndrome Model)—going beyond elicited suicide ideation—need to be incorporated into current clinical suicide assessment training and become the legal standard of care for expected clinical behavior. Without this expanded clinical assessment perspective, the standard of care for suicide assessment is out of sync with current knowledge—an emerging dilemma for the forensic evaluation of suicide wrongful death cases.
Keywords: Forensic evaluation, standard of care, suicide, suicide assessment, wrongful death.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2562916 Determining the Maximum Lateral Displacement Due to Sever Earthquakes without Using Nonlinear Analysis
Authors: Mussa Mahmoudi
Abstract:
For Seismic design, it is important to estimate, maximum lateral displacement (inelastic displacement) of the structures due to sever earthquakes for several reasons. Seismic design provisions estimate the maximum roof and storey drifts occurring in major earthquakes by amplifying the drifts of the structures obtained by elastic analysis subjected to seismic design load, with a coefficient named “displacement amplification factor" which is greater than one. Here, this coefficient depends on various parameters, such as ductility and overstrength factors. The present research aims to evaluate the value of the displacement amplification factor in seismic design codes and then tries to propose a value to estimate the maximum lateral structural displacement from sever earthquakes, without using non-linear analysis. In seismic codes, since the displacement amplification is related to “force reduction factor" hence; this aspect has been accepted in the current study. Meanwhile, two methodologies are applied to evaluate the value of displacement amplification factor and its relation with the force reduction factor. In the first methodology, which is applied for all structures, the ratio of displacement amplification and force reduction factors is determined directly. Whereas, in the second methodology that is applicable just for R/C moment resisting frame, the ratio is obtained by calculating both factors, separately. The acquired results of these methodologies are alike and estimate the ratio of two factors from 1 to 1.2. The results indicate that the ratio of the displacement amplification factor and the force reduction factor differs to those proposed by seismic provisions such as NEHRP, IBC and Iranian seismic code (standard no. 2800).Keywords: Displacement amplification factor, Ductility factor, Force reduction factor, Maximum lateral displacement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28862915 Pavement Roughness Prediction Systems: A Bump Integrator Approach
Authors: Manish Pal, Rumi Sutradhar
Abstract:
Pavement surface unevenness plays a pivotal role on roughness index of road which affects on riding comfort ability. Comfort ability refers to the degree of protection offered to vehicle occupants from uneven elements in the road surface. So, it is preferable to have a lower roughness index value for a better riding quality of road users. Roughness is generally defined as an expression of irregularities in the pavement surface which can be measured using different equipments like MERLIN, Bump integrator, Profilometer etc. Among them Bump Integrator is quite simple and less time consuming in case of long road sections. A case study is conducted on low volume roads in West District in Tripura to determine roughness index (RI) using Bump Integrator at the standard speed of 32 km/h. But it becomes too tough to maintain the requisite standard speed throughout the road section. The speed of Bump Integrator (BI) has to lower or higher in some distinctive situations. So, it becomes necessary to convert these roughness index values of other speeds to the standard speed of 32 km/h. This paper highlights on that roughness index conversional model. Using SPSS (Statistical Package of Social Sciences) software a generalized equation is derived among the RI value at standard speed of 32 km/h and RI value at other speed conditions.
Keywords: Bump Integrator, Pavement Distresses, Roughness Index, SPSS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 66702914 LAYMOD; A Layered and Modular Platform for CAx Collaboration Management and Supporting Product data Integration based on STEP Standard
Authors: Omid F. Valilai, Mahmoud Houshmand
Abstract:
Nowadays companies strive to survive in a competitive global environment. To speed up product development/modifications, it is suggested to adopt a collaborative product development approach. However, despite the advantages of new IT improvements still many CAx systems work separately and locally. Collaborative design and manufacture requires a product information model that supports related CAx product data models. To solve this problem many solutions are proposed, which the most successful one is adopting the STEP standard as a product data model to develop a collaborative CAx platform. However, the improvement of the STEP-s Application Protocols (APs) over the time, huge number of STEP AP-s and cc-s, the high costs of implementation, costly process for conversion of older CAx software files to the STEP neutral file format; and lack of STEP knowledge, that usually slows down the implementation of the STEP standard in collaborative data exchange, management and integration should be considered. In this paper the requirements for a successful collaborative CAx system is discussed. The STEP standard capability for product data integration and its shortcomings as well as the dominant platforms for supporting CAx collaboration management and product data integration are reviewed. Finally a platform named LAYMOD to fulfil the requirements of CAx collaborative environment and integrating the product data is proposed. The platform is a layered platform to enable global collaboration among different CAx software packages/developers. It also adopts the STEP modular architecture and the XML data structures to enable collaboration between CAx software packages as well as overcoming the STEP standard limitations. The architecture and procedures of LAYMOD platform to manage collaboration and avoid contradicts in product data integration are introduced.Keywords: CAx, Collaboration management, STEP applicationmodules, STEP standard, XML data structures
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22182913 A Software-Supported Methodology for Designing General-Purpose Interconnection Networks for Reconfigurable Architectures
Authors: Kostas Siozios, Dimitrios Soudris, Antonios Thanailakis
Abstract:
Modern applications realized onto FPGAs exhibit high connectivity demands. Throughout this paper we study the routing constraints of Virtex devices and we propose a systematic methodology for designing a novel general-purpose interconnection network targeting to reconfigurable architectures. This network consists of multiple segment wires and SB patterns, appropriately selected and assigned across the device. The goal of our proposed methodology is to maximize the hardware utilization of fabricated routing resources. The derived interconnection scheme is integrated on a Virtex style FPGA. This device is characterized both for its high-performance, as well as for its low-energy requirements. Due to this, the design criterion that guides our architecture selections was the minimal Energy×Delay Product (EDP). The methodology is fully-supported by three new software tools, which belong to MEANDER Design Framework. Using a typical set of MCNC benchmarks, extensive comparison study in terms of several critical parameters proves the effectiveness of the derived interconnection network. More specifically, we achieve average Energy×Delay Product reduction by 63%, performance increase by 26%, reduction in leakage power by 21%, reduction in total energy consumption by 11%, at the expense of increase of channel width by 20%.
Keywords: Design Methodology, FPGA, Interconnection, Low-Energy, High-Performance, CAD tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17202912 SPH Method used for Flow Predictions at a Turgo Impulse Turbine: Comparison with Fluent
Authors: Phoevos K. Koukouvinis, John S. Anagnostopoulos, Dimitris E. Papantonis
Abstract:
This work is an attempt to use the standard Smoothed Particle Hydrodynamics methodology for the simulation of the complex unsteady, free-surface flow in a rotating Turgo impulse water turbine. A comparison of two different geometries was conducted. The SPH method due to its mesh-less nature is capable of capturing the flow features appearing in the turbine, without diffusion at the water/air interface. Furthermore results are compared with a commercial CFD package (Fluent®) and the SPH algorithm proves to be capable of providing similar results, in much less time than the mesh based CFD program. A parametric study was also performed regarding the turbine inlet angle.Keywords: Smoothed Particle Hydrodynamics, Mesh-lessmethods, Impulse turbines, Turgo turbine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26402911 Towards a Systematic Evaluation of Web Design
Authors: Ivayla Trifonova, Naoum Jamous, Holger Schrödl
Abstract:
A good web design is a prerequisite for a successful business nowadays, especially since the internet is the most common way for people to inform themselves. Web design includes the optical composition, the structure, and the user guidance of websites. The importance of each website leads to the question if there is a way to measure its usefulness. The aim of this paper is to suggest a methodology for the evaluation of web design. The desired outcome is to have an evaluation that is concentrated on a specific website and its target group.
Keywords: Evaluation methodology, factor analysis, target group, web design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12662910 A New Cut–Through Mechanism in IEEE 802.16 Mesh Networks
Authors: Yi-Ting Mai, Chun-Chuan Yang, Cheng-Jung Wen
Abstract:
IEEE 802.16 is a new wireless technology standard, it has some advantages, including wider coverage, higher bandwidth, and QoS support. As the new wireless technology for last mile solution, there are designed two models in IEEE 802.16 standard. One is PMP (point to multipoint) and the other is Mesh. In this paper we only focus on IEEE 802.16 Mesh model. According to the IEEE 802.16 standard description, Mesh model has two scheduling modes, centralized and distributed. Considering the pros and cons of the two scheduling, we present the combined scheduling QoS framework that the BS (Base Station) controls time frame scheduling and selects the shortest path from source to destination directly. On the other hand, we propose the Expedited Queue mechanism to cut down the transmission time. The EQ mechanism can reduce a lot of end-to-end delay in our QoS framework. Simulation study has shown that the average delay is smaller than contrasts. Furthermore, our proposed scheme can also achieve higher performance.Keywords: IEEE 802.16 Mesh, Scheduling, Expedited Queue, QoS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18432909 Reformulations of Big Bang-Big Crunch Algorithm for Discrete Structural Design Optimization
Authors: O. Hasançebi, S. Kazemzadeh Azad
Abstract:
In the present study the efficiency of Big Bang-Big Crunch (BB-BC) algorithm is investigated in discrete structural design optimization. It is shown that a standard version of the BB-BC algorithm is sometimes unable to produce reasonable solutions to problems from discrete structural design optimization. Two reformulations of the algorithm, which are referred to as modified BB-BC (MBB-BC) and exponential BB-BC (EBB-BC), are introduced to enhance the capability of the standard algorithm in locating good solutions for steel truss and frame type structures, respectively. The performances of the proposed algorithms are experimented and compared to its standard version as well as some other algorithms over several practical design examples. In these examples, steel structures are sized for minimum weight subject to stress, stability and displacement limitations according to the provisions of AISC-ASD.Keywords: Structural optimization, discrete optimization, metaheuristics, big bang-big crunch (BB-BC) algorithm, design optimization of steel trusses and frames.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23892908 Study of Aerodynamic Characteristics of the Unmanned Aircraft in the Wake
Authors: O. Solovyov, S. Eryomenko, V. Kobrin, V. Chmovzh
Abstract:
The methodology of numerical simulation and calculation of aerodynamic characteristics of aircraft taking into account impact of wake on it has been developed. The results of numerical experiment in comparison with the data obtained in the wind tunnel are presented. Efficiency of methodology of calculation and the reliability of the results is shown.
Keywords: Unmanned aircraft, vortex wake, aerodynamic characteristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17872907 Thermodynamic Approach of Lanthanide-Iron Double Oxides Formation
Authors: Vera Varazashvili, Murman Tsarakhov, Tamar Mirianashvili, Teimuraz Pavlenishvili, Tengiz Machaladze, Mzia Khundadze
Abstract:
Standard Gibbs energy of formation ΔGfor(298.15) of lanthanide-iron double oxides of garnet-type crystal structure R3Fe5O12 - RIG (R – are rare earth ions) from initial oxides are evaluated. The calculation is based on the data of standard entropies S298.15 and standard enthalpies ΔH298.15 of formation of compounds which are involved in the process of garnets synthesis. Gibbs energy of formation is presented as temperature function ΔGfor(T) for the range 300-1600K. The necessary starting thermodynamic data were obtained from calorimetric study of heat capacity – temperature functions and by using the semi-empirical method for calculation of ΔH298.15 of formation. Thermodynamic functions for standard temperature – enthalpy, entropy and Gibbs energy - are recommended as reference data for technological evaluations. Through the structural series of rare earth-iron garnets the correlation between thermodynamic properties and characteristics of lanthanide ions are elucidated.Keywords: Calorimetry, entropy, enthalpy, heat capacity, gibbs energy of formation, rare earth iron garnets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19272906 A Multi-Objective Methodology for Selecting Lean Initiatives in Modular Construction Companies
Authors: Saba Shams Bidhendi, Steven Goh, Andrew Wandel
Abstract:
The implementation of lean manufacturing initiatives has produced significant impacts in improving operational performance and reducing manufacturing wastes in the production process. However, selecting an appropriate set of lean strategies is critical to avoid misapplication of the lean manufacturing techniques and consequential increase in non-value-adding activities. To the author’s best knowledge, there is currently no methodology to select lean strategies that considers their impacts on manufacturing wastes and performance metrics simultaneously. In this research, a multi-objective methodology is proposed that suggests an appropriate set of lean initiatives based on their impacts on performance metrics and manufacturing wastes and within manufacturers’ resource limitation. The proposed methodology in this research suggests the best set of lean initiatives for implementation that have highest impacts on identified critical performance metrics and manufacturing wastes. Therefore, manufacturers can assure that implementing suggested lean tools improves their production performance and reduces manufacturing wastes at the same time. A case study was conducted to show the effectiveness and validate the proposed model and methodologies.
Keywords: Lean manufacturing, Lean strategies, manufacturing wastes, manufacturing performance metrics, decision making, optimisation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7942905 A Methodology to Analyze Technology Convergence: Patent-Citation Based Technology Input-Output Analysis
Authors: Jeeeun Kim, Sungjoo Lee
Abstract:
This research proposes a methodology for patent-citation-based technology input-output analysis by applying the patent information to input-output analysis developed for the dependencies among different industries. For this analysis, a technology relationship matrix and its components, as well as input and technology inducement coefficients, are constructed using patent information. Then, a technology inducement coefficient is calculated by normalizing the degree of citation from certain IPCs to the different IPCs (International patent classification) or to the same IPCs. Finally, we construct a Dependency Structure Matrix (DSM) based on the technology inducement coefficient to suggest a useful application for this methodology.
Keywords: Technology spillover effect, technology relationship, IO table, technology inducement coefficients, patent analysis, patent citation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25732904 A Novel Solution to Restricted Earth Fault Low Impedance Relay Maloperation
Authors: K. N. Dinesh Babu, R. Ramaprabha, V. Rajini, V. Nagarajan
Abstract:
In this paper, various methods of providing restricted earth fault protection are discussed. The proper operation of high and low impedance Restricted Earth Fault (REF) protection for various applications has been discussed. The maloperation of a relay due to improper placement of CTs has been identified and a simple/unique solution has been proposed in this work with a case study. Moreover, it is found that the proper placement of CT in high impedance method will provide the same result with reduced CT. This methodology has been successfully implemented in Al Takreer refinery for a 2000 KVA transformer. The outcome of the paper may be included in IEEEC37.91 standard to give the proper guidance for protection engineers to sort out the issues related to mal functioning of REF relays.Keywords: Relay maloperation, transformer, low impedance REF, MatLab, 64R, IEEE C37.91.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39242903 The International Labor Standard on the Elimination of Discrimination in Employment: Response and Prospect of Malaysia
Authors: Harlida Abdul Wahab
Abstract:
Discrimination in employment has its wider social and economic consequences other than mere violating a basic human right. Discrimination involves treating people differently because of certain grounds such as race, color, or sex, which results in the impairment of equality of opportunity and treatment. As an essential part of promoting decent work, combating discrimination through the principle of non-discrimination has been established by the International Labor Organization (ILO) through the Declaration on Fundamental Principles and Rights at Work 1998. Considering elimination of discrimination in employment as a core labor standard, member states are expected to respect, promote and implement it to their national laws and policies. Being a member state, Malaysia has to position herself align with this international requirement. The author discusses the related convention together with Malaysia-s responses on the matter. At the closing stage, the prospect of Malaysia is presumed taking into account of the current positions and reports submitted to the ILO.Keywords: Discrimination, employment, international labor standard, Malaysia.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1871