Search results for: particle based method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16304

Search results for: particle based method

1154 Energy Management System with Temperature Rise Prevention on Hybrid Ships

Authors: Asser S. Abdelwahab, Nabil H. Abbasy, Ragi A. Hamdy

Abstract:

Marine shipping has now become one of the major worldwide contributors to pollution and greenhouse gas emissions. Hybrid ships technology based on multiple energy sources has taken a great scope of research to get rid of ship emissions and cut down fuel expenses. Insufficiency between power generated and the demand load to withstand the transient behavior on ships during severe climate conditions will lead to a blackout. Thus, an efficient energy management system (EMS) is a mandatory scope for achieving higher system efficiency while enhancing the lifetime of the onboard storage systems is another salient EMS scope. Considering energy storage system conditions, both the battery state of charge (SOC) and temperature represent important parameters to prevent any malfunction of the storage system that eventually degrades the whole system. In this paper, a two battery packs ratio fuzzy logic control model is proposed. The overall aim is to control the charging/discharging current while including both the battery SOC and temperature in the energy management system. The full designs of the proposed controllers are described and simulated using Matlab. The results prove the successfulness of the proposed controller in stabilizing the system voltage during both loading and unloading while keeping the energy storage system in a healthy condition.

Keywords: energy storage system, fuzzy logic control, hybrid ship, thermal runaway

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 575
1153 Epic Consciousness: A Possibility for Epic Expression in Post-War American Literature During the Age of Late Capitalism

Authors: Safwa Yargui

Abstract:

This research is about the quest for a post-war American epic poem in the age of late capitalism. This paper explores the possibility of an epic poem in the context of post-war late capitalist America, despite the prevailing scholarly scepticism regarding the existence of epic poetry after Milton’s Paradise Lost. The aim of this paper is to demonstrate the possibility of a post-war American epic through the argument of epic consciousness by distinguishing the scholarly views on the impossibility of an epic poem and the counterargument of epic consciousness which contends that American epic poetry is possible. Epic consciousness provides a significant nuance to the reading of the post-war American epic by focusing on the epic’s responsiveness to late capitalism, via various language forms; cultural manifestations; and conscious distortions of late capitalist media-related language; in addition to the epic’ conscious inclusion of the process of writing a post-war epic that requires a direct engagement with American-based materials. By focusing on interdisciplinary theoretical approaches, this paper includes both socio-cultural literary theories as well as literary and epic approaches developed by scholars that respectively contextualise the late capitalist situation and the question of post-war American epic poetry. In examining the role of consciousness, this paper aims to suggest a re-thinking of the post-war American epic that is capable of self-commitment, for the purpose of achieving a new sense of epic poetry in post-war late capitalist America.

Keywords: American epic, epic consciousness, late capitalism, post-war poetry.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 211
1152 Vibroacoustic Modulation of Wideband Vibrations and Its Possible Application for Windmill Blade Diagnostics

Authors: Abdullah Alnutayfat, Alexander Sutin, Dong Liu

Abstract:

Wind turbine has become one of the most popular energy production methods. However, failure of blades and maintenance costs evolve into significant issues in the wind power industry, so it is essential to detect the initial blade defects to avoid the collapse of the blades and structure. This paper aims to apply modulation of high-frequency blade vibrations by low-frequency blade rotation, which is close to the known Vibro-Acoustic Modulation (VAM) method. The high-frequency wideband blade vibration is produced by the interaction of the surface blades with the environment air turbulence, and the low-frequency modulation is produced by alternating bending stress due to gravity. The low-frequency load of rotational wind turbine blades ranges between 0.2-0.4 Hz and can reach up to 2 Hz for strong wind. The main difference between this study and previous ones on VAM methods is the use of a wideband vibration signal from the blade's natural vibrations. Different features of the VAM are considered using a simple model of breathing crack. This model considers the simple mechanical oscillator, where the parameters of the oscillator are varied due to low-frequency blade rotation. During the blade's operation, the internal stress caused by the weight of the blade modifies the crack's elasticity and damping. The laboratory experiment using steel samples demonstrates the possibility of VAM using a probe wideband noise signal. A cycle load with a small amplitude was used as a pump wave to damage the tested sample, and a small transducer generated a wideband probe wave. The received signal demodulation was conducted using the Detecting of Envelope Modulation on Noise (DEMON) approach. In addition, the experimental results were compared with the modulation index (MI) technique regarding the harmonic pump wave. The wideband and traditional VAM methods demonstrated similar sensitivity for earlier detection of invisible cracks. Importantly, employing a wideband probe signal with the DEMON approach speeds up and simplifies testing since it eliminates the need to conduct tests repeatedly for various harmonic probe frequencies and to adjust the probe frequency.

Keywords: Damage detection, turbine blades, Vibro-Acoustic Structural Health Monitoring, SHM, Detecting of Envelope Modulation on Noise.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 410
1151 Data Mining to Capture User-Experience: A Case Study in Notebook Product Appearance Design

Authors: Rhoann Kerh, Chen-Fu Chien, Kuo-Yi Lin

Abstract:

In the era of rapidly increasing notebook market, consumer electronics manufacturers are facing a highly dynamic and competitive environment. In particular, the product appearance is the first part for user to distinguish the product from the product of other brands. Notebook product should differ in its appearance to engage users and contribute to the user experience (UX). The UX evaluates various product concepts to find the design for user needs; in addition, help the designer to further understand the product appearance preference of different market segment. However, few studies have been done for exploring the relationship between consumer background and the reaction of product appearance. This study aims to propose a data mining framework to capture the user’s information and the important relation between product appearance factors. The proposed framework consists of problem definition and structuring, data preparation, rules generation, and results evaluation and interpretation. An empirical study has been done in Taiwan that recruited 168 subjects from different background to experience the appearance performance of 11 different portable computers. The results assist the designers to develop product strategies based on the characteristics of consumers and the product concept that related to the UX, which help to launch the products to the right customers and increase the market shares. The results have shown the practical feasibility of the proposed framework.

Keywords: Consumers Decision Making, Product Design, Rough Set Theory, User Experience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3499
1150 Trimmed Mean as an Adaptive Robust Estimator of a Location Parameter for Weibull Distribution

Authors: Carolina B. Baguio

Abstract:

One of the purposes of the robust method of estimation is to reduce the influence of outliers in the data, on the estimates. The outliers arise from gross errors or contamination from distributions with long tails. The trimmed mean is a robust estimate. This means that it is not sensitive to violation of distributional assumptions of the data. It is called an adaptive estimate when the trimming proportion is determined from the data rather than being fixed a “priori-. The main objective of this study is to find out the robustness properties of the adaptive trimmed means in terms of efficiency, high breakdown point and influence function. Specifically, it seeks to find out the magnitude of the trimming proportion of the adaptive trimmed mean which will yield efficient and robust estimates of the parameter for data which follow a modified Weibull distribution with parameter λ = 1/2 , where the trimming proportion is determined by a ratio of two trimmed means defined as the tail length. Secondly, the asymptotic properties of the tail length and the trimmed means are also investigated. Finally, a comparison is made on the efficiency of the adaptive trimmed means in terms of the standard deviation for the trimming proportions and when these were fixed a “priori". The asymptotic tail lengths defined as the ratio of two trimmed means and the asymptotic variances were computed by using the formulas derived. While the values of the standard deviations for the derived tail lengths for data of size 40 simulated from a Weibull distribution were computed for 100 iterations using a computer program written in Pascal language. The findings of the study revealed that the tail lengths of the Weibull distribution increase in magnitudes as the trimming proportions increase, the measure of the tail length and the adaptive trimmed mean are asymptotically independent as the number of observations n becomes very large or approaching infinity, the tail length is asymptotically distributed as the ratio of two independent normal random variables, and the asymptotic variances decrease as the trimming proportions increase. The simulation study revealed empirically that the standard error of the adaptive trimmed mean using the ratio of tail lengths is relatively smaller for different values of trimming proportions than its counterpart when the trimming proportions were fixed a 'priori'.

Keywords: Adaptive robust estimate, asymptotic efficiency, breakdown point, influence function, L-estimates, location parameter, tail length, Weibull distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2058
1149 Two-Stage Launch Vehicle Trajectory Modeling for Low Earth Orbit Applications

Authors: Assem M. F. Sallam, Ah. El-S. Makled

Abstract:

This paper presents a study on the trajectory of a two stage launch vehicle. The study includes dynamic responses of motion parameters as well as the variation of angles affecting the orientation of the launch vehicle (LV). LV dynamic characteristics including state vector variation with corresponding altitude and velocity for the different LV stages separation, as well as the angle of attack and flight path angles are also discussed. A flight trajectory study for the drop zone of first stage and the jettisoning of fairing are introduced in the mathematical modeling to study their effect. To increase the accuracy of the LV model, atmospheric model is used taking into consideration geographical location and the values of solar flux related to the date and time of launch, accurate atmospheric model leads to enhancement of the calculation of Mach number, which affects the drag force over the LV. The mathematical model is implemented on MATLAB based software (Simulink). The real available experimental data are compared with results obtained from the theoretical computation model. The comparison shows good agreement, which proves the validity of the developed simulation model; the maximum error noticed was generally less than 10%, which is a result that can lead to future works and enhancement to decrease this level of error.

Keywords: Launch vehicle modeling, launch vehicle trajectory, mathematical modeling, MATLAB-Simulink.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3267
1148 Mining Network Data for Intrusion Detection through Naïve Bayesian with Clustering

Authors: Dewan Md. Farid, Nouria Harbi, Suman Ahmmed, Md. Zahidur Rahman, Chowdhury Mofizur Rahman

Abstract:

Network security attacks are the violation of information security policy that received much attention to the computational intelligence society in the last decades. Data mining has become a very useful technique for detecting network intrusions by extracting useful knowledge from large number of network data or logs. Naïve Bayesian classifier is one of the most popular data mining algorithm for classification, which provides an optimal way to predict the class of an unknown example. It has been tested that one set of probability derived from data is not good enough to have good classification rate. In this paper, we proposed a new learning algorithm for mining network logs to detect network intrusions through naïve Bayesian classifier, which first clusters the network logs into several groups based on similarity of logs, and then calculates the prior and conditional probabilities for each group of logs. For classifying a new log, the algorithm checks in which cluster the log belongs and then use that cluster-s probability set to classify the new log. We tested the performance of our proposed algorithm by employing KDD99 benchmark network intrusion detection dataset, and the experimental results proved that it improves detection rates as well as reduces false positives for different types of network intrusions.

Keywords: Clustering, detection rate, false positive, naïveBayesian classifier, network intrusion detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5525
1147 Application of Seismic Isolators in Kutahya City Hospital Project Utilizing Double Friction Pendulum Type Devices

Authors: Kaan Yamanturk, Cihan Dogruoz

Abstract:

Seismic isolators have been utilized around the world to protect the structures, nonstructural components and contents from the damaging effects of earthquakes. In Structural Engineering, seismic isolation is used for protecting buildings and its vibration-sensitive contents from earthquakes. Seismic isolation is a passive control system that lowers effective earthquake forces by utilizing flexible bearings. One of the most significant isolation systems is seismic isolators. In this paper, double pendulum type Teflon coated seismic isolators utilized in a city hospital project by Guris Construction and Engineering Co. Inc, located in Kutahya, Turkey, have been investigated. Totally, 498 seismic isolators were applied in the project. These isolators are double friction pendulum type seismic isolation devices. The review of current practices is also examined in this study. The focus of this study is related to the application of passive seismic isolation systems for buildings as practiced in Kutahya City Hospital Project. Based on the study, the acceleration at the top floor will be 0.18 g and it will decrease 0.01 g in every floor. Therefore, seismic isolators are very important for buildings located in earthquake zones.

Keywords: Maximum considered earthquake, moment resisting frame, seismic isolator, seismic design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 656
1146 Elections, Checks and Balances, and Government Expenditures: Empirical Evidence for Japan, South Korea, and Taiwan

Authors: Yuan-Hong Ho, Chiung-Ju Huang

Abstract:

Previous studies on political budget cycles (PBCs) implicitly assume the executive has full discretion power over fiscal policy, neglecting the role of checks and balances of the legislature. This paper goes beyond traditional PBCs models and sheds light on the case study of Japan, South Korea, and Taiwan over the 1988-2007 periods. Based on the results, we find no evidence of electoral impacts on the public expenditures in South Korean and Taiwan's congressional elections. We also noted that PBCs are found on Taiwan-s government expenditures during our sample periods. Furthermore, the results also show that Japan-s legislature has a significant checks and balances on government-s expenditures. However, empirical results show that the legislature veto player in Taiwan neither has effect on the reduction of public expenditures, nor has the moderating effect over Taiwan-s political budget cycles, albeit that they are statistically insignificant.We suggest that the existence of PBCs in Taiwan is due to a weaker systemof checks and balances. Our conjecture is that Taiwan either has no legislative veto player or has observed low compliance to the law during the time period examined in our study.

Keywords: Checks and balances, compliance to the law, political budget cycles, veto player.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1719
1145 Extraction of Bran Protein Using Enzymes and Polysaccharide Precipitation

Authors: Sudarat Jiamyangyuen, Tipawan Thongsook, Riantong Singanusong, Chanida Saengtubtim

Abstract:

Rice bran is normally used as a raw material for rice bran oil production or sold as feed with a low price. Conventionally, the protein in defatted rice bran was extracted using alkaline extraction and acid precipitation, which involves in chemical usage and lowering some nutritious component. This study was conducted in order to extract of rice bran protein concentrate (RBPC) from defatted rice bran using enzymes and employing polysaccharides in a precipitating step. The properties of RBPC obtained will be compared to those of a control sample extracted using a conventional method. The results showed that extraction of protein from rice bran using enzymes exhibited the higher protein recovery compared to that extraction with alkaline. The extraction conditions using alcalase 2% (v/w) at 50 C, pH 9.5 gave the highest protein (2.44%) and yield (32.09%) in extracted solution compared to other enzymes. Rice bran protein concentrate powder prepared by a precipitation step using alginate (protein in solution: alginate 1:0.016) exhibited the highest protein (27.55%) and yield (6.84%). Precipitation using alginate was better than that of acid. RBPC extracted with alkaline (ALK) or enzyme alcalase (ALC), then precipitated with alginate (AL) (samples RBP-ALK-AL and RBP-ALC-AL) yielded the precipitation rate of 75% and 91.30%, respectively. Therefore, protein precipitation using alginate was then selected. Amino acid profile of control sample, and sample precipitated with alginate, as compared to casein and soy protein isolated, showed that control sample showed the highest content among all sample. Functional property study of RBP showed that the highest nitrogen solubility occurred in pH 8-10. There was no statically significant between emulsion capacity and emulsion stability of control and sample precipitated by alginate. However, control sample showed a higher of foaming capacity and foaming stability compared to those of sample precipitated with alginate. The finding was successful in terms of minimizing chemicals used in extraction and precipitation steps in preparation of rice bran protein concentrate. This research involves in a production of value-added product in which the double amount of protein (28%) compared to original amount (14%) contained in rice bran could be beneficial in terms of adding to food products e.g. healthy drink with high protein and fiber. In addition, the basic knowledge of functional property of rice bran protein concentrate was obtained, which can be used to appropriately select the application of this value-added product from rice bran.

Keywords: Alginate, carrageenan, rice bran, rice bran protein.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2649
1144 Kinematic Modeling and Workspace Analysis of a Spatial Cable Suspended Robot as Incompletely Restrained Positioning Mechanism

Authors: Jahanbakhsh Hamedi, Hassan Zohoor

Abstract:

This article proposes modeling, simulation and kinematic and workspace analysis of a spatial cable suspended robot as incompletely Restrained Positioning Mechanism (IRPM). These types of robots have six cables equal to the number of degrees of freedom. After modeling, the kinds of workspace are defined then an statically reachable combined workspace for different geometric structures of fixed and moving platform is obtained. This workspace is defined as the situations of reference point of the moving platform (center of mass) which under external forces such as weight and with ignorance of inertial effects, the moving platform should be in static equilibrium under conditions that length of all cables must not be exceeded from the maximum value and all of cables must be at tension (they must have non-negative tension forces). Then the effect of various parameters such as the size of moving platform, the size of fixed platform, geometric configuration of robots, magnitude of applied forces and moments to moving platform on workspace of these robots with different geometric configuration are investigated. Obtained results should be effective in employing these robots under different conditions of applied wrench for increasing the workspace volume.

Keywords: Kinematic modeling, applied wrench, workspace, cable based robot.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1684
1143 Timescape-Based Panoramic View for Historic Landmarks

Authors: H. Ali, A. Whitehead

Abstract:

Providing a panoramic view of famous landmarks around the world offers artistic and historic value for historians, tourists, and researchers. Exploring the history of famous landmarks by presenting a comprehensive view of a temporal panorama merged with geographical and historical information presents a unique challenge of dealing with images that span a long period, from the 1800’s up to the present. This work presents the concept of temporal panorama through a timeline display of aligned historic and modern images for many famous landmarks. Utilization of this panorama requires a collection of hundreds of thousands of landmark images from the Internet comprised of historic images and modern images of the digital age. These images have to be classified for subset selection to keep the more suitable images that chronologically document a landmark’s history. Processing of historic images captured using older analog technology under various different capturing conditions represents a big challenge when they have to be used with modern digital images. Successful processing of historic images to prepare them for next steps of temporal panorama creation represents an active contribution in cultural heritage preservation through the fulfillment of one of UNESCO goals in preservation and displaying famous worldwide landmarks.

Keywords: Cultural heritage, image registration, image subset selection, registered image similarity, temporal panorama, timescapes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1032
1142 Extending BDI Multiagent Systems with Agent Norms

Authors: Francisco José Plácido da Cunha, Tassio Ferenzini Martins Sirqueira, Marx Leles Viana, Carlos José Pereira de Lucena

Abstract:

Open Multiagent Systems (MASs) are societies in which heterogeneous and independently designed entities (agents) work towards similar, or different ends. Software agents are autonomous and the diversity of interests among different members living in the same society is a fact. In order to deal with this autonomy, these open systems use mechanisms of social control (norms) to ensure a desirable social order. This paper considers the following types of norms: (i) obligation — agents must accomplish a specific outcome; (ii) permission — agents may act in a particular way, and (iii) prohibition — agents must not act in a specific way. All of these characteristics mean to encourage the fulfillment of norms through rewards and to discourage norm violation by pointing out the punishments. Once the software agent decides that its priority is the satisfaction of its own desires and goals, each agent must evaluate the effects associated to the fulfillment of one or more norms before choosing which one should be fulfilled. The same applies when agents decide to violate a norm. This paper also introduces a framework for the development of MASs that provide support mechanisms to the agent’s decision-making, using norm-based reasoning. The applicability and validation of this approach is demonstrated applying a traffic intersection scenario.

Keywords: BDI aAgent, BDI4JADE framework, multiagent system, normative agents.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 981
1141 Emergence of Fluoroquinolone Resistance in Pigs, Nigeria

Authors: Igbakura I. Luga, Alex A. Adikwu

Abstract:

A comparison of resistance to quinolones was carried out on isolates of Shiga toxin-producing Escherichia coliO157:H7 from cattle and mecA and nuc genes harbouring Staphylococcus aureus from pigs. The isolates were separately tested in the first and current decades of the 21st century. The objective was to demonstrate the dissemination of resistance to this frontline class of antibiotic by bacteria from food animals and bring to the limelight the spread of antibiotic resistance in Nigeria. A total of 10 isolates of the E. coli O157:H7 and 9 of mecA and nuc genes harbouring S. aureus were obtained following isolation, biochemical testing, and serological identification using the Remel Wellcolex E. coli O157:H7 test. Shiga toxin-production screening in the E. coli O157:H7 using the verotoxin E. coli reverse passive latex agglutination (VTEC-RPLA) test; and molecular identification of the mecA and nuc genes in S. aureus. Detection of the mecA and nuc genes were carried out using the protocol by the Danish Technical University (DTU) using the following primers mecA-1:5'-GGGATCATAGCGTCATTATTC-3', mecA-2: 5'-AACGATTGTGACACGATAGCC-3', nuc-1: 5'-TCAGCAAATGCATCACAAACAG-3', nuc-2: 5'-CGTAAATGCACTTGCTTCAGG-3' for the mecA and nuc genes, respectively. The nuc genes confirm the S. aureus isolates and the mecA genes as being methicillin-resistant and so pathogenic to man. The fluoroquinolones used in the antibiotic resistance testing were norfloxacin (10 µg) and ciprofloxacin (5 µg) in the E. coli O157:H7 isolates and ciprofloxacin (5 µg) in the S. aureus isolates. Susceptibility was tested using the disk diffusion method on Muller-Hinton agar. Fluoroquinolone resistance was not detected from isolates of E. coli O157:H7 from cattle. However, 44% (4/9) of the S. aureus were resistant to ciprofloxacin. Resistance of up to 44% in isolates of mecA and nuc genes harbouring S. aureus is a compelling evidence for the rapid spread of antibiotic resistance from bacteria in food animals from Nigeria. Ciprofloxacin is the drug of choice for the treatment of Typhoid fever, therefore widespread resistance to it in pathogenic bacteria is of great public health significance. The study concludes that antibiotic resistance in bacteria from food animals is on the increase in Nigeria. The National Food and Drug Administration and Control (NAFDAC) agency in Nigeria should implement the World Health Organization (WHO) global action plan on antimicrobial resistance. A good starting point can be coordinating the WHO, Office of International Epizootics (OIE), Food and Agricultural Organization (FAO) tripartite draft antimicrobial resistance monitoring and evaluation (M&E) framework in Nigeria.

Keywords: Fluoroquinolone, Nigeria, resistance, Staphylococcus aureus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1004
1140 Theoretical Modeling and Experimental Study of Combustion and Performance Characteristics of Biodiesel in Turbocharged Low Heat Rejection D.I Diesel Engine

Authors: B.Rajendra Prasath, P.Tamilporai, Mohd.F.Shabir

Abstract:

An effort has been taken to simulate the combustion and performance characteristics of biodiesel fuel in direct injection (D.I) low heat rejection (LHR) diesel engine. Comprehensive analyses on combustion characteristics such as cylinder pressure, peak cylinder pressure, heat release and performance characteristics such as specific fuel consumption and brake thermal efficiency are carried out. Compression ignition (C.I) engine cycle simulation was developed and modified in to LHR engine for both diesel and biodiesel fuel. On the basis of first law of thermodynamics the properties at each degree crank angle was calculated. Preparation and reaction rate model was used to calculate the instantaneous heat release rate. A gas-wall heat transfer calculations are based on the ANNAND-s combined heat transfer model with instantaneous wall temperature to analyze the effect of coating on heat transfer. The simulated results are validated by conducting the experiments on the test engine under identical operating condition on a turbocharged D.I diesel engine. In this analysis 20% of biodiesel (derived from Jatropha oil) blended with diesel and used in both conventional and LHR engine. The simulated combustion and performance characteristics results are found satisfactory with the experimental value.

Keywords: Biodiesel, Direct injection, Low heat rejection, Turbocharged engine

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2761
1139 Large Eddy Simulation of Compartment Fire with Gas Combustible

Authors: Mliki Bouchmel, Abbassi Mohamed Ammar, Kamel Geudri, Chrigui Mouldi, Omri Ahmed

Abstract:

The objective of this work is to use the Fire Dynamics Simulator (FDS) to investigate the behavior of a kerosene small-scale fire. FDS is a Computational Fluid Dynamics (CFD) tool developed specifically for fire applications. Throughout its development, FDS is used for the resolution of practical problems in fire protection engineering. At the same time FDS is used to study fundamental fire dynamics and combustion. Predictions are based on Large Eddy Simulation (LES) with a Smagorinsky turbulence model. LES directly computes the large-scale eddies and the sub-grid scale dissipative processes are modeled. This technique is the default turbulence model which was used in this study. The validation of the numerical prediction is done using a direct comparison of combustion output variables to experimental measurements. Effect of the mesh size on the temperature evolutions is investigated and optimum grid size is suggested. Effect of width openings is investigated. Temperature distribution and species flow are presented for different operating conditions. The effect of the composition of the used fuel on atmospheric pollution is also a focus point within this work. Good predictions are obtained where the size of the computational cells within the fire compartment is less than 1/10th of the characteristic fire diameter.

Keywords: Large eddy simulation, Radiation, Turbulence, combustion, pollution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2168
1138 Optimal Manufacturing Scheduling for Dependent Details Processing

Authors: Ivan C. Mustakerov, Daniela I. Borissova

Abstract:

The increasing competitiveness in manufacturing industry is forcing manufacturers to seek effective processing schedules. The paper presents an optimization manufacture scheduling approach for dependent details processing with given processing sequences and times on multiple machines. By defining decision variables as start and end moments of details processing it is possible to use straightforward variables restrictions to satisfy different technological requirements and to formulate easy to understand and solve optimization tasks for multiple numbers of details and machines. A case study example is solved for seven base moldings for CNC metalworking machines processed on five different machines with given processing order among details and machines and known processing time-s duration. As a result of linear optimization task solution the optimal manufacturing schedule minimizing the overall processing time is obtained. The manufacturing schedule defines the moments of moldings delivery thus minimizing storage costs and provides mounting due-time satisfaction. The proposed optimization approach is based on real manufacturing plant problem. Different processing schedules variants for different technological restrictions were defined and implemented in the practice of Bulgarian company RAIS Ltd. The proposed approach could be generalized for other job shop scheduling problems for different applications.

Keywords: Optimal manufacturing scheduling, linear programming, metalworking machines production, dependant details processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1470
1137 A User Friendly Tool for Performance Evaluation of Different Reference Evapotranspiration Methods

Authors: Vijay Shankar

Abstract:

Evapotranspiration (ET) is a major component of the hydrologic cycle and its accurate estimation is essential for hydrological studies. In past, various estimation methods have been developed for different climatological data, and the accuracy of these methods varies with climatic conditions. Reference crop evapotranspiration (ET0) is a key variable in procedures established for estimating evapotranspiration rates of agricultural crops. Values of ET0 are used with crop coefficients for many aspects of irrigation and water resources planning and management. Numerous methods are used for estimating ET0. As per internationally accepted procedures outlined in the United Nations Food and Agriculture Organization-s Irrigation and Drainage Paper No. 56(FAO-56), use of Penman-Monteith equation is recommended for computing ET0 from ground based climatological observations. In the present study, seven methods have been selected for performance evaluation. User friendly software has been developed using programming language visual basic. The visual basic has ability to create graphical environment using less coding. For given data availability the developed software estimates reference evapotranspiration for any given area and period for which data is available. The accuracy of the software has been checked by the examples given in FAO-56.The developed software is a user friendly tool for estimating ET0 under different data availability and climatic conditions.

Keywords: Crop coefficient, Crop evapotranspiration, Field moisture, Irrigation Scheduling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1642
1136 Evaluation of Exerting Force on the Heating Surface Due to Bubble Ebullition in Subcooled Flow Boiling

Authors: M. R. Nematollahi

Abstract:

Vibration characteristics of subcooled flow boiling on thin and long structures such as a heating rod were recently investigated by the author. The results show that the intensity of the subcooled boiling-induced vibration (SBIV) was influenced strongly by the conditions of the subcooling temperature, linear power density and flow velocity. Implosive bubble formation and collapse are the main nature of subcooled boiling, and their behaviors are the only sources to originate from SBIV. Therefore, in order to explain the phenomenon of SBIV, it is essential to obtain reliable information about bubble behavior in subcooled boiling conditions. This was investigated at different conditions of coolant subcooling temperatures of 25 to 75°C, coolant flow velocities of 0.16 to 0.53m/s, and linear power densities of 100 to 600 W/cm. High speed photography at 13,500 frames per second was performed at these conditions. The results show that even at the highest subcooling condition, the absolute majority of bubbles collapse very close to the surface after detaching from the heating surface. Based on these observations, a simple model of surface tension and momentum change is introduced to offer a rough quantitative estimate of the force exerted on the heating surface during the bubble ebullition. The formation of a typical bubble in subcooled boiling is predicted to exert an excitation force in the order of 10-4 N.

Keywords: Subcooled boiling, vibration mechanism, bubble behavior.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1530
1135 Increase of Organization in Complex Systems

Authors: Georgi Yordanov Georgiev, Michael Daly, Erin Gombos, Amrit Vinod, Gajinder Hoonjan

Abstract:

Measures of complexity and entropy have not converged to a single quantitative description of levels of organization of complex systems. The need for such a measure is increasingly necessary in all disciplines studying complex systems. To address this problem, starting from the most fundamental principle in Physics, here a new measure for quantity of organization and rate of self-organization in complex systems based on the principle of least (stationary) action is applied to a model system - the central processing unit (CPU) of computers. The quantity of organization for several generations of CPUs shows a double exponential rate of change of organization with time. The exact functional dependence has a fine, S-shaped structure, revealing some of the mechanisms of self-organization. The principle of least action helps to explain the mechanism of increase of organization through quantity accumulation and constraint and curvature minimization with an attractor, the least average sum of actions of all elements and for all motions. This approach can help describe, quantify, measure, manage, design and predict future behavior of complex systems to achieve the highest rates of self organization to improve their quality. It can be applied to other complex systems from Physics, Chemistry, Biology, Ecology, Economics, Cities, network theory and others where complex systems are present.

Keywords: Organization, self-organization, complex system, complexification, quantitative measure, principle of least action, principle of stationary action, attractor, progressive development, acceleration, stochastic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1625
1134 On Pooling Different Levels of Data in Estimating Parameters of Continuous Meta-Analysis

Authors: N. R. N. Idris, S. Baharom

Abstract:

A meta-analysis may be performed using aggregate data (AD) or an individual patient data (IPD). In practice, studies may be available at both IPD and AD level. In this situation, both the IPD and AD should be utilised in order to maximize the available information. Statistical advantages of combining the studies from different level have not been fully explored. This study aims to quantify the statistical benefits of including available IPD when conducting a conventional summary-level meta-analysis. Simulated meta-analysis were used to assess the influence of the levels of data on overall meta-analysis estimates based on IPD-only, AD-only and the combination of IPD and AD (mixed data, MD), under different study scenario. The percentage relative bias (PRB), root mean-square-error (RMSE) and coverage probability were used to assess the efficiency of the overall estimates. The results demonstrate that available IPD should always be included in a conventional meta-analysis using summary level data as they would significantly increased the accuracy of the estimates.On the other hand, if more than 80% of the available data are at IPD level, including the AD does not provide significant differences in terms of accuracy of the estimates. Additionally, combining the IPD and AD has moderating effects on the biasness of the estimates of the treatment effects as the IPD tends to overestimate the treatment effects, while the AD has the tendency to produce underestimated effect estimates. These results may provide some guide in deciding if significant benefit is gained by pooling the two levels of data when conducting meta-analysis.

Keywords: Aggregate data, combined-level data, Individual patient data, meta analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1729
1133 Towards an Enhanced Stochastic Simulation Model for Risk Analysis in Highway Construction

Authors: Anshu Manik, William G. Buttlar, Kasthurirangan Gopalakrishnan

Abstract:

Over the years, there is a growing trend towards quality-based specifications in highway construction. In many Quality Control/Quality Assurance (QC/QA) specifications, the contractor is primarily responsible for quality control of the process, whereas the highway agency is responsible for testing the acceptance of the product. A cooperative investigation was conducted in Illinois over several years to develop a prototype End-Result Specification (ERS) for asphalt pavement construction. The final characteristics of the product are stipulated in the ERS and the contractor is given considerable freedom in achieving those characteristics. The risk for the contractor or agency depends on how the acceptance limits and processes are specified. Stochastic simulation models are very useful in estimating and analyzing payment risk in ERS systems and these form an integral part of the Illinois-s prototype ERS system. This paper describes the development of an innovative methodology to estimate the variability components in in-situ density, air voids and asphalt content data from ERS projects. The information gained from this would be crucial in simulating these ERS projects for estimation and analysis of payment risks associated with asphalt pavement construction. However, these methods require at least two parties to conduct tests on all the split samples obtained according to the sampling scheme prescribed in present ERS implemented in Illinois.

Keywords: Asphalt Pavement, Risk Analysis, StochasticSimulation, QC/QA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1500
1132 Experimental Study of Geotextile Effect on Improving Soil Bearing Capacity in Aggregate Surfaced Roads

Authors: Mahdi Taghipour Masoumi, Ali Abdi Kordani, Mahmoud Nazirizad

Abstract:

Geosynthetics utilization plays an important role in the construction of highways with no additive layers, such as asphalt concrete or cement concrete, or in a subgrade layer which affects the bearing capacity of unbounded layers. This laboratory experimental study was carried out to evaluate changes in the load bearing capacity of reinforced soil with these materials in highway roadbed with regard to geotextile properties. California Bearing Ratio (CBR) test samples were prepared with two types of soil: Clayey and sandy containing non-reinforced and reinforced soil. The samples comprised three types of geotextiles with different characteristics (150, 200, 300 g/m2) and depths (H= 5, 10, 20, 30, 50, 100 mm), and were grouped into two forms, one-layered and two-layered, based on the sample materials in order to perform defined tests. The results showed that the soil bearing characteristics increased when one layer of geotextile was used in clayey and sandy samples reinforced by geotextile. However, the bearing capacity of the soil, in the presence of a geotextile layer material with depth of more than 30 mm, had no remarkable effect. Furthermore, when the two-layered geotextile was applied in material samples, although it increased the soil resistance, it also showed that through the addition of a number or weights of geotextile into samples, the natural composition of the soil changed and the results are unreliable.

Keywords: Reinforced soil, geosynthetics, geotextile, transportation capacity, CBR experiments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2571
1131 Evaluation of Forage Yield and Competition Indices for Intercropped Barley and Legumes

Authors: Abdollah Javanmard, Fariborz Shekari, Hasan Dehghanian

Abstract:

Barley (Hordeum vulgare L.), vetch (Vicia villosa), and grass pea (Lathyrus sativus L.) monocultures as well as mixtures of barley with each of the above legumes, in three seeding ratios (i.e., barley: legume 75:25, 50:50 and 25:75, based on seed numbers) were used to investigated forage yield and competition indices. The results showed that intercropping reduced the dry matter yield of the three component plants, compared with their respective monocrops. The greatest value of total dry matter yield was obtained from barley25- grasspea75 (5.44 t ha-1) mixture, followed by grass pea sole crop (4.99 t ha-1). The total actual yield loss (AYL) values were positive and greater than 0 in all mixtures, indicating an advantage from intercropping over sole crops. Intercropped barley had a higher relative crowding coefficient (K=1.64) than intercropped legumes (K=1.20), indicating that barley was more competitive than legumes in mixtures. Furthermore, grass pea was more competitive than vetch in mixtures with barley. The highest land equivalent ratio (LER), system productivity index (SPI) and monetary advantage index (MAI) were obtained when barley was mixed at a rate of 25% with 75% seed rate of grass pea. It is concluded that intercropping of barley with grass pea has a good potential to improve the performance of forage with high land-use efficiency.

Keywords: Forage, grass pea, intercropping, land equivalent ratio (LER), monetary advantage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1845
1130 Novel D- glucose Based Glycomonomers Synthesis and Characterization

Authors: M.S. Mazăre, A. M. Pană, L. M. Ştefan, M. Silion, M. Bălan, G. Bandur, L. M. Rusnac

Abstract:

In the last decade, carbohydrates have attracted great attention as renewable resources for the chemical industry. Carbohydrates are abundantly found in nature in the form of monomers, oligomers and polymers, or as components of biopolymers and other naturally occurring substances. As natural products, they play important roles in conferring certain physical, chemical, and biological properties to their carrier molecules.The synthesis of this particular carbohydrate glycomonomer is part of our work to obtain biodegradable polymers. Our current paper describes the synthesis and characterization of a novel carbohydrate glycomonomer starting from D-glucose, in several synthesis steps, that involve the protection/deprotection of the D-glucose ring via acetylation, tritylation, then selective deprotection of the aromaticaliphatic protective group, in order to obtain 1,2,3,4-tetra-O-acetyl- 6-O-allyl-β-D-glucopyranose. The glycomonomer was then obtained by the allylation in drastic conditions of 1,2,3,4-tetra-O-acetyl-6-Oallyl- β-D-glucopyranose with allylic alcohol in the presence of stannic chloride, in methylene chloride, at room temperature. The proposed structure of the glycomonomer, 2,3,4-tri-O-acetyl-1,6-di- O-allyl-β-D-glucopyranose, was confirmed by FTIR, NMR and HPLC-MS spectrometry. This glycomonomer will be further submitted to copolymerization with certain acrylic or methacrylic monomers in order to obtain competitive plastic materials for applications in the biomedical field.

Keywords: allylation, D-glucose, glycomonomer, trityl chloride

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2027
1129 Normalizing Scientometric Indicators of Individual Publications Using Local Cluster Detection Methods on Citation Networks

Authors: Levente Varga, Dávid Deritei, Mária Ercsey-Ravasz, Răzvan Florian, Zsolt I. Lázár, István Papp, Ferenc Járai-Szabó

Abstract:

One of the major shortcomings of widely used scientometric indicators is that different disciplines cannot be compared with each other. The issue of cross-disciplinary normalization has been long discussed, but even the classification of publications into scientific domains poses problems. Structural properties of citation networks offer new possibilities, however, the large size and constant growth of these networks asks for precaution. Here we present a new tool that in order to perform cross-field normalization of scientometric indicators of individual publications relays on the structural properties of citation networks. Due to the large size of the networks, a systematic procedure for identifying scientific domains based on a local community detection algorithm is proposed. The algorithm is tested with different benchmark and real-world networks. Then, by the use of this algorithm, the mechanism of the scientometric indicator normalization process is shown for a few indicators like the citation number, P-index and a local version of the PageRank indicator. The fat-tail trend of the article indicator distribution enables us to successfully perform the indicator normalization process.

Keywords: Citation networks, scientometric indicator, cross-field normalization, local cluster detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 716
1128 Resistance and Sub-Resistances of RC Beams Subjected to Multiple Failure Modes

Authors: F. Sangiorgio, J. Silfwerbrand, G. Mancini

Abstract:

Geometric and mechanical properties all influence the resistance of RC structures and may, in certain combination of property values, increase the risk of a brittle failure of the whole system. This paper presents a statistical and probabilistic investigation on the resistance of RC beams designed according to Eurocodes 2 and 8, and subjected to multiple failure modes, under both the natural variation of material properties and the uncertainty associated with cross-section and transverse reinforcement geometry. A full probabilistic model based on JCSS Probabilistic Model Code is derived. Different beams are studied through material nonlinear analysis via Monte Carlo simulations. The resistance model is consistent with Eurocode 2. Both a multivariate statistical evaluation and the data clustering analysis of outcomes are then performed. Results show that the ultimate load behaviour of RC beams subjected to flexural and shear failure modes seems to be mainly influenced by the combination of the mechanical properties of both longitudinal reinforcement and stirrups, and the tensile strength of concrete, of which the latter appears to affect the overall response of the system in a nonlinear way. The model uncertainty of the resistance model used in the analysis plays undoubtedly an important role in interpreting results.

Keywords: Modelling, Monte Carlo Simulations, Probabilistic Models, Data Clustering, Reinforced Concrete Members, Structural Design.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2100
1127 Choosing Local Organic Food: Consumer Motivations and Ethical Spaces

Authors: Artur Saraiva, Moritz von Schwedler, Emília Fernandes

Abstract:

In recent years, the organic sector has increased significantly. However, with the ‘conventionalization’ of these products, it has been questioned whether these products have been losing their original vision. Accordingly, this research based on 31 phenomenological interviews with committed organic consumers in urban and rural areas of Portugal, aims to analyse how ethical motivations and ecological awareness are related to organic food consumption. The content thematic analysis highlights aspects related to society and environmental concerns. On an individual level, the importance of internal coherence, peace of mind and balance that these consumers find in the consumption of local organic products was stressed. For these consumers, local organic products consumption made for significant changes in their lives, aiding in the establishment of a green identity, and involves a certain philosophy of life. This vision of an organic lifestyle is grounded in a political and ecological perspective, beyond the usual organic definition, as a ‘post-organic era’. The paper contributes to better understand how an ideological environmental discourse allows highlighting the relationship between consumers’ environmental concerns and the politics of food, resulting in a possible transition to new sustainable consumption practices.

Keywords: Organic consumption, localism, content thematic analysis, pro-environmental discourse, political consumption, Portugal.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1418
1126 Temperature Related Alterations to Mineral Levels and Crystalline Structure in Porcine Long Bone: Intense Heat vs. Open Flame

Authors: Caighley Logan, Suzzanne McColl

Abstract:

The outcome of fire related fatalities, along with other research, has found fires can have a detrimental effect to the mineral and crystalline structures within bone. This study focused on the mineral and crystalline structures within porcine bone samples to analyse the changes caused, with the intent of effectively ‘reverse engineering’ the data collected from burned bone samples to discover what may have happened. Using Fourier Transform Infrared (FTIR), and X-Ray Fluorescence (XRF), the data were collected from a controlled source of intense heat (muffle furnace) and an open fire, based in a living room setting in a standard size shipping container (2.5 m x 2.4 m) of a similar temperature with a known ignition source, a gasoline lighter. This approach is to analyse the changes to the samples and how the changes differ depending on the heat source. Results have found significant differences in the levels of remaining minerals for each type of heat/burning (p =< 0.001), particularly Phosphorus and Calcium, this also includes notable additions of absorbed elements and minerals from the surrounding materials, i.e., Cerium (Ce), Bromine (Br) and Neodymium (Ne). The analysis techniques included provide validated results in conjunction with previous studies.

Keywords: Forensic anthropology, thermal alterations, porcine bone, FTIR, XRF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 169
1125 Comparison between Separable and Irreducible Goppa Code in McEliece Cryptosystem

Authors: Thuraya M. Qaradaghi, Newroz N. Abdulrazaq

Abstract:

The McEliece cryptosystem is an asymmetric type of cryptography based on error correction code. The classical McEliece used irreducible binary Goppa code which considered unbreakable until now especially with parameter [1024, 524, and 101], but it is suffering from large public key matrix which leads to be difficult to be used practically. In this work Irreducible and Separable Goppa codes have been introduced. The Irreducible and Separable Goppa codes used are with flexible parameters and dynamic error vectors. A Comparison between Separable and Irreducible Goppa code in McEliece Cryptosystem has been done. For encryption stage, to get better result for comparison, two types of testing have been chosen; in the first one the random message is constant while the parameters of Goppa code have been changed. But for the second test, the parameters of Goppa code are constant (m=8 and t=10) while the random message have been changed. The results show that the time needed to calculate parity check matrix in separable are higher than the one for irreducible McEliece cryptosystem, which is considered expected results due to calculate extra parity check matrix in decryption process for g2(z) in separable type, and the time needed to execute error locator in decryption stage in separable type is better than the time needed to calculate it in irreducible type. The proposed implementation has been done by Visual studio C#.

Keywords: McEliece cryptosystem, Goppa code, separable, irreducible.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2200