Search results for: The Kernel density estimate.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1939

Search results for: The Kernel density estimate.

79 Sea Level Characteristics Referenced to Specific Geodetic Datum in Alexandria, Egypt

Authors: Ahmed M. Khedr, Saad M. Abdelrahman, Kareem M. Tonbol

Abstract:

Two geo-referenced sea level datasets (September 2008 – November 2010) and (April 2012 – January 2014) were recorded at Alexandria Western Harbour (AWH). Accurate re-definition of tidal datum, referred to the latest International Terrestrial Reference Frame (ITRF-2014), was discussed and updated to improve our understanding of the old predefined tidal datum at Alexandria. Tidal and non-tidal components of sea level were separated with the use of Delft-3D hydrodynamic model-tide suit (Delft-3D, 2015). Tidal characteristics at AWH were investigated and harmonic analysis showed the most significant 34 constituents with their amplitudes and phases. Tide was identified as semi-diurnal pattern as indicated by a “Form Factor” of 0.24 and 0.25, respectively. Principle tidal datums related to major tidal phenomena were recalculated referred to a meaningful geodetic height datum. The portion of residual energy (surge) out of the total sea level energy was computed for each dataset and found 77% and 72%, respectively. Power spectral density (PSD) showed accurate resolvability in high band (1–6) cycle/days for the nominated independent constituents, except some neighbouring constituents, which are too close in frequency. Wind and atmospheric pressure data, during the recorded sea level time, were analysed and cross-correlated with the surge signals. Moderate association between surge and wind and atmospheric pressure data were obtained. In addition, long-term sea level rise trend at AWH was computed and showed good agreement with earlier estimated rates.

Keywords: Alexandria, Delft-3D, Egypt, geodetic reference, harmonic analysis, sea level.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1360
78 Maximization of Lifetime for Wireless Sensor Networks Based on Energy Efficient Clustering Algorithm

Authors: Frodouard Minani

Abstract:

Since last decade, wireless sensor networks (WSNs) have been used in many areas like health care, agriculture, defense, military, disaster hit areas and so on. Wireless Sensor Networks consist of a Base Station (BS) and more number of wireless sensors in order to monitor temperature, pressure, motion in different environment conditions. The key parameter that plays a major role in designing a protocol for Wireless Sensor Networks is energy efficiency which is a scarcest resource of sensor nodes and it determines the lifetime of sensor nodes. Maximizing sensor node’s lifetime is an important issue in the design of applications and protocols for Wireless Sensor Networks. Clustering sensor nodes mechanism is an effective topology control approach for helping to achieve the goal of this research. In this paper, the researcher presents an energy efficiency protocol to prolong the network lifetime based on Energy efficient clustering algorithm. The Low Energy Adaptive Clustering Hierarchy (LEACH) is a routing protocol for clusters which is used to lower the energy consumption and also to improve the lifetime of the Wireless Sensor Networks. Maximizing energy dissipation and network lifetime are important matters in the design of applications and protocols for wireless sensor networks. Proposed system is to maximize the lifetime of the Wireless Sensor Networks by choosing the farthest cluster head (CH) instead of the closest CH and forming the cluster by considering the following parameter metrics such as Node’s density, residual-energy and distance between clusters (inter-cluster distance). In this paper, comparisons between the proposed protocol and comparative protocols in different scenarios have been done and the simulation results showed that the proposed protocol performs well over other comparative protocols in various scenarios.

Keywords: Base station, clustering algorithm, energy efficient, wireless sensor networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 844
77 Aircraft Selection Using Multiple Criteria Decision Making Analysis Method with Different Data Normalization Techniques

Authors: C. Ardil

Abstract:

This paper presents an original application of multiple criteria decision making analysis theory to the evaluation of aircraft selection problem. The selection of an optimal, efficient and reliable fleet, network and operations planning policy is one of the most important factors in aircraft selection problem. Given that decision making in aircraft selection involves the consideration of a number of opposite criteria and possible solutions, such a selection can be considered as a multiple criteria decision making analysis problem. This study presents a new integrated approach to decision making by considering the multiple criteria utility theory and the maximal regret minimization theory methods as well as aircraft technical, economical, and environmental aspects. Multiple criteria decision making analysis method uses different normalization techniques to allow criteria to be aggregated with qualitative and quantitative data of the decision problem. Therefore, selecting a suitable normalization technique for the model is also a challenge to provide data aggregation for the aircraft selection problem. To compare the impact of different normalization techniques on the decision problem, the vector, linear (sum), linear (max), and linear (max-min) data normalization techniques were identified to evaluate aircraft selection problem. As a logical implication of the proposed approach, it enhances the decision making process through enabling the decision maker to: (i) use higher level knowledge regarding the selection of criteria weights and the proposed technique, (ii) estimate the ranking of an alternative, under different data normalization techniques and integrated criteria weights after a posteriori analysis of the final rankings of alternatives. A set of commercial passenger aircraft were considered in order to illustrate the proposed approach. The obtained results of the proposed approach were compared using Spearman's rho tests. An analysis of the final rank stability with respect to the changes in criteria weights was also performed so as to assess the sensitivity of the alternative rankings obtained by the application of different data normalization techniques and the proposed approach.

Keywords: Normalization Techniques, Aircraft Selection, Multiple Criteria Decision Making, Multiple Criteria Decision Making Analysis, MCDMA

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 586
76 A Settlement Strategy for Health Facilities in Emerging Countries: A Case Study in Brazil

Authors: Domenico Chizzoniti, Monica Moscatelli, Letizia Cattani, Piero Favino, Luca Preis

Abstract:

A settlement strategy is to anticipate and respond the needs of existing and future communities through the provision of primary health care facilities in marginalized areas. Access to a health care network is important to improving healthcare coverage, often lacking, in developing countries. The study explores that a good sanitary system strategy of rural contexts brings advantages to an existing settlement: improving transport, communication, water and social facilities. The objective of this paper is to define a possible methodology to implement primary health care facilities in disadvantaged areas of emerging countries. In this research, we analyze the case study of Lauro de Freitas, a municipality in the Brazilian state of Bahia, part of the Metropolitan Region of Salvador, with an area of 57,662 km² and 194.641 inhabitants. The health localization system in Lauro de Freitas is an integrated process that involves not only geographical aspects, but also a set of factors: population density, epidemiological data, allocation of services, road networks, and more. Data were collected also using semi-structured interviews and questionnaires to the local population. Synthesized data suggest that moving away from the coast where there is the greatest concentration of population and services, a network of primary health care facilities is able to improve the living conditions of small-dispersed communities. Based on the health service needs of populations, we have developed a methodological approach that is particularly useful in rural and remote contexts in emerging countries.

Keywords: Primary health care, developing countries, policy health planning, settlement strategy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 981
75 Oil-Water Two-Phase Flow Characteristics in Horizontal Pipeline – A Comprehensive CFD Study

Authors: Anand B. Desamala, Ashok Kumar Dasamahapatra, Tapas K. Mandal

Abstract:

In the present work, detailed analysis on flow characteristics of a pair of immiscible liquids through horizontal pipeline is simulated by using ANSYS FLUENT 6.2. Moderately viscous oil and water (viscosity ratio = 107, density ratio = 0.89 and interfacial tension = 0.024 N/m) have been taken as system fluids for the study. Volume of Fluid (VOF) method has been employed by assuming unsteady flow, immiscible liquid pair, constant liquid properties, and co-axial flow. Meshing has been done using GAMBIT. Quadrilateral mesh type has been chosen to account for the surface tension effect more accurately. From the grid independent study, we have selected 47037 number of mesh elements for the entire geometry. Simulation successfully predicts slug, stratified wavy, stratified mixed and annular flow, except dispersion of oil in water, and dispersion of water in oil. Simulation results are validated with horizontal literature data and good conformity is observed. Subsequently, we have simulated the hydrodynamics (viz., velocity profile, area average pressure across a cross section and volume fraction profile along the radius) of stratified wavy and annular flow at different phase velocities. The simulation results show that in the annular flow, total pressure of the mixture decreases with increase in oil velocity due to the fact that pipe cross section is completely wetted with water. Simulated oil volume fraction shows maximum at the centre in core annular flow, whereas, in stratified flow, maximum value appears at upper side of the pipeline. These results are in accord with the actual flow configuration. Our findings could be useful in designing pipeline for transportation of crude oil.

Keywords: CFD, Horizontal pipeline, Oil-water flow, VOF technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5711
74 Analytical Slope Stability Analysis Based on the Statistical Characterization of Soil Shear Strength

Authors: Bernardo C. P. Albuquerque, Darym J. F. Campos

Abstract:

Increasing our ability to solve complex engineering problems is directly related to the processing capacity of computers. By means of such equipments, one is able to fast and accurately run numerical algorithms. Besides the increasing interest in numerical simulations, probabilistic approaches are also of great importance. This way, statistical tools have shown their relevance to the modelling of practical engineering problems. In general, statistical approaches to such problems consider that the random variables involved follow a normal distribution. This assumption tends to provide incorrect results when skew data is present since normal distributions are symmetric about their means. Thus, in order to visualize and quantify this aspect, 9 statistical distributions (symmetric and skew) have been considered to model a hypothetical slope stability problem. The data modeled is the friction angle of a superficial soil in Brasilia, Brazil. Despite the apparent universality, the normal distribution did not qualify as the best fit. In the present effort, data obtained in consolidated-drained triaxial tests and saturated direct shear tests have been modeled and used to analytically derive the probability density function (PDF) of the safety factor of a hypothetical slope based on Mohr-Coulomb rupture criterion. Therefore, based on this analysis, it is possible to explicitly derive the failure probability considering the friction angle as a random variable. Furthermore, it is possible to compare the stability analysis when the friction angle is modelled as a Dagum distribution (distribution that presented the best fit to the histogram) and as a Normal distribution. This comparison leads to relevant differences when analyzed in light of the risk management.

Keywords: Statistical slope stability analysis, Skew distributions, Probability of failure, Functions of random variables.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1545
73 Suicide Wrongful Death: Standard of Care Problems Involving the Inaccurate Discernment of Lethal Risk When Focusing on the Elicitation of Suicide Ideation

Authors: Bill D. Geis, Frederick Newman

Abstract:

Suicide and wrongful death forensic cases are the fastest rising tort in mental health law. Most suicide-related personal injury claims fall into the legal category of “wrongful death.” Though mental health experts may be called on to address a range of forensic questions in wrongful death cases, the central consultation that most experts provide is about the negligence element—specifically, the issue of whether the clinician met the clinical standard of care in assessing, treating, and managing the deceased person’s mental health care. Standards of care, varying from US state to state, are broad and address what a reasonable clinician might do in a similar circumstance. This fact leaves the issue of the suicide standard of care, in each case, up to forensic experts to put forth a reasoned estimate of what the standard of care should have been in the specific case under litigation. Because the general state guidelines for standard of care are broad, forensic experts are readily retained to provide scientific and clinical opinions about whether or not a clinician met the standard of care in their suicide assessment, treatment, and management of the case. In the past and in much of current practice, the assessment of suicide has centered on the elicitation of verbalized suicide ideation. But suicide ideation, in the matter of suicide risk determination, may be a necessary but insufficient target of lethal suicide risk assessment. Assessment of near-term suicide risk—assessment that goes beyond verbalized suicide ideation and relates to acute crisis variables—is likely needed. Specifically, such other or additional suicide risk variable assessment may be required in the context of lethal suicide risk situations, as opposed to the discernment of general, nonlethal suicide behavior as a standard of practice (whether a patient is having suicidal thoughts or exhibiting an ambivalent suicide attempt potential). In the current study, verbalized suicide ideation information was unhelpful in the assessment of lethal risk. The Lethal Suicide Risk Assessment, Acute Model, and other dynamic, near-term risk models (such as the Acute Suicide Affective Disorder Model and the Suicide Crisis Syndrome Model)—going beyond elicited suicide ideation—need to be incorporated into current clinical suicide assessment training and become the legal standard of care for expected clinical behavior. Without this expanded clinical assessment perspective, the standard of care for suicide assessment is out of sync with current knowledge—an emerging dilemma for the forensic evaluation of suicide wrongful death cases.

Keywords: Forensic evaluation, standard of care, suicide, suicide assessment, wrongful death.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 256
72 Reinforcing Effects of Natural Micro-Particles on the Dynamic Impact Behaviour of Hybrid Bio-Composites Made of Short Kevlar Fibers Reinforced Thermoplastic Composite Armor

Authors: Edison E. Haro, Akindele G. Odeshi, Jerzy A. Szpunar

Abstract:

Hybrid bio-composites are developed for use in protective armor through positive hybridization offered by reinforcement of high-density polyethylene (HDPE) with Kevlar short fibers and palm wood micro-fillers. The manufacturing process involved a combination of extrusion and compression molding techniques. The mechanical behavior of Kevlar fiber reinforced HDPE with and without palm wood filler additions are compared. The effect of the weight fraction of the added palm wood micro-fillers is also determined. The Young modulus was found to increase as the weight fraction of organic micro-particles increased. However, the flexural strength decreased with increasing weight fraction of added micro-fillers. The interfacial interactions between the components were investigated using scanning electron microscopy. The influence of the size, random alignment and distribution of the natural micro-particles was evaluated. Ballistic impact and dynamic shock loading tests were performed to determine the optimum proportion of Kevlar short fibers and organic micro-fillers needed to improve impact strength of the HDPE. These results indicate a positive hybridization by deposition of organic micro-fillers on the surface of short Kevlar fibers used in reinforcing the thermoplastic matrix leading to enhancement of the mechanical strength and dynamic impact behavior of these materials. Therefore, these hybrid bio-composites can be promising materials for different applications against high velocity impacts.

Keywords: Hybrid bio-composites, organic nano-fillers, dynamic shocking loading, ballistic impacts, energy absorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 756
71 Biological Methods to Control Parasitic Weed Phelipanche ramosa L. Pomel in the Field Tomato Crop

Authors: F. Lops, G. Disciglio, A. Carlucci, G. Gatta, L. Frabboni, A. Tarantino, E. Tarantino

Abstract:

Phelipanche ramosa L. Pomel is a root holoparasitic weed plant of many cultivations, particularly of tomato (Lycopersicum esculentum L.) crop. In Italy, Phelipanche problem is increasing, both in density and in acreage. The biological control of this parasitic weed involves the use of living organisms as numerous fungi and bacteria that can infect the parasitic weed, while it may improve the crop growth. This paper deals with the biocontrol with microorganism, including Arbuscular mycorrhizal (AM) fungi and fungal pathogens as Fusarium oxisporum spp. Colonization of crop roots by AM fungi can provide protection of crops against parasitic weeds because of a reduction in their seed germination and attachment, while F. oxisporum, isolated from diseased broomrape tubercles, proved to be highly virulent on P. ramosa. The experimental trial was carried out in open field at Foggia province (Apulia Region, Southern Italy), during the spring-summer season 2016, in order to evaluate the effect of four biological treatments: AM fungi and Fusarium oxisporum applied in the soil alone or combined together, and Rizosum Max® product, compared with the untreated control, to reduce the P. ramosa infestation in processing tomato crop. The principal results to be drawn from this study under field condition, in contrast of those reported previously under laboratory and greenhouse conditions, show that both AM fungi and F. oxisporum do not provide the reduction of the number of emerged shoots of P. ramosa. This can arise probably from the low efficacy seedling of the agent pathogens for the control of this parasite in the field. On the contrary, the Rizosum Max® product, containing AM fungi and some rizophere bacteria combined with several minerals and organic substances, appears to be most effective for the reduction of P. ramosa infestation.

Keywords: Arbuscular mycorrhizal fungi, biocontrol methods, Phelipanche ramosa, F. oxisporum spp.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1066
70 Ports and Airports: Gateways to Vector-Borne Diseases in Portugal Mainland

Authors: Maria C. Proença, Maria T. Rebelo, Maria J. Alves, Sofia Cunha

Abstract:

Vector-borne diseases are transmitted to humans by mosquitos, sandflies, bugs, ticks, and other vectors. Some are re-transmitted between vectors, if the infected human has a new contact when his levels of infection are high. The vector is infected for lifetime and can transmit infectious diseases not only between humans but also from animals to humans. Some vector borne diseases are very disabling and globally account for more than one million deaths worldwide. The mosquitoes from the complex Culex pipiens sl. are the most abundant in Portugal, and we dispose in this moment of a data set from the surveillance program that has been carried on since 2006 across the country. All mosquitos’ species are included, but the large coverage of Culex pipiens sl. and its importance for public health make this vector an interesting candidate to assess risk of disease amplification. This work focus on ports and airports identified as key areas of high density of vectors. Mosquitoes being ectothermic organisms, the main factor for vector survival and pathogen development is temperature. Minima and maxima local air temperatures for each area of interest are averaged by month from data gathered on a daily basis at the national network of meteorological stations, and interpolated in a geographic information system (GIS). The range of temperatures ideal for several pathogens are known and this work shows how to use it with the meteorological data in each port and airport facility, to focus an efficient implementation of countermeasures and reduce simultaneously risk transmission and mitigation costs. The results show an increased alert with decreasing latitude, which corresponds to higher minimum and maximum temperatures and a lower amplitude range of the daily temperature.

Keywords: Human health, risk assessment, risk management, vector-borne diseases.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2053
69 Laboratory Investigations on the Utilization of Recycled Construction Aggregates in Asphalt Mixtures

Authors: Farzaneh Tahmoorian, Bijan Samali, John Yeaman

Abstract:

Road networks are increasingly expanding all over the world. The construction and maintenance of the road pavements require large amounts of aggregates. Considerable usage of various natural aggregates for constructing roads as well as the increasing rate at which solid waste is generated have attracted the attention of many researchers in the pavement industry to investigate the feasibility of the application of some of the waste materials as alternative materials in pavement construction. Among various waste materials, construction and demolition wastes, including Recycled Construction Aggregate (RCA) constitute a major part of the municipal solid wastes in Australia. Creating opportunities for the application of RCA in civil and geotechnical engineering applications is an efficient way to increase the market value of RCA. However, in spite of such promising potentials, insufficient and inconclusive data and information on the engineering properties of RCA had limited the reliability and design specifications of RCA to date. In light of this, this paper, as a first step of a comprehensive research, aims to investigate the feasibility of the application of RCA obtained from construction and demolition wastes for the replacement of part of coarse aggregates in asphalt mixture. As the suitability of aggregates for using in asphalt mixtures is determined based on the aggregate characteristics, including physical and mechanical properties of the aggregates, an experimental program is set up to evaluate the physical and mechanical properties of RCA. This laboratory investigation included the measurement of compressive strength and workability of RCA, particle shape, water absorption, flakiness index, crushing value, deleterious materials and weak particles, wet/dry strength variation, and particle density. In addition, the comparison of RCA properties with virgin aggregates has been included as part of this investigation and this paper presents the results of these investigations on RCA, basalt, and the mix of RCA/basalt.

Keywords: Asphalt, basalt, pavement, recycled aggregate.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 970
68 TNFRSF11B Gene Polymorphisms A163G and G11811C in Prediction of Osteoporosis Risk

Authors: Boroňová I., Bernasovská J., Kľoc J., Tomková Z., Petrejčíková E., Gabriková D., Mačeková S.

Abstract:

Osteoporosis is a complex health disease characterized by low bone mineral density, which is determined by an interaction of genetics with metabolic and environmental factors. Current research in genetics of osteoporosis is focused on identification of responsible genes and polymorphisms. TNFRSF11B gene plays a key role in bone remodeling. The aim of this study was to investigate the genotype and allele distribution of A163G (rs3102735) osteoprotegerin gene promoter and G1181C (rs2073618) osteoprotegerin first exon polymorphisms in the group of 180 unrelated postmenopausal women with diagnosed osteoporosis and 180 normal controls. Genomic DNA was isolated from peripheral blood leukocytes using standard methodology. Genotyping for presence of different polymorphisms was performed using the Custom Taqman®SNP Genotyping assays. Hardy-Weinberg equilibrium was tested for each SNP in the groups of participants using the chi-square (χ2) test. The distribution of investigated genotypes in the group of patients with osteoporosis were as follows: AA (66.7%), AG (32.2%), GG (1.1%) for A163G polymorphism; GG (19.4%), CG (44.4%), CC (36.1%) for G1181C polymorphism. The distribution of genotypes in normal controls were follows: AA (71.1%), AG (26.1%), GG (2.8%) for A163G polymorphism; GG (22.2%), CG (48.9%), CC (28.9%) for G1181C polymorphism. In A163G polymorphism the variant G allele was more common among patients with osteoporosis: 17.2% versus 15.8% in normal controls. Also, in G1181C polymorphism the phenomenon of more frequent occurrence of C allele in the group of patients with osteoporosis was observed (58.3% versus 53.3%). Genotype and allele distributions showed no significant differences (A163G: χ2=0.270, p=0.605; χ2=0.250, p=0.616; G1181C: χ2= 1.730, p=0.188; χ2=1.820, p=0.177). Our results represents an initial study, further studies of more numerous file and associations studies will be carried out. Knowing the distribution of genotypes is important for assessing the impact of these polymorphisms on various parameters associated with osteoporosis. Screening for identification of “at-risk” women likely to develop osteoporosis and initiating subsequent early intervention appears to be most effective strategy to substantially reduce the risks of osteoporosis.

Keywords: Osteoporosis, Real-time PCR method, SNP polymorphisms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2244
67 Some Morphological Characteristics of Perennial Ryegrass Genotypes and Correlations among Their Characteristics

Authors: A. Özköse, A. Tamkoç

Abstract:

The present study involved analysis of certain characteristics of the perennial ryegrass (Lolium perenne L.) genotypes collected from the natural flora of Ankara, and explores a correlation among them. In order to evaluate the plants for breeding purpose as per Turkey's environmental conditions, the perennial ryegrass plants were collected from natural pasture of Ankara in 2004 and were utilized for the study. Seeds of the collected plants were sown in pots and seedlings were prepared in a greenhouse. In 2005, the seedlings were transplanted at 50 × 50 cm2 intervals in Randomized Complete Blocks Design in an experimental field. In 2007 and 2008, data were recorded from the observations and measurements of 568 perennial ryegrasses. The plant characteristics, which were investigated, included re-growth time in spring, color, density, growth habit, tendency to form inflorescence, time of inflorescence, plant height, length of upper internode, spike length, leaf length, leaf width, leaf area, leaf shape, number of spikelets per spike, seed yield per spike and 1000 grain weight and the correlation analyses were made using this data. Correlation coefficients were estimated between all paired combinations of the studied traits. The yield components exhibited varying trends of association among themselves. Seed yield per spike showed significant and positive association with the number of spikelets per spike, 1000 grain weight, plant height, length of upper internode, spike length, leaf length, leaf width, leaf area and color, but significant and negative association with the growth habit and re-growth time in spring.

Keywords: Correlation, morphological traits, Lolium perenne.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942
66 Frequency Response of Complex Systems with Localized Nonlinearities

Authors: E. Menga, S. Hernandez

Abstract:

Finite Element Models (FEMs) are widely used in order to study and predict the dynamic properties of structures and usually, the prediction can be obtained with much more accuracy in the case of a single component than in the case of assemblies. Especially for structural dynamics studies, in the low and middle frequency range, most complex FEMs can be seen as assemblies made by linear components joined together at interfaces. From a modelling and computational point of view, these types of joints can be seen as localized sources of stiffness and damping and can be modelled as lumped spring/damper elements, most of time, characterized by nonlinear constitutive laws. On the other side, most of FE programs are able to run nonlinear analysis in time-domain. They treat the whole structure as nonlinear, even if there is one nonlinear degree of freedom (DOF) out of thousands of linear ones, making the analysis unnecessarily expensive from a computational point of view. In this work, a methodology in order to obtain the nonlinear frequency response of structures, whose nonlinearities can be considered as localized sources, is presented. The work extends the well-known Structural Dynamic Modification Method (SDMM) to a nonlinear set of modifications, and allows getting the Nonlinear Frequency Response Functions (NLFRFs), through an ‘updating’ process of the Linear Frequency Response Functions (LFRFs). A brief summary of the analytical concepts is given, starting from the linear formulation and understanding what the implications of the nonlinear one, are. The response of the system is formulated in both: time and frequency domain. First the Modal Database is extracted and the linear response is calculated. Secondly the nonlinear response is obtained thru the NL SDMM, by updating the underlying linear behavior of the system. The methodology, implemented in MATLAB, has been successfully applied to estimate the nonlinear frequency response of two systems. The first one is a two DOFs spring-mass-damper system, and the second example takes into account a full aircraft FE Model. In spite of the different levels of complexity, both examples show the reliability and effectiveness of the method. The results highlight a feasible and robust procedure, which allows a quick estimation of the effect of localized nonlinearities on the dynamic behavior. The method is particularly powerful when most of the FE Model can be considered as acting linearly and the nonlinear behavior is restricted to few degrees of freedom. The procedure is very attractive from a computational point of view because the FEM needs to be run just once, which allows faster nonlinear sensitivity analysis and easier implementation of optimization procedures for the calibration of nonlinear models.

Keywords: Frequency response, nonlinear dynamics, structural dynamic modification, softening effect, rubber.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1312
65 Complex Network Approach to International Trade of Fossil Fuel

Authors: Semanur Soyyiğit Kaya, Ercan Eren

Abstract:

Energy has a prominent role for development of nations. Countries which have energy resources also have strategic power in the international trade of energy since it is essential for all stages of production in the economy. Thus, it is important for countries to analyze the weaknesses and strength of the system. On the other side, international trade is one of the fields that are analyzed as a complex network via network analysis. Complex network is one of the tools to analyze complex systems with heterogeneous agents and interaction between them. A complex network consists of nodes and the interactions between these nodes. Total properties which emerge as a result of these interactions are distinct from the sum of small parts (more or less) in complex systems. Thus, standard approaches to international trade are superficial to analyze these systems. Network analysis provides a new approach to analyze international trade as a network. In this network, countries constitute nodes and trade relations (export or import) constitute edges. It becomes possible to analyze international trade network in terms of high degree indicators which are specific to complex networks such as connectivity, clustering, assortativity/disassortativity, centrality, etc. In this analysis, international trade of crude oil and coal which are types of fossil fuel has been analyzed from 2005 to 2014 via network analysis. First, it has been analyzed in terms of some topological parameters such as density, transitivity, clustering etc. Afterwards, fitness to Pareto distribution has been analyzed via Kolmogorov-Smirnov test. Finally, weighted HITS algorithm has been applied to the data as a centrality measure to determine the real prominence of countries in these trade networks. Weighted HITS algorithm is a strong tool to analyze the network by ranking countries with regards to prominence of their trade partners. We have calculated both an export centrality and an import centrality by applying w-HITS algorithm to the data. As a result, impacts of the trading countries have been presented in terms of high-degree indicators.

Keywords: Complex network approach, fossil fuel, international trade, network theory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2386
64 Cluster Based Energy Efficient and Fault Tolerant n-Coverage in Wireless Sensor Network

Authors: D. Satish Kumar, N. Nagarajan

Abstract:

Coverage conservation and extend the network lifetime are the primary issues in wireless sensor networks. Due to the large variety of applications, coverage is focus to a wide range of interpretations. The applications necessitate that each point in the area is observed by only one sensor while other applications may require that each point is enclosed by at least sensors (n>1) to achieve fault tolerance. Sensor scheduling activities in existing Transparent and non- Transparent relay modes (T-NT) Mobile Multi-Hop relay networks fails to guarantee area coverage with minimal energy consumption and fault tolerance. To overcome these issues, Cluster based Energy Competent n- coverage scheme called (CEC n-coverage scheme) to ensure the full coverage of a monitored area while saving energy. CEC n-coverage scheme uses a novel sensor scheduling scheme based on the n-density and the remaining energy of each sensor to determine the state of all the deployed sensors to be either active or sleep as well as the state durations. Hence, it is attractive to trigger a minimum number of sensors that are able to ensure coverage area and turn off some redundant sensors to save energy and therefore extend network lifetime. In addition, decisive a smallest amount of active sensors based on the degree coverage required and its level. A variety of numerical parameters are computed using ns2 simulator on existing (T-NT) Mobile Multi-Hop relay networks and CEC n-coverage scheme. Simulation results showed that CEC n-coverage scheme in wireless sensor network provides better performance in terms of the energy efficiency, 6.61% reduced fault tolerant in terms of seconds and the percentage of active sensors to guarantee the area coverage compared to exiting algorithm.

Keywords: Wireless Sensor network, Mobile Multi-Hop relay networks, n-coverage, Cluster based Energy Competent, Transparent and non- Transparent relay modes, Fault Tolerant, sensor scheduling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2147
63 Designing a Pre-Assessment Tool to Support the Achievement of Green Building Certifications

Authors: Jisun Mo, Paola Boarin

Abstract:

The impact of common buildings on climate and environment has prompted people to get involved in the green building standards aimed at implementing rating tools or certifications. Thus, green building rating systems were introduced to the construction industry, and the demand for certified green buildings has increased gradually and succeeded considerably in enhancing people’s environmental awareness. However, the existing certification process has been unsatisfactory in attracting stakeholders and/or professionals who are actively engaged in adopting a rating system. It is because they have faced recurring barriers regarding limited information in understanding the rating process, time-consuming procedures and higher costs, which have a direct influence on pursuing green building rating systems. To promote the achievement of green building certifications within the building industry more successfully, this paper aims at designing a Pre-Assessment Tool (PAT) framework that can help stakeholders and/or professionals engaged in the construction industry to clarify their basic knowledge, timeframe and extra costs needed to activate a green building certification. First, taking the first steps towards the rating tool seems to be complicated because of upfront commitment to understanding the overall rating procedure is required. This conceptual PAT framework can increase basic knowledge of the rating tool and the certification process, mainly in terms of all resources or information of each credit requirements. Second, the assessment process of rating tools is generally known as a “lengthy and time-consuming system”, contributing to unenthusiastic reactions concerning green building projects. The proposed framework can predict the timeframe needed to identify how long it will take for a green project to process each credit requirement and the documentation required from the beginning of the certification process to final approval. Finally, most people often have the initial perception that pursuing green building certification costs more than constructing a non-green building, which makes it more difficult to execute rating tools. To overcome this issue, this PAT will help users to estimate the extra expenses such as certification fees and third-party contributions based on the track of the amount of time it takes to implement the rating tool throughout all the related stages. Also, it can prevent unexpected or hidden costs occurring in the process of assessment. Therefore, this proposed PAT framework can be recommended as an effective method to support the decision-making of inexperienced users and play an important role in promoting green building certification.

Keywords: Barriers, certification process, green building rating systems, pre-assessment tool.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 827
62 Effect of Heat Treatment on Mechanical Properties and Wear Behavior of Al7075 Alloy Reinforced with Beryl and Graphene Hybrid Metal Matrix Composites

Authors: Shanawaz Patil, Mohamed Haneef, K. S. Narayanaswamy

Abstract:

In the recent years, aluminum metal matrix composites were most widely used, which are finding wide applications in various field such as automobile, aerospace defense etc., due to their outstanding mechanical properties like low density, light weight, exceptional high levels of strength, stiffness, wear resistance, high temperature resistance, low coefficient of thermal expansion and good formability. In the present work, an effort is made to study the effect of heat treatment on mechanical properties of aluminum 7075 alloy reinforced with constant weight percentage of naturally occurring mineral beryl and varying weight percentage of graphene. The hybrid composites are developed with 0.5 wt. %, 1wt.%, 1.5 wt.% and 2 wt.% of graphene and 6 wt.% of beryl  by stir casting liquid metallurgy route. The cast specimens of unreinforced aluminum alloy and hybrid composite samples were prepared for heat treatment process and subjected to solutionizing treatment (T6) at a temperature of 490±5 oC for 8 hours in a muffle furnace followed by quenching in boiling water. The microstructure analysis of as cast and heat treated hybrid composite specimens are examined by scanning electron microscope (SEM). The tensile test and hardness test of unreinforced aluminum alloy and hybrid composites are examined. The wear behavior is examined by pin-on disc apparatus. The results of as cast specimens and heat treated specimens were compared. The heat treated Al7075-Beryl-Graphene hybrid composite had better properties and significantly improved the ultimate tensile strength, hardness and reduced wear loss when compared to aluminum alloy and  as cast hybrid composites.

Keywords: Beryl, graphene, heat treatment, mechanical properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1053
61 Performance Analysis of HSDPA Systems using Low-Density Parity-Check (LDPC)Coding as Compared to Turbo Coding

Authors: K. Anitha Sheela, J. Tarun Kumar

Abstract:

HSDPA is a new feature which is introduced in Release-5 specifications of the 3GPP WCDMA/UTRA standard to realize higher speed data rate together with lower round-trip times. Moreover, the HSDPA concept offers outstanding improvement of packet throughput and also significantly reduces the packet call transfer delay as compared to Release -99 DSCH. Till now the HSDPA system uses turbo coding which is the best coding technique to achieve the Shannon limit. However, the main drawbacks of turbo coding are high decoding complexity and high latency which makes it unsuitable for some applications like satellite communications, since the transmission distance itself introduces latency due to limited speed of light. Hence in this paper it is proposed to use LDPC coding in place of Turbo coding for HSDPA system which decreases the latency and decoding complexity. But LDPC coding increases the Encoding complexity. Though the complexity of transmitter increases at NodeB, the End user is at an advantage in terms of receiver complexity and Bit- error rate. In this paper LDPC Encoder is implemented using “sparse parity check matrix" H to generate a codeword at Encoder and “Belief Propagation algorithm "for LDPC decoding .Simulation results shows that in LDPC coding the BER suddenly drops as the number of iterations increase with a small increase in Eb/No. Which is not possible in Turbo coding. Also same BER was achieved using less number of iterations and hence the latency and receiver complexity has decreased for LDPC coding. HSDPA increases the downlink data rate within a cell to a theoretical maximum of 14Mbps, with 2Mbps on the uplink. The changes that HSDPA enables includes better quality, more reliable and more robust data services. In other words, while realistic data rates are only a few Mbps, the actual quality and number of users achieved will improve significantly.

Keywords: AMC, HSDPA, LDPC, WCDMA, 3GPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2048
60 Time Domain and Frequency Domain Analyses of Measured Metocean Data for Malaysian Waters

Authors: Duong Vannak, Mohd Shahir Liew, Guo Zheng Yew

Abstract:

Data of wave height and wind speed were collected from three existing oil fields in South China Sea – offshore Peninsular Malaysia, Sarawak and Sabah regions. Extreme values and other significant data were employed for analysis. The data were recorded from 1999 until 2008. The results show that offshore structures are susceptible to unacceptable motions initiated by wind and waves with worst structural impacts caused by extreme wave heights. To protect offshore structures from damage, there is a need to quantify descriptive statistics and determine spectra envelope of wind speed and wave height, and to ascertain the frequency content of each spectrum for offshore structures in the South China Sea shallow waters using measured time series. The results indicate that the process is nonstationary; it is converted to stationary process by first differencing the time series. For descriptive statistical analysis, both wind speed and wave height have significant influence on the offshore structure during the northeast monsoon with high mean wind speed of 13.5195 knots ( = 6.3566 knots) and the high mean wave height of 2.3597 m ( = 0.8690 m). Through observation of the spectra, there is no clear dominant peak and the peaks fluctuate randomly. Each wind speed spectrum and wave height spectrum has its individual identifiable pattern. The wind speed spectrum tends to grow gradually at the lower frequency range and increasing till it doubles at the higher frequency range with the mean peak frequency range of 0.4104 Hz to 0.4721 Hz, while the wave height tends to grow drastically at the low frequency range, which then fluctuates and decreases slightly at the high frequency range with the mean peak frequency range of 0.2911 Hz to 0.3425 Hz.

Keywords: Metocean, Offshore Engineering, Time Series, Descriptive Statistics, Autospectral Density Function, Wind, Wave.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3679
59 An Approach for Ensuring Data Flow in Freight Delivery and Management Systems

Authors: Aurelija Burinskienė, Dalė Dzemydienė, Arūnas Miliauskas

Abstract:

This research aims at developing the approach for more effective freight delivery and transportation process management. The road congestions and the identification of causes are important, as well as the context information recognition and management. The measure of many parameters during the transportation period and proper control of driver work became the problem. The number of vehicles per time unit passing at a given time and point for drivers can be evaluated in some situations. The collection of data is mainly used to establish new trips. The flow of the data is more complex in urban areas. Herein, the movement of freight is reported in detail, including the information on street level. When traffic density is extremely high in congestion cases, and the traffic speed is incredibly low, data transmission reaches the peak. Different data sets are generated, which depend on the type of freight delivery network. There are three types of networks: long-distance delivery networks, last-mile delivery networks and mode-based delivery networks; the last one includes different modes, in particular, railways and other networks. When freight delivery is switched from one type of the above-stated network to another, more data could be included for reporting purposes and vice versa. In this case, a significant amount of these data is used for control operations, and the problem requires an integrated methodological approach. The paper presents an approach for providing e-services for drivers by including the assessment of the multi-component infrastructure needed for delivery of freights following the network type. The construction of such a methodology is required to evaluate data flow conditions and overloads, and to minimize the time gaps in data reporting. The results obtained show the possibilities of the proposing methodological approach to support the management and decision-making processes with functionality of incorporating networking specifics, by helping to minimize the overloads in data reporting.

Keywords: Transportation networks, freight delivery, data flow, monitoring, e-services.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 607
58 High Quality Colored Wind Chimes by Anodization on Aluminum Alloy

Authors: Chia-Chih Wei, Yun-Qi Li, Ssu-Ying Chen, Hsuan-Jung Chen, Hsi-Wen Yang, Chih-Yuan Chen, Chien-Chon Chen

Abstract:

In this paper, we used a high-quality anodization technique to make a colored wind chime with a nano-tube structure anodic film, which controls the length-to-diameter ratio of an aluminum rod and controls the oxide film structure on the surface of the aluminum rod by an anodizing method. The research experiment used hard anodization to grow a controllable thickness of anodic film on an aluminum alloy surface. The hard anodization film has high hardness, high insulation, high-temperature resistance, good corrosion resistance, colors, and mass production properties that can be further applied to transportation, electronic products, biomedical fields, or energy industry applications. This study also provides in-depth research and a detailed discussion of the related process of aluminum alloy surface hard anodizing, including pre-anodization, anodization, and post-anodization. The experiment parameters of anodization include using a mixed acid solution of sulfuric acid and oxalic acid as an anodization electrolyte and controlling the temperature, time, current density, and final voltage to obtain the anodic film. In the results of the experiments, the properties of the anodic film, including thickness, hardness, insulation, and corrosion characteristics, the microstructure of the anode film were measured, and the hard anodization efficiency was calculated. Thereby it can obtain different transmission speeds of sound in the aluminum rod. And, different audio sounds can present on the aluminum rod. Another feature of the present experiment result is the use of the anodizing method and dyeing method, laser engraving patterning and electrophoresis method to make good-quality colored aluminum wind chimes.

Keywords: Anodization, aluminum, wind chime, nano-tube.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82
57 Microstructure and Corrosion Behavior of Laser Welded Magnesium Alloys with Silver Nanoparticles

Authors: M. Ishak, K. Yamasaki, K. Maekawa

Abstract:

Magnesium alloys have gained increased attention in recent years in automotive, electronics, and medical industry. This because of magnesium alloys have better properties than aluminum alloys and steels in respects of their low density and high strength to weight ratio. However, the main problems of magnesium alloy welding are the crack formation and the appearance of porosity during the solidification. This paper proposes a unique technique to weld two thin sheets of AZ31B magnesium alloy using a paste containing Ag nanoparticles. The paste containing Ag nanoparticles of 5 nm in average diameter and an organic solvent was used to coat the surface of AZ31B thin sheet. The coated sheet was heated at 100 °C for 60 s to evaporate the solvent. The dried sheet was set as a lower AZ31B sheet on the jig, and then lap fillet welding was carried out by using a pulsed Nd:YAG laser in a closed box filled with argon gas. The characteristics of the microstructure and the corrosion behavior of the joints were analyzed by opticalmicroscopy (OM), energy dispersive spectrometry (EDS), electron probe micro-analyzer (EPMA), scanning electron microscopy (SEM), and immersion corrosion test. The experimental results show that the wrought AZ31B magnesium alloy can be joined successfully using Ag nanoparticles. Ag nanoparticles insert promote grain refinement, narrower the HAZ width and wider bond width compared to weld without and insert. Corrosion rate of welded AZ31B with Ag nanoparticles reduced up to 44 % compared to base metal. The improvement of corrosion resistance of welded AZ31B with Ag nanoparticles due to finer grains and large grain boundaries area which consist of high Al content. β-phase Mg17Al12 could serve as effective barrier and suppressed further propagation of corrosion. Furthermore, Ag distribution in fusion zone provide much more finer grains and may stabilize the magnesium solid solution making it less soluble or less anodic in aqueous

Keywords: Laser welding, magnesium alloys, nanoparticles, mechanical property

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2073
56 Identification of the Antimicrobial Effect of Liquorice Extracts on Gram-Positive Bacteria: Determination of Minimum Inhibitory Concentration and Mechanism of Action Using a luxABCDE Reporter Strain

Authors: Madiha El Awamie, Catherine Rees

Abstract:

Natural preservatives have been used as alternatives to traditional chemical preservatives; however, a limited number have been commercially developed and many remain to be investigated as sources of safer and effective antimicrobials. In this study, we have been investigating the antimicrobial activity of an extract of Glycyrrhiza glabra (liquorice) that was provided as a waste material from the production of liquorice flavourings for the food industry, and to investigate if this retained the expected antimicrobial activity so it could be used as a natural preservative. Antibacterial activity of liquorice extract was screened for evidence of growth inhibition against eight species of Gram-negative and Gram-positive bacteria, including Listeria monocytogenes, Listeria innocua, Staphylococcus aureus, Enterococcus faecalis and Bacillus subtilis. The Gram-negative bacteria tested include Pseudomonas aeruginosa, Escherichia coli and Salmonella typhimurium but none of these were affected by the extract. In contrast, for all of the Gram-positive bacteria tested, growth was inhibited as monitored using optical density. However parallel studies using viable count indicated that the cells were not killed meaning that the extract was bacteriostatic rather than bacteriocidal. The Minimum Inhibitory Concentration [MIC] and Minimum Bactericidal Concentration [MBC] of the extract was also determined and a concentration of 50 µg ml-1 was found to have a strong bacteriostatic effect on Gram-positive bacteria. Microscopic analysis indicated that there were changes in cell shape suggesting the cell wall was affected. In addition, the use of a reporter strain of Listeria transformed with the bioluminescence genes luxABCDE indicated that cell energy levels were reduced when treated with either 12.5 or 50 µg ml-1 of the extract, with the reduction in light output being proportional to the concentration of the extract used. Together these results suggest that the extract is inhibiting the growth of Gram-positive bacteria only by damaging the cell wall and/or membrane.

Keywords: Antibacterial activity, bioluminescence, Glycyrrhiza glabra, natural preservative.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1682
55 The Effect of Information vs. Reasoning Gap Tasks on the Frequency of Conversational Strategies and Accuracy in Speaking among Iranian Intermediate EFL Learners

Authors: Hooriya Sadr Dadras, Shiva Seyed Erfani

Abstract:

Speaking skills merit meticulous attention both on the side of the learners and the teachers. In particular, accuracy is a critical component to guarantee the messages to be conveyed through conversation because a wrongful change may adversely alter the content and purpose of the talk. Different types of tasks have served teachers to meet numerous educational objectives. Besides, negotiation of meaning and the use of different strategies have been areas of concern in socio-cultural theories of SLA. Negotiation of meaning is among the conversational processes which have a crucial role in facilitating the understanding and expression of meaning in a given second language. Conversational strategies are used during interaction when there is a breakdown in communication that leads to the interlocutor attempting to remedy the gap through talk. Therefore, this study was an attempt to investigate if there was any significant difference between the effect of reasoning gap tasks and information gap tasks on the frequency of conversational strategies used in negotiation of meaning in classrooms on one hand, and on the accuracy in speaking of Iranian intermediate EFL learners on the other. After a pilot study to check the practicality of the treatments, at the outset of the main study, the Preliminary English Test was administered to ensure the homogeneity of 87 out of 107 participants who attended the intact classes of a 15 session term in one control and two experimental groups. Also, speaking sections of PET were used as pretest and posttest to examine their speaking accuracy. The tests were recorded and transcribed to estimate the percentage of the number of the clauses with no grammatical errors in the total produced clauses to measure the speaking accuracy. In all groups, the grammatical points of accuracy were instructed and the use of conversational strategies was practiced. Then, different kinds of reasoning gap tasks (matchmaking, deciding on the course of action, and working out a time table) and information gap tasks (restoring an incomplete chart, spot the differences, arranging sentences into stories, and guessing game) were manipulated in experimental groups during treatment sessions, and the students were required to practice conversational strategies when doing speaking tasks. The conversations throughout the terms were recorded and transcribed to count the frequency of the conversational strategies used in all groups. The results of statistical analysis demonstrated that applying both the reasoning gap tasks and information gap tasks significantly affected the frequency of conversational strategies through negotiation. In the face of the improvements, the reasoning gap tasks had a more significant impact on encouraging the negotiation of meaning and increasing the number of conversational frequencies every session. The findings also indicated both task types could help learners significantly improve their speaking accuracy. Here, applying the reasoning gap tasks was more effective than the information gap tasks in improving the level of learners’ speaking accuracy.

Keywords: Accuracy in speaking, conversational strategies, information gap tasks, reasoning gap tasks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1170
54 Catalytic Pyrolysis of Sewage Sludge for Upgrading Bio-Oil Quality Using Sludge-Based Activated Char as an Alternative to HZSM5

Authors: Ali Zaker, Zhi Chen

Abstract:

Due to the concerns about the depletion of fossil fuel sources and the deteriorating environment, the attempt to investigate the production of renewable energy will play a crucial role as a potential to alleviate the dependency on mineral fuels. One particular area of interest is generation of bio-oil through sewage sludge (SS) pyrolysis. SS can be a potential candidate in contrast to other types of biomasses due to its availability and low cost. However, the presence of high molecular weight hydrocarbons and oxygenated compounds in the SS bio-oil hinders some of its fuel applications. In this context, catalytic pyrolysis is another attainable route to upgrade bio-oil quality. Among different catalysts (i.e., zeolites) studied for SS pyrolysis, activated chars (AC) are eco-friendly alternatives. The beneficial features of AC derived from SS comprise the comparatively large surface area, porosity, enriched surface functional groups and presence of a high amount of metal species that can improve the catalytic activity. Hence, a sludge-based AC catalyst was fabricated in a single-step pyrolysis reaction with NaOH as the activation agent and was compared with HZSM5 zeolite in this study. The thermal decomposition and kinetics were invested via thermogravimetric analysis (TGA) for guidance and control of pyrolysis and catalytic pyrolysis and the design of the pyrolysis setup. The results indicated that the pyrolysis and catalytic pyrolysis contain four obvious stages and the main decomposition reaction occurred in the range of 200-600 °C. Coats-Redfern method was applied in the 2nd and 3rd devolatilization stages to estimate the reaction order and activation energy (E) from the mass loss data. The average activation energy (Em) values for the reaction orders n = 1, 2 and 3 were in the range of 6.67-20.37 kJ/mol for SS; 1.51-6.87 kJ/mol for HZSM5; and 2.29-9.17 kJ/mol for AC, respectively. According to the results, AC and HZSM5 both were able to improve the reaction rate of SS pyrolysis by abridging the Em value. Moreover, to generate and examine the effect of the catalysts on the quality of bio-oil, a fixed-bed pyrolysis system was designed and implemented. The composition analysis of the produced bio-oil was carried out via gas chromatography/mass spectrometry (GC/MS). The selected SS to catalyst ratios were 1:1, 2:1 and 4:1. The optimum ratio in terms of cracking the long-chain hydrocarbons and removing oxygen-containing compounds was 1:1 for both catalysts. The upgraded bio-oils with HZSM5 and AC were in the total range of C4-C17 with around 72% in the range of C4-C9. The bio-oil from pyrolysis of SS contained 49.27% oxygenated compounds while the presence of HZSM5 and AC dropped to 7.3% and 13.02%, respectively. Meanwhile, generation of value-added chemicals such as light aromatic compounds were significantly improved in the catalytic process. Furthermore, the fabricated AC catalyst was characterized by BET, SEM-EDX, FT-IR and TGA techniques. Overall, this research demonstrated that AC is an efficient catalyst in the pyrolysis of SS and can be used as a cost-competitive catalyst in contrast to HZSM5.

Keywords: Activated char, bio-oil, catalytic pyrolysis, HZSM5, sewage sludge.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 713
53 Application of Recycled Tungsten Carbide Powder for Fabrication of Iron Based Powder Metallurgy Alloy

Authors: Yukinori Taniguchi, Kazuyoshi Kurita, Kohei Mizuta, Keigo Nishitani, Ryuichi Fukuda

Abstract:

Tungsten carbide is widely used as a tool material in metal manufacturing process. Since tungsten is typical rare metal, establishment of recycle process of tungsten carbide tools and restore into cemented carbide material bring great impact to metal manufacturing industry. Recently, recycle process of tungsten carbide has been developed and established gradually. However, the demands for quality of cemented carbide tool are quite severe because hardness, toughness, anti-wear ability, heat resistance, fatigue strength and so on should be guaranteed for precision machining and tool life. Currently, it is hard to restore the recycled tungsten carbide powder entirely as raw material for new processed cemented carbide tool. In this study, to suggest positive use of recycled tungsten carbide powder, we have tried to fabricate a carbon based sintered steel which shows reinforced mechanical properties with recycled tungsten carbide powder. We have made set of newly designed sintered steels. Compression test of sintered specimen in density ratio of 0.85 (which means 15% porosity inside) has been conducted. As results, at least 1.7 times higher in nominal strength in the amount of 7.0 wt.% was shown in recycled WC powder. The strength reached to over 600 MPa for the Fe-WC-Co-Cu sintered alloy. Wear test has been conducted by using ball-on-disk type friction tester using 5 mm diameter ball with normal force of 2 N in the dry conditions. Wear amount after 1,000 m running distance shows that about 1.5 times longer life was shown in designed sintered alloy. Since results of tensile test showed that same tendency in previous testing, it is concluded that designed sintered alloy can be used for several mechanical parts with special strength and anti-wear ability in relatively low cost due to recycled tungsten carbide powder.

Keywords: Tungsten carbide, recycle process, compression test, powder metallurgy, anti-wear ability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1477
52 Effect of Different Tillage Systems on Soil Properties and Production on Wheat, Maize and Soybean Crop

Authors: P. I. Moraru, T. Rusu

Abstract:

Soil tillage systems can be able to influence soil compaction, water dynamics, soil temperature and crop yield. These processes can be expressed as changes of soil microbiological activity, soil respiration and sustainability of agriculture. Objectives of this study were: 1 - to assess the effects of tillage systems (Conventional System (CS), Minimum Tillage (MT), No-Tillage (NT)) on soil compaction, soil temperature, soil moisture and soil respiration and 2- to establish the effect of the changes on the production of wheat, maize and soybean. Five treatments were installed: CS-plough; MT-paraplow, chisel, rotary grape; NT-direct sowing. The study was conducted on an Argic-Stagnic Faeoziom. The MT and NT applications reduce or completely eliminate the soil mobilization, due to this; soil is compacted in the first year of application. The degree of compaction is directly related to soil type and its state of degradation. The state of soil compaction diminished over time, tending toward a specific type of soil density. Soil moisture was higher in NT and MT at the time of sowing and in the early stages of vegetation and differences diminished over time. Moisture determinations showed statistically significant differences. The MT and NT applications reduced the thermal amplitude in the first 15cm of soil depth and increased the soil temperature by 0.5-2.20C. Water dynamics and soil temperature showed no differences on the effect of crop yields. The determinations confirm the effect of soil tillage system on soil respiration; the daily average was lower at NT (315-1914 mmoli m-2s-1) and followed by MT (318-2395 mmoli m-2s-1) and is higher in the CS (321-2480 mmol m-2s-1). Comparing with CS, all the four conservation tillage measures decreased soil respiration, with the best effects of no-tillage. Although wheat production at MT and NT applications, had no significant differences soybean production was significantly affected from MT and NT applications. The differences in crop yields are recorded at maize and can be a direct consequence of loosening, mineralization and intensive mobilization of soil fertility.

Keywords: Soil tillage, soil properties, yield.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3914
51 Analysis of Non-Conventional Roundabout Performance in Mixed Traffic Conditions

Authors: Guneet Saini, Shahrukh, Sunil Sharma

Abstract:

Traffic congestion is the most critical issue faced by those in the transportation profession today. Over the past few years, roundabouts have been recognized as a measure to promote efficiency at intersections globally. In developing countries like India, this type of intersection still faces a lot of issues, such as bottleneck situations, long queues and increased waiting times, due to increasing traffic which in turn affect the performance of the entire urban network. This research is a case study of a non-conventional roundabout, in terms of geometric design, in a small town in India. These types of roundabouts should be analyzed for their functionality in mixed traffic conditions, prevalent in many developing countries. Microscopic traffic simulation is an effective tool to analyze traffic conditions and estimate various measures of operational performance of intersections such as capacity, vehicle delay, queue length and Level of Service (LOS) of urban roadway network. This study involves analyzation of an unsymmetrical non-circular 6-legged roundabout known as “Kala Aam Chauraha” in a small town Bulandshahr in Uttar Pradesh, India using VISSIM simulation package which is the most widely used software for microscopic traffic simulation. For coding in VISSIM, data are collected from the site during morning and evening peak hours of a weekday and then analyzed for base model building. The model is calibrated on driving behavior and vehicle parameters and an optimal set of calibrated parameters is obtained followed by validation of the model to obtain the base model which can replicate the real field conditions. This calibrated and validated model is then used to analyze the prevailing operational traffic performance of the roundabout which is then compared with a proposed alternative to improve efficiency of roundabout network and to accommodate pedestrians in the geometry. The study results show that the alternative proposed is an advantage over the present roundabout as it considerably reduces congestion, vehicle delay and queue length and hence, successfully improves roundabout performance without compromising on pedestrian safety. The study proposes similar designs for modification of existing non-conventional roundabouts experiencing excessive delays and queues in order to improve their efficiency especially in the case of developing countries. From this study, it can be concluded that there is a need to improve the current geometry of such roundabouts to ensure better traffic performance and safety of drivers and pedestrians negotiating the intersection and hence this proposal may be considered as a best fit.

Keywords: Operational performance, roundabout, simulation, VISSIM, traffic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 782
50 Ultrasonic Investigation of Molecular Interaction in Binary Liquid Mixture of Polyethylene Glycol with Ethanol

Authors: S. Grace Sahaya Sheba, R. Omegala Priakumari

Abstract:

Polyethylene glycol (PEG) is a condensation polymer of ethylene oxide and water. It is soluble in water and in many organic solvents. PEG is used to make emulsifying agents, detergents, soaps, plasticizers, ointments etc. Ethanol (C2H5OH) also known as ethyl alcohol is a well-known organic compound and has wide applications in chemical industry as it is used as a solvent for paint, varnish, in preserving biological specimens, used as a fuel mixed with petrol etc. Though their chemical and physical properties are already studied, still because of their uses in day to day life the authors thought it is better to study some more of their physical properties like ultrasonic velocity and hence adiabatic compressibility, free length, etc. A detailed study of such properties and some excess parameters like excess adiabatic compressibility, excess free volume and few more in the liquid mixtures of these two compounds with PEG as a solute and Ethanol as a solvent at various mole fractions may throw some light on deeper understanding of molecular interaction between the solute and the solvent supported by NMR, IR etc. Hence the present research work is on ultrasonics/allied studies on these two liquid mixtures. Ultrasonic velocity (U), density (ρ) and viscosity (η) at room temperature and at different mole fraction from 0 to 0.055 of ethanol in PEG have been experimentally carried out by the authors. Acoustical parameters such as adiabatic compressibility (β), free volume (Vf), acoustic impedance (Z), internal pressure (πi), intermolecular free length (Lf) and relaxation time (τ) were calculated from the experimental data. We have calculated excess parameters like excess adiabatic compressibility (βE), excess internal pressure (πiE) free length (LfE) and excess acoustic impedance (ZE) etc for these two chosen liquid mixtures. The excess compressibility is positive and maximum around a mole fraction 0.007 and excess internal pressure is negative and maximum at the same mole fraction and longer free length. The results are analyzed and it may be concluded that the molecular interactions between the solute and the solvent is not strong and it may be weak. Appropriate graphs are drawn.

Keywords: Adiabatic Compressibility, Binary mixture, Induce dipole, Polarizability, Ultrasonic.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2784