Search results for: variable gravitational coupling (G) and Cosmological Constant term (β).
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2816

Search results for: variable gravitational coupling (G) and Cosmological Constant term (β).

296 Incorporating Semantic Similarity Measure in Genetic Algorithm : An Approach for Searching the Gene Ontology Terms

Authors: Razib M. Othman, Safaai Deris, Rosli M. Illias, Hany T. Alashwal, Rohayanti Hassan, FarhanMohamed

Abstract:

The most important property of the Gene Ontology is the terms. These control vocabularies are defined to provide consistent descriptions of gene products that are shareable and computationally accessible by humans, software agent, or other machine-readable meta-data. Each term is associated with information such as definition, synonyms, database references, amino acid sequences, and relationships to other terms. This information has made the Gene Ontology broadly applied in microarray and proteomic analysis. However, the process of searching the terms is still carried out using traditional approach which is based on keyword matching. The weaknesses of this approach are: ignoring semantic relationships between terms, and highly depending on a specialist to find similar terms. Therefore, this study combines semantic similarity measure and genetic algorithm to perform a better retrieval process for searching semantically similar terms. The semantic similarity measure is used to compute similitude strength between two terms. Then, the genetic algorithm is employed to perform batch retrievals and to handle the situation of the large search space of the Gene Ontology graph. The computational results are presented to show the effectiveness of the proposed algorithm.

Keywords: Gene Ontology, Semantic similarity measure, Genetic algorithm, Ontology search

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1465
295 Mango (Mangifera indica L.) Lyophilization Using Vacuum-Induced Freezing

Authors: Natalia A. Salazar, Erika K. Méndez, Catalina Álvarez, Carlos E. Orrego

Abstract:

Lyophilization, also called freeze-drying, is an important dehydration technique mainly used for pharmaceuticals. Food industry also uses lyophilization when it is important to retain most of the nutritional quality, taste, shape and size of dried products and to extend their shelf life. Vacuum-Induced during freezing cycle (VI) has been used in order to control ice nucleation and, consequently, to reduce the time of primary drying cycle of pharmaceuticals preserving quality properties of the final product. This procedure has not been applied in freeze drying of foods. The present work aims to investigate the effect of VI on the lyophilization drying time, final moisture content, density and reconstitutional properties of mango (Mangifera indica L.) slices (MS) and mango pulp-maltodextrin dispersions (MPM) (30% concentration of total solids). Control samples were run at each freezing rate without using induced vacuum. The lyophilization endpoint was the same for all treatments (constant difference between capacitance and Pirani vacuum gauges). From the experimental results it can be concluded that at the high freezing rate (0.4°C/min) reduced the overall process time up to 30% comparing process time required for the control and VI of the lower freeze rate (0.1°C/min) without affecting the quality characteristics of the dried product, which yields a reduction in costs and energy consumption for MS and MPM freeze drying. Controls and samples treated with VI at freezing rate of 0.4°C/min in MS showed similar results in moisture and density parameters. Furthermore, results from MPM dispersion showed favorable values when VI was applied because dried product with low moisture content and low density was obtained at shorter process time compared with the control. There were not found significant differences between reconstitutional properties (rehydration for MS and solubility for MPM) of freeze dried mango resulting from controls, and VI treatments.

Keywords: Drying time, lyophilization, mango, vacuum induced freezing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2244
294 An Efficient Backward Semi-Lagrangian Scheme for Nonlinear Advection-Diffusion Equation

Authors: Soyoon Bak, Sunyoung Bu, Philsu Kim

Abstract:

In this paper, a backward semi-Lagrangian scheme combined with the second-order backward difference formula is designed to calculate the numerical solutions of nonlinear advection-diffusion equations. The primary aims of this paper are to remove any iteration process and to get an efficient algorithm with the convergence order of accuracy 2 in time. In order to achieve these objects, we use the second-order central finite difference and the B-spline approximations of degree 2 and 3 in order to approximate the diffusion term and the spatial discretization, respectively. For the temporal discretization, the second order backward difference formula is applied. To calculate the numerical solution of the starting point of the characteristic curves, we use the error correction methodology developed by the authors recently. The proposed algorithm turns out to be completely iteration free, which resolves the main weakness of the conventional backward semi-Lagrangian method. Also, the adaptability of the proposed method is indicated by numerical simulations for Burgers’ equations. Throughout these numerical simulations, it is shown that the numerical results is in good agreement with the analytic solution and the present scheme offer better accuracy in comparison with other existing numerical schemes.

Keywords: Semi-Lagrangian method, Iteration free method, Nonlinear advection-diffusion equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2459
293 FEM Models of Glued Laminated Timber Beams Enhanced by Bayesian Updating of Elastic Moduli

Authors: L. Melzerová, T. Janda, M. Šejnoha, J. Šejnoha

Abstract:

Two finite element (FEM) models are presented in this paper to address the random nature of the response of glued timber structures made of wood segments with variable elastic moduli evaluated from 3600 indentation measurements. This total database served to create the same number of ensembles as was the number of segments in the tested beam. Statistics of these ensembles were then assigned to given segments of beams and the Latin Hypercube Sampling (LHS) method was called to perform 100 simulations resulting into the ensemble of 100 deflections subjected to statistical evaluation. Here, a detailed geometrical arrangement of individual segments in the laminated beam was considered in the construction of two-dimensional FEM model subjected to in fourpoint bending to comply with the laboratory tests. Since laboratory measurements of local elastic moduli may in general suffer from a significant experimental error, it appears advantageous to exploit the full scale measurements of timber beams, i.e. deflections, to improve their prior distributions with the help of the Bayesian statistical method. This, however, requires an efficient computational model when simulating the laboratory tests numerically. To this end, a simplified model based on Mindlin’s beam theory was established. The improved posterior distributions show that the most significant change of the Young’s modulus distribution takes place in laminae in the most strained zones, i.e. in the top and bottom layers within the beam center region. Posterior distributions of moduli of elasticity were subsequently utilized in the 2D FEM model and compared with the original simulations.

Keywords: Bayesian inference, FEM, four point bending test, laminated timber, parameter estimation, prior and posterior distribution, Young’s modulus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2182
292 The Concept of the Aesthetic Features in Architectural Structures of the Museums

Authors: D. Moussazadeh, A. Aytug

Abstract:

The focus of this study is to analyze and elaborate the formal factors in the architectural features of the museums. From aesthetic vantage point, this study has scrutinized the formal aesthetic values and identity-related features of the museums. Furthermore, the importance of the museums as the centers of knowledge, science and arts has gradually increased in the last century, whereby they have shifted from an elite standing to the pluralist approach as to address every sections of the community. This study will focus on the museum structures that are designed with the aesthetic apprehension, and presented as the artistic works on the basis of an objective attitude to elaborate the formal aesthetic factors on the formal aesthetics. It is of great importance to increase such studies for getting some concrete results to perceive the recent term aesthetic approaches and improve the forms in line with such approaches. This study elaborates the aesthetic facts solely on the basis of visual dimensions, but ignores the subjective effects to evaluate it in formal, subjective and conceptual aspects. The main material of this study comprises of the descriptive works on the conceptual substructure, and a number of schedules drawn on such concepts, which are applied on the example museum structures. Such works cover many several existing sources such as the design, philosophy, artistic philosophy, shape, form, design elements and principles as well as the museums.

Keywords: Aesthetics, design principles and elements, Gestalt.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1047
291 A Cumulative Learning Approach to Data Mining Employing Censored Production Rules (CPRs)

Authors: Rekha Kandwal, Kamal K.Bharadwaj

Abstract:

Knowledge is indispensable but voluminous knowledge becomes a bottleneck for efficient processing. A great challenge for data mining activity is the generation of large number of potential rules as a result of mining process. In fact sometimes result size is comparable to the original data. Traditional data mining pruning activities such as support do not sufficiently reduce the huge rule space. Moreover, many practical applications are characterized by continual change of data and knowledge, thereby making knowledge voluminous with each change. The most predominant representation of the discovered knowledge is the standard Production Rules (PRs) in the form If P Then D. Michalski & Winston proposed Censored Production Rules (CPRs), as an extension of production rules, that exhibit variable precision and supports an efficient mechanism for handling exceptions. A CPR is an augmented production rule of the form: If P Then D Unless C, where C (Censor) is an exception to the rule. Such rules are employed in situations in which the conditional statement 'If P Then D' holds frequently and the assertion C holds rarely. By using a rule of this type we are free to ignore the exception conditions, when the resources needed to establish its presence, are tight or there is simply no information available as to whether it holds or not. Thus the 'If P Then D' part of the CPR expresses important information while the Unless C part acts only as a switch changes the polarity of D to ~D. In this paper a scheme based on Dempster-Shafer Theory (DST) interpretation of a CPR is suggested for discovering CPRs from the discovered flat PRs. The discovery of CPRs from flat rules would result in considerable reduction of the already discovered rules. The proposed scheme incrementally incorporates new knowledge and also reduces the size of knowledge base considerably with each episode. Examples are given to demonstrate the behaviour of the proposed scheme. The suggested cumulative learning scheme would be useful in mining data streams.

Keywords: Censored production rules, cumulative learning, data mining, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1459
290 Flexible Wormhole-Switched Network-on-chip with Two-Level Priority Data Delivery Service

Authors: Faizal A. Samman, Thomas Hollstein, Manfred Glesner

Abstract:

A synchronous network-on-chip using wormhole packet switching and supporting guaranteed-completion best-effort with low-priority (LP) and high-priority (HP) wormhole packet delivery service is presented in this paper. Both our proposed LP and HP message services deliver a good quality of service in term of lossless packet completion and in-order message data delivery. However, the LP message service does not guarantee minimal completion bound. The HP packets will absolutely use 100% bandwidth of their reserved links if the HP packets are injected from the source node with maximum injection. Hence, the service are suitable for small size messages (less than hundred bytes). Otherwise the other HP and LP messages, which require also the links, will experience relatively high latency depending on the size of the HP message. The LP packets are routed using a minimal adaptive routing, while the HP packets are routed using a non-minimal adaptive routing algorithm. Therefore, an additional 3-bit field, identifying the packet type, is introduced in their packet headers to classify and to determine the type of service committed to the packet. Our NoC prototypes have been also synthesized using a 180-nm CMOS standard-cell technology to evaluate the cost of implementing the combination of both services.

Keywords: Network-on-Chip, Parallel Pipeline Router Architecture, Wormhole Switching, Two-Level Priority Service.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1738
289 Effect of Heat Treatment on Mechanical Properties and Wear Behavior of Al7075 Alloy Reinforced with Beryl and Graphene Hybrid Metal Matrix Composites

Authors: Shanawaz Patil, Mohamed Haneef, K. S. Narayanaswamy

Abstract:

In the recent years, aluminum metal matrix composites were most widely used, which are finding wide applications in various field such as automobile, aerospace defense etc., due to their outstanding mechanical properties like low density, light weight, exceptional high levels of strength, stiffness, wear resistance, high temperature resistance, low coefficient of thermal expansion and good formability. In the present work, an effort is made to study the effect of heat treatment on mechanical properties of aluminum 7075 alloy reinforced with constant weight percentage of naturally occurring mineral beryl and varying weight percentage of graphene. The hybrid composites are developed with 0.5 wt. %, 1wt.%, 1.5 wt.% and 2 wt.% of graphene and 6 wt.% of beryl  by stir casting liquid metallurgy route. The cast specimens of unreinforced aluminum alloy and hybrid composite samples were prepared for heat treatment process and subjected to solutionizing treatment (T6) at a temperature of 490±5 oC for 8 hours in a muffle furnace followed by quenching in boiling water. The microstructure analysis of as cast and heat treated hybrid composite specimens are examined by scanning electron microscope (SEM). The tensile test and hardness test of unreinforced aluminum alloy and hybrid composites are examined. The wear behavior is examined by pin-on disc apparatus. The results of as cast specimens and heat treated specimens were compared. The heat treated Al7075-Beryl-Graphene hybrid composite had better properties and significantly improved the ultimate tensile strength, hardness and reduced wear loss when compared to aluminum alloy and  as cast hybrid composites.

Keywords: Beryl, graphene, heat treatment, mechanical properties.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1012
288 Long-Term Structural Behavior of Resilient Materials for Reduction of Floor Impact Sound

Authors: J. Y. Lee, J. Kim, H. J. Chang, J. M. Kim

Abstract:

People’s tendency towards living in apartment houses is increasing in a densely populated country. However, some residents living in apartment houses are bothered by noise coming from the houses above. In order to reduce noise pollution, the communities are increasingly imposing a bylaw, including the limitation of floor impact sound, minimum thickness of floors, and floor soundproofing solutions. This research effort focused on the specific long-time deflection of resilient materials in the floor sound insulation systems of apartment houses. The experimental program consisted of testing nine floor sound insulation specimens subjected to sustained load for 45 days. Two main parameters were considered in the experimental investigation: three types of resilient materials and magnitudes of loads. The test results indicated that the structural behavior of the floor sound insulation systems under long-time load was quite different from that the systems under short-time load. The loading period increased the deflection of floor sound insulation systems and the increasing rate of the long-time deflection of the systems with ethylene vinyl acetate was smaller than that of the systems with low density ethylene polystyrene.

Keywords: Resilient materials, floor sound insulation systems, long-time deflection, sustained load, noise pollution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2330
287 Laboratory Analysis of Stormwater Runoff Hydraulic and Pollutant Removal Performance of Pervious Concrete Based on Seashell By-Products

Authors: Jean-Jacques Randrianarimanana, Nassim Sebaibi, Mohamed Boutouil

Abstract:

In order to solve problems associated with stormwater runoff in urban areas and their effects on natural and artificial water bodies, the integration of new technical solutions to the rainwater drainage becomes even more essential. Permeable pavement systems are one of the most widely used techniques. This paper presents a laboratory analysis of stormwater runoff hydraulic and pollutant removal performance of permeable pavement system using pervious pavements based on seashell products. The laboratory prototype is a square column of 25 cm of side and consists of the surface in pervious concrete, a bedding of 3 cm in height, a geotextile and a subbase layer of 50 cm in height. A series of constant simulated rain events using semi-synthetic runoff which varied in intensity and duration were carried out. The initial vertical saturated hydraulic conductivity of the entire pervious pavement system was 0.25 cm/s (148 L/m2/min). The hydraulic functioning was influenced by both the inlet flow rate value and the test duration. The total water losses including evaporation ranged between 9% to 20% for all hydraulic experiments. The temporal and vertical variability of the pollutant removal efficiency (PRE) of the system were studied for total suspended solids (TSS). The results showed that the PRE along the vertical profile was influenced by the size of the suspended solids, and the pervious paver has the highest capacity to trap pollutant than the other porous layers of the permeable pavement system after the geotextile. The TSS removal efficiency was about 80% for the entire system. The first-flush effect of TSS was observed, but it appeared only at the beginning (2 to 6 min) of the experiments. It has been shown that the PPS can capture first-flush. The project in which this study is integrated aims to contribute to both the valorization of shellfish waste and the sustainable management of rainwater.

Keywords: Hydraulic, pervious concrete, pollutant removal efficiency, seashell by-products, stormwater runoff.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 921
286 Leachate Generation from Landfill Lysimeter using Different Types of Soil Cover

Authors: S. Karnchanawong, P. Yongpisalpop

Abstract:

The objectives of this study are to determine the effects of soil cover type on characteristics of leachates generated from landfill lysimeters. Four lysimeters with diameter and height of 0.15 and 3.00 m, respectively, were prepared. Three lysimeters were filled with municipal waste and three different cover soil types i.e. sandy loam soil, silty loam soil and clay soil while another lysimeter was filled solely with municipal waste. The study was conducted in the rainy season. Leachate quantities were measured every day and leachate characteristics were determined once a week. The cumulative leachate quantity from the lysimeter filled solely with municipal waste was found to be around 27% higher than the lysimeters using cover soils. There were no any differences of the cumulative leachate amounts generated from the lysimeters using three types of soils. The comparison of the total mass of pollutants generated from all lysimeters showed that the lysimeter filled solely with municipal waste generated the maximum quantities of pollutants. Among the lysimeters using different types of soils, the lysimeter using sandy loam soil generated the lowest amount of most of pollutants, compared with the lysimeters using silty loam and clay soils. It can be concluded that in term of pollutant attenuation in the leachate, a sandy loam is the most suitable soil to be used as a cover soil in the landfill.

Keywords: cover soil, leachate, sandy loam soil, silty loam soil, clay soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2108
285 Hydrogen and Diesel Combustion on a Single Cylinder Four Stroke Diesel Engine in Dual Fuel mode with Varying Injection Strategies

Authors: Probir Kumar Bose, Rahul Banerjee, Madhujit Deb

Abstract:

The present energy situation and the concerns about global warming has stimulated active research interest in non-petroleum, carbon free compounds and non-polluting fuels, particularly for transportation, power generation, and agricultural sectors. Environmental concerns and limited amount of petroleum fuels have caused interests in the development of alternative fuels for internal combustion (IC) engines. The petroleum crude reserves however, are declining and consumption of transport fuels particularly in the developing countries is increasing at high rates. Severe shortage of liquid fuels derived from petroleum may be faced in the second half of this century. Recently more and more stringent environmental regulations being enacted in the USA and Europe have led to the research and development activities on clean alternative fuels. Among the gaseous fuels hydrogen is considered to be one of the clean alternative fuel. Hydrogen is an interesting candidate for future internal combustion engine based power trains. In this experimental investigation, the performance and combustion analysis were carried out on a direct injection (DI) diesel engine using hydrogen with diesel following the TMI(Time Manifold Injection) technique at different injection timings of 10 degree,45 degree and 80 degree ATDC using an electronic control unit (ECU) and injection durations were controlled. Further, the tests have been carried out at a constant speed of 1500rpm at different load conditions and it can be observed that brake thermal efficiency increases with increase in load conditions with a maximum gain of 15% at full load conditions during all injection strategies of hydrogen. It was also observed that with the increase in hydrogen energy share BSEC started reducing and it reduced to a maximum of 9% as compared to baseline diesel at 10deg ATDC injection during maximum injection proving the exceptional combustion properties of hydrogen.

Keywords: Hydrogen, performance, combustion, alternative fuels.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3381
284 Anomaly Detection in a Data Center with a Reconstruction Method Using a Multi-Autoencoders Model

Authors: Victor Breux, Jérôme Boutet, Alain Goret, Viviane Cattin

Abstract:

Early detection of anomalies in data centers is important to reduce downtimes and the costs of periodic maintenance. However, there is little research on this topic and even fewer on the fusion of sensor data for the detection of abnormal events. The goal of this paper is to propose a method for anomaly detection in data centers by combining sensor data (temperature, humidity, power) and deep learning models. The model described in the paper uses one autoencoder per sensor to reconstruct the inputs. The auto-encoders contain Long-Short Term Memory (LSTM) layers and are trained using the normal samples of the relevant sensors selected by correlation analysis. The difference signal between the input and its reconstruction is then used to classify the samples using feature extraction and a random forest classifier. The data measured by the sensors of a data center between January 2019 and May 2020 are used to train the model, while the data between June 2020 and May 2021 are used to assess it. Performances of the model are assessed a posteriori through F1-score by comparing detected anomalies with the data center’s history. The proposed model outperforms the state-of-the-art reconstruction method, which uses only one autoencoder taking multivariate sequences and detects an anomaly with a threshold on the reconstruction error, with an F1-score of 83.60% compared to 24.16%.

Keywords: Anomaly detection, autoencoder, data centers, deep learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 696
283 Armed Groups and Intra State Conflict: A Study on the Egyptian Case

Authors: Ghzlan Mahmoud Abdel Aziz

Abstract:

This case study aims to identify the intrastate conflicts between the nation state and armed groups. Nowadays, most wars weaken states against armed groups. Thus, it is very important to negotiate with such groups in order to reinforce the law for the protection of victims. These armed groups are the cause of conflicts and they are related with many of humanitarian issues that result out of conflicts. In this age of rivalry; terrorists, insurgents, or transnational criminal parties have surfaced to the top as a reaction to these armed groups in an effort to set up a new world order. Moreover, the intra state conflicts became increasingly treacherous than the interstate conflicts, particularly when nation state systems deal with armed groups which try to influence the state. The unexpected upraising of the Arab Spring during 2011 in parts of the Middle East and North Africa formed various patterns of conflicts. The events of the Arab Spring resulted in current and long term change across the region. Significant modifications in the level, strength and period of armed conflict around the world have been made. Egypt was in the center of these events. It has fought back the armed groups under the name of terrorism and spread common disorder and violence among civilians. On this note, this study focuses on the problem of the transformation in the methods of organized violence within one state rather than between two state or more and analyzes the objectives, strategies, and internal composition of armed groups and the environments that foster them, with a focus on the Egyptian case.

Keywords: Armed groups, conflicts, Egyptian armed forces, intrastate conflicts.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1304
282 Strategic Mine Planning: A SWOT Analysis Applied to KOV Open Pit Mine in the Democratic Republic of Congo

Authors: Patrick May Mukonki

Abstract:

KOV pit (Kamoto Oliveira Virgule) is located 10 km from Kolwezi town, one of the mineral rich town in the Lualaba province of the Democratic Republic of Congo. The KOV pit is currently operating under the Katanga Mining Limited (KML), a Glencore-Gecamines (a State Owned Company) join venture. Recently, the mine optimization process provided a life of mine of approximately 10 years withnice pushbacks using the Datamine NPV Scheduler software. In previous KOV pit studies, we recently outlined the impact of the accuracy of the geological information on a long-term mine plan for a big copper mine such as KOV pit. The approach taken, discussed three main scenarios and outlined some weaknesses on the geological information side, and now, in this paper that we are going to develop here, we are going to highlight, as an overview, those weaknesses, strengths and opportunities, in a global SWOT analysis. The approach we are taking here is essentially descriptive in terms of steps taken to optimize KOV pit and, at every step, we categorized the challenges we faced to have a better tradeoff between what we called strengths and what we called weaknesses. The same logic is applied in terms of the opportunities and threats. The SWOT analysis conducted in this paper demonstrates that, despite a general poor ore body definition, and very rude ground water conditions, there is room for improvement for such high grade ore body.

Keywords: Mine planning, mine optimization, mine scheduling, SWOT analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1550
281 Comparison of Different Techniques to Estimate Surface Soil Moisture

Authors: S. Farid F. Mojtahedi, Ali Khosravi, Behnaz Naeimian, S. Adel A. Hosseini

Abstract:

Land subsidence is a gradual settling or sudden sinking of the land surface from changes that take place underground. There are different causes of land subsidence; most notably, ground-water overdraft and severe weather conditions. Subsidence of the land surface due to ground water overdraft is caused by an increase in the intergranular pressure in unconsolidated aquifers, which results in a loss of buoyancy of solid particles in the zone dewatered by the falling water table and accordingly compaction of the aquifer. On the other hand, exploitation of underground water may result in significant changes in degree of saturation of soil layers above the water table, increasing the effective stress in these layers, and considerable soil settlements. This study focuses on estimation of soil moisture at surface using different methods. Specifically, different methods for the estimation of moisture content at the soil surface, as an important term to solve Richard’s equation and estimate soil moisture profile are presented, and their results are discussed through comparison with field measurements obtained from Yanco1 station in south-eastern Australia. Surface soil moisture is not easy to measure at the spatial scale of a catchment. Due to the heterogeneity of soil type, land use, and topography, surface soil moisture may change considerably in space and time.

Keywords: Artificial neural network, empirical method, remote sensing, surface soil moisture, unsaturated soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2107
280 Integrated Modeling Approach for Energy Planning and Climate Change Mitigation Assessment in the State of Florida

Authors: Kuntal Thakkar, Chaouki Ghenai, Ahmed Hachicha

Abstract:

An integrated modeling approach was used in this study for energy planning and climate change mitigation assessment. The main objective of this study was to develop various green-house gas (GHG) mitigations scenarios in the energy demand and supply sectors for the state of Florida. The Long range energy alternative planning (LEAP) model was used in this study to examine the energy alternative and GHG emissions reduction scenarios for short and long term (2010-2050). One of the energy analysis and GHG mitigation scenarios was developed by taking into account the available renewable energy resources potential for power generation in the state of Florida. This will help to compare and analyze the GHG reduction measure against “Business As Usual” and ‘State of Florida Policy” scenarios. Two master scenarios: “Electrification” and “Energy efficiency and Lifestyle” were developed through combination of various mitigation scenarios: technological changes and energy efficiency and conservation. The results show a net reduction of the energy demand and GHG emissions by adopting these two energy scenarios compared to the business as usual.

Keywords: Integrated modeling, energy planning, climate change mitigation assessment, greenhouse gas emissions, renewable energy, energy efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1760
279 Design and Development of 5-DOF Color Sorting Manipulator for Industrial Applications

Authors: Atef. A. Ata, Sohair F. Rezeka, Ahmed El-Shenawy, Mohammed Diab

Abstract:

Image processing in today’s world grabs massive attentions as it leads to possibilities of broaden application in many fields of high technology. The real challenge is how to improve existing sorting system applications which consists of two integrated stations of processing and handling with a new image processing feature. Existing color sorting techniques use a set of inductive, capacitive, and optical sensors to differentiate object color. This research presents a mechatronic color sorting system solution with the application of image processing. A 5-DOF robot arm is designed and developed with pick and place operation to act as the main part of the color sorting system. Image processing procedure senses the circular objects in an image captured in real time by a webcam fixed at the end-effector then extracts color and position information out of it. This information is passed as a sequence of sorting commands to the manipulator that has pick-and-place mechanism. Performance analysis proves that this color based object sorting system works accurately under ideal condition in term of adequate illumination, circular objects shape and color. The circular objects tested for sorting are red, green and blue. For non-ideal condition, such as unspecified color the accuracy reduces to 80%.

Keywords: Robotics manipulator, 5-DOF manipulator, image processing, Color sorting, Pick-and-place.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4182
278 Seamless Handover in Urban 5G-UAV Systems Using Entropy Weighted Method

Authors: Anirudh Sunil Warrier, Saba Al-Rubaye, Dimitrios Panagiotakopoulos, Gokhan Inalhan, Antonios Tsourdos

Abstract:

The demand for increased data transfer rate and network traffic capacity has given rise to the concept of heterogeneous networks. Heterogeneous networks are wireless networks, consisting of devices using different underlying radio access technologies (RAT). For Unmanned Aerial Vehicles (UAVs) this enhanced data rate and network capacity are even more critical especially in their applications of medicine, delivery missions and military. In an urban heterogeneous network environment, the UAVs must be able switch seamlessly from one base station (BS) to another for maintaining a reliable link. Therefore, seamless handover in such urban environments has become a major challenge. In this paper, a scheme to achieve seamless handover is developed, an algorithm based on Received Signal Strength (RSS) criterion for network selection is used and Entropy Weighted Method (EWM) is implemented for decision making. Seamless handover using EWM decision-making is demonstrated successfully for a UAV moving across fifth generation (5G) and long-term evolution (LTE) networks via a simulation level analysis. Thus, a solution for UAV-5G communication, specifically the mobility challenge in heterogeneous networks is solved and this work could act as step forward in making UAV-5G architecture integration a possibility.

Keywords: Air to ground, A2G, fifth generation, 5G, handover, mobility, unmanned aerial vehicle, UAV, urban environments.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 376
277 Using the Nerlovian Adjustment Model to Assess the Response of Farmers to Price and Other Related Factors: Evidence from Sierra Leone Rice Cultivation

Authors: Alhaji M. H. Conteh, Xiangbin Yan, Alfred V. Gborie

Abstract:

The goal of this study was to increase the awareness of the description and assessments of rice acreage response and to offer mechanisms for agricultural policy scrutiny. The ordinary least square (OLS) technique was utilized to determine the coefficients of acreage response models for the rice varieties. The magnitudes of the coefficients (λ) of both the ROK lagged and NERICA lagged acreages were found positive and highly significant, which indicates that farmers’ adjustment rate was very low. Regarding lagged actual price for both the ROK and NERICE rice varieties, the short-run price elasticitieswere lower than long-run, which is suggesting a long term adjustment of the acreage under the crop.

However, the apparent recommendations for policy transformation are to open farm gate prices and to decrease government’s involvement in agricultural sector especially in the acquisition of agricultural inputs. Impending research have to be centered on how this might be better realized. Necessary conditions should be made available to the private sector by means of minimizing price volatility. In accordance with structural reforms, it is necessary to convey output prices to farmers with minimum distortion. There is need to eradicate price subsidies and control, which generate distortion in the market in addition to huge financial costs.

Keywords: Acreage response, rate of adjustment, rice varieties, Sierra Leone.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3756
276 SAF: A Substitution and Alignment Free Similarity Measure for Protein Sequences

Authors: Abdellali Kelil, Shengrui Wang, Ryszard Brzezinski

Abstract:

The literature reports a large number of approaches for measuring the similarity between protein sequences. Most of these approaches estimate this similarity using alignment-based techniques that do not necessarily yield biologically plausible results, for two reasons. First, for the case of non-alignable (i.e., not yet definitively aligned and biologically approved) sequences such as multi-domain, circular permutation and tandem repeat protein sequences, alignment-based approaches do not succeed in producing biologically plausible results. This is due to the nature of the alignment, which is based on the matching of subsequences in equivalent positions, while non-alignable proteins often have similar and conserved domains in non-equivalent positions. Second, the alignment-based approaches lead to similarity measures that depend heavily on the parameters set by the user for the alignment (e.g., gap penalties and substitution matrices). For easily alignable protein sequences, it's possible to supply a suitable combination of input parameters that allows such an approach to yield biologically plausible results. However, for difficult-to-align protein sequences, supplying different combinations of input parameters yields different results. Such variable results create ambiguities and complicate the similarity measurement task. To overcome these drawbacks, this paper describes a novel and effective approach for measuring the similarity between protein sequences, called SAF for Substitution and Alignment Free. Without resorting either to the alignment of protein sequences or to substitution relations between amino acids, SAF is able to efficiently detect the significant subsequences that best represent the intrinsic properties of protein sequences, those underlying the chronological dependencies of structural features and biochemical activities of protein sequences. Moreover, by using a new efficient subsequence matching scheme, SAF more efficiently handles protein sequences that contain similar structural features with significant meaning in chronologically non-equivalent positions. To show the effectiveness of SAF, extensive experiments were performed on protein datasets from different databases, and the results were compared with those obtained by several mainstream algorithms.

Keywords: Protein, Similarity, Substitution, Alignment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1382
275 Mineral Nitrogen Retention, Nitrogen Availability and Plant Growth in the Soil Influenced by Addition of Organic and Mineral Fertilizers – Lysimetric Experiment

Authors: Lukáš Plošek, Jaroslav Hynšt, Jaroslav Záhora, Jakub Elbl, Antonín Kintl, Ivana Charousová, Silvia Kovácsová

Abstract:

Compost can influence soil fertility and plant health. At the same time compost can play an important role in the nitrogen cycle and it can influence leaching of mineral nitrogen from soil to underground water.

This paper deals with the influence of compost addition and mineral nitrogen fertilizer on leaching of mineral nitrogen, nitrogen availability in microbial biomass and plant biomass production in the lysimetric experiment. Twenty one lysimeters were filed with topsoil and subsoil collected in the area of protection zone of underground source of drinking water - Březová nad Svitavou. The highest leaching of mineral nitrogen was detected in the variant fertilized only mineral nitrogen fertilizer (624.58 mg m-2), the lowest leaching was recorded in the variant with high addition of compost (315.51 mg m-2). On the other hand, losses of mineral nitrogen are not in connection with the losses of available form of nitrogen in microbial biomass. Because lost of mineral nitrogen was detected in variant with the least change in the availability of N in microbial biomass.

The leaching of mineral nitrogen, yields as well as the results concerning nitrogen availability from the first year of long term experiment suggest that compost can positive influence the leaching of nitrogen into underground water.

Keywords: Nitrogen, Compost, Biomass production, Lysimeter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2743
274 A Multi-layer Artificial Neural Network Architecture Design for Load Forecasting in Power Systems

Authors: Axay J Mehta, Hema A Mehta, T.C.Manjunath, C. Ardil

Abstract:

In this paper, the modelling and design of artificial neural network architecture for load forecasting purposes is investigated. The primary pre-requisite for power system planning is to arrive at realistic estimates of future demand of power, which is known as Load Forecasting. Short Term Load Forecasting (STLF) helps in determining the economic, reliable and secure operating strategies for power system. The dependence of load on several factors makes the load forecasting a very challenging job. An over estimation of the load may cause premature investment and unnecessary blocking of the capital where as under estimation of load may result in shortage of equipment and circuits. It is always better to plan the system for the load slightly higher than expected one so that no exigency may arise. In this paper, a load-forecasting model is proposed using a multilayer neural network with an appropriately modified back propagation learning algorithm. Once the neural network model is designed and trained, it can forecast the load of the power system 24 hours ahead on daily basis and can also forecast the cumulative load on daily basis. The real load data that is used for the Artificial Neural Network training was taken from LDC, Gujarat Electricity Board, Jambuva, Gujarat, India. The results show that the load forecasting of the ANN model follows the actual load pattern more accurately throughout the forecasted period.

Keywords: Power system, Load forecasting, Neural Network, Neuron, Stabilization, Network structure, Load.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3382
273 Implementing an Intuitive Reasoner with a Large Weather Database

Authors: Yung-Chien Sun, O. Grant Clark

Abstract:

In this paper, the implementation of a rule-based intuitive reasoner is presented. The implementation included two parts: the rule induction module and the intuitive reasoner. A large weather database was acquired as the data source. Twelve weather variables from those data were chosen as the “target variables" whose values were predicted by the intuitive reasoner. A “complex" situation was simulated by making only subsets of the data available to the rule induction module. As a result, the rules induced were based on incomplete information with variable levels of certainty. The certainty level was modeled by a metric called "Strength of Belief", which was assigned to each rule or datum as ancillary information about the confidence in its accuracy. Two techniques were employed to induce rules from the data subsets: decision tree and multi-polynomial regression, respectively for the discrete and the continuous type of target variables. The intuitive reasoner was tested for its ability to use the induced rules to predict the classes of the discrete target variables and the values of the continuous target variables. The intuitive reasoner implemented two types of reasoning: fast and broad where, by analogy to human thought, the former corresponds to fast decision making and the latter to deeper contemplation. . For reference, a weather data analysis approach which had been applied on similar tasks was adopted to analyze the complete database and create predictive models for the same 12 target variables. The values predicted by the intuitive reasoner and the reference approach were compared with actual data. The intuitive reasoner reached near-100% accuracy for two continuous target variables. For the discrete target variables, the intuitive reasoner predicted at least 70% as accurately as the reference reasoner. Since the intuitive reasoner operated on rules derived from only about 10% of the total data, it demonstrated the potential advantages in dealing with sparse data sets as compared with conventional methods.

Keywords: Artificial intelligence, intuition, knowledge acquisition, limited certainty.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1359
272 Performance Assessment of Computational Gridon Weather Indices from HOAPS Data

Authors: Madhuri Bhavsar, Anupam K Singh, Shrikant Pradhan

Abstract:

Long term rainfall analysis and prediction is a challenging task especially in the modern world where the impact of global warming is creating complications in environmental issues. These factors which are data intensive require high performance computational modeling for accurate prediction. This research paper describes a prototype which is designed and developed on grid environment using a number of coupled software infrastructural building blocks. This grid enabled system provides the demanding computational power, efficiency, resources, user-friendly interface, secured job submission and high throughput. The results obtained using sequential execution and grid enabled execution shows that computational performance has enhanced among 36% to 75%, for decade of climate parameters. Large variation in performance can be attributed to varying degree of computational resources available for job execution. Grid Computing enables the dynamic runtime selection, sharing and aggregation of distributed and autonomous resources which plays an important role not only in business, but also in scientific implications and social surroundings. This research paper attempts to explore the grid enabled computing capabilities on weather indices from HOAPS data for climate impact modeling and change detection.

Keywords: Climate model, Computational Grid, GridApplication, Heterogeneous Grid

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1411
271 The Influence of Travel Experience within Perceived Public Transport Quality

Authors: Armando Cartenì, Ilaria Henke

Abstract:

The perceived public transport quality is an important driver that influences both customer satisfaction and mobility choices. The competition among transport operators needs to improve the quality of the services and identify which attributes are perceived as relevant by passengers. Among the “traditional” public transport quality attributes there are, for example: travel and waiting time, regularity of the services, and ticket price. By contrast, there are some “non-conventional” attributes that could significantly influence customer satisfaction jointly with the “traditional” ones. Among these, the beauty/aesthetics of the transport terminals (e.g. rail station and bus terminal) is probably one of the most impacting on user perception. Starting from these considerations, the point stressed in this paper was if (and how munch) the travel experience of the overall travel (e.g. how long is the travel, how many transport modes must be used) influences the perception of the public transport quality. The aim of this paper was to investigate the weight of the terminal quality (e.g. aesthetic, comfort and service offered) within the overall travel experience. The case study was the extra-urban Italian bus network. The passengers of the major Italian terminal bus were interviewed and the analysis of the results shows that about the 75% of the travelers, are available to pay up to 30% more for the ticket price for having a high quality terminal. A travel experience effect was observed: the average perceived transport quality varies with the characteristic of the overall trip. The passengers that have a “long trip” (travel time greater than 2 hours) perceived as “low” the overall quality of the trip even if they pass through a high quality terminal. The opposite occurs for the “short trip” passengers. This means that if a traveler passes through a high quality station, the overall perception of that terminal could be significantly reduced if he is tired from a long trip. This result is important and if confirmed through other case studies, will allow to conclude that the “travel experience impact" must be considered as an explicit design variable for public transport services and planning.

Keywords: Transportation planning, sustainable mobility, decision support system, discrete choice model, design problem.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1143
270 The Risk Assessment of Nano-particles and Investigation of Their Environmental Impact

Authors: Nader Nabhani, Amir Tofighi

Abstract:

Nanotechnology is the science of creating, using and manipulating objects which have at least one dimension in range of 0.1 to 100 nanometers. In other words, nanotechnology is reconstructing a substance using its individual atoms and arranging them in a way that is desirable for our purpose. The main reason that nanotechnology has been attracting attentions is the unique properties that objects show when they are formed at nano-scale. These differing characteristics that nano-scale materials show compared to their nature-existing form is both useful in creating high quality products and dangerous when being in contact with body or spread in environment. In order to control and lower the risk of such nano-scale particles, the main following three topics should be considered: 1) First of all, these materials would cause long term diseases that may show their effects on body years after being penetrated in human organs and since this science has become recently developed in industrial scale not enough information is available about their hazards on body. 2) The second is that these particles can easily spread out in environment and remain in air, soil or water for very long time, besides their high ability to penetrate body skin and causing new kinds of diseases. 3) The third one is that to protect body and environment against the danger of these particles, the protective barriers must be finer than these small objects and such defenses are hard to accomplish. This paper will review, discuss and assess the risks that human and environment face as this new science develops at a high rate.

Keywords: Nanotechnology, risk assessment, environment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959
269 Managing Business Processes in the Age of Digital Transformation: A Literature Review

Authors: Ana-Marija Stjepić, Dalia Suša Vugec

Abstract:

Today, digital transformation is one of the leading topics that occupy the attention of scientific circles and business experts. Organizational success is most often reflected through the successful managing of business processes. Given the growing market for digital innovations and its ever-increasing impact on business, organizations need to be prepared for organizational changes that come with the digital era. In order to maintain their competitive advantage in the global market, organizations must adapt their processes to new digitalization conditions. The main goal of this study is to point out the link between the digital transformation and the business process management concept. Therefore, in order to contribute to the scientific field that explores the potential relation between business process management concept and digital transformation, a literature review has been conducted. Papers have been searched within the Business Process Management Journal by keywords related to the term digital transformation. Selected papers have been analyzed according to the topic, type of publication, year of publication, keywords, etc. The results reveal a growing number of papers published on the topic of digital transformation to the Business Process Management Journal, but the lack of case studies. This paper contributes to the extension of academic literature in this important, yet insufficiently researched, scientific field that creates the bond between two strong concepts of digital transformation and business process management.

Keywords: Business process management, digital transformation, digitalization, process change.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1685
268 Contaminant Transport in Soil from a Point Source

Authors: S. A. Nta, M. J. Ayotamuno, A. H. Igoni, R. N. Okparanma

Abstract:

The work sought to understand the pattern of movement of contaminant from a continuous point source through soil. The soil used was sandy-loam in texture. The contaminant used was municipal solid waste landfill leachate, introduced as a point source through an entry point located at the center of top layer of the soil tank. Analyses were conducted after maturity periods of 50 and 80 days. The maximum change in chemical concentration was observed on soil samples at a radial distance of 0.25 m. Finite element approximation based model was used to assess the future prediction, management and remediation in the polluted area. The actual field data collected for the case study were used to calibrate the modeling and thus simulated the flow pattern of the pollutants through soil. MATLAB R2015a was used to visualize the flow of pollutant through the soil. Dispersion coefficient at 0.25 and 0.50 m radial distance from the point of application of leachate shows a measure of the spreading of a flowing leachate due to the nature of the soil medium, with its interconnected channels distributed at random in all directions. Surface plots of metals on soil after maturity period of 80 days shows a functional relationship between a designated dependent variable (Y), and two independent variables (X and Z). Comparison of measured and predicted profile transport along the depth after 50 and 80 days of leachate application and end of the experiment shows that there were no much difference between the predicted and measured concentrations as they were all lying close to each other. For the analysis of contaminant transport, finite difference approximation based model was very effective in assessing the future prediction, management and remediation in the polluted area. The experiment gave insight into the most likely pattern of movement of contaminant as a result of continuous percolations of the leachate on soil. This is important for contaminant movement prediction and subsequent remediation of such soils.

Keywords: Contaminant, dispersion, point or leaky source, surface plot, soil.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 488
267 Prediction on Housing Price Based on Deep Learning

Authors: Li Yu, Chenlu Jiao, Hongrun Xin, Yan Wang, Kaiyang Wang

Abstract:

In order to study the impact of various factors on the housing price, we propose to build different prediction models based on deep learning to determine the existing data of the real estate in order to more accurately predict the housing price or its changing trend in the future. Considering that the factors which affect the housing price vary widely, the proposed prediction models include two categories. The first one is based on multiple characteristic factors of the real estate. We built Convolution Neural Network (CNN) prediction model and Long Short-Term Memory (LSTM) neural network prediction model based on deep learning, and logical regression model was implemented to make a comparison between these three models. Another prediction model is time series model. Based on deep learning, we proposed an LSTM-1 model purely regard to time series, then implementing and comparing the LSTM model and the Auto-Regressive and Moving Average (ARMA) model. In this paper, comprehensive study of the second-hand housing price in Beijing has been conducted from three aspects: crawling and analyzing, housing price predicting, and the result comparing. Ultimately the best model program was produced, which is of great significance to evaluation and prediction of the housing price in the real estate industry.

Keywords: Deep learning, convolutional neural network, LSTM, housing prediction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4940