Search results for: P and S waves arrival time.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6772

Search results for: P and S waves arrival time.

4192 In vitro Control of Aedes aegypti Larvae Using Beauveria bassiana

Authors: R. O. B. Bitencourt, F. S. Farias, M. C. Freitas, C. J. R. Balduino, E.S. Mesquita, A. R. C. Corval, P. S. Gôlo, E. G. Pontes, V. R. E. P. Bittencourt, I. C. Angelo

Abstract:

Aedes aegypti larval survival rate was assessed after exposure to blastopores or conidia (mineral oil-in-water formulation or aqueous suspension) of Beauveria bassiana CG 479 propagules (blastospores or conidia). Here, mineral oil was used in the fungal formulation to control Aedes aegypti larvae. 1%, 0.5% or 0.1% mineral oil-in-water solutions were used to evaluate mineral oil toxicity for mosquito larvae. In the oil toxicity test, 0.1% mineral oil solution reduced only 4.5% larval survival; accordingly, this concentration was chosen for fungal oil-in-water formulations. Aqueous suspensions were prepared using 0.01% Tween 80® in sterile dechlorinated water. A. aegypti larvae (L2) were exposed in aqueous suspensions or mineral oil-in-water fungal formulations at 1×107 propagules mL-1; the survival rate (assessed daily, for 7 days) and the median survival time (S50) were calculated. Seven days after the treatment, mosquito larvae survival rates were 8.56%, 16.22%, 58%, and 42.56% after exposure to oil-in-water blastospores, oil-in-water conidia, blastospores aqueous suspension and conidia aqueous suspension (respectively). Larvae exposed to 0.01% Tween 80® had 100% survival rate and the ones treated with 0.1% mineral oil-in-water had 95.11% survival rate. Larvae treated with conidia (regardless the presence of oil) or treated with blastospores formulation had survival median time (S50) ranging from one to two days. S50 was not determined (ND) when larvae were exposed to blastospores aqueous suspension, 0.01% Tween 80® (aqueous control) or 0.1% mineral oil-in-water formulation (oil control). B. bassiana conidia and blastospores (mineral oil-in-water formulated or suspended in water) had potential to control A. aegypti mosquito larvae, despite mineral oil-in-water formulation yielded better results in comparison to aqueous suspensions. Here, B. bassiana CG 479 isolate is suggested as a potential biocontrol agent of A. aegypti mosquito larvae.

Keywords: Blastospores, formulation, mosquitoes, conidia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 930
4191 An Estimation of Rice Output Supply Response in Sierra Leone: A Nerlovian Model Approach

Authors: Alhaji M. H. Conteh, Xiangbin Yan, Issa Fofana, Brima Gegbe, Tamba I. Isaac

Abstract:

Rice grain is Sierra Leone’s staple food and the nation imports over 120,000 metric tons annually due to a shortfall in its cultivation. Thus, the insufficient level of the crop's cultivation in Sierra Leone is caused by many problems and this led to the endlessly widening supply and demand for the crop within the country. Consequently, this has instigated the government to spend huge money on the importation of this grain that would have been otherwise cultivated domestically at a cheaper cost. Hence, this research attempts to explore the response of rice supply with respect to its demand in Sierra Leone within the period 1980-2010. The Nerlovian adjustment model to the Sierra Leone rice data set within the period 1980-2010 was used. The estimated trend equations revealed that time had significant effect on output, productivity (yield) and area (acreage) of rice grain within the period 1980-2010 and this occurred generally at the 1% level of significance. The results showed that, almost the entire growth in output had the tendency to increase in the area cultivated to the crop. The time trend variable that was included for government policy intervention showed an insignificant effect on all the variables considered in this research. Therefore, both the short-run and long-run price response was inelastic since all their values were less than one. From the findings above, immediate actions that will lead to productivity growth in rice cultivation are required. To achieve the above, the responsible agencies should provide extension service schemes to farmers as well as motivating them on the adoption of modern rice varieties and technology in their rice cultivation ventures.

Keywords: Nerlovian adjustment model, price elasticities, Sierra Leone, Trend equations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2796
4190 Comparison of Different Hydrograph Routing Techniques in XPSTORM Modelling Software: A Case Study

Authors: Fatema Akram, Mohammad Golam Rasul, Mohammad Masud Kamal Khan, Md. Sharif Imam Ibne Amir

Abstract:

A variety of routing techniques are available to develop surface runoff hydrographs from rainfall. The selection of runoff routing method is very vital as it is directly related to the type of watershed and the required degree of accuracy. There are different modelling softwares available to explore the rainfall-runoff process in urban areas. XPSTORM, a link-node based, integrated stormwater modelling software, has been used in this study for developing surface runoff hydrograph for a Golf course area located in Rockhampton in Central Queensland in Australia. Four commonly used methods, namely SWMM runoff, Kinematic wave, Laurenson, and Time-Area are employed to generate runoff hydrograph for design storm of this study area. In runoff mode of XPSTORM, the rainfall, infiltration, evaporation and depression storage for subcatchments were simulated and the runoff from the subcatchment to collection node was calculated. The simulation results are presented, discussed and compared. The total surface runoff generated by SWMM runoff, Kinematic wave and Time-Area methods are found to be reasonably close, which indicates any of these methods can be used for developing runoff hydrograph of the study area. Laurenson method produces a comparatively less amount of surface runoff, however, it creates highest peak of surface runoff among all which may be suitable for hilly region. Although the Laurenson hydrograph technique is widely acceptable surface runoff routing technique in Queensland (Australia), extensive investigation is recommended with detailed topographic and hydrologic data in order to assess its suitability for use in the case study area.

Keywords: ARI, design storm, IFD, rainfall temporal pattern, routing techniques, surface runoff, XPSTORM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5047
4189 Unveiling the Mathematical Essence of Machine Learning: A Comprehensive Exploration

Authors: Randhir Singh Baghel

Abstract:

In this study, the fundamental ideas guiding the dynamic area of machine learning—where models thrive and algorithms change over time—are rooted in an innate mathematical link. This study explores the fundamental ideas that drive the development of intelligent systems, providing light on the mutually beneficial link between mathematics and machine learning.

Keywords: Machine Learning, deep learning, Neural Network, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166
4188 Typical Day Prediction Model for Output Power and Energy Efficiency of a Grid-Connected Solar Photovoltaic System

Authors: Yan Su, L. C. Chan

Abstract:

A novel typical day prediction model have been built and validated by the measured data of a grid-connected solar photovoltaic (PV) system in Macau. Unlike conventional statistical method used by previous study on PV systems which get results by averaging nearby continuous points, the present typical day statistical method obtain the value at every minute in a typical day by averaging discontinuous points at the same minute in different days. This typical day statistical method based on discontinuous point averaging makes it possible for us to obtain the Gaussian shape dynamical distributions for solar irradiance and output power in a yearly or monthly typical day. Based on the yearly typical day statistical analysis results, the maximum possible accumulated output energy in a year with on site climate conditions and the corresponding optimal PV system running time are obtained. Periodic Gaussian shape prediction models for solar irradiance, output energy and system energy efficiency have been built and their coefficients have been determined based on the yearly, maximum and minimum monthly typical day Gaussian distribution parameters, which are obtained from iterations for minimum Root Mean Squared Deviation (RMSD). With the present model, the dynamical effects due to time difference in a day are kept and the day to day uncertainty due to weather changing are smoothed but still included. The periodic Gaussian shape correlations for solar irradiance, output power and system energy efficiency have been compared favorably with data of the PV system in Macau and proved to be an improvement than previous models.

Keywords: Grid Connected, RMSD, Solar PV System, Typical Day.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1679
4187 Evaluation of Chromium Fortified Parboiled Rice Coated with Herbal Extracts: Cooking Quality and Sensory Properties

Authors: Wisnu Adi Yulianto, Agus Slamet, Sri Luwihana, Septian Albar Dwi Suprayogi

Abstract:

Parboiled rice was developed to produce rice, which has a low glycemic index for diabetics. However, diabetics also have a chromium (Cr) deficiency. Thus, it is important to fortify rice with Cr to increase the Cr content. Moreover, parboiled rice becomes rancid easily and has a musty odor, rendering the rice unfavorable. Natural herbs such as pandan leaves (Pandanus amaryllifolius Roxb.), bay leaves (Syzygium polyanthum [Wigh] Walp) and cinnamon bark powder (Cinnamomon cassia) are commonly added to food as aroma enhancers. Previous research has shown that these herbs could improve insulin sensitivity. The purpose of this study was to evaluate the effect of herbal extract coatings on the cooking quality and the preference level of chromium fortified - parboiled rice (CFPR). The rice grain variety used for this experiment was Ciherang and the fortificant was CrCl3. The three herbal extracts used for coating the CFPR were cinnamon, pandan and bay leaf, with concentration variations of 3%, 6%, and 9% (w/w) for each of the extracts. The samples were analyzed for their alkali spreading value, cooking time, elongation, water uptake ratio, solid loss, colour and lightness; and their sensory properties were determined by means of an organoleptic test. The research showed that coating the CFPR with pandan and cinnamon extracts at a concentration of 3% each produced a preferred CFPR. When coated with those herbal extracts the CFPR had the following cooking quality properties: alkali spreading value 5 (intermediate gelatinization temperature), cooking time, 26-27 min, color value, 14.95-15.00, lightness, 42.30 – 44.06, elongation, 1.53 – 1.54, water uptake ratio , 4.05-4.06, and solid loss, 0.09/100 g – 0.13 g/100 g.

Keywords: Bay leaves, chromium, cinnamon, pandan leaves, parboiled rice.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1981
4186 Optimization of Quercus cerris Bark Liquefaction

Authors: Luísa P. Cruz-Lopes, Hugo Costa e Silva, Idalina Domingos, José Ferreira, Luís Teixeira de Lemos, Bruno Esteves

Abstract:

The liquefaction process of cork based tree barks has led to an increase of interest due to its potential innovation in the lumber and wood industries. In this particular study the bark of Quercus cerris (Turkish oak) is used due to its appreciable amount of cork tissue, although of inferior quality when compared to the cork provided by other Quercus trees. This study aims to optimize alkaline catalysis liquefaction conditions, regarding several parameters. To better comprehend the possible chemical characteristics of the bark of Quercus cerris, a complete chemical analysis was performed. The liquefaction process was performed in a double-jacket reactor heated with oil, using glycerol and a mixture of glycerol/ethylene glycol as solvents, potassium hydroxide as a catalyst, and varying the temperature, liquefaction time and granulometry. Due to low liquefaction efficiency resulting from the first experimental procedures a study was made regarding different washing techniques after the filtration process using methanol and methanol/water. The chemical analysis stated that the bark of Quercus cerris is mostly composed by suberin (ca. 30%) and lignin (ca. 24%) as well as insolvent hemicelluloses in hot water (ca. 23%). On the liquefaction stage, the results that led to higher yields were: using a mixture of methanol/ethylene glycol as reagents and a time and temperature of 120 minutes and 200 ºC, respectively. It is concluded that using a granulometry of <80 mesh leads to better results, even if this parameter barely influences the liquefaction efficiency. Regarding the filtration stage, washing the residue with methanol and then distilled water leads to a considerable increase on final liquefaction percentages, which proves that this procedure is effective at liquefying suberin content and lignocellulose fraction.

Keywords: Liquefaction, alkaline catalysis, optimization, Quercus cerris bark.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
4185 Uncontrollable Inaccuracy in Inverse Problems

Authors: Yu. Menshikov

Abstract:

In this paper the influence of errors of function derivatives in initial time which have been obtained by experiment (uncontrollable inaccuracy) to the results of inverse problem solution was investigated. It was shown that these errors distort the inverse problem solution as a rule near the beginning of interval where the solutions are analyzed. Several methods for removing the influence of uncontrollable inaccuracy have been suggested. 

Keywords: Inverse problems, uncontrollable inaccuracy, filtration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1170
4184 The Direct Drivers of Ethnocentric Consumer, Intention and Actual Purchasing Behavior in Malaysia

Authors: Nik Kamariah Nik-Mat, Noor Hasmini Abd-Ghani, Jamal Mohammed Esmail Al-Ekam

Abstract:

The Malaysian government had consistently revived its campaign for “Buy Malaysian Goods” from time to time. The purpose of the campaign is to remind consumers to be ethnocentric and patriotic when purchasing product and services. This is necessary to ensure high demand for local products and services compared to foreign products. However, the decline of domestic investment in 2012 has triggered concern for the Malaysian economy. Hence, this study attempts to determine the drivers of actual purchasing behavior, intention to purchase domestic products and ethnocentrism. The study employs the cross-sectional primary data, self-administered on household, selected using stratified random sampling in four Malaysian regions. A nine factor driver of actual domestic purchasing behavior (culture openness, conservatism, collectivism, patriotism, control belief, interest in foreign travel, attitude, ethnocentrism and intention) were measured utilizing 60 items, using 7-point Likertscale. From 1000 questionnaires distributed, a sample of 486 were returned representing 48.6 percent response rate. From the fit generated structural model (SEM analysis), it was found that the drivers of actual purchase behavior are collectivism, cultural openness and patriotism; the drivers of intention to purchase domestic product are attitude, control belief, collectivism and conservatism; and drivers of ethnocentrism are cultural openness, control belief, foreign travel and patriotism. It also shows that Malaysian consumers scored high in ethnocentrism and patriotism. The findings are discussed in the perspective of its implication to Malaysian National Agenda.

Keywords: Actual purchase, ethnocentrism, culture openness, conservatism, collectivism, patriotism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3159
4183 Game-Tree Simplification by Pattern Matching and Its Acceleration Approach using an FPGA

Authors: Suguru Ochiai, Toru Yabuki, Yoshiki Yamaguchi, Yuetsu Kodama

Abstract:

In this paper, we propose a Connect6 solver which adopts a hybrid approach based on a tree-search algorithm and image processing techniques. The solver must deal with the complicated computation and provide high performance in order to make real-time decisions. The proposed approach enables the solver to be implemented on a single Spartan-6 XC6SLX45 FPGA produced by XILINX without using any external devices. The compact implementation is achieved through image processing techniques to optimize a tree-search algorithm of the Connect6 game. The tree search is widely used in computer games and the optimal search brings the best move in every turn of a computer game. Thus, many tree-search algorithms such as Minimax algorithm and artificial intelligence approaches have been widely proposed in this field. However, there is one fundamental problem in this area; the computation time increases rapidly in response to the growth of the game tree. It means the larger the game tree is, the bigger the circuit size is because of their highly parallel computation characteristics. Here, this paper aims to reduce the size of a Connect6 game tree using image processing techniques and its position symmetric property. The proposed solver is composed of four computational modules: a two-dimensional checkmate strategy checker, a template matching module, a skilful-line predictor, and a next-move selector. These modules work well together in selecting next moves from some candidates and the total amount of their circuits is small. The details of the hardware design for an FPGA implementation are described and the performance of this design is also shown in this paper.

Keywords: Connect6, pattern matching, game-tree reduction, hardware direct computation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1974
4182 Ultra-Light Overhead Conveyor Systems for Logistics Applications

Authors: Batin Latif Aylak, Bernd Noche

Abstract:

Overhead conveyor systems satisfy by their simple
construction, wide application range and their full compatibility with
other manufacturing systems, which are designed according to
international standards. Ultra-light overhead conveyor systems are
rope-based conveying systems with individually driven vehicles. The
vehicles can move automatically on the rope and this can be realized
by energy and signals. Crossings are realized by switches. Overhead
conveyor systems are particularly used in the automotive industry but
also at post offices. Overhead conveyor systems always must be
integrated with a logistical process by finding the best way for a
cheaper material flow and in order to guarantee precise and fast
workflows. With their help, any transport can take place without
wasting ground and space, without excessive company capacity, lost
or damaged products, erroneous delivery, endless travels and without
wasting time. Ultra-light overhead conveyor systems provide optimal
material flow, which produces profit and saves time. This article
illustrates the advantages of the structure of the ultra-light overhead
conveyor systems in logistics applications and explains the steps of
their system design. After an illustration of the steps, currently
available systems on the market will be shown by means of their
technical characteristics. Due to their simple construction, demands
to an ultra-light overhead conveyor system will be illustrated.

Keywords: Logistics, material flow, overhead conveyor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2066
4181 A Secure Auditing Framework for Load Balancing in Cloud Environment

Authors: R. Geetha, T. Padmavathy

Abstract:

Security audit is an important aspect or feature to be considered in cloud service customer. It is basically a certification process to audit the controls that deliver the security requirements. Security audits are conducted by trained and qualified staffs that belong to an independent auditing organization. Security audits must be carried as a standard of security controls. Proper check to be made that the cloud user has a proper reporting and logging facilities with the customer's system and hence ensuring appropriate business and operational flow of data through cloud service. We propose a cloud-based secure auditing framework, which enables confided in power to safely store their mystery information on the semi-believed cloud specialist co-ops, and specifically share their mystery information with a wide scope of information recipient, to diminish the key administration intricacy for power proprietors and information collectors. Unique in relation to past cloud-based information framework, data proprietors transfer their mystery information into cloud utilizing static and dynamic evaluating plan. Another propelled determination is, if any information beneficiary needs individual record to download, the information collector will send the solicitation to the expert. The specialist proprietor has the Access Control. At the off probability, the businessman must impart the primary record to the knowledge collector, acknowledge statistics beneficiary solicitation. Once the acknowledgement for the records is over, the recipient downloads the first record and this record shifting time with date and downloading time with date are monitored by the inspector. In addition to deduplication concept, diminished cloud memory area using dynamic document distribution has been proposed.

Keywords: Cloud computing, cloud storage auditing, data integrity, key exposure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1168
4180 Dynamics of a Discrete Three Species Food Chain System

Authors: Kejun Zhuang, Zhaohui Wen

Abstract:

The main purpose of this paper is to investigate a discrete time three–species food chain system with ratio dependence. By employing coincidence degree theory and analysis techniques, sufficient conditions for existence of periodic solutions are established.

Keywords: Food chain, ratio–dependent, coincidence degree, periodic solutions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513
4179 On Bianchi Type Cosmological Models in Lyra’s Geometry

Authors: R. K. Dubey

Abstract:

Bianchi type cosmological models have been studied on the basis of Lyra’s geometry. Exact solution has been obtained by considering a time dependent displacement field for constant deceleration parameter and varying cosmological term of the universe. The physical behavior of the different models has been examined for different cases.

Keywords: Bianchi type-I cosmological model, variable gravitational coupling (G) and Cosmological Constant term (β).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1250
4178 Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters

Authors: Svetlana Cojocaru, Mircea Petic, Inga Titchiev

Abstract:

Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.

Keywords: Lexicon of disasters, modelling, Petri nets, text annotation, social disasters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1157
4177 Effect of Flour Concentration and Retrogradation Treatment on Physical Properties of Instant Sinlek Brown Rice

Authors: Supat Chaiyakul, Direk Sukkasem, Patnachapa Natthapanpaisith

Abstract:

Sinlek rice flour beverage or instant product is a dietary supplement for dysphagia, or difficulty swallowing. It is also consumed by individuals who need to consume supplements to maintain their calorific needs. This product provides protein, fat, iron, and a high concentration of carbohydrate from rice flour. However, the application of native flour is limited due to its high viscosity. Starch modification by controlling starch retrogradation was used in this study. The research studies the effects of rice flour concentration and retrogradation treatment on the physical properties of instant Sinlek brown rice. The native rice flour, gelatinized rice flour, and flour gels retrograded under 4 °C for 3 and 7 days were investigated. From the statistical results, significant differences between native and retrograded flour were observed. The concentration of rice flour was the main factor influencing the swelling power, solubility, and pasting properties. With the increase in rice flour content from 10 to 15%, swelling power, peak viscosity, trough, and final viscosity decreased; but, solubility, pasting temperature, peak time, breakdown, and setback increased. The peak time, pasting temperature, peak viscosity, trough, and final viscosity decreased as the storage period increased from 3 to 7 days. The retrograded rice flour powders had lower pasting temperature, peak viscosity, breakdown, and final viscosity than the gelatinized and native flour powders. Reduction of starch viscosity by gelatinization and controlling starch retrogradation could allow for increased quantities of rice flour in instant rice beverages. Also, the treatment could increase the energy and nutrient densities of rice beverages without affecting the viscosity of this product.

Keywords: Instant rice, pasting properties, pregelatinization, retrogradation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1626
4176 Stability Analysis of Linear Switched Systems with Mixed Delays

Authors: Xiuyong Ding, Lan Shu

Abstract:

This paper addresses the stability of the switched systems with discrete and distributed time delays. By applying Lyapunov functional and function method, we show that, if the norm of system matrices Bi is small enough, the asymptotic stability is always achieved. Finally, a example is provided to verify technically feasibility and operability of the developed results.

Keywords: Switched system, stability, Lyapunov function, Lyapunov functional, delays.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1782
4175 Phosphine Mortality Estimation for Simulation of Controlling Pest of Stored Grain: Lesser Grain Borer (Rhyzopertha dominica)

Authors: Mingren Shi, Michael Renton

Abstract:

There is a world-wide need for the development of sustainable management strategies to control pest infestation and the development of phosphine (PH3) resistance in lesser grain borer (Rhyzopertha dominica). Computer simulation models can provide a relatively fast, safe and inexpensive way to weigh the merits of various management options. However, the usefulness of simulation models relies on the accurate estimation of important model parameters, such as mortality. Concentration and time of exposure are both important in determining mortality in response to a toxic agent. Recent research indicated the existence of two resistance phenotypes in R. dominica in Australia, weak and strong, and revealed that the presence of resistance alleles at two loci confers strong resistance, thus motivating the construction of a two-locus model of resistance. Experimental data sets on purified pest strains, each corresponding to a single genotype of our two-locus model, were also available. Hence it became possible to explicitly include mortalities of the different genotypes in the model. In this paper we described how we used two generalized linear models (GLM), probit and logistic models, to fit the available experimental data sets. We used a direct algebraic approach generalized inverse matrix technique, rather than the traditional maximum likelihood estimation, to estimate the model parameters. The results show that both probit and logistic models fit the data sets well but the former is much better in terms of small least squares (numerical) errors. Meanwhile, the generalized inverse matrix technique achieved similar accuracy results to those from the maximum likelihood estimation, but is less time consuming and computationally demanding.

Keywords: mortality estimation, probit models, logistic model, generalized inverse matrix approach, pest control simulation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1584
4174 Towards a Framework for Embedded Weight Comparison Algorithm with Business Intelligence in the Plantation Domain

Authors: M. Pushparani, A. Sagaya

Abstract:

Embedded systems have emerged as important elements in various domains with extensive applications in automotive, commercial, consumer, healthcare and transportation markets, as there is emphasis on intelligent devices. On the other hand, Business Intelligence (BI) has also been extensively used in a range of applications, especially in the agriculture domain which is the area of this research. The aim of this research is to create a framework for Embedded Weight Comparison Algorithm with Business Intelligence (EWCA-BI). The weight comparison algorithm will be embedded within the plantation management system and the weighbridge system. This algorithm will be used to estimate the weight at the site and will be compared with the actual weight at the plantation. The algorithm will be used to build the necessary alerts when there is a discrepancy in the weight, thus enabling better decision making. In the current practice, data are collected from various locations in various forms. It is a challenge to consolidate data to obtain timely and accurate information for effective decision making. Adding to this, the unstable network connection leads to difficulty in getting timely accurate information. To overcome the challenges embedding is done on a portable device that will have the embedded weight comparison algorithm to also assist in data capture and synchronize data at various locations overcoming the network short comings at collection points. The EWCA-BI will provide real-time information at any given point of time, thus enabling non-latent BI reports that will provide crucial information to enable efficient operational decision making. This research has a high potential in bringing embedded system into the agriculture industry. EWCA-BI will provide BI reports with accurate information with uncompromised data using an embedded system and provide alerts, therefore, enabling effective operation management decision-making at the site.

Keywords: Embedded business intelligence, weight comparison algorithm, oil palm plantation, embedded systems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1182
4173 'Performance-Based' Seismic Methodology and Its Application in Seismic Design of Reinforced Concrete Structures

Authors: Jelena R. Pejović, Nina N. Serdar

Abstract:

This paper presents an analysis of the “Performance-Based” seismic design method, in order to overcome the perceived disadvantages and limitations of the existing seismic design approach based on force, in engineering practice. Bearing in mind, the specificity of the earthquake as a load and the fact that the seismic resistance of the structures solely depends on its behaviour in the nonlinear field, traditional seismic design approach based on force and linear analysis is not adequate. “Performance-Based” seismic design method is based on nonlinear analysis and can be used in everyday engineering practice. This paper presents the application of this method to eight-story high reinforced concrete building with combined structural system (reinforced concrete frame structural system in one direction and reinforced concrete ductile wall system in other direction). The nonlinear time-history analysis is performed on the spatial model of the structure using program Perform 3D, where the structure is exposed to forty real earthquake records. For considered building, large number of results were obtained. It was concluded that using this method we could, with a high degree of reliability, evaluate structural behavior under earthquake. It is obtained significant differences in the response of structures to various earthquake records. Also analysis showed that frame structural system had not performed well at the effect of earthquake records on soil like sand and gravel, while a ductile wall system had a satisfactory behavior on different types of soils.

Keywords: Ductile wall, frame system, nonlinear time-history analysis, performance-based methodology, RC building.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1495
4172 Data Centers’ Temperature Profile Simulation Optimized by Finite Elements and Discretization Methods

Authors: José Alberto García Fernández, Zhimin Du, Xinqiao Jin

Abstract:

Nowadays, data center industry faces strong challenges for increasing the speed and data processing capacities while at the same time is trying to keep their devices a suitable working temperature without penalizing that capacity. Consequently, the cooling systems of this kind of facilities use a large amount of energy to dissipate the heat generated inside the servers, and developing new cooling techniques or perfecting those already existing would be a great advance in this type of industry. The installation of a temperature sensor matrix distributed in the structure of each server would provide the necessary information for collecting the required data for obtaining a temperature profile instantly inside them. However, the number of temperature probes required to obtain the temperature profiles with sufficient accuracy is very high and expensive. Therefore, other less intrusive techniques are employed where each point that characterizes the server temperature profile is obtained by solving differential equations through simulation methods, simplifying data collection techniques but increasing the time to obtain results. In order to reduce these calculation times, complicated and slow computational fluid dynamics simulations are replaced by simpler and faster finite element method simulations which solve the Burgers‘ equations by backward, forward and central discretization techniques after simplifying the energy and enthalpy conservation differential equations. The discretization methods employed for solving the first and second order derivatives of the obtained Burgers‘ equation after these simplifications are the key for obtaining results with greater or lesser accuracy regardless of the characteristic truncation error.

Keywords: Burgers’ equations, CFD simulation, data center, discretization methods, FEM simulation, temperature profile.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 516
4171 Optical Properties of Some A2BCl4 Type Chlorides

Authors: D. H. Gahane, B. M. Bahirwar, S. V. Moharil

Abstract:

Efficient luminescence is reported for the first time in Eu2+ activated double Chlorides A2BCl4 (A=Alkali metal, B=Alkaline earth element). A simple wet-chemical preparation is described. Emission intensities are comparable to that of the commercial phosphor. Excitation covers near UV region. These phosphors may be useful for applications like solid state lighting, scintillation detectors and X-ray storage using photo-stimulable phosphors.

Keywords: Alkaline Earth, Chloride, Luminescence.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1801
4170 A Temporal QoS Ontology for ERTMS/ETCS

Authors: Marc Sango, Olimpia Hoinaru, Christophe Gransart, Laurence Duchien

Abstract:

Ontologies offer a means for representing and sharing information in many domains, particularly in complex domains. For example, it can be used for representing and sharing information of System Requirement Specification (SRS) of complex systems like the SRS of ERTMS/ETCS written in natural language. Since this system is a real-time and critical system, generic ontologies, such as OWL and generic ERTMS ontologies provide minimal support for modeling temporal information omnipresent in these SRS documents. To support the modeling of temporal information, one of the challenges is to enable representation of dynamic features evolving in time within a generic ontology with a minimal redesign of it. The separation of temporal information from other information can help to predict system runtime operation and to properly design and implement them. In addition, it is helpful to provide a reasoning and querying techniques to reason and query temporal information represented in the ontology in order to detect potential temporal inconsistencies. To address this challenge, we propose a lightweight 3-layer temporal Quality of Service (QoS) ontology for representing, reasoning and querying over temporal and non-temporal information in a complex domain ontology. Representing QoS entities in separated layers can clarify the distinction between the non QoS entities and the QoS entities in an ontology. The upper generic layer of the proposed ontology provides an intuitive knowledge of domain components, specially ERTMS/ETCS components. The separation of the intermediate QoS layer from the lower QoS layer allows us to focus on specific QoS Characteristics, such as temporal or integrity characteristics. In this paper, we focus on temporal information that can be used to predict system runtime operation. To evaluate our approach, an example of the proposed domain ontology for handover operation, as well as a reasoning rule over temporal relations in this domain-specific ontology, are presented.

Keywords: System Requirement Specification, ERTMS/ETCS, Temporal Ontologies, Domain Ontologies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3135
4169 Constructing a Simple Polygonalizations

Authors: V. Tereshchenko, V. Muravitskiy

Abstract:

We consider the methods of construction simple polygons for a set S of n points and applying them for searching the minimal area polygon. In this paper we propose the approximate algorithm, which generates the simple polygonalizations of a fixed set of points and finds the minimal area polygon, in O (n3) time and using O(n2) memory.

Keywords: simple polygon, approximate algorithm, minimal area polygon, polygonalizations

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2222
4168 In situ Real-Time Multivariate Analysis of Methanolysis Monitoring of Sunflower Oil Using FTIR

Authors: Pascal Mwenge, Tumisang Seodigeng

Abstract:

The combination of world population and the third industrial revolution led to high demand for fuels. On the other hand, the decrease of global fossil 8fuels deposits and the environmental air pollution caused by these fuels has compounded the challenges the world faces due to its need for energy. Therefore, new forms of environmentally friendly and renewable fuels such as biodiesel are needed. The primary analytical techniques for methanolysis yield monitoring have been chromatography and spectroscopy, these methods have been proven reliable but are more demanding, costly and do not provide real-time monitoring. In this work, the in situ monitoring of biodiesel from sunflower oil using FTIR (Fourier Transform Infrared) has been studied; the study was performed using EasyMax Mettler Toledo reactor equipped with a DiComp (Diamond) probe. The quantitative monitoring of methanolysis was performed by building a quantitative model with multivariate calibration using iC Quant module from iC IR 7.0 software. 15 samples of known concentrations were used for the modelling which were taken in duplicate for model calibration and cross-validation, data were pre-processed using mean centering and variance scale, spectrum math square root and solvent subtraction. These pre-processing methods improved the performance indexes from 7.98 to 0.0096, 11.2 to 3.41, 6.32 to 2.72, 0.9416 to 0.9999, RMSEC, RMSECV, RMSEP and R2Cum, respectively. The R2 value of 1 (training), 0.9918 (test), 0.9946 (cross-validation) indicated the fitness of the model built. The model was tested against univariate model; small discrepancies were observed at low concentration due to unmodelled intermediates but were quite close at concentrations above 18%. The software eliminated the complexity of the Partial Least Square (PLS) chemometrics. It was concluded that the model obtained could be used to monitor methanol of sunflower oil at industrial and lab scale.

Keywords: Biodiesel, calibration, chemometrics, FTIR, methanolysis, multivariate analysis, transesterification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 934
4167 Thailand Throne Hall Architecture in the Grand Palace in the Early Days of Ratthanakosin Era

Authors: Somchai Seviset, Lin Jian Qun

Abstract:

Amarindra-vinitchai-mahaisuraya Bhiman throne hall is one of the most significant throne halls in the grand palace in the Ratthanakosin city situated in Bangkok, Thailand. This is the first group of throne halls built in order to serve as a place for meetings, performing state affairs and royal duties until the present time. The structure and pattern of architectural design including the decoration and interior design of the throne hall obviously exhibits and convey the status of the king under the context of Thai society in the early period of Ratthanakosin era. According to the tradition of ruling the kingdom in absolute monarchy which had been in place since Ayutthaya era (A.D.1350-1767), the king was deemed as Deva Raja, the highest power and authority over the kingdom and as the greatest emperor of the universe (Chakkravatin). The architectural design adopted the concept of “Prasada" or Viman which served as the dwelling place of the gods and was presented in the form of “Thai traditional architecture" For the interior design of the throne hall, it had been adopted to be the heaven and the centre of the Universe in line with the cosmological beliefs of ancient people described in scripture Tribhumikatha (Tri Bhumi) written by Phra Maha Thamma Raja (Phraya Lithai) of the Sukhothai era (A.D.1347-1368). According to this belief, the throne hall had been designed to represent mount Meru, the central of the universe. On the top end of Mount Meru is situated the Viman and dwelling place of Indra who is the king of gods according to the idea of Deva Raja (the king god Avatar). At the same time, Indra also existed as the king of the universe simultaneously.

Keywords: Amarindra-vinitchai-mahaisuraya Bhiman throne hall, throne hall architecture, grand palace, Thai traditional architecture, Ratthanakosin era

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2519
4166 An Expansion Method for Schrödinger Equation of Quantum Billiards with Arbitrary Shapes

Authors: İnci M. Erhan

Abstract:

A numerical method for solving the time-independent Schrödinger equation of a particle moving freely in a three-dimensional axisymmetric region is developed. The boundary of the region is defined by an arbitrary analytic function. The method uses a coordinate transformation and an expansion in eigenfunctions. The effectiveness is checked and confirmed by applying the method to a particular example, which is a prolate spheroid.

Keywords: Bessel functions, Eigenfunction expansion, Quantum billiard, Schrödinger equation, Spherical harmonics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5215
4165 Probabilistic Graphical Model for the Web

Authors: M. Nekri, A. Khelladi

Abstract:

The world wide web network is a network with a complex topology, the main properties of which are the distribution of degrees in power law, A low clustering coefficient and a weak average distance. Modeling the web as a graph allows locating the information in little time and consequently offering a help in the construction of the research engine. Here, we present a model based on the already existing probabilistic graphs with all the aforesaid characteristics. This work will consist in studying the web in order to know its structuring thus it will enable us to modelize it more easily and propose a possible algorithm for its exploration.

Keywords: Clustering coefficient, preferential attachment, small world, Web community.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1604
4164 A Mark-Up Approach to Add Value

Authors: Ivaylo I. Atanasov, Evelina N.Pencheva

Abstract:

This paper presents a mark-up approach to service creation in Next Generation Networks. The approach allows deriving added value from network functions exposed by Parlay/OSA (Open Service Access) interfaces. With OSA interfaces service logic scripts might be executed both on callrelated and call-unrelated events. To illustrate the approach XMLbased language constructions for data and method definitions, flow control, time measuring and supervision and database access are given and an example of OSA application is considered.

Keywords: Service creation, mark-up approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1687
4163 Multi-Sensor Image Fusion for Visible and Infrared Thermal Images

Authors: Amit Kr. Happy

Abstract:

This paper is motivated by the importance of multi-sensor image fusion with specific focus on Infrared (IR) and Visible image (VI) fusion for various applications including military reconnaissance. Image fusion can be defined as the process of combining two or more source images into a single composite image with extended information content that improves visual perception or feature extraction. These images can be from different modalities like Visible camera & IR Thermal Imager. While visible images are captured by reflected radiations in the visible spectrum, the thermal images are formed from thermal radiation (IR) that may be reflected or self-emitted. A digital color camera captures the visible source image and a thermal IR camera acquires the thermal source image. In this paper, some image fusion algorithms based upon Multi-Scale Transform (MST) and region-based selection rule with consistency verification have been proposed and presented. This research includes implementation of the proposed image fusion algorithm in MATLAB along with a comparative analysis to decide the optimum number of levels for MST and the coefficient fusion rule. The results are presented, and several commonly used evaluation metrics are used to assess the suggested method's validity. Experiments show that the proposed approach is capable of producing good fusion results. While deploying our image fusion algorithm approaches, we observe several challenges from the popular image fusion methods. While high computational cost and complex processing steps of image fusion algorithms provide accurate fused results, but they also make it hard to become deployed in system and applications that require real-time operation, high flexibility and low computation ability. So, the methods presented in this paper offer good results with minimum time complexity.

Keywords: Image fusion, IR thermal imager, multi-sensor, Multi-Scale Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 430