Search results for: minimum fuzziness criterion
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2613

Search results for: minimum fuzziness criterion

1503 Effect of Correlation of Random Variables on Structural Reliability Index

Authors: Agnieszka Dudzik

Abstract:

The problem of correlation between random variables in the structural reliability analysis has been extensively discussed in literature on the subject. The cases taken under consideration were usually related to correlation between random variables from one side of ultimate limit state: correlation between particular loads applied on structure or correlation between resistance of particular members of a structure as a system. It has been proved that positive correlation between these random variables reduces the reliability of structure and increases the probability of failure. In the paper, the problem of correlation between random variables from both side of the limit state equation will be taken under consideration. The simplest case where these random variables are of the normal distributions will be concerned. The case when a degree of that correlation is described by the covariance or the coefficient of correlation will be used. Special attention will be paid on questions: how much that correlation changes the reliability level and can it be ignored. In reliability analysis will be used well-known methods for assessment of the failure probability: based on the Hasofer-Lind reliability index and Monte Carlo method adapted to the problem of correlation. The main purpose of this work will be a presentation how correlation of random variables influence on reliability index of steel bar structures. Structural design parameters will be defined as deterministic values and random variables. The latter will be correlated. The criterion of structural failure will be expressed by limit functions related to the ultimate and serviceability limit state. In the description of random variables will be used only for the normal distribution. Sensitivity of reliability index to the random variables will be defined. If the reliability index sensitivity due to the random variable X will be low when compared with other variables, it can be stated that the impact of this variable on failure probability is small. Therefore, in successive computations, it can be treated as a deterministic parameter. Sensitivity analysis leads to simplify the description of the mathematical model, determine the new limit functions and values of the Hasofer-Lind reliability index. In the examples, the NUMPRESS software will be used in the reliability analysis.

Keywords: correlation of random variables, reliability index, sensitivity of reliability index, steel structure

Procedia PDF Downloads 237
1502 Software Verification of Systematic Resampling for Optimization of Particle Filters

Authors: Osiris Terry, Kenneth Hopkinson, Laura Humphrey

Abstract:

Systematic resampling is the most popularly used resampling method in particle filters. This paper seeks to further the understanding of systematic resampling by defining a formula made up of variables from the sampling equation and the particle weights. The formula is then verified via SPARK, a software verification language. The verified systematic resampling formula states that the minimum/maximum number of possible samples taken of a particle is equal to the floor/ceiling value of particle weight divided by the sampling interval, respectively. This allows for the creation of a randomness spectrum that each resampling method can fall within. Methods on the lower end, e.g., systematic resampling, have less randomness and, thus, are quicker to reach an estimate. Although lower randomness allows for error by having a larger bias towards the size of the weight, having this bias creates vulnerabilities to the noise in the environment, e.g., jamming. Conclusively, this is the first step in characterizing each resampling method. This will allow target-tracking engineers to pick the best resampling method for their environment instead of choosing the most popularly used one.

Keywords: SPARK, software verification, resampling, systematic resampling, particle filter, tracking

Procedia PDF Downloads 84
1501 The Role of Hausa Oral Praise Singer in Conflict Management and Social Mobilization in Nigeria

Authors: Ladan Surajo

Abstract:

Nigeria as a third world country is full of people who cannot read and write, thereby constituting a stumbling block to the modern way of communication. It is a well known fact that Nigeria is a heterogeneous country with an estimated 450 or more ethnic groups communicating in divergent languages. Despite this scenario, English, Hausa, Igbo and Yoruba languages are predominantly used in the country. Apart from English language, Hausa has a wider coverage of usage among the indigenous languages in Nigeria, thereby using it in the area of social mobilization and conflict management cannot be overemphasized. Hausa Oral Singers are depicting their artistic and God endowed talents through singing to mobilize and sensitize the local communities about government programmes and the ills of other social problems of the society. It is the belief of this researcher that if used properly, the Hausa Oral Singers will assist immensely in reducing to the barest minimum some social ills of the society in Nigeria. More so that music is the food of the heart and has a resounding impact in changing the behaviour of individuals and groups.

Keywords: oral, singers, praise, social mobilization, conflict management

Procedia PDF Downloads 462
1500 Harmonic Data Preparation for Clustering and Classification

Authors: Ali Asheibi

Abstract:

The rapid increase in the size of databases required to store power quality monitoring data has demanded new techniques for analysing and understanding the data. One suggested technique to assist in analysis is data mining. Preparing raw data to be ready for data mining exploration take up most of the effort and time spent in the whole data mining process. Clustering is an important technique in data mining and machine learning in which underlying and meaningful groups of data are discovered. Large amounts of harmonic data have been collected from an actual harmonic monitoring system in a distribution system in Australia for three years. This amount of acquired data makes it difficult to identify operational events that significantly impact the harmonics generated on the system. In this paper, harmonic data preparation processes to better understanding of the data have been presented. Underlying classes in this data has then been identified using clustering technique based on the Minimum Message Length (MML) method. The underlying operational information contained within the clusters can be rapidly visualised by the engineers. The C5.0 algorithm was used for classification and interpretation of the generated clusters.

Keywords: data mining, harmonic data, clustering, classification

Procedia PDF Downloads 247
1499 Maintaining Experimental Consistency in Geomechanical Studies of Methane Hydrate Bearing Soils

Authors: Lior Rake, Shmulik Pinkert

Abstract:

Methane hydrate has been found in significant quantities in soils offshore within continental margins and in permafrost within arctic regions where low temperature and high pressure are present. The mechanical parameters for geotechnical engineering are commonly evaluated in geomechanical laboratories adapted to simulate the environmental conditions of methane hydrate-bearing sediments (MHBS). Due to the complexity and high cost of natural MHBS sampling, most laboratory investigations are conducted on artificially formed samples. MHBS artificial samples can be formed using different hydrate formation methods in the laboratory, where methane gas and water are supplied into the soil pore space under the methane hydrate phase conditions. The most commonly used formation method is the excess gas method which is considered a relatively simple, time-saving, and repeatable testing method. However, there are several differences in the procedures and techniques used to produce the hydrate using the excess gas method. As a result of the difference between the test facilities and the experimental approaches that were carried out in previous studies, different measurement criteria and analyses were proposed for MHBS geomechanics. The lack of uniformity among the various experimental investigations may adversely impact the reliability of integrating different data sets for unified mechanical model development. In this work, we address some fundamental aspects relevant to reliable MHBS geomechanical investigations, such as hydrate homogeneity in the sample, the hydrate formation duration criterion, the hydrate-saturation evaluation method, and the effect of temperature measurement accuracy. Finally, a set of recommendations for repeatable and reliable MHBS formation will be suggested for future standardization of MHBS geomechanical investigation.

Keywords: experimental study, laboratory investigation, excess gas, hydrate formation, standardization, methane hydrate-bearing sediment

Procedia PDF Downloads 58
1498 Reduction in Population Growth under Various Contraceptive Strategies in Uttar Pradesh, India

Authors: Prashant Verma, K. K. Singh, Anjali Singh, Ujjaval Srivastava

Abstract:

Contraceptive policies have been derived to achieve desired reductions in the growth rate and also, applied to the data of Uttar-Pradesh, India for illustration. Using the Lotka’s integral equation for the stable population, expressions for the proportion of contraceptive users at different ages have been obtained. At the age of 20 years, 42% of contraceptive users is imperative to reduce the present annual growth rate of 0.036 to 0.02, assuming that 40% of the contraceptive users discontinue at the age of 25 years and 30% again continue contraceptive use at age 30 years. Further, presuming that 75% of women start using contraceptives at the age of 23 years, and 50% of the remaining women start using contraceptives at the age of 28 years, while the rest of them start using it at the age of 32 years. If we set a minimum age of marriage as 20 years, a reduction of 0.019 in growth rate will be obtained. This study describes how the level of contraceptive use at different age groups of women reduces the growth rate in the state of Uttar Pradesh. The article also promotes delayed marriage in the region.

Keywords: child bearing, contraceptive devices, contraceptive policies, population growth, stable population

Procedia PDF Downloads 255
1497 Growth Performance and Critical Supersaturation of Heterogeneous Condensation for High Concentration of Insoluble Sub-Micron Particles

Authors: Jie Yin, Jun Zhang

Abstract:

Measuring the growth performance and critical supersaturation of particle group have a high reference value for constructing a supersaturated water vapor environment that can improve the removal efficiency of the high-concentration particle group. The critical supersaturation and the variation of the growth performance with supersaturation for high-concentration particles were measured by a flow cloud chamber. Findings suggest that the influence of particle concentration on the growth performance will reduce with the increase of supersaturation. Reducing residence time and increasing particle concentration have similar effects on the growth performance of the high-concentration particle group. Increasing particle concentration and shortening residence time will increase the critical supersaturation of the particle group. The critical supersaturation required to activate a high-concentration particle group is lower than that of the single-particle when the minimum particle size in the particle group is the same as that of a single particle.

Keywords: sub-micron particles, heterogeneous condensation, critical supersaturation, nucleation

Procedia PDF Downloads 157
1496 Study of the Antimicrobial Activity of Aminoreductone against Pathogenic Bacteria in Comparison with Other Antibiotics

Authors: Vu Thu Trang, Lam Xuan Thanh, Samira Sarter, Tomoko Shimamura, Hiroaki Takeuchi  

Abstract:

Antimicrobial activities of aminoreductone (AR), a product formed in the initial stage of Maillard reaction, were screened against pathogenic bacteria. A significant growth inhibition of AR against all 7 isolates (Staphylococcus aureus ATCC® 25923™, Salmonella Typhimurium ATCC® 14028™, Bacillus cereus ATCC® 13061™, Bacillus subtilis ATCC® 11774™, Escherichia coli ATCC® 25922™, Enterococcus faecalis ATCC® 29212™, Listeria innocua ATCC® 33090™) were observed by the standard disc diffusion methods. The inhibition zone for each isolate by AR (2.5 mg) ranged from 15±0 mm to 28.3±0.4 mm in diameter. The minimum inhibitory concentration (MIC) of AR ranging from 20 mM to 26 mM was proven in the seven isolates tested. AR also showed the similar effect of growth inhibition in comparison with antibiotics frequently used for the treatment of infections bacteria, such as amikacin, ciprofloxacin, meropennem, and levofloxacin. The results indicated that foods containing AR are valuable sources of bioactive compounds towards pathogenic bacteria.

Keywords: pathogenic bacteria, aminoreductone, Maillard reaction, antimicrobial activity

Procedia PDF Downloads 384
1495 Efficient Antenna Array Beamforming with Robustness against Random Steering Mismatch

Authors: Ju-Hong Lee, Ching-Wei Liao, Kun-Che Lee

Abstract:

This paper deals with the problem of using antenna sensors for adaptive beamforming in the presence of random steering mismatch. We present an efficient adaptive array beamformer with robustness to deal with the considered problem. The robustness of the proposed beamformer comes from the efficient designation of the steering vector. Using the received array data vector, we construct an appropriate correlation matrix associated with the received array data vector and a correlation matrix associated with signal sources. Then, the eigenvector associated with the largest eigenvalue of the constructed signal correlation matrix is designated as an appropriate estimate of the steering vector. Finally, the adaptive weight vector required for adaptive beamforming is obtained by using the estimated steering vector and the constructed correlation matrix of the array data vector. Simulation results confirm the effectiveness of the proposed method.

Keywords: adaptive beamforming, antenna array, linearly constrained minimum variance, robustness, steering vector

Procedia PDF Downloads 199
1494 Computational Fluid Dynamics and Experimental Evaluation of Two Batch Type Electrocoagulation Stirred Tank Reactors Used in the Removal of Cr (VI) from Waste Water

Authors: Phanindra Prasad Thummala, Umran Tezcan Un

Abstract:

In this study, hydrodynamics analysis of two batch type electrocoagulation stirred tank reactors, used for the electrocoagulation treatment of Cr(VI) wastewater, was carried using computational fluid dynamics (CFD). The aim of the study was to evaluate the impact of mixing characteristics on overall performance of electrocoagulation reactor. The CFD simulations were performed using ANSYS FLUENT 14.4 software. The mixing performance of each reactor was evaluated by numerically modelling tracer dispersion in each reactor configuration. The uniformity in tracer dispersion was assumed when 90% of the ratio of the maximum to minimum concentration of the tracer was realized. In parallel, experimental evaluation of both the electrocoagulation reactors for removal of Cr(VI) from wastewater was also carried out. The results of CFD and experimental analysis clearly show that the reactor which can give higher uniformity in lesser time, will perform better as an electrocoagulation reactor for removal of Cr(VI) from wastewater.

Keywords: CFD, stirred tank reactors, electrocoagulation, Cr(VI) wastewater

Procedia PDF Downloads 461
1493 Experimental Studies on Prestressed Precast Concrete Bridge Piers

Authors: C. Shim, C. Koem, S. Park, S. Lee

Abstract:

This paper deals with experimental studies on pre stressed precast concrete columns with continuous reinforcing bars and pre stressing tendons. Design requirements on minimum transverse reinforcement ratio are not included in current design codes. Pre stressing introduces additional compression to the column. Precast columns with different transverse reinforcement ratios were tested to derive adequate design requirement. Displacement ductility of the pre stressed precast columns was evaluated and compared with previous studies. Design of axial steels including reinforcing bars and pre stressing tendons influenced on the seismic performance. Without significant increase of transverse reinforcement ratio, the specimens showed required displacement ductility without reduction of their flexural strength. Design recommendations for precast bridge piers were derived.

Keywords: displacement ductility, flexural strength, prestressed precast column, transverse reinforcement

Procedia PDF Downloads 278
1492 Industrial Ecology Perspectives of Food Supply Chains: A Framework of Analysis

Authors: Luciano Batista, Sylvia Saes, Nuno Fouto, Liam Fassam

Abstract:

This paper introduces the theoretical and methodological basis of an analytical framework conceived with the purpose of bringing industrial ecology perspectives into the core of the underlying disciplines supporting analyses in studies concerned with environmental sustainability aspects beyond the product cycle in a supply chain. Given the pressing challenges faced by the food sector, the framework focuses upon waste minimization through industrial linkages in food supply chains. The combination of industrial ecology practice with basic LCA elements, the waste hierarchy model, and the spatial scale of industrial symbiosis allows the standardization of qualitative analyses and associated outcomes. Such standardization enables comparative analysis not only between different stages of a supply chain, but also between different supply chains. The analytical approach proposed contributes more coherently to the wider circular economy aspiration of optimizing the flow of goods to get the most out of raw materials and cuts wastes to a minimum.

Keywords: by-product synergy, food supply chain, industrial ecology, industrial symbiosis

Procedia PDF Downloads 420
1491 Surface Sterilization Of Aquatic Plant, Cryptocoryne affinis by Using Clorox and Mercury Chloride

Authors: Sridevi Devadas

Abstract:

This study was aimed to examine the combination efficiency of Clorox (5.25% Sodium Hypochlorite) and mercury chloride (HgCl2) as a reagent for surface sterilization process of aquatic plant and cryptocoryne affinis (C. affinis). The treatment applied 10% of the Clorox and 0.1ppm of mercury chloride. The maximum exposure time for clorox and mercury chloride was 10min and 60sec respectively. After exposed to the treatments protocols (T1-T15) the explants were transferred to culture room under control temperature at 25°C ± 2°C and subjected to 16 hours fluorescence light (2000 lumens) for 30 days. The both sterilizing agents were not applied on control specimens. Upon analysis, The result indicates all of the treatments protocols produced sterile explants at range of minimum 1.5 ± 0.7 (30%) to maximum 5.0 ± 0.0 (100%). Meanwhile, maximum 1.0 ± 0.7 numbers of leaves and 1.4 ± 0.6 numbers of roots have been produced. The optimized exposure time was 0 to 15 min for Clorox and 30 sec for HgCl2 whereby 90% to 100% sterilization was archived at this condition.

Keywords: Cryptocoryne affinis, surface sterilization, tissue culture, clorox, mercury chloride

Procedia PDF Downloads 380
1490 Comparison of Allowable Stress Method and Time History Response Analysis for Seismic Design of Buildings

Authors: Sayuri Inoue, Naohiro Nakamura, Tsubasa Hamada

Abstract:

The seismic design method of buildings is classified into two types: static design and dynamic design. The static design is a design method that exerts static force as seismic force and is a relatively simple design method created based on the experience of seismic motion in the past 100 years. At present, static design is used for most of the Japanese buildings. Dynamic design mainly refers to the time history response analysis. It is a comparatively difficult design method that input the earthquake motion assumed in the building model and examine the response. Currently, it is only used for skyscrapers and specific buildings. In the present design standard in Japan, it is good to use either the design method of the static design and the dynamic design in the medium and high-rise buildings. However, when actually designing middle and high-rise buildings by two kinds of design methods, the relatively simple static design method satisfies the criteria, but in the case of a little difficult dynamic design method, the criterion isn't often satisfied. This is because the dynamic design method was built with the intention of designing super high-rise buildings. In short, higher safety is required as compared with general buildings, and criteria become stricter. The authors consider applying the dynamic design method to general buildings designed by the static design method so far. The reason is that application of the dynamic design method is reasonable for buildings that are out of the conventional standard structural form such as emphasizing design. For the purpose, it is important to compare the design results when the criteria of both design methods are arranged side by side. In this study, we performed time history response analysis to medium-rise buildings that were actually designed with allowable stress method. Quantitative comparison between static design and dynamic design was conducted, and characteristics of both design methods were examined.

Keywords: buildings, seismic design, allowable stress design, time history response analysis, Japanese seismic code

Procedia PDF Downloads 155
1489 An Exponential Field Path Planning Method for Mobile Robots Integrated with Visual Perception

Authors: Magdy Roman, Mostafa Shoeib, Mostafa Rostom

Abstract:

Global vision, whether provided by overhead fixed cameras, on-board aerial vehicle cameras, or satellite images can always provide detailed information on the environment around mobile robots. In this paper, an intelligent vision-based method of path planning and obstacle avoidance for mobile robots is presented. The method integrates visual perception with a new proposed field-based path-planning method to overcome common path-planning problems such as local minima, unreachable destination and unnecessary lengthy paths around obstacles. The method proposes an exponential angle deviation field around each obstacle that affects the orientation of a close robot. As the robot directs toward, the goal point obstacles are classified into right and left groups, and a deviation angle is exponentially added or subtracted to the orientation of the robot. Exponential field parameters are chosen based on Lyapunov stability criterion to guarantee robot convergence to the destination. The proposed method uses obstacles' shape and location, extracted from global vision system, through a collision prediction mechanism to decide whether to activate or deactivate obstacles field. In addition, a search mechanism is developed in case of robot or goal point is trapped among obstacles to find suitable exit or entrance. The proposed algorithm is validated both in simulation and through experiments. The algorithm shows effectiveness in obstacles' avoidance and destination convergence, overcoming common path planning problems found in classical methods.

Keywords: path planning, collision avoidance, convergence, computer vision, mobile robots

Procedia PDF Downloads 194
1488 Storage System Validation Study for Raw Cocoa Beans Using Minitab® 17 and R (R-3.3.1)

Authors: Anthony Oppong Kyekyeku, Sussana Antwi-Boasiako, Emmanuel De-Graft Johnson Owusu Ansah

Abstract:

In this observational study, the performance of a known conventional storage system was tested and evaluated for fitness for its intended purpose. The system has a scope extended for the storage of dry cocoa beans. System sensitivity, reproducibility and uncertainties are not known in details. This study discusses the system performance in the context of existing literature on factors that influence the quality of cocoa beans during storage. Controlled conditions were defined precisely for the system to give reliable base line within specific established procedures. Minitab® 17 and R statistical software (R-3.3.1) were used for the statistical analyses. The approach to the storage system testing was to observe and compare through laboratory test methods the quality of the cocoa beans samples before and after storage. The samples were kept in Kilner jars and the temperature of the storage environment controlled and monitored over a period of 408 days. Standard test methods use in international trade of cocoa such as the cut test analysis, moisture determination with Aqua boy KAM III model and bean count determination were used for quality assessment. The data analysis assumed the entire population as a sample in order to establish a reliable baseline to the data collected. The study concluded a statistically significant mean value at 95% Confidence Interval (CI) for the performance data analysed before and after storage for all variables observed. Correlational graphs showed a strong positive correlation for all variables investigated with the exception of All Other Defect (AOD). The weak relationship between the before and after data for AOD had an explained variability of 51.8% with the unexplained variability attributable to the uncontrolled condition of hidden infestation before storage. The current study concluded with a high-performance criterion for the storage system.

Keywords: benchmarking performance data, cocoa beans, hidden infestation, storage system validation

Procedia PDF Downloads 174
1487 Optimization of Flip Bucket Dents in Order to Reduce Scour Hole Depth (Plunge Pool) Using a Comprehensive Physical Model

Authors: Majid Galoie, Khodadad Safavi, Abdolreza Karami Nejad, Reza Roshan

Abstract:

Scour downstream of a flip bucket in a plunge pool is caused by impingement of water jet force. In order to reduce this force and consequently reduce scour hole depth, flip buckets may equip by dents. The minimum scour hole depth might be occurred by optimization of dents (number, shape, placement) on flip buckets. In this study, a comprehensive physical model has been developed and various options for dents have been investigated. The experimental data for each dent option such as scour hole depth, angle of impingement jet, piezometric pressure in tail-water and jet trajectory have been measured for various discharges. Finally, the best option can be found by analysis of the experimental results which has been expressed in this paper.

Keywords: scouring process, plunge pool, scour hole depth, physical model, flip bucket

Procedia PDF Downloads 394
1486 Evaluation and Analysis of Light Emitting Diode Distribution in an Indoor Visible Light Communication

Authors: Olawale J. Olaluyi, Ayodele S. Oluwole, O. Akinsanmi, Johnson O. Adeogo

Abstract:

Communication using visible light VLC is considered a cutting-edge technology used for data transmission and illumination since it uses less energy than radio frequency (RF) technology and has a large bandwidth, extended lifespan, and high security. The room's irregular distribution of small base stations, or LED array distribution, is the cause of the obscured area, minimum signal-to-noise ratio (SNR), and received power. In order to maximize the received power distribution and SNR at the center of the room for an indoor VLC system, the researchers offer an innovative model for the placement of eight LED array distributions in this work. We have investigated the arrangement of the LED array distribution with regard to receiving power to fill the open space in the center of the room. The suggested LED array distribution saved 36.2% of the transmitted power, according to the simulation findings. Aside from that, the entire room was equally covered. This leads to an increase in both received power and SNR.

Keywords: visible light communication (VLC), light emitted diodes (LED), optical power distribution, signal-to-noise ratio (SNR).

Procedia PDF Downloads 88
1485 Prediction of Vapor Liquid Equilibrium for Dilute Solutions of Components in Ionic Liquid by Neural Networks

Authors: S. Mousavian, A. Abedianpour, A. Khanmohammadi, S. Hematian, Gh. Eidi Veisi

Abstract:

Ionic liquids are finding a wide range of applications from reaction media to separations and materials processing. In these applications, Vapor–Liquid equilibrium (VLE) is the most important one. VLE for six systems at 353 K and activity coefficients at infinite dilution 〖(γ〗_i^∞) for various solutes (alkanes, alkenes, cycloalkanes, cycloalkenes, aromatics, alcohols, ketones, esters, ethers, and water) in the ionic liquids (1-ethyl-3-methylimidazolium bis (trifluoromethylsulfonyl)imide [EMIM][BTI], 1-hexyl-3-methyl imidazolium bis (trifluoromethylsulfonyl) imide [HMIM][BTI], 1-octyl-3-methylimidazolium bis(trifluoromethylsulfonyl) imide [OMIM][BTI], and 1-butyl-1-methylpyrrolidinium bis (trifluoromethylsulfonyl) imide [BMPYR][BTI]) have been used to train neural networks in the temperature range from (303 to 333) K. Densities of the ionic liquids, Hildebrant constant of substances, and temperature were selected as input of neural networks. The networks with different hidden layers were examined. Networks with seven neurons in one hidden layer have minimum error and good agreement with experimental data.

Keywords: ionic liquid, neural networks, VLE, dilute solution

Procedia PDF Downloads 300
1484 Efficient Reduction of Organophosphate Pesticide from Fruits and Vegetables Using Cost Effective Neutralizer

Authors: Debjani Dasgupta, Aman Zalawadia, Anuj Thapa, Pranjali Sing, Ashish Dabade

Abstract:

Organophosphate group pesticides are common pesticide group, which gain entry into food product due to incomplete removal of pesticide residues. The current food industry raw material handling process is not sufficient to eliminate pesticide residues. A neutralizer was used to neutralize the residues of pesticide on Vitis vinifera (Grapes). The water based dilution of neutralizer was demonstrated on fruits like grapes. Analysis for pesticides in water wash and neutralizer wash was carried out using GCMS. Fruits washed with neutralizer exhibited 72.95% removal of pesticides compared with normal water wash method. An economical chemical neutralizer can be used to remove such residues in raw material handling at industrial scale with minor modification in process to achieve minimum pesticide entry into final food products.

Keywords: GCMS, organophosphate, raw material handling, Vitis vinifera, pesticide neutralizer

Procedia PDF Downloads 273
1483 Multiple Images Stitching Based on Gradually Changing Matrix

Authors: Shangdong Zhu, Yunzhou Zhang, Jie Zhang, Hang Hu, Yazhou Zhang

Abstract:

Image stitching is a very important branch in the field of computer vision, especially for panoramic map. In order to eliminate shape distortion, a novel stitching method is proposed based on gradually changing matrix when images are horizontal. For images captured horizontally, this paper assumes that there is only translational operation in image stitching. By analyzing each parameter of the homography matrix, the global homography matrix is gradually transferred to translation matrix so as to eliminate the effects of scaling, rotation, etc. in the image transformation. This paper adopts matrix approximation to get the minimum value of the energy function so that the shape distortion at those regions corresponding to the homography can be minimized. The proposed method can avoid multiple horizontal images stitching failure caused by accumulated shape distortion. At the same time, it can be combined with As-Projective-As-Possible algorithm to ensure precise alignment of overlapping area.

Keywords: image stitching, gradually changing matrix, horizontal direction, matrix approximation, homography matrix

Procedia PDF Downloads 319
1482 Natural Ventilation around and through Building: A Numerical Study

Authors: A. Kaddour, S. M. A. Bekkouche

Abstract:

Limiting heat losses during ventilation of indoor building spaces has become a basic aim for architects. Much experience has been gained in terms of ventilation of indoor spaces. Nevertheless, due to the complex applications, attempts to create a theoretical base for solving the problems related to the issue are limited, especially determining the minimum ventilation period required within a designated space. In this paper we have approached this matter, both theoretically and computationally. The conclusion we reached was that controlled ventilation of spaces through vent holes that successively open and close at regular time intervals can limit the excessive circulation of air masses, which in turn limits heat losses. Air change rates through open and tilted windows in rooms of residential buildings driven by atmospheric motions are investigated to evaluate natural ventilation concepts. Model of thermal building simulations is used. A separated sample storey and a sample single room in larger scales were used to measure air transport through window openings under the influence of the external pressure distribution.

Keywords: natural ventilation, temperature factor, air change rates, air circulation

Procedia PDF Downloads 442
1481 Analysis of the Impact of Refractivity on Ultra High Frequency Signal Strength over Gusau, North West, Nigeria

Authors: B. G. Ayantunji, B. Musa, H. Mai-Unguwa, L. A. Sunmonu, A. S. Adewumi, L. Sa'ad, A. Kado

Abstract:

For achieving reliable and efficient communication system, both terrestrial and satellite communication, surface refractivity is critical in planning and design of radio links. This study analyzed the impact of atmospheric parameters on Ultra High Frequency (UHF) signal strength over Gusau, North West, Nigeria. The analysis exploited meteorological data measured simultaneously with UHF signal strength for the month of June 2017 using a Davis Vantage Pro2 automatic weather station and UHF signal strength measuring devices respectively. The instruments were situated at the premise of Federal University, Gusau (6° 78' N, 12° 13' E). The refractivity values were computed using ITU-R model. The result shows that the refractivity value attained the highest value of 366.28 at 2200hr and a minimum value of 350.66 at 2100hr local time. The correlation between signal strength and refractivity is 0.350; Humidity is 0.532 and a negative correlation of -0.515 for temperature.

Keywords: refractivity, UHF (ultra high frequency) signal strength, free space, automatic weather station

Procedia PDF Downloads 197
1480 The Influence of Green Supply Chain Management Practices' Implementation on Organizational Performance: An Empirical Case Study in Spain

Authors: Keivan Amirbagheri, Ana Nuñez-Carballosa, Laura Guitart-Tarrés

Abstract:

Over the last couple of decades, enterprises have begun to accept the need for environmental management and have started to implement environmental management programs to compete in the markets. The implementation of green supply chain management (GSCM) practices can provide valuable opportunities to improve firm performance. Through the prior investigations, the ascending tendency of the numbers of published papers in the field of green supply chain management practices has been reported and it shows the high interest level of the authors to work in this area. Besides, there is still a gap to study more about the relationship of GSCM to the organizational performance (OP). So, the purpose of this research is to study the practices related to green supply chain management that influence the results of the company as an organizational performance. Based on our previous works, from one part we have collected these GSCM practices (planning, operational, and communication practices) and classified them through conducting some literature reviews to analyze their effects on the OP’s factors (balanced scorecard’s perspectives). To do so we design a case study methodology through semi-structured interviews and secondary data from some multinational well-known companies based in Spain. The cases have been selected with the criterion of trying to collect members of the entire supply chain to have a vision as global as possible. The results report the considerable influence of green supply chain management practices on the organizational performance of the companies of the study. In addition, they represent that the implementation of green supply chain management practices especially in a long-term perspective can be economically justified. From the point of view of the personal, they feel better about being a member of this type of company that has been structured on environmental issues. Also, for these companies, the image that has been created by the implementation of these practices helps them to facilitate their marketing program.

Keywords: green supply chain management, organizational performance, case study, Spain

Procedia PDF Downloads 190
1479 Execution of Optimization Algorithm in Cascaded H-Bridge Multilevel Inverter

Authors: M. Suresh Kumar, K. Ramani

Abstract:

This paper proposed the harmonic elimination of Cascaded H-Bridge Multi-Level Inverter by using Selective Harmonic Elimination-Pulse Width Modulation method programmed with Particle Swarm Optimization algorithm. PSO method determine proficiently the required switching angles to eliminate low order harmonics up to the 11th order from the inverter output voltage waveform while keeping the magnitude of the fundamental harmonics at the desired value. Results demonstrate that the proposed method does efficiently eliminate a great number of specific harmonics and the output voltage is resulted in minimum Total Harmonic Distortion. The results shown that the PSO algorithm attain successfully to the global solution faster than other algorithms.

Keywords: multi-level inverter, Selective Harmonic Elimination Pulse Width Modulation (SHEPWM), Particle Swarm Optimization (PSO), Total Harmonic Distortion (THD)

Procedia PDF Downloads 603
1478 Experimental Investigation and Optimization of Nanoparticle Mass Concentration and Heat Input of Loop Heat Pipe

Authors: P. Gunnasegaran, M. Z. Abdullah, M. Z. Yusoff, Nur Irmawati

Abstract:

This study presents experimental and optimization of nanoparticle mass concentration and heat input based on the total thermal resistance (Rth) of loop heat pipe (LHP), employed for PC-CPU cooling. In this study, silica nanoparticles (SiO2) in water with particle mass concentration ranged from 0% (pure water) to 1% is considered as the working fluid within the LHP. The experimental design and optimization is accomplished by the design of the experimental tool, Response Surface Methodology (RSM). The results show that the nanoparticle mass concentration and the heat input have a significant effect on the Rth of LHP. For a given heat input, the Rth is found to decrease with the increase of the nanoparticle mass concentration up to 0.5% and increased thereafter. It is also found that the Rth is decreased when the heat input is increased from 20W to 60W. The results are optimized with the objective of minimizing the Rt, using Design-Expert software, and the optimized nanoparticle mass concentration and heat input are 0.48% and 59.97W, respectively, the minimum thermal resistance being 2.66(ºC/W).

Keywords: loop heat pipe, nanofluid, optimization, thermal resistance

Procedia PDF Downloads 460
1477 Numerical Modeling of Air Pollution with PM-Particles and Dust

Authors: N. Gigauri, A. Surmava, L. Intskirveli, V. Kukhalashvili, S. Mdivani

Abstract:

The subject of our study is atmospheric air pollution with numerical modeling. In the presented article, as the object of research, there is chosen city Tbilisi, the capital of Georgia, with a population of one and a half million and a difficult terrain. The main source of pollution in Tbilisi is currently vehicles and construction dust. The concentrations of dust and PM (Particulate Matter) were determined in the air of Tbilisi and in its vicinity. There are estimated their monthly maximum, minimum, and average concentrations. Processes of dust propagation in the atmosphere of the city and its surrounding territory are modelled using a 3D regional model of atmospheric processes and an admixture transfer-diffusion equation. There were taken figures of distribution of the polluted cloud and dust concentrations in different areas of the city at different heights and at different time intervals with the background stationary westward and eastward wind. It is accepted that the difficult terrain and mountain-bar circulation affect the deformation of the cloud and its spread, there are determined time periods when the dust concentration in the city is greater than MAC (Maximum Allowable Concentration, MAC=0.5 mg/m³).

Keywords: air pollution, dust, numerical modeling, PM-particles

Procedia PDF Downloads 140
1476 A Deletion-Cost Based Fast Compression Algorithm for Linear Vector Data

Authors: Qiuxiao Chen, Yan Hou, Ning Wu

Abstract:

As there are deficiencies of the classic Douglas-Peucker Algorithm (DPA), such as high risks of deleting key nodes by mistake, high complexity, time consumption and relatively slow execution speed, a new Deletion-Cost Based Compression Algorithm (DCA) for linear vector data was proposed. For each curve — the basic element of linear vector data, all the deletion costs of its middle nodes were calculated, and the minimum deletion cost was compared with the pre-defined threshold. If the former was greater than or equal to the latter, all remaining nodes were reserved and the curve’s compression process was finished. Otherwise, the node with the minimal deletion cost was deleted, its two neighbors' deletion costs were updated, and the same loop on the compressed curve was repeated till the termination. By several comparative experiments using different types of linear vector data, the comparison between DPA and DCA was performed from the aspects of compression quality and computing efficiency. Experiment results showed that DCA outperformed DPA in compression accuracy and execution efficiency as well.

Keywords: Douglas-Peucker algorithm, linear vector data, compression, deletion cost

Procedia PDF Downloads 251
1475 Spectrum Assignment Algorithms in Optical Networks with Protection

Authors: Qusay Alghazali, Tibor Cinkler, Abdulhalim Fayad

Abstract:

In modern optical networks, the flex grid spectrum usage is most widespread, where higher bit rate streams get larger spectrum slices while lower bit rate traffic streams get smaller spectrum slices. To our practice, under the ITU-T recommendation, G.694.1, spectrum slices of 50, 75, and 100 GHz are being used with central frequency at 193.1 THz. However, when these spectrum slices are not sufficient, multiple spectrum slices can use either one next to another or anywhere in the optical wavelength. In this paper, we propose the analysis of the wavelength assignment problem. We compare different algorithms for this spectrum assignment with and without protection. As a reference for comparisons, we concluded that the Integer Linear Programming (ILP) provides the global optimum for all cases. The most scalable algorithm is the greedy one, which yields results in subsequent ranges even for more significant network instances. The algorithms’ benchmark implemented using the LEMON C++ optimization library and simulation runs based on a minimum number of spectrum slices assigned to lightpaths and their execution time.

Keywords: spectrum assignment, integer linear programming, greedy algorithm, international telecommunication union, library for efficient modeling and optimization in networks

Procedia PDF Downloads 169
1474 Evaluating the Understanding of the University Students (Basic Sciences and Engineering) about the Numerical Representation of the Average Rate of Change

Authors: Saeid Haghjoo, Ebrahim Reyhani, Fahimeh Kolahdouz

Abstract:

The present study aimed to evaluate the understanding of the students in Tehran universities (Iran) about the numerical representation of the average rate of change based on the Structure of Observed Learning Outcomes (SOLO). In the present descriptive-survey research, the statistical population included undergraduate students (basic sciences and engineering) in the universities of Tehran. The samples were 604 students selected by random multi-stage clustering. The measurement tool was a task whose face and content validity was confirmed by math and mathematics education professors. Using Cronbach's Alpha criterion, the reliability coefficient of the task was obtained 0.95, which verified its reliability. The collected data were analyzed by descriptive statistics and inferential statistics (chi-squared and independent t-tests) under SPSS-24 software. According to the SOLO model in the prestructural, unistructural, and multistructural levels, basic science students had a higher percentage of understanding than that of engineering students, although the outcome was inverse at the relational level. However, there was no significant difference in the average understanding of both groups. The results indicated that students failed to have a proper understanding of the numerical representation of the average rate of change, in addition to missconceptions when using physics formulas in solving the problem. In addition, multiple solutions were derived along with their dominant methods during the qualitative analysis. The current research proposed to focus on the context problems with approximate calculations and numerical representation, using software and connection common relations between math and physics in the teaching process of teachers and professors.

Keywords: average rate of change, context problems, derivative, numerical representation, SOLO taxonomy

Procedia PDF Downloads 92