Search results for: Loss Reduction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2373

Search results for: Loss Reduction

393 A PIM (Processor-In-Memory) for Computer Graphics : Data Partitioning and Placement Schemes

Authors: Jae Chul Cha, Sandeep K. Gupta

Abstract:

The demand for higher performance graphics continues to grow because of the incessant desire towards realism. And, rapid advances in fabrication technology have enabled us to build several processor cores on a single die. Hence, it is important to develop single chip parallel architectures for such data-intensive applications. In this paper, we propose an efficient PIM architectures tailored for computer graphics which requires a large number of memory accesses. We then address the two important tasks necessary for maximally exploiting the parallelism provided by the architecture, namely, partitioning and placement of graphic data, which affect respectively load balances and communication costs. Under the constraints of uniform partitioning, we develop approaches for optimal partitioning and placement, which significantly reduce search space. We also present heuristics for identifying near-optimal placement, since the search space for placement is impractically large despite our optimization. We then demonstrate the effectiveness of our partitioning and placement approaches via analysis of example scenes; simulation results show considerable search space reductions, and our heuristics for placement performs close to optimal – the average ratio of communication overheads between our heuristics and the optimal was 1.05. Our uniform partitioning showed average load-balance ratio of 1.47 for geometry processing and 1.44 for rasterization, which is reasonable.

Keywords: Data Partitioning and Placement, Graphics, PIM, Search Space Reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492
392 Logistics Outsourcing: Performance Models and Financial and Operational Indicators

Authors: Carlos Sanchís-Pedregosa, José A. D. M achuca, María del Mar González-Zamora

Abstract:

The growing outsourcing of logistics services resulting from the ongoing current in firms of costs reduction/increased efficiency means that it is becoming more and more important for the companies doing the outsourcing to carry out a proper evaluation. The multiple definitions and measures of logistics service performance found in research on the topic create a certain degree of confusion and do not clear the way towards the proper measurement of their performance. Do a model and a specific set of indicators exist that can be considered appropriate for measuring the performance of logistics services outsourcing in industrial environments? Are said indicators in keeping with the objectives pursued by outsourcing? We aim to answer these and other research questions in the study we have initiated in the field within the framework of the international High Performance Manufacturing (HPM) project of which this paper forms part. As the first stage of this research, this paper reviews articles dealing with the topic published in the last 15 years with the aim of detecting the models most used to make this measurement and determining which performance indicators are proposed as part of said models and which are most used. The first steps are also taken in determining whether these indicators, financial and operational, cover the aims that are being pursued when outsourcing logistics services. The findings show there is a wide variety of both models and indicators used. This would seem to testify to the need to continue with our research in order to try to propose a model and a set of indicators for measuring the performance of logistics services outsourcing in industrial environments.

Keywords: Logistics, objectives, outsourcing, performancemeasurement systems

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2134
391 Feasibility Study of MongoDB and Radio Frequency Identification Technology in Asset Tracking System

Authors: Mohd Noah A. Rahman, Afzaal H. Seyal, Sharul T. Tajuddin, Hartiny Md Azmi

Abstract:

Taking into consideration the real time situation specifically the higher academic institutions, small, medium to large companies, public to private sectors and the remaining sectors, do experience the inventory or asset shrinkages due to theft, loss or even inventory tracking errors. This happening is due to a zero or poor security systems and measures being taken and implemented in their organizations. Henceforth, implementing the Radio Frequency Identification (RFID) technology into any manual or existing web-based system or web application can simply deter and will eventually solve certain major issues to serve better data retrieval and data access. Having said, this manual or existing system can be enhanced into a mobile-based system or application. In addition to that, the availability of internet connections can aid better services of the system. Such involvement of various technologies resulting various privileges to individuals or organizations in terms of accessibility, availability, mobility, efficiency, effectiveness, real-time information and also security. This paper will look deeper into the integration of mobile devices with RFID technologies with the purpose of asset tracking and control. Next, it is to be followed by the development and utilization of MongoDB as the main database to store data and its association with RFID technology. Finally, the development of a web based system which can be viewed in a mobile based formation with the aid of Hypertext Preprocessor (PHP), MongoDB, Hyper-Text Markup Language 5 (HTML5), Android, JavaScript and AJAX programming language.

Keywords: RFID, asset tracking system, MongoDB, NoSQL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1648
390 Applying the Extreme-Based Teaching Model in Post-Secondary Online Classroom Setting: A Field Experiment

Authors: Leon Pan

Abstract:

The first programming course within post-secondary education has long been recognized as a challenging endeavor for both educators and students alike. Historically, these courses have exhibited high failure rates and a notable number of dropouts. Instructors often lament students' lack of effort on their coursework, and students often express frustration that the teaching methods employed are not effective. Drawing inspiration from the successful principles of Extreme Programming, this study introduces an approach—the Extremes-based teaching model—aimed at enhancing the teaching of introductory programming courses. To empirically determine the effectiveness of the model, a comparison was made between a section taught using the extreme-based model and another utilizing traditional teaching methods. Notably, the extreme-based teaching class required students to work collaboratively on projects, while also demanding continuous assessment and performance enhancement within groups. This paper details the application of the extreme-based model within the post-secondary online classroom context and presents the compelling results that emphasize its effectiveness in advancing the teaching and learning experiences. The extreme-based model led to a significant increase of 13.46 points in the weighted total average and a commendable 10% reduction in the failure rate.

Keywords: Extreme-based teaching model, innovative pedagogical methods, project-based learning, team-based learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 128
389 The Investigation of Green Roof and White Roof Cooling Potential on Single Storey Residential Building in the Malaysian Climate

Authors: Asmat Ismail, Muna Hanim Abdul Samad, Abdul Malek Abdul Rahman

Abstract:

The phenomenon of global warming or climate change has led to many environmental issues including higher atmospheric temperatures, intense precipitation, increased greenhouse gaseous emissions and increased indoor discomfort. Studies have shown that bringing nature to the roof such as constructing green roof and implementing high-reflective roof may give positive impact in mitigating the effects of global warming and in increasing thermal comfort sensation inside buildings. However, no study has been conducted to compare both types of passive roof treatments in Malaysia in order to increase thermal comfort in buildings. Therefore, this study is conducted to investigate the effect of green roof and white painted roof as passive roof treatment in improving indoor comfort of Malaysian homes. This study uses an experimental approach in which the measurements of temperatures are conducted on the case study building. The measurements of outdoor and indoor environments were conducted on the flat roof with two different types of roof treatment that are green roof and white roof. The measurement of existing black bare roof was also conducted to act as a control for this study.

Keywords: global warming, green roof, white painted roof, indoor temperature reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2739
388 Grid Coordination with Marketmaker Agents

Authors: Xin Bai, Kresimir Sivoncik, Damla Turgut, Ladislau Bölöni

Abstract:

Market based models are frequently used in the resource allocation on the computational grid. However, as the size of the grid grows, it becomes difficult for the customer to negotiate directly with all the providers. Middle agents are introduced to mediate between the providers and customers and facilitate the resource allocation process. The most frequently deployed middle agents are the matchmakers and the brokers. The matchmaking agent finds possible candidate providers who can satisfy the requirements of the consumers, after which the customer directly negotiates with the candidates. The broker agents are mediating the negotiation with the providers in real time. In this paper we present a new type of middle agent, the marketmaker. Its operation is based on two parallel operations - through the investment process the marketmaker is acquiring resources and resource reservations in large quantities, while through the resale process it sells them to the customers. The operation of the marketmaker is based on the fact that through its global view of the grid it can perform a more efficient resource allocation than the one possible in one-to-one negotiations between the customers and providers. We present the operation and algorithms governing the operation of the marketmaker agent, contrasting it with the matchmaker and broker agents. Through a series of simulations in the task oriented domain we compare the operation of the three agents types. We find that the use of marketmaker agent leads to a better performance in the allocation of large tasks and a significant reduction of the messaging overhead.

Keywords: grid computing, autonomous agents, market-basedgrid

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1529
387 Antimicrobial, Antiplasmid and Cytotoxicity Potentials of Marine Algae Halimeda opuntia and Sarconema filiforme Collected from Red Sea Coast

Authors: Samy A. Selim

Abstract:

The antimicrobial, antiplasmid and cytotoxic activities of marine algae Halimeda opuntia and Sarconema filiforme were investigated. Antimicrobial bioassay against some human pathogenic bacteria and yeast were conducted using disc diffusion method. Halimeda extract exhibited antibacterial activity against six species of microrganisms, with significant inhibition against Staphylococcus aureus. While Sarconema extract was better potent as antifungal against Candida albicans. Comparative antibacterial studies showed that Halimeda extract showed equivalent or better activity as compared with commercial antibiotic when tested against Staphylococcus aureus. Further tests conducted using dilution method showed both extracts as having bacteriostatic mode of action against the tested microorganisms. Methanol extract of two species showed significant cytotoxicity (LC50 <500μg) on brine shrimp. Halimeda opuntia showed highest cytotoxic activity (LC50 =192.3μg). Also, the present investigation was undertaken to investigate the ability of methanolic extract of the algal extracts to cure R-plasmids from certain clinical E. coli isolates. The active fraction of Halimeda and Sarconema could cure plasmids from E. coli at curing efficiencies of approximately 78%. The active fraction mediated plasmid curing resulted in the subsequent loss of antibiotic resistance encoded in the plasmids as revealed by antibiotic resistance profile of cured strains. The screening results confirm the possible use of marine algae Halimeda opuntia and Sarconema filiforme as a source of pharmacological benefits.

Keywords: Antimicrobial, antiplasmid Cytotoxicity, Marine Algae.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3079
386 Numerical Simulation of Heat Exchanger Area of R410A-R23 and R404A-R508B Cascade Refrigeration System at Various Evaporating and Condensing Temperature

Authors: A. D. Parekh, P. R. Tailor

Abstract:

Capacity and efficiency of any refrigerating system diminish rapidly as the difference between the evaporating and condensing temperature is increased by reduction in the evaporator temperature. The single stage vapour compression refrigeration system is limited to an evaporator temperature of -40 0C. Below temperature of -40 0C the either cascade refrigeration system or multi stage vapour compression system is employed. Present work describes thermal design of main three heat exchangers namely condenser (HTS), cascade condenser and evaporator (LTS) of R404A-R508B and R410A-R23 cascade refrigeration system. Heat transfer area of condenser (HTS), cascade condenser and evaporator (LTS) for both systems have been compared and the effect of condensing and evaporating temperature on heat-transfer area for both systems have been studied under same operating condition. The results shows that the required heat-transfer area of condenser and cascade condenser for R410A-R23 cascade system is lower than the R404A-R508B cascade system but heat transfer area of evaporator is similar for both the system. The heat transfer area of condenser and cascade condenser decreases with increase in condensing temperature (Tc), whereas the heat transfer area of cascade condenser and evaporator increases with increase in evaporating temperature (Te).

Keywords: Heat-transfer area, R410A, R404A, R508B, R23, Refrigeration system, Thermal design

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2437
385 A CFD Study of Turbulent Convective Heat Transfer Enhancement in Circular Pipeflow

Authors: Perumal Kumar, Rajamohan Ganesan

Abstract:

Addition of milli or micro sized particles to the heat transfer fluid is one of the many techniques employed for improving heat transfer rate. Though this looks simple, this method has practical problems such as high pressure loss, clogging and erosion of the material of construction. These problems can be overcome by using nanofluids, which is a dispersion of nanosized particles in a base fluid. Nanoparticles increase the thermal conductivity of the base fluid manifold which in turn increases the heat transfer rate. Nanoparticles also increase the viscosity of the basefluid resulting in higher pressure drop for the nanofluid compared to the base fluid. So it is imperative that the Reynolds number (Re) and the volume fraction have to be optimum for better thermal hydraulic effectiveness. In this work, the heat transfer enhancement using aluminium oxide nanofluid using low and high volume fraction nanofluids in turbulent pipe flow with constant wall temperature has been studied by computational fluid dynamic modeling of the nanofluid flow adopting the single phase approach. Nanofluid, up till a volume fraction of 1% is found to be an effective heat transfer enhancement technique. The Nusselt number (Nu) and friction factor predictions for the low volume fractions (i.e. 0.02%, 0.1 and 0.5%) agree very well with the experimental values of Sundar and Sharma (2010). While, predictions for the high volume fraction nanofluids (i.e. 1%, 4% and 6%) are found to have reasonable agreement with both experimental and numerical results available in the literature. So the computationally inexpensive single phase approach can be used for heat transfer and pressure drop prediction of new nanofluids.

Keywords: Heat transfer intensification, nanofluid, CFD, friction factor

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2874
384 Effects of Various Wavelet Transforms in Dynamic Analysis of Structures

Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar

Abstract:

Time history dynamic analysis of structures is considered as an exact method while being computationally intensive. Filtration of earthquake strong ground motions applying wavelet transform is an approach towards reduction of computational efforts, particularly in optimization of structures against seismic effects. Wavelet transforms are categorized into continuum and discrete transforms. Since earthquake strong ground motion is a discrete function, the discrete wavelet transform is applied in the present paper. Wavelet transform reduces analysis time by filtration of non-effective frequencies of strong ground motion. Filtration process may be repeated several times while the approximation induces more errors. In this paper, strong ground motion of earthquake has been filtered once applying each wavelet. Strong ground motion of Northridge earthquake is filtered applying various wavelets and dynamic analysis of sampled shear and moment frames is implemented. The error, regarding application of each wavelet, is computed based on comparison of dynamic response of sampled structures with exact responses. Exact responses are computed by dynamic analysis of structures applying non-filtered strong ground motion.

Keywords: Wavelet transform, computational error, computational duration, strong ground motion data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
383 An Experimental Study to Mitigate Swelling Pressure of Expansive Tabuk Shale, Saudi Arabia

Authors: A. A. Embaby, A. Abu Halawa, M. Ramadan

Abstract:

In Kingdom of Saudi Arabia, there are several areas where expansive soil exists in the form of variable-thicknesses layers in the developed regions. Severe distress to infrastructures can be caused by the development of heave and swelling pressure in this kind of expansive shale. Among the various techniques for expansive soil mitigation, the removal and replacement technique is very popular for lightly loaded structures and shallow foundations. This paper presents the result of an experimental study conducted for evaluating the effect of type and thickness of the cushion soils on mitigation of swelling characteristics of expanded shale. Seven undisturbed shale samples collected from Al Qadsiyah district, which is located in the Tabuk town north Kingdom of Saudi Arabia, are treated with two types of cushion coarse-grained sediments (CCS); sand and gravel. Each type is represented with three thicknesses, 22%, 33% and 44% in relation to the depth of the active zone. The test results indicated that the replacement of expansive shale by CCS reduces the swelling potential and pressure. It is found that the reduction in swelling depends on the type and thickness of CCS. The treatment by removing the original expansive shale and replacing it by cushion sand with 44% thickness reduced the swelling potential and pressure of about 53.29% and 62.78 %, respectively.

Keywords: Cushion coarse-grained sediments, expansive soil, Saudi Arabia, swelling pressure, Tabuk Shale.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1540
382 The Optimal Placement of Capacitor in Order to Reduce Losses and the Profile of Distribution Network Voltage with GA, SA

Authors: Limouzade E., Joorabian M.

Abstract:

Most of the losses in a power system relate to the distribution sector which always has been considered. From the important factors which contribute to increase losses in the distribution system is the existence of radioactive flows. The most common way to compensate the radioactive power in the system is the power to use parallel capacitors. In addition to reducing the losses, the advantages of capacitor placement are the reduction of the losses in the release peak of network capacity and improving the voltage profile. The point which should be considered in capacitor placement is the optimal placement and specification of the amount of the capacitor in order to maximize the advantages of capacitor placement. In this paper, a new technique has been offered for the placement and the specification of the amount of the constant capacitors in the radius distribution network on the basis of Genetic Algorithm (GA). The existing optimal methods for capacitor placement are mostly including those which reduce the losses and voltage profile simultaneously. But the retaliation cost and load changes have not been considered as influential UN the target function .In this article, a holistic approach has been considered for the optimal response to this problem which includes all the parameters in the distribution network: The price of the phase voltage and load changes. So, a vast inquiry is required for all the possible responses. So, in this article, we use Genetic Algorithm (GA) as the most powerful method for optimal inquiry.

Keywords: Genetic Algorithm (GA), capacitor placement, voltage profile, network losses, Simulating Annealing (SA), distribution network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1535
381 A Case Study on the Numerical-Probability Approach for Deep Excavation Analysis

Authors: Komeil Valipourian

Abstract:

Urban advances and the growing need for developing infrastructures has increased the importance of deep excavations. In this study, after the introducing probability analysis as an important issue, an attempt has been made to apply it for the deep excavation project of Bangkok’s Metro as a case study. For this, the numerical probability model has been developed based on the Finite Difference Method and Monte Carlo sampling approach. The results indicate that disregarding the issue of probability in this project will result in an inappropriate design of the retaining structure. Therefore, probabilistic redesign of the support is proposed and carried out as one of the applications of probability analysis. A 50% reduction in the flexural strength of the structure increases the failure probability just by 8% in the allowable range and helps improve economic conditions, while maintaining mechanical efficiency. With regard to the lack of efficient design in most deep excavations, by considering geometrical and geotechnical variability, an attempt was made to develop an optimum practical design standard for deep excavations based on failure probability. On this basis, a practical relationship is presented for estimating the maximum allowable horizontal displacement, which can help improve design conditions without developing the probability analysis.

Keywords: Numerical probability modeling, deep excavation, allowable maximum displacement, finite difference method, FDM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 691
380 Stabilizing Effects of Deep Eutectic Solvents on Alcohol Dehydrogenase Mediated Systems

Authors: Fatima Zohra Ibn Majdoub Hassani, Ivan Lavandera, Joseph Kreit

Abstract:

This study explored the effects of different organic solvents, temperature, and the amount of glycerol on the alcohol dehydrogenase (ADH)-catalysed stereoselective reduction of different ketones. These conversions were then analyzed by gas chromatography. It was found that when the amount of deep eutectic solvents (DES) increases, it can improve the stereoselectivity of the enzyme although reducing its ability to convert the substrate into the corresponding alcohol. Moreover, glycerol was found to have a strong stabilizing effect on the ADH from Ralstonia sp. (E. coli/ RasADH). In the case of organic solvents, it was observed that the best conversions into the alcohols were achieved with DMSO and hexane. It was also observed that temperature decreased the ability of the enzyme to convert the substrates into the products and also affected the selectivity. In addition to that, the recycling of DES up to three times gave good conversions and enantiomeric excess results and glycerol showed a positive effect in the stability of various ADHs. Using RasADH, a good conversion and enantiomeric excess into the S-alcohol were obtained. It was found that an enhancement of the temperature disabled the stabilizing effect of glycerol and decreased the stereoselectivity of the enzyme. However, for other ADHs a temperature increase had an opposite positive effect, especially with ADH-T from Thermoanaerobium sp. One of the objectives of this study was to see the effect of cofactors such as NAD(P) on the biocatlysis activities of ADHs.

Keywords: Alcohol dehydrogenases, DES, gas chromatography, RasADH.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1140
379 A Trainable Neural Network Ensemble for ECG Beat Classification

Authors: Atena Sajedin, Shokoufeh Zakernejad, Soheil Faridi, Mehrdad Javadi, Reza Ebrahimpour

Abstract:

This paper illustrates the use of a combined neural network model for classification of electrocardiogram (ECG) beats. We present a trainable neural network ensemble approach to develop customized electrocardiogram beat classifier in an effort to further improve the performance of ECG processing and to offer individualized health care. We process a three stage technique for detection of premature ventricular contraction (PVC) from normal beats and other heart diseases. This method includes a denoising, a feature extraction and a classification. At first we investigate the application of stationary wavelet transform (SWT) for noise reduction of the electrocardiogram (ECG) signals. Then feature extraction module extracts 10 ECG morphological features and one timing interval feature. Then a number of multilayer perceptrons (MLPs) neural networks with different topologies are designed. The performance of the different combination methods as well as the efficiency of the whole system is presented. Among them, Stacked Generalization as a proposed trainable combined neural network model possesses the highest recognition rate of around 95%. Therefore, this network proves to be a suitable candidate in ECG signal diagnosis systems. ECG samples attributing to the different ECG beat types were extracted from the MIT-BIH arrhythmia database for the study.

Keywords: ECG beat Classification; Combining Classifiers;Premature Ventricular Contraction (PVC); Multi Layer Perceptrons;Wavelet Transform

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2214
378 Thermodynamic Cycle Analysis for Overall Efficiency Improvement and Temperature Reduction in Gas Turbines

Authors: Jeni A. Popescu, Ionut Porumbel, Valeriu A. Vilag, Cleopatra F. Cuciumita

Abstract:

The paper presents a thermodynamic cycle analysis for three turboshaft engines. The first cycle is a Brayton cycle, describing the evolution of a classical turboshaft, based on the Klimov TV2 engine. The other four cycles aim at approaching an Ericsson cycle, by replacing the Brayton cycle adiabatic expansion in the turbine by quasi-isothermal expansion. The maximum quasi- Ericsson cycles temperature is set to a lower value than the maximum Brayton cycle temperature, equal to the Brayton cycle power turbine inlet temperature, in order to decrease the engine NOx emissions. Also, the power/expansion ratio distribution over the stages of the gas generator turbine is maintained the same. In two of the considered quasi-Ericsson cycles, the efficiencies of the gas generator turbine, as well as the power/expansion ratio distribution over the stages of the gas generator turbine are maintained the same as for the reference case, while for the other two cases, the efficiencies are increased in order to obtain the same shaft power as in the reference case. For the two cases respecting the first condition, both the shaft power and the thermodynamic efficiency of the engine decrease, while for the other two, the power and efficiency are maintained, as a result of assuming new, more efficient gas generator turbines.

Keywords: Combustion, Ericsson, thermodynamic analysis, turbine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2462
377 Effect of Twin Cavities on the Axially Loaded Pile in Clay

Authors: Ali A. Al-Jazaairry, Tahsin T. Sabbagh

Abstract:

Presence of cavities in soil predictably induces ground deformation and changes in soil stress, which might influence adjacent existing pile foundations, though the effect of twin cavities on a nearby pile needs to be understood. This research is an attempt to identify the behaviour of piles subjected to axial load and embedded in cavitied clayey soil. A series of finite element modelling were conducted to investigate the performance of piled foundation located in such soils. The validity of the numerical simulation was evaluated by comparing it with available field test and alternative analytical model. The study involved many parameters such as twin cavities size, depth, spacing between cavities, and eccentricity of cavities from the pile axis on the pile performance subjected to axial load. The study involved many cases; in each case, a critical value has been found in which cavities’ presence has shown minimum impact on the behaviour of pile. Load-displacement relationships of the affecting parameters on the pile behaviour were presented to provide helpful information for designing piled foundation situated near twin underground cavities. It was concluded that the presence of the cavities within the soil mass reduces the ultimate capacity of pile. This reduction differs according to the size and location of the cavity.

Keywords: Axial load, clay, finite element, pile, twin cavities, ultimate capacity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1252
376 Reduction of Chloride Dioxide in Paper Bleaching using Peroxide Activation

Authors: Ayo Afolabi, Rudzani Sigwadi, Ambali Abdulkareem

Abstract:

All around the world pulp and paper industries are the biggest plant production with the environmental pollution as the biggest challenge facing the pulp manufacturing operations. The concern among these industries is to produce a high volume of papers with the high quality standard and of low cost without affecting the environment. This result obtained from this bleaching study show that the activation of peroxide was an effective method of reducing the total applied charge of chlorine dioxide which is harmful to our environment and also show that softwood and hardwood Kraft pulps responded linearly to the peroxide treatments. During the bleaching process the production plant produce chlorines. Under the trial stages chloride dioxide has been reduced by 3 kg/ton to reduce the brightness from 65% ISO to 60% ISO of pulp and the dosing point returned to the E stage charges by pre-treating Kraft pulps with hydrogen peroxide. The pulp and paper industry has developed elemental chlorine free (ECF) and totally chlorine free (TCF) bleaching, in their quest for being environmental friendly, they have been looking at ways to turn their ECF process into a TCF process while still being competitive. This prompted the research to investigate the capability of the hydrogen peroxide as catalyst to reduce chloride dioxide.

Keywords: bleaching, chloride dioxide, elemental chlorine free (ECF), hydrogen peroxide, totally chlorine free (TCF).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2824
375 Motion Analysis for Duplicate Frame Removal in Wireless Capsule Endoscope Video

Authors: Min Kook Choi, Hyun Gyu Lee, Ryan You, Byeong-Seok Shin, Sang-Chul Lee

Abstract:

Wireless capsule Endoscopy (WCE) has rapidly shown its wide applications in medical domain last ten years thanks to its noninvasiveness for patients and support for thorough inspection through a patient-s entire digestive system including small intestine. However, one of the main barriers to efficient clinical inspection procedure is that it requires large amount of effort for clinicians to inspect huge data collected during the examination, i.e., over 55,000 frames in video. In this paper, we propose a method to compute meaningful motion changes of WCE by analyzing the obtained video frames based on regional optical flow estimations. The computed motion vectors are used to remove duplicate video frames caused by WCE-s imaging nature, such as repetitive forward-backward motions from peristaltic movements. The motion vectors are derived by calculating directional component vectors in four local regions. Our experiments are performed on small intestine area, which is of main interest to clinical experts when using WCEs, and our experimental results show significant frame reductions comparing with a simple frame-to-frame similarity-based image reduction method.

Keywords: Wireless capsule endoscopy, optical flow, duplicated image, duplicated frame.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1691
374 The effect of Gamma Irradiation on the Nutritional Properties of Functional Products of the Green Banana

Authors: Magda S. Taipina, Maria L. Garbelotti, Mariana G.B. Cadioli

Abstract:

Banana is one of the most consumed fruits in the tropics and subtropics. Brazil accounts for about 9% of the world banana production. However, the production losses are as high as 30 to 40% and even much higher in some developing countries. The green banana flour is a complex carbohydrate source, including a high total starch (73.4%), resistant starch (17.5%) with functional properties. Gamma irradiation is considered to be an alternative method for food preservation. It has been performed due to the need of extending the shelf - life of foods, whilst maintaining their safety and avoiding one of the main concerns: the nutrient loss. In this work data about on the effects of ionizing radiation on the physicochemical analysis (carbohydrate, proteins, lipids, alimentary fiber, moistures and ashes) of Brazilian functional products (biscuits and bread) of the green banana pulp are presented. The caloric value was calculated. No significant difference was observed between the samples of irradiated and non – irradiated green banana biscuits with the following determinations: carbohydrates, proteins, alimentary fiber and ashes. Only a small significant difference was found in lipids (macronutrients). The results of physical chemical analysis of the irradiated and non- irradiated green banana bread non- irradiated showed no significant difference with the following determinations: carbohydrates, lipids (macronutrients), moisture, ashes and caloric value. A small difference was found in proteins (macronutrients). Irradiation of functional products (biscuits and bread) with doses of 1 and 3kGy maintained their original macronutrients content, showing good radioresistance.

Keywords: Irradiation, Functional Food, Nutritional value.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1667
373 Flood Modeling in Urban Area Using a Well-Balanced Discontinuous Galerkin Scheme on Unstructured Triangular Grids

Authors: Rabih Ghostine, Craig Kapfer, Viswanathan Kannan, Ibrahim Hoteit

Abstract:

Urban flooding resulting from a sudden release of water due to dam-break or excessive rainfall is a serious threatening environment hazard, which causes loss of human life and large economic losses. Anticipating floods before they occur could minimize human and economic losses through the implementation of appropriate protection, provision, and rescue plans. This work reports on the numerical modelling of flash flood propagation in urban areas after an excessive rainfall event or dam-break. A two-dimensional (2D) depth-averaged shallow water model is used with a refined unstructured grid of triangles for representing the urban area topography. The 2D shallow water equations are solved using a second-order well-balanced discontinuous Galerkin scheme. Theoretical test case and three flood events are described to demonstrate the potential benefits of the scheme: (i) wetting and drying in a parabolic basin (ii) flash flood over a physical model of the urbanized Toce River valley in Italy; (iii) wave propagation on the Reyran river valley in consequence of the Malpasset dam-break in 1959 (France); and (iv) dam-break flood in October 1982 at the town of Sumacarcel (Spain). The capability of the scheme is also verified against alternative models. Computational results compare well with recorded data and show that the scheme is at least as efficient as comparable second-order finite volume schemes, with notable efficiency speedup due to parallelization.

Keywords: Flood modeling, dam-break, shallow water equations, Discontinuous Galerkin scheme, MUSCL scheme.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 946
372 Arteriosclerosis and Periodontitis: Correlation Expressed in the Amount of Fibrinogen in Blood

Authors: Nevila Alliu, Saimir Heta, Ilma Robo, Vera Ostreni

Abstract:

Periodontitis as an oral pathology caused by specific bacterial flora functions as a focal infection for the onset and aggravation of arteriosclerosis. These two distant pathologies, since they affect organs at a distance from each other, communicate with each other with correlation at the level of markers of inflammation in the blood. Fluctuations in the level of fibrinogen in the blood, depending on the active or passive phase of the existing periodontitis, affect the promotion of arteriosclerosis. The study is of the brief communication article type with the aim to analyze the effect of non-surgical periodontal treatment on fluctuations in the level of fibrinogen in the blood. The reduction of fibrinogen's level in blood after non-surgical periodontal treatment of periodontitis in the patient's oral cavity, is a common consequence supported by literature sources. Also, the influence of a high amount of fibrinogen in blood on the occurrence of arteriosclerosis at the same patient, is also another important data that again rely on many sources of literature. Thromboembolism and arteriosclerosis, as risk factors expressed in clinical data, have temporary bacteremia in the blood, which can occur significantly and often between phases of non-surgical periodontal treatment of periodontitis, treatments performed with treatment phases and protocols of predetermined treatment. Arterial thromboembolism has a significant factor, such as high levels of fibrinogen in the blood, which are significantly reduced during the period of non-surgical periodontal treatment.

Keywords: Fibrinogen, refractory periodontitis, atherosclerosis, non-surgical, periodontal treatment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 127
371 Can Exams Be Shortened? Using a New Empirical Approach to Test in Finance Courses

Authors: Eric S. Lee, Connie Bygrave, Jordan Mahar, Naina Garg, Suzanne Cottreau

Abstract:

Marking exams is universally detested by lecturers. Final exams in many higher education courses often last 3.0 hrs. Do exams really need to be so long? Can we justifiably reduce the number of questions on them? Surprisingly few have researched these questions, arguably because of the complexity and difficulty of using traditional methods. To answer these questions empirically, we used a new approach based on three key elements: Use of an unusual variation of a true experimental design, equivalence hypothesis testing, and an expanded set of six psychometric criteria to be met by any shortened exam if it is to replace a current 3.0-hr exam (reliability, validity, justifiability, number of exam questions, correspondence, and equivalence). We compared student performance on each official 3.0-hr exam with that on five shortened exams having proportionately fewer questions (2.5, 2.0, 1.5, 1.0, and 0.5 hours) in a series of four experiments conducted in two classes in each of two finance courses (224 students in total). We found strong evidence that, in these courses, shortening of final exams to 2.0 hrs was warranted on all six psychometric criteria. Shortening these exams by one hour should result in a substantial one-third reduction in lecturer time and effort spent marking, lower student stress, and more time for students to prepare for other exams. Our approach provides a relatively simple, easy-to-use methodology that lecturers can use to examine the effect of shortening their own exams.

Keywords: Exam length, psychometric criteria, synthetic experimental designs, test length.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1502
370 Improving Fault Resilience and Reconstruction of Overlay Multicast Tree Using Leaving Time of Participants

Authors: Bhed Bahadur Bista

Abstract:

Network layer multicast, i.e. IP multicast, even after many years of research, development and standardization, is not deployed in large scale due to both technical (e.g. upgrading of routers) and political (e.g. policy making and negotiation) issues. Researchers looked for alternatives and proposed application/overlay multicast where multicast functions are handled by end hosts, not network layer routers. Member hosts wishing to receive multicast data form a multicast delivery tree. The intermediate hosts in the tree act as routers also, i.e. they forward data to the lower hosts in the tree. Unlike IP multicast, where a router cannot leave the tree until all members below it leave, in overlay multicast any member can leave the tree at any time thus disjoining the tree and disrupting the data dissemination. All the disrupted hosts have to rejoin the tree. This characteristic of the overlay multicast causes multicast tree unstable, data loss and rejoin overhead. In this paper, we propose that each node sets its leaving time from the tree and sends join request to a number of nodes in the tree. The nodes in the tree will reject the request if their leaving time is earlier than the requesting node otherwise they will accept the request. The node can join at one of the accepting nodes. This makes the tree more stable as the nodes will join the tree according to their leaving time, earliest leaving time node being at the leaf of the tree. Some intermediate nodes may not follow their leaving time and leave earlier than their leaving time thus disrupting the tree. For this, we propose a proactive recovery mechanism so that disrupted nodes can rejoin the tree at predetermined nodes immediately. We have shown by simulation that there is less overhead when joining the multicast tree and the recovery time of the disrupted nodes is much less than the previous works. Keywords

Keywords: Network layer multicast, Fault Resilience, IP multicast

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1386
369 Development of a System for Measuring the Three-Axis Pedal Force in Cycling and Its Applications

Authors: Joo-Hack Lee, Jin-Seung Choi, Dong-Won Kang, Jeong-Woo Seo, Ju-Young Kim, Dae-Hyeok Kim, Seung-Tae Yang, Gye-Rae Tack

Abstract:

For cycling, the analysis of the pedal force is one of the important factors in the study of exercise ability assessment and overuse injuries. In past studies, a two-axis measurement sensor was used at the sagittal plane to measure the force only in the anterior, posterior, and vertical directions and to analyze the loss of force and the injury on the frontal plane due to the forces in the right and left directions. In this study, which is a basic study on diverse analyses of the pedal force that consider the forces on the sagittal plane and the frontal plane, a three-axis pedal force measurement sensor was developed to measure the anterior-posterior (Fx), medio-lateral (Fz), and vertical (Fy) forces. The sensor was fabricated with a size and shape similar to those of the general flat pedal, and had a 550g weight that allowed smooth pedaling. Its measurement range was ±1000 N for Fx and Fz and ±2000 N for Fy, and its non-linearity, hysteresis, and repeatability were approximately 0.5%. The data were sampled at 1000 Hz using a signal collector. To use the developed sensor, the pedaling efficiency (index of efficiency, IE) and the range of left and right (medio-lateral, ML) forces were measured with two seat heights (low and high). The results of the measurement showed that the IE was higher and the force range in the ML direction was lower with the high position than with the low position. The developed measurement sensor and its application results will be useful in understanding and explaining the complicated pedaling technique, and will enable diverse kinematic analyses of the pedal force on the sagittal plane and the frontal plane.

Keywords: Cycling, Index of effectiveness, Pedal force.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2572
368 Comparison of Current Chinese and Japanese Design Specification for Bridge Pile in Liquefied Ground

Authors: Baydaa H. Maula, Ling Zhang, Tang Liang, Gao Xia, Xu Peng-Ju, Zhang Yong-Qiang, Kang Jie, Su Lei

Abstract:

Firstly, this study briefly presents the current situation that there exists a vast gap between current Chinese and Japanese seismic design specification for bridge pile foundation in liquefiable and liquefaction-induced lateral spreading ground; The Chinese and Japanese seismic design method and technical detail for bridge pile foundation in liquefying and lateral spreading ground are described and compared systematically and comprehensively, the methods of determining coefficient of subgrade reaction and its reduction factor as well as the computing mode of the applied force on pile foundation due to liquefaction-induced lateral spreading soil in Japanese design specification are especially introduced. Subsequently, the comparison indicates that the content of Chinese seismic design specification for bridge pile foundation in liquefiable and liquefaction-induced lateral spreading ground, just presenting some qualitative items, is too general and lacks systematicness and maneuverability. Finally, some defects of seismic design specification in China are summarized, so the improvement and revision of specification in the field turns out to be imperative for China, some key problems of current Chinese specifications are generalized and the corresponding improvement suggestions are proposed.

Keywords: liquefying soil, laterally spreading ground, seismic design specification for bridge pile foundation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3652
367 Improving Worm Detection with Artificial Neural Networks through Feature Selection and Temporal Analysis Techniques

Authors: Dima Stopel, Zvi Boger, Robert Moskovitch, Yuval Shahar, Yuval Elovici

Abstract:

Computer worm detection is commonly performed by antivirus software tools that rely on prior explicit knowledge of the worm-s code (detection based on code signatures). We present an approach for detection of the presence of computer worms based on Artificial Neural Networks (ANN) using the computer's behavioral measures. Identification of significant features, which describe the activity of a worm within a host, is commonly acquired from security experts. We suggest acquiring these features by applying feature selection methods. We compare three different feature selection techniques for the dimensionality reduction and identification of the most prominent features to capture efficiently the computer behavior in the context of worm activity. Additionally, we explore three different temporal representation techniques for the most prominent features. In order to evaluate the different techniques, several computers were infected with five different worms and 323 different features of the infected computers were measured. We evaluated each technique by preprocessing the dataset according to each one and training the ANN model with the preprocessed data. We then evaluated the ability of the model to detect the presence of a new computer worm, in particular, during heavy user activity on the infected computers.

Keywords: Artificial Neural Networks, Feature Selection, Temporal Analysis, Worm Detection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1726
366 Construction and Validation of a Hybrid Lumbar Spine Model for the Fast Evaluation of Intradiscal Pressure and Mobility

Authors: Ali Hamadi Dicko, Nicolas Tong-Yette, Benjamin Gilles, François Faure, Olivier Palombi

Abstract:

A novel hybrid model of the lumbar spine, allowing fast static and dynamic simulations of the disc pressure and the spine mobility, is introduced in this work. Our contribution is to combine rigid bodies, deformable finite elements, articular constraints, and springs into a unique model of the spine. Each vertebra is represented by a rigid body controlling a surface mesh to model contacts on the facet joints and the spinous process. The discs are modeled using a heterogeneous tetrahedral finite element model. The facet joints are represented as elastic joints with six degrees of freedom, while the ligaments are modeled using non-linear one-dimensional elastic elements. The challenge we tackle is to make these different models efficiently interact while respecting the principles of Anatomy and Mechanics. The mobility, the intradiscal pressure, the facet joint force and the instantaneous center of rotation of the lumbar spine are validated against the experimental and theoretical results of the literature on flexion, extension, lateral bending as well as axial rotation. Our hybrid model greatly simplifies the modeling task and dramatically accelerates the simulation of pressure within the discs, as well as the evaluation of the range of motion and the instantaneous centers of rotation, without penalizing precision. These results suggest that for some types of biomechanical simulations, simplified models allow far easier modeling and faster simulations compared to usual full-FEM approaches without any loss of accuracy.

Keywords: Hybrid, modeling, fast simulation, lumbar spine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2381
365 Night-Time Traffic Light Detection Based On SVM with Geometric Moment Features

Authors: Hyun-Koo Kim, Young-Nam Shin, Sa-gong Kuk, Ju H. Park, Ho-Youl Jung

Abstract:

This paper presents an effective traffic lights detection method at the night-time. First, candidate blobs of traffic lights are extracted from RGB color image. Input image is represented on the dominant color domain by using color transform proposed by Ruta, then red and green color dominant regions are selected as candidates. After candidate blob selection, we carry out shape filter for noise reduction using information of blobs such as length, area, area of boundary box, etc. A multi-class classifier based on SVM (Support Vector Machine) applies into the candidates. Three kinds of features are used. We use basic features such as blob width, height, center coordinate, area, area of blob. Bright based stochastic features are also used. In particular, geometric based moment-s values between candidate region and adjacent region are proposed and used to improve the detection performance. The proposed system is implemented on Intel Core CPU with 2.80 GHz and 4 GB RAM and tested with the urban and rural road videos. Through the test, we show that the proposed method using PF, BMF, and GMF reaches up to 93 % of detection rate with computation time of in average 15 ms/frame.

Keywords: Night-time traffic light detection, multi-class classification, driving assistance system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3884
364 Precision Grinding of Titanium (Ti-6Al-4V) Alloy Using Nanolubrication

Authors: Ahmed A. D. Sarhan, Hong Wan Ping, M. Sayuti

Abstract:

In this current era of competitive machinery productions, the industries are designed to place more emphasis on the product quality and reduction of cost whilst abiding by the pollution-preventing policy. In attempting to delve into the concerns, the industries are aware that the effectiveness of existing lubrication systems must be improved to achieve power-efficient and pollution-preventing machining processes. As such, this research is targeted to study on a plausible solution to the issue in grinding titanium alloy (Ti-6Al-4V) by using nanolubrication, as an alternative to flood grinding. The aim of this research is to evaluate the optimum condition of grinding force and surface roughness using MQL lubricating system to deliver nano-oil at different level of weight concentration of Silicon Dioxide (SiO2) mixed normal mineral oil. Taguchi Design of Experiment (DoE) method is carried out using a standard Taguchi orthogonal array of L16(43) to find the optimized combination of weight concentration mixture of SiO2, nozzle orientation and pressure of MQL. Surface roughness and grinding force are also analyzed using signal-to-noise(S/N) ratio to determine the best level of each factor that are tested. Consequently, the best combination of parameters is tested for a period of time and the results are compared with conventional grinding method of dry and flood condition. The results show a positive performance of MQL nanolubrication.  

Keywords: Grinding, MQL, precision grinding, Taguchi optimization, titanium alloy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1883