Search results for: Data Reduction
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8667

Search results for: Data Reduction

8337 Edible Oil Industry Wastewater Treatment by Microfiltration with Ceramic Membrane

Authors: Zita Šereš, Dragana Šoronja Simović, Ljubica Dokić, Lidietta Giorno, Biljana Pajin, Cecilia Hodur, Nikola Maravić

Abstract:

Membrane technology is convenient for separation of suspended solids, colloids and high molecular weight materials that are present. The idea is that the waste stream from edible oil industry, after the separation of oil by using skimmers is subjected to microfiltration and the obtained permeate can be used again in the production process. The wastewater from edible oil industry was used for the microfiltration. For the microfiltration of this effluent a tubular membrane was used with a pore size of 200 nm at transmembrane pressure in range up to 3 bar and in range of flow rate up to 300 L/h. Box–Behnken design was selected for the experimental work and the responses considered were permeate flux and chemical oxygen demand (COD) reduction. The reduction of the permeate COD was in the range 40-60% according to the feed. The highest permeate flux achieved during the process of microfiltration was 160 L/m2h.

Keywords: Ceramic membrane, edible oil, microfiltration, wastewater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1619
8336 The Application of Data Mining Technology in Building Energy Consumption Data Analysis

Authors: Liang Zhao, Jili Zhang, Chongquan Zhong

Abstract:

Energy consumption data, in particular those involving public buildings, are impacted by many factors: the building structure, climate/environmental parameters, construction, system operating condition, and user behavior patterns. Traditional methods for data analysis are insufficient. This paper delves into the data mining technology to determine its application in the analysis of building energy consumption data including energy consumption prediction, fault diagnosis, and optimal operation. Recent literature are reviewed and summarized, the problems faced by data mining technology in the area of energy consumption data analysis are enumerated, and research points for future studies are given.

Keywords: Data mining, data analysis, prediction, optimization, building operational performance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3696
8335 Reduction of Search Space by Applying Controlled Genetic Operators for Weight Constrained Shortest Path Problem

Authors: A.K.M. Khaled Ahsan Talukder, Taibun Nessa, Kaushik Roy

Abstract:

The weight constrained shortest path problem (WCSPP) is one of most several known basic problems in combinatorial optimization. Because of its importance in many areas of applications such as computer science, engineering and operations research, many researchers have extensively studied the WCSPP. This paper mainly concentrates on the reduction of total search space for finding WCSP using some existing Genetic Algorithm (GA). For this purpose, some controlled schemes of genetic operators are adopted on list chromosome representation. This approach gives a near optimum solution with smaller elapsed generation than classical GA technique. From further analysis on the matter, a new generalized schema theorem is also developed from the philosophy of Holland-s theorem.

Keywords: Genetic Algorithm, Evolutionary Optimization, Multi Objective Optimization, Non-linear Schema Theorem, WCSPP.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1405
8334 Query Algebra for Semistuctured Data

Authors: Ei Ei Myat, Ni Lar Thein

Abstract:

With the tremendous growth of World Wide Web (WWW) data, there is an emerging need for effective information retrieval at the document level. Several query languages such as XML-QL, XPath, XQL, Quilt and XQuery are proposed in recent years to provide faster way of querying XML data, but they still lack of generality and efficiency. Our approach towards evolving a framework for querying semistructured documents is based on formal query algebra. Two elements are introduced in the proposed framework: first, a generic and flexible data model for logical representation of semistructured data and second, a set of operators for the manipulation of objects defined in the data model. In additional to accommodating several peculiarities of semistructured data, our model offers novel features such as bidirectional paths for navigational querying and partitions for data transformation that are not available in other proposals.

Keywords: Algebra, Semistructured data, Query Algebra.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1366
8333 Improving Performance of World Wide Web by Adaptive Web Traffic Reduction

Authors: Achuthsankar S. Nair, J. S. Jayasudha

Abstract:

The ever increasing use of World Wide Web in the existing network, results in poor performance. Several techniques have been developed for reducing web traffic by compressing the size of the file, saving the web pages at the client side, changing the burst nature of traffic into constant rate etc. No single method was adequate enough to access the document instantly through the Internet. In this paper, adaptive hybrid algorithms are developed for reducing web traffic. Intelligent agents are used for monitoring the web traffic. Depending upon the bandwidth usage, user-s preferences, server and browser capabilities, intelligent agents use the best techniques to achieve maximum traffic reduction. Web caching, compression, filtering, optimization of HTML tags, and traffic dispersion are incorporated into this adaptive selection. Using this new hybrid technique, latency is reduced to 20 – 60 % and cache hit ratio is increased 40 – 82 %.

Keywords: Bandwidth, Congestion, Intelligent Agents, Prefetching, Web Caching.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737
8332 Simulation Data Summarization Based on Spatial Histograms

Authors: Jing Zhao, Yoshiharu Ishikawa, Chuan Xiao, Kento Sugiura

Abstract:

In order to analyze large-scale scientific data, research on data exploration and visualization has gained popularity. In this paper, we focus on the exploration and visualization of scientific simulation data, and define a spatial V-Optimal histogram for data summarization. We propose histogram construction algorithms based on a general binary hierarchical partitioning as well as a more specific one, the l-grid partitioning. For effective data summarization and efficient data visualization in scientific data analysis, we propose an optimal algorithm as well as a heuristic algorithm for histogram construction. To verify the effectiveness and efficiency of the proposed methods, we conduct experiments on the massive evacuation simulation data.

Keywords: Simulation data, data summarization, spatial histograms, exploration and visualization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 743
8331 The Effect of Global Solar Variations on the Performance of n-AlGaAs/p-GaAs Solar Cells

Authors: A. Guechi, M. Chegaar

Abstract:

This study investigates how AlGaAs/GaAs thin film solar cells perform under varying global solar spectrum due to the changes of environmental parameters such as the air mass and the atmospheric turbidity. The solar irradiance striking the solar cell is simulated using the spectral irradiance model SMARTS2 (Simple Model of the Atmospheric Radiative Transfer of Sunshine) for clear skies on the site of Setif (Algeria). The results show a reduction in the short circuit current due to increasing atmospheric turbidity, it is 63.09% under global radiation. However increasing air mass leads to a reduction in the short circuit current of 81.73%. The efficiency decreases with increasing atmospheric turbidity and air mass.

Keywords: AlGaAs/GaAs, Solar Cells, Environmental parameters, Spectral Variation, SMARTS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2202
8330 Polyacrylate Modified Copper Nanoparticles with Controlled Size

Authors: Robert Prucek, Aleš Panáček, Jan Filip, Libor Kvítek, Radek Zbořil

Abstract:

The preparation of Cu nanoparticles (NPs) through the reduction of copper ions by sodium borohydride in the presence of sodium polyacrylate with a molecular weight of 1200 is reported. Cu NPs were synthesized at a concentration of copper salt equal to 2.5, 5, and 10 mM, and at a molar ratio of copper ions and monomeric unit of polyacrylate equal to 1:2. The as-prepared Cu NPs have diameters of about 2.5–3 nm for copper concentrations of 2.5 and 5 mM, and 6 nm for copper concentration of 10 mM. Depending on the copper salt concentration and concentration of additionally added polyacrylate to Cu particle dispersion, primarily formed NPs grow through the process of aggregation and/or coalescence into clusters and/or particles with a diameter between 20–100 nm. The amount of additionally added sodium polyacrylate influences the stability of Cu particles against air oxidation. The catalytic efficiency of the prepared Cu particles for the reduction of 4-nitrophenol is discussed.

Keywords: Copper, nanoparticles, sodium polyacrylate, catalyst, 4-nitrophenol.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3441
8329 Effect of Two Entomopathogenic Fungi Beauveria bassiana and Metarhizium anisopliae var. acridum on the Haemolymph of the Desert Locust Schistocerca gregaria

Authors: Fatima Zohra Bissaad, Farid Bounaceur, Nassima Behidj, Nadjiba Chebouti, Fatma Halouane, Bahia Doumandji-Mitiche

Abstract:

Effect of Beauveria bassiana and Metarhizium anisopliae var. acridum on the 5th instar nymphs of Schistocerca gregaria was studied in the laboratory. Infection by these both entomopathogenic fungi caused reduction in the hemolymph total protein. The average amounts of total proteins were 2.3, 2.07, 2.09 µg/100 ml of haemolymph in the control and M. anisopliae var. acridum, and B. bassiana based-treatments, respectively. Three types of haemocytes were recognized and identified as prohaemocytes, plasmatocytes and granulocytes. The treatment caused significant reduction in the total haemocyte count and in each haemocyte type on the 9th day after its application.

Keywords: Beauveria bassiana, haemolymph picture, haemolymph protein, Metarhizium anisopliae var. acridum, Schistocerca gregaria.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1460
8328 Interest of the Sequences Pseudo Noises Codes of Different Lengths for the Reduction from the Interference between Users of CDMA Network

Authors: Nerguè Kassahan Kone, Souleymane Oumtanaga

Abstract:

The third generation (3G) of cellular system adopted the spread spectrum as solution for the transmission of the data in the physical layer. Contrary to systems IS-95 or CDMAOne (systems with spread spectrum of the preceding generation), the new standard, called Universal Mobil Telecommunications System (UMTS), uses long codes in the down link. The system is conceived for the vocal communication and the transmission of the data. In particular, the down link is very important, because of the asymmetrical request of the data, i.e., more remote loading towards the mobiles than towards the basic station. Moreover, the UMTS uses for the down link an orthogonal spreading out with a variable factor of spreading out (OVSF for Orthogonal Variable Spreading Factor). This characteristic makes it possible to increase the flow of data of one or more users by reducing their factor of spreading out without changing the factor of spreading out of other users. In the current standard of the UMTS, two techniques to increase the performances of the down link were proposed, the diversity of sending antenna and the codes space-time. These two techniques fight only fainding. The receiver proposed for the mobil station is the RAKE, but one can imagine a receiver more sophisticated, able to reduce the interference between users and the impact of the coloured noise and interferences to narrow band. In this context, where the users have long codes synchronized with variable factor of spreading out and ignorance by the mobile of the other active codes/users, the use of the sequences of code pseudo-noises different lengths is presented in the form of one of the most appropriate solutions.

Keywords: DS-CDMA, multiple access interference, ratio Signal / interference + Noise.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1346
8327 A Power Reduction Technique for Built-In-Self Testing Using Modified Linear Feedback Shift Register

Authors: Mayank Shakya, Soundra Pandian. K. K

Abstract:

A linear feedback shift register (LFSR) is proposed which targets to reduce the power consumption from within. It reduces the power consumption during testing of a Circuit Under Test (CUT) at two stages. At first stage, Control Logic (CL) makes the clocks of the switching units of the register inactive for a time period when output from them is going to be same as previous one and thus reducing unnecessary switching of the flip-flops. And at second stage, the LFSR reorders the test vectors by interchanging the bit with its next and closest neighbor bit. It keeps fault coverage capacity of the vectors unchanged but reduces the Total Hamming Distance (THD) so that there is reduction in power while shifting operation.

Keywords: Linear Feedback Shift Register, Total Hamming Distance, Fault Coverage, Control Logic

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2017
8326 Reduction of Content of Lead and Zinc from Wastewater by Using of Metallurgical Waste

Authors: L. Rozumová, J. Seidlerová

Abstract:

The aim of this paper was to study the sorption properties of a blast furnace sludge used as the sorbent. The sorbent was utilized for reduction of content of lead and zinc ions. Sorbent utilized in this work was obtained from metallurgical industry from process of wet gas treatment in iron production. The blast furnace sludge was characterized by X-Ray diffraction, scanning electron microscopy, and XRFS spectroscopy. Sorption experiments were conducted in batch mode. The sorption of metal ions in the sludge was determined by correlation of adsorption isotherm models. The adsorption of lead and zinc ions was best fitted with Langmuir adsorption isotherms. The adsorption capacity of lead and zinc ions was 53.8 mg.g-1 and 10.7 mg.g-1, respectively. The results indicated that blast furnace sludge could be effectively used as secondary material and could be also employed as a low-cost alternative for the removal of heavy metals ions from wastewater.

Keywords: Blast furnace sludge, lead, zinc, sorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 894
8325 Hydrochemical Contamination Profiling and Spatial-Temporal Mapping with the Support of Multivariate and Cluster Statistical Analysis

Authors: S. Barbosa, M. Pinto, J. A. Almeida, E. Carvalho, C. Diamantino

Abstract:

The aim of this work was to test a methodology able to generate spatial-temporal maps that can synthesize simultaneously the trends of distinct hydrochemical indicators in an old radium-uranium tailings dam deposit. Multidimensionality reduction derived from principal component analysis and subsequent data aggregation derived from clustering analysis allow to identify distinct hydrochemical behavioral profiles and generate synthetic evolutionary hydrochemical maps.

Keywords: Contamination plume migration, K-means of PCA scores, groundwater and mine water monitoring, spatial-temporal hydrochemical trends.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 610
8324 Epileptic Seizure Prediction by Exploiting Signal Transitions Phenomena

Authors: Mohammad Zavid Parvez, Manoranjan Paul

Abstract:

A seizure prediction method is proposed by extracting global features using phase correlation between adjacent epochs for detecting relative changes and local features using fluctuation/ deviation within an epoch for determining fine changes of different EEG signals. A classifier and a regularization technique are applied for the reduction of false alarms and improvement of the overall prediction accuracy. The experiments show that the proposed method outperforms the state-of-the-art methods and provides high prediction accuracy (i.e., 97.70%) with low false alarm using EEG signals in different brain locations from a benchmark data set.

Keywords: Epilepsy, Seizure, Phase Correlation, Fluctuation, Deviation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2461
8323 Improving the Effectiveness of Software Testing through Test Case Reduction

Authors: R. P. Mahapatra, Jitendra Singh

Abstract:

This paper proposes a new technique for improving the efficiency of software testing, which is based on a conventional attempt to reduce test cases that have to be tested for any given software. The approach utilizes the advantage of Regression Testing where fewer test cases would lessen time consumption of the testing as a whole. The technique also offers a means to perform test case generation automatically. Compared to one of the techniques in the literature where the tester has no option but to perform the test case generation manually, the proposed technique provides a better option. As for the test cases reduction, the technique uses simple algebraic conditions to assign fixed values to variables (Maximum, minimum and constant variables). By doing this, the variables values would be limited within a definite range, resulting in fewer numbers of possible test cases to process. The technique can also be used in program loops and arrays.

Keywords: Software Testing, Test Case Generation, Test CaseReduction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3004
8322 Lookup Table Reduction and Its Error Analysis of Hall Sensor-Based Rotation Angle Measurement

Authors: Young-San Shin, Seongsoo Lee

Abstract:

Hall sensor is widely used to measure rotation angle. When the Hall voltage is measured for linear displacement, it is converted to angular displacement using arctangent function, which requires a large lookup table. In this paper, a lookup table reduction technique is presented for angle measurement. When the input of the lookup table is small within a certain threshold, the change of the outputs with respect to the change of the inputs is relatively small. Thus, several inputs can share same output, which significantly reduce the lookup table size. Its error analysis was also performed, and the threshold was determined so as to maintain the error less than 1°. When the Hall voltage has 11-bit resolution, the lookup table size is reduced from 1,024 samples to 279 samples.

Keywords: Hall sensor, angle measurement, lookup table, arctangent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1519
8321 Potential of Detailed Environmental Data Produced by Information and Communication Technology Tools for Better Consideration of Microclimatology Issues in Urban Planning to Promote Active Mobility

Authors: Živa Ravnikar, Alfonso Bahillo Martinez, Barbara Goličnik Marušić

Abstract:

Climate change mitigation has been formally adopted and announced by countries over the globe, where cities are targeting carbon neutrality through various more or less successful, systematic, and fragmentary actions. The article is based on the fact that environmental conditions affect human comfort and the usage of space. Urban planning can, with its sustainable solutions, not only support climate mitigation in terms of a planet reduction of global warming but as well enabling natural processes that in the immediate vicinity produce environmental conditions that encourage people to walk or cycle. However, the article draws attention to the importance of integrating climate consideration into urban planning, where detailed environmental data play a key role, enabling urban planners to improve or monitor environmental conditions on cycle paths. In a practical aspect, this paper tests a particular ICT tool, a prototype used for environmental data. Data gathering was performed along the cycling lanes in Ljubljana (Slovenia), where the main objective was to assess the tool's data applicable value within the planning of comfortable cycling lanes. The results suggest that such transportable devices for in-situ measurements can help a researcher interpret detailed environmental information, characterized by fine granularity and precise data spatial and temporal resolution. Data can be interpreted within human comfort zones, where graphical representation is in the form of a map, enabling the link of the environmental conditions with a spatial context. The paper also provides preliminary results in terms of the potential of such tools for identifying the correlations between environmental conditions and different spatial settings, which can help urban planners to prioritize interventions in places. The paper contributes to multidisciplinary approaches as it demonstrates the usefulness of such fine-grained data for better consideration of microclimatology in urban planning, which is a prerequisite for creating climate-comfortable cycling lanes promoting active mobility.

Keywords: Information and communication technology tools, urban planning, human comfort, microclimate, cycling lanes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 461
8320 Feeder Reconfiguration for Loss Reduction in Unbalanced Distribution System Using Genetic Algorithm

Authors: Ganesh. Vulasala, Sivanagaraju. Sirigiri, Ramana. Thiruveedula

Abstract:

This paper presents an efficient approach to feeder reconfiguration for power loss reduction and voltage profile imprvement in unbalanced radial distribution systems (URDS). In this paper Genetic Algorithm (GA) is used to obtain solution for reconfiguration of radial distribution systems to minimize the losses. A forward and backward algorithm is used to calculate load flows in unbalanced distribution systems. By simulating the survival of the fittest among the strings, the optimum string is searched by randomized information exchange between strings by performing crossover and mutation. Results have shown that proposed algorithm has advantages over previous algorithms The proposed method is effectively tested on 19 node and 25 node unbalanced radial distribution systems.

Keywords: Distribution system, Load flows, Reconfiguration, Genetic Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3239
8319 Optimal Feature Extraction Dimension in Finger Vein Recognition Using Kernel Principal Component Analysis

Authors: Amir Hajian, Sepehr Damavandinejadmonfared

Abstract:

In this paper the issue of dimensionality reduction is investigated in finger vein recognition systems using kernel Principal Component Analysis (KPCA). One aspect of KPCA is to find the most appropriate kernel function on finger vein recognition as there are several kernel functions which can be used within PCA-based algorithms. In this paper, however, another side of PCA-based algorithms -particularly KPCA- is investigated. The aspect of dimension of feature vector in PCA-based algorithms is of importance especially when it comes to the real-world applications and usage of such algorithms. It means that a fixed dimension of feature vector has to be set to reduce the dimension of the input and output data and extract the features from them. Then a classifier is performed to classify the data and make the final decision. We analyze KPCA (Polynomial, Gaussian, and Laplacian) in details in this paper and investigate the optimal feature extraction dimension in finger vein recognition using KPCA.

Keywords: Biometrics, finger vein recognition, Principal Component Analysis (PCA), Kernel Principal Component Analysis (KPCA).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957
8318 Wavelet-Based Despeckling of Synthetic Aperture Radar Images Using Adaptive and Mean Filters

Authors: Syed Musharaf Ali, Muhammad Younus Javed, Naveed Sarfraz Khattak

Abstract:

In this paper we introduced new wavelet based algorithm for speckle reduction of synthetic aperture radar images, which uses combination of undecimated wavelet transformation, wiener filter (which is an adaptive filter) and mean filter. Further more instead of using existing thresholding techniques such as sure shrinkage, Bayesian shrinkage, universal thresholding, normal thresholding, visu thresholding, soft and hard thresholding, we use brute force thresholding, which iteratively run the whole algorithm for each possible candidate value of threshold and saves each result in array and finally selects the value for threshold that gives best possible results. That is why it is slow as compared to existing thresholding techniques but gives best results under the given algorithm for speckle reduction.

Keywords: Brute force thresholding, directional smoothing, direction dependent mask, undecimated wavelet transformation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2870
8317 Bitrate Reduction Using FMO for Video Streaming over Packet Networks

Authors: Le Thanh Ha, Hye-Soo Kim, Chun-Su Park, Seung-Won Jung, Sung-Jea Ko

Abstract:

Flexible macroblock ordering (FMO), adopted in the H.264 standard, allows to partition all macroblocks (MBs) in a frame into separate groups of MBs called Slice Groups (SGs). FMO can not only support error-resilience, but also control the size of video packets for different network types. However, it is well-known that the number of bits required for encoding the frame is increased by adopting FMO. In this paper, we propose a novel algorithm that can reduce the bitrate overhead caused by utilizing FMO. In the proposed algorithm, all MBs are grouped in SGs based on the similarity of the transform coefficients. Experimental results show that our algorithm can reduce the bitrate as compared with conventional FMO.

Keywords: Data Partition, Entropy Coding, Greedy Algorithm, H.264/AVC, Slice Group.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1300
8316 Treatment of Tannery Effluents by the Process of Coagulation

Authors: G. Shegani

Abstract:

Coagulation is a process that sanitizes leather effluents. It aims to reduce pollutants such as Chemical Oxygen Demand (COD), chloride, sulfate, chromium, suspended solids, and other dissolved solids. The current study aimed to evaluate coagulation efficiency of tannery wastewater by analyzing the change in organic matter, odor, color, ammonium ions, nutrients, chloride, H2S, sulfate, suspended solids, total dissolved solids, fecal pollution, and chromium hexavalent before and after treatment. Effluent samples were treated with coagulants Ca(OH)2 and FeSO4 .7H2O. The best advantages of this treatment included the removal of: COD (81.60%); ammonia ions (98.34%); nitrate ions (92%); chromium hexavalent (75.00%); phosphate (70.00%); chloride (69.20%); and H₂S (50%). Results also indicated a high level of efficiency in the reduction of fecal pollution indicators. Unfortunately, only a modest reduction of sulfate (19.00%) and TSS (13.00%) and an increase in TDS (15.60%) was observed. 

Keywords: Coagulation, Effluent, Tannery, Treatment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4151
8315 A Novel Model for Simultaneously Minimising Costs and Risks in Just-in-Time Systems Using Multi-Backup Suppliers: Part 2- Results

Authors: Faraj El Dabee, Romeo Marian, Yousef Amer

Abstract:

This paper implements the inventory model developed in the first part of this paper in a simplified problem to simultaneously reduce costs and risks in JIT systems. This model is developed to ascertain an optimal ordering strategy for procuring raw materials by using regular multi-external and local backup suppliers to reduce the total cost of the products, and at the same time to reduce the risks arising from this cost reduction within production systems. A comparison between the cost of using the JIT system and using the proposed inventory model shows the superiority of the use of the inventory model.

Keywords: Lean manufacturing, Just-in-Time (JIT), production system, cost-risk reduction, inventory model, eternal supplier, local backup supplier.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1537
8314 Development and Control of Deep Seated Gravitational Slope Deformation: The Case of Colzate-Vertova Landslide, Bergamo, Northern Italy

Authors: Paola Comella, Vincenzo Francani, Paola Gattinoni

Abstract:

This paper presents the Colzate-Vertova landslide, a Deep Seated Gravitational Slope Deformation (DSGSD) located in the Seriana Valley, Northern Italy. The paper aims at describing the development as well as evaluating the factors that influence the evolution of the landslide. After defining the conceptual model of the landslide, numerical simulations were developed using a finite element numerical model, first with a two-dimensional domain, and later with a three-dimensional one. The results of the 2-D model showed a displacement field typical of a sackung, as a consequence of the erosion along the Seriana Valley. The analysis also showed that the groundwater flow could locally affect the slope stability, bringing about a reduction in the safety factor, but without reaching failure conditions. The sensitivity analysis carried out on the strength parameters pointed out that slope failures could be reached only for relevant reduction of the geotechnical characteristics. Such a result does not fit the real conditions observed on site, where a number of small failures often develop all along the hillslope. The 3-D model gave a more comprehensive analysis of the evolution of the DSGSD, also considering the border effects. The results showed that the convex profile of the slope favors the development of displacements along the lateral valley, with a relevant reduction in the safety factor, justifying the existing landslides.

Keywords: Deep seated gravitational slope deformation, Italy, landslide, numerical modeling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1019
8313 Optimizing Telehealth Internet of Things Integration: A Sustainable Approach through Fog and Cloud Computing Platforms for Energy Efficiency

Authors: Yunyong Guo, Sudhakar Ganti, Bryan Guo

Abstract:

The swift proliferation of telehealth Internet of Things (IoT) devices has sparked concerns regarding energy consumption and the need for streamlined data processing. This paper presents an energy-efficient model that integrates telehealth IoT devices into a platform based on fog and cloud computing. This integrated system provides a sustainable and robust solution to address the challenges. Our model strategically utilizes fog computing as a localized data processing layer and leverages cloud computing for resource-intensive tasks, resulting in a significant reduction in overall energy consumption. The incorporation of adaptive energy-saving strategies further enhances the efficiency of our approach. Simulation analysis validates the effectiveness of our model in improving energy efficiency for telehealth IoT systems, particularly when integrated with localized fog nodes and both private and public cloud infrastructures. Subsequent research endeavors will concentrate on refining the energy-saving model, exploring additional functional enhancements, and assessing its broader applicability across various healthcare and industry sectors.

Keywords: Energy-efficient, fog computing, IoT, telehealth.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 56
8312 Effect of Windrow Management on Ammonia and Nitrous Oxide Emissions from Swine Manure Composting

Authors: Nanh Lovanh, John Loughrin, Kimberly Cook, Phil Silva, Byung-Taek Oh

Abstract:

In the era of sustainability, utilization of livestock wastes as soil amendment to provide micronutrients for crops is very economical and sustainable. It is well understood that livestock wastes are comparable, if not better, nutrient sources for crops as chemical fertilizers. However, the large concentrated volumes of animal manure produced from livestock operations and the limited amount of available nearby agricultural land areas necessitated the need for volume reduction of these animal wastes. Composting of these animal manures is a viable option for biomass and pathogenic reduction in the environment. Nevertheless, composting also increases the potential loss of available nutrients for crop production as well as unwanted emission of anthropogenic air pollutants due to the loss of ammonia and other compounds via volatilization. In this study, we examine the emission of ammonia and nitrous oxide from swine manure windrows to evaluate the benefit of biomass reduction in conjunction with the potential loss of available nutrients. The feedstock for the windrows was obtained from swine farm in Kentucky where swine manure was mixed with wood shaving as absorbent material. Static flux chambers along with photoacoustic gas analyzer were used to monitor ammonia and nitrous oxide concentrations during the composting process. The results show that ammonia and nitrous oxide fluxes were quite high during the initial composting process and after the turning of each compost pile. Over the period of roughly three months of composting, the biochemical oxygen demand (BOD) decreased by about 90%. Although composting of animal waste is quite beneficial for biomass reduction, composting may not be economically feasible from an agronomical point of view due to time, nutrient loss (N loss), and potential environmental pollution (ammonia and greenhouse gas emissions). Therefore, additional studies are needed to assess and validate the economics and environmental impact of animal (swine) manure composting (e.g., crop yield or impact on climate change).

Keywords: Windrow, swine manure, ammonia, nitrous oxide, fluxes, management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1966
8311 Observations about the Principal Components Analysis and Data Clustering Techniques in the Study of Medical Data

Authors: Cristina G. Dascâlu, Corina Dima Cozma, Elena Carmen Cotrutz

Abstract:

The medical data statistical analysis often requires the using of some special techniques, because of the particularities of these data. The principal components analysis and the data clustering are two statistical methods for data mining very useful in the medical field, the first one as a method to decrease the number of studied parameters, and the second one as a method to analyze the connections between diagnosis and the data about the patient-s condition. In this paper we investigate the implications obtained from a specific data analysis technique: the data clustering preceded by a selection of the most relevant parameters, made using the principal components analysis. Our assumption was that, using the principal components analysis before data clustering - in order to select and to classify only the most relevant parameters – the accuracy of clustering is improved, but the practical results showed the opposite fact: the clustering accuracy decreases, with a percentage approximately equal with the percentage of information loss reported by the principal components analysis.

Keywords: Data clustering, medical data, principal components analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1492
8310 Dynamic Features Selection for Heart Disease Classification

Authors: Walid MOUDANI

Abstract:

The healthcare environment is generally perceived as being information rich yet knowledge poor. However, there is a lack of effective analysis tools to discover hidden relationships and trends in data. In fact, valuable knowledge can be discovered from application of data mining techniques in healthcare system. In this study, a proficient methodology for the extraction of significant patterns from the Coronary Heart Disease warehouses for heart attack prediction, which unfortunately continues to be a leading cause of mortality in the whole world, has been presented. For this purpose, we propose to enumerate dynamically the optimal subsets of the reduced features of high interest by using rough sets technique associated to dynamic programming. Therefore, we propose to validate the classification using Random Forest (RF) decision tree to identify the risky heart disease cases. This work is based on a large amount of data collected from several clinical institutions based on the medical profile of patient. Moreover, the experts- knowledge in this field has been taken into consideration in order to define the disease, its risk factors, and to establish significant knowledge relationships among the medical factors. A computer-aided system is developed for this purpose based on a population of 525 adults. The performance of the proposed model is analyzed and evaluated based on set of benchmark techniques applied in this classification problem.

Keywords: Multi-Classifier Decisions Tree, Features Reduction, Dynamic Programming, Rough Sets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2524
8309 Improving Sales through Inventory Reduction: A Retail Chain Case Study

Authors: M. G. Mattos, J. E. Pécora Jr, T. A. Briso

Abstract:

Today's challenging business environment, with unpredictable demand and volatility, requires a supply chain strategy that handles uncertainty and risks in the right way. Even though inventory models have been previously explored, this paper seeks to apply these concepts on a practical situation. This study involves the inventory replenishment problem, applying techniques that are mainly based on mathematical assumptions and modeling. The primary goal is to improve the retailer’s supply chain processes taking store differences when setting the various target stock levels. Through inventory review policy, picking piece implementation and minimum exposure definition, we were able not only to promote the inventory reduction as well as improve sales results. The inventory management theory from literature review was then tested on a single case study regarding a particular department in one of the largest Latam retail chains.

Keywords: Inventory, distribution, retail, risk, safety stock, sales, uncertainty.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1806
8308 Investigation of Tool Temperature and Surface Quality in Hot Machining of Hard-to-Cut Materials

Authors: M.Davami, M.Zadshakoyan

Abstract:

Production of hard-to-cut materials with uncoated carbide cutting tools in turning, not only cause tool life reduction but also, impairs the product surface roughness. In this paper, influence of hot machining method were studied and presented in two cases. Case1-Workpiece surface roughness quality with constant cutting parameter and 300 ºC initial workpiece surface temperature. Case 2- Tool temperature variation when cutting with two speeds 78.5 (m/min) and 51 (m/min). The workpiece material and tool used in this study were AISI 1060 steel (45HRC) and uncoated carbide TNNM 120408-SP10(SANDVIK Coromant) respectively. A gas flam heating source was used to preheating of the workpiece surface up to 300 ºC, causing reduction of yield stress about 15%. Results obtained experimentally, show that the method used can considerably improved surface quality of the workpiece.

Keywords: Hard-to-cut material, Hot machining, Surfaceroughness, Tool Temperature

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2189