Search results for: Minimizing Cost
634 Solid Waste Management through Mushroom Cultivation – An Eco Friendly Approach
Authors: Mary Josephine
Abstract:
Waste of certain process can be the input source of other sectors in order to reduce environmental pollution. Today there are more and more solid wastes are generated, but only very small amount of those are recycled. So, the threatening of environmental pressure to public health is very serious. The methods considered for the treatment of solid waste are biogas tanks or processing to make animal feed and fertilizer, however, they did not perform well. An alternative approach is growing mushrooms on waste residues. This is regarded as an environmental friendly solution with potential economical benefit. The substrate producers do their best to produce quality substrate at low cost. Apart from other methods, this can be achieved by employing biologically degradable wastes used as the resource material component of the substrate. Mushroom growing is a significant tool for the restoration, replenishment and remediation of Earth’s overburdened ecosphere. One of the rational methods of waste utilization involves locally available wastes. The present study aims to find out the yield of mushroom grown on locally available waste for free and to conserve our environment by recycling wastes.
Keywords: Biodegradable, environment, mushroom, remediation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5313633 Low Cost Real Time Robust Identification of Impulsive Signals
Authors: R. Biondi, G. Dys, G. Ferone, T. Renard, M. Zysman
Abstract:
This paper describes an automated implementable system for impulsive signals detection and recognition. The system uses a Digital Signal Processing device for the detection and identification process. Here the system analyses the signals in real time in order to produce a particular response if needed. The system analyses the signals in real time in order to produce a specific output if needed. Detection is achieved through normalizing the inputs and comparing the read signals to a dynamic threshold and thus avoiding detections linked to loud or fluctuating environing noise. Identification is done through neuronal network algorithms. As a setup our system can receive signals to “learn” certain patterns. Through “learning” the system can recognize signals faster, inducing flexibility to new patterns similar to those known. Sound is captured through a simple jack input, and could be changed for an enhanced recording surface such as a wide-area recorder. Furthermore a communication module can be added to the apparatus to send alerts to another interface if needed.
Keywords: Sound Detection, Impulsive Signal, Background Noise, Neural Network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2335632 Artificial Intelligence in Penetration Testing of a Connected and Autonomous Vehicle Network
Authors: Phillip Garrad, Saritha Unnikrishnan
Abstract:
The increase in connected and autonomous vehicles (CAV) creates more opportunities for cyber-attacks. Cyber-attacks can be performed with malicious intent or for research and testing purposes. As connected vehicles approach full autonomy, the possible impact of these cyber-attacks also grows. This review analyses the challenges faced in CAV cybersecurity testing. This includes access and cost of the representative test setup and lack of experts in the field A review of potential solutions to overcome these challenges is presented. Studies have demonstrated Artificial Intelligence (AI) as a promising technique to reduce runtime, enhance effectiveness and comprehensively cover all the standard test aspects in penetration testing in other industries. However, this review has identified a significant gap in the systematic implementation of AI for penetration testing in the CAV cybersecurity domain. The expectation from this review is to investigate potential AI algorithms, which can demonstrate similar improvements in runtime and efficiency for a CAV model. If proven to be an effective means of penetration test for CAV, this methodology may be used on a full CAV test network.
Keywords: Cybersecurity, connected vehicles, software simulation, artificial intelligence, penetration testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 490631 Comparison of Electrical Parameters of Oil-Immersed and Dry-Type Transformer Using Finite Element Method
Authors: U. Amin, A. Talib, S. A. Qureshi, M. J. Hossain, G. Ahmad
Abstract:
The choice evaluation between oil-immersed and dry-type transformers is often controlled by cost, location, and application. This paper compares the electrical performance of liquid- filled and dry-type transformers, which will assist the customer to choose the right and efficient ones for particular applications. An accurate assessment of the time-average flux density, electric field intensity and voltage distribution in an oil-insulated and a dry-type transformer have been computed and investigated. The detailed transformer modeling and analysis has been carried out to determine electrical parameter distributions. The models of oil-immersed and dry-type transformers are developed and solved by using the finite element method (FEM) to compare the electrical parameters. The effects of non-uniform and non-coherent voltage gradient, flux density and electric field distribution on the power losses and insulation properties of transformers are studied in detail. The results show that, for the same voltage and kilo-volt-ampere (kVA) rating, oil-immersed transformers have better insulation properties and less hysteresis losses than the dry-type.
Keywords: Finite element method, flux density, transformer, voltage gradient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1233630 Using the Simple Fixed Rate Approach to Solve Economic Lot Scheduling Problem under the Basic Period Approach
Authors: Yu-Jen Chang, Yun Chen, Hei-Lam Wong
Abstract:
The Economic Lot Scheduling Problem (ELSP) is a valuable mathematical model that can support decision-makers to make scheduling decisions. The basic period approach is effective for solving the ELSP. The assumption for applying the basic period approach is that a product must use its maximum production rate to be produced. However, a product can lower its production rate to reduce the average total cost when a facility has extra idle time. The past researches discussed how a product adjusts its production rate under the common cycle approach. To the best of our knowledge, no studies have addressed how a product lowers its production rate under the basic period approach. This research is the first paper to discuss this topic. The research develops a simple fixed rate approach that adjusts the production rate of a product under the basic period approach to solve the ELSP. Our numerical example shows our approach can find a better solution than the traditional basic period approach. Our mathematical model that applies the fixed rate approach under the basic period approach can serve as a reference for other related researches.Keywords: Economic Lot, Basic Period, Genetic Algorithm, Fixed Rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1940629 An Off-the-Shelf Scheme for Dependable Grid Systems Using Virtualization
Authors: Toshinori Takabatake
Abstract:
Recently, grid computing has been widely focused on the science, industry, and business fields, which are required a vast amount of computing. Grid computing is to provide the environment that many nodes (i.e., many computers) are connected with each other through a local/global network and it is available for many users. In the environment, to achieve data processing among nodes for any applications, each node executes mutual authentication by using certificates which published from the Certificate Authority (for short, CA). However, if a failure or fault has occurred in the CA, any new certificates cannot be published from the CA. As a result, a new node cannot participate in the gird environment. In this paper, an off-the-shelf scheme for dependable grid systems using virtualization techniques is proposed and its implementation is verified. The proposed approach using the virtualization techniques is to restart an application, e.g., the CA, if it has failed. The system can tolerate a failure or fault if it has occurred in the CA. Since the proposed scheme is implemented at the application level easily, the cost of its implementation by the system builder hardly takes compared it with other methods. Simulation results show that the CA in the system can recover from its failure or fault.Keywords: grid computing, restarting application, certificate authority, virtualization, dependability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1378628 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Disease
Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang
Abstract:
Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.
Keywords: Alzheimer’s disease, Speech Emotion Recognition, longitudinal biomarker, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 275627 Mobile Multicast Support using Old Foreign Agent (MMOFA)
Authors: Hamed Rajabi, Naser Nematbakhsh, Naser Movahediniya
Abstract:
IP multicasting is a key technology for many existing and emerging applications on the Internet. Furthermore, with increasing popularity of wireless devices and mobile equipment, it is necessary to determine the best way to provide this service in a wireless environment. IETF Mobile IP, that provides mobility for hosts in IP networks, proposes two approaches for mobile multicasting, namely, remote subscription (MIP-RS) and bi-directional tunneling (MIP-BT). In MIP-RS, a mobile host re-subscribes to the multicast groups each time it moves to a new foreign network. MIP-RS suffers from serious packet losses while mobile host handoff occurs. In MIP-BT, mobile hosts send and receive multicast packets by way of their home agents (HAs), using Mobile IP tunnels. Therefore, it suffers from inefficient routing and wastage of system resources. In this paper, we propose a protocol called Mobile Multicast support using Old Foreign Agent (MMOFA) for Mobile Hosts. MMOFA is derived from MIP-RS and with the assistance of Mobile host's Old foreign agent, routes the missing datagrams due to handoff in adjacent network via tunneling. Also, we studied the performance of the proposed protocol by simulation under ns-2.27. The results demonstrate that MMOFA has optimal routing efficiency and low delivery cost, as compared to other approaches.Keywords: Mobile Multicast, Mobile IP, MMOFA, NS-2. 27.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1469626 Production of Eicosapentaenoic Acid and Fucoxanthin in Cold-Tolerant Diatom Strains
Authors: Nga Phuong Dang, Terje Vasskog, Ashwiny Pandey, Rajnish Kaur Calay
Abstract:
Diatoms hold great potential for nutraceutical production as they are source of high value products such as eicosapentaenoic acid (EPA) and pigment fucoxanthin. EPA has proven human health benefits, and fucoxanthin can be used as both medicinal and nutritional ingredient to prevent and treat chronic diseases. The economic perspective of commercial production of a single product from microalgae is not attractive due to the high production cost. To improve the economic viability, we explore the concept of combining the production of both EPA and fucoxanthin in a single process. In our current study, we isolated twelve new microalgae isolates from Ofotfjord. Eight of them are diatoms and 4 of them are cyanobacteria and microalgae. Screening the new diatom isolates revealed that two strains are cold-tolerant diatom which can grow at 10 °C. They accumulated significant amount of lipid, which was up to 40-60% of the dried mass. The EPA contents from the two strains ranged from 15-19% of the total fatty acid, while fucoxanthin concentrations were between 1-1.4% of the dried biomass. Comparing with other studied diatom, this is promising result. The two strains hold promise as source for EPA and fucoxanthin production.
Keywords: Microalgae, fucoxanthin, eicosapentaenoic acid, diatom, fatty acid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 538625 Reliability Verification of the Performance Evaluation of Multiphase Pump
Authors: Joon-Hyung Kim, Him-Chan Lee, Jin-Hyuk Kim, Yong-Kab Lee, Young-Seok Choi
Abstract:
The crude oil in an oil well exists in various phases such as gas, seawater, and sand, as well as oil. Therefore, a phase separator is needed at the front of a single-phase pump for pressurization and transfer. On the other hand, the application of a multiphase pump can provide such advantages as simplification of the equipment structure and cost savings, because there is no need for a phase separation process. Therefore, the crude oil transfer method using a multiphase pump is being applied to recently developed oil wells. Due to this increase in demand, technical demands for the development of multiphase pumps are sharply increasing, but the progress of research into related technologies is insufficient, due to the nature of multiphase pumps that require high levels of skills. This study was conducted to verify the reliability of pump performance evaluation using numerical analysis, which is the basis of the development of a multiphase pump. For this study, a model was designed by selecting the specifications of this study. The performance of the designed model was evaluated through numerical analysis and experiment. The results of the performance evaluation were compared to verify the reliability of the result using numerical analysis.
Keywords: Multiphase pump, Numerical analysis, Experiment, Performance evaluation, Reliability verification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3134624 Analyzing the Market Growth in API Economy Using Time-Evolving Model
Authors: Hiroki Yoshikai, Shin’ichi Arakawa, Tetsuya Takine, Masayuki Murata
Abstract:
API (Application Programming Interface) economy is expected to create new value by converting corporate services such as information processing and data provision into APIs and using these APIs to connect services. Understanding dynamics of a market of API economy under strategies of participants is crucial to fully maximize the values of API economy. To capture the behavior of a market in which the number of participants changes over time, we present a time-evolving market model for a platform in which API providers who provide APIs to service providers participate in addition to service providers and consumers. Then, we use the market model to clarify the role API providers play in expanding market participants and forming ecosystems. The results show that the platform with API providers increased the number of market participants by 67% and decreased the cost to develop services by 25% compared to the platform without API providers. Furthermore, during the expansion phase of the market, it is found that the profits of participants are mostly the same when 70% of the revenue from consumers is distributed to service providers and API providers. It is also found that, when the market is mature, the profits of the service provider and API provider will decrease significantly due to their competitions and the profit of the platform increases.
Keywords: API Economy, ecosystem, platform, API providers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 244623 Optimal Design of Airfoil Platform Shapes with High Aspect Ratio Using Genetic Algorithm
Authors: Kyoungwoo Park, Byeong-Sam Kim
Abstract:
Unmanned aerial vehicles (UAVs) performing their operations for a long time have been attracting much attention in military and civil aviation industries for the past decade. The applicable field of UAV is changing from the military purpose only to the civil one. Because of their low operation cost, high reliability and the necessity of various application areas, numerous development programs have been initiated around the world. To obtain the optimal solutions of the design variable (i.e., sectional airfoil profile, wing taper ratio and sweep) for high performance of UAVs, both the lift and lift-to-drag ratio are maximized whereas the pitching moment should be minimized, simultaneously. It is found that the lift force and lift-to-drag ratio are linearly dependent and a unique and dominant solution are existed. However, a trade-off phenomenon is observed between the lift-to-drag ratio and pitching moment. As the result of optimization, sixty-five (65) non-dominated Pareto individuals at the cutting edge of design spaces that are decided by airfoil shapes can be obtained.Keywords: Unmanned aerial vehicle (UAV), Airfoil, CFD, Shape optimization, Genetic Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959622 How the Conversations in Social Media Concern in Sales in the Automobile Industry in Spain
Authors: González Carreño G., Rejas Muslera Rj., Padilla Fernández-Vega J., Cepeda González M. I.
Abstract:
Automobile Industry has great importance in the Spanish economy (8,7 % of the active Spanish population is employed in this sector).The above mentioned sector has been one of the principal sectors affected by the current economic crisis, consistently, the budgets in advertising have been severely limited (46,9 % less in the period of reference), these needs of reduction have originated a substantial change in the advertising strategy (from 2007 the increase of the advertising investment in Internet is 251,6 %), and increase profitability. The growing use of social media by consumers therefore makes online consumer conversations an attractive additional format for Automobile firms to promote products at a lower cost. This research analyzes the relation between the activity in Social Media and the design in the car industry, looking for relations between strategies of design based on Social Media and sales and a channel of information for companies to know what the consumer preferences. For this ongoing research we used a longitudinal withdrawal of information has been used using information of panel. Managerial and research implications of the finding are discussed.Keywords: Automobile Industry, Design, Economics Crisis, Innovation, Internet, Social Media.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930621 Effect of Alkali Treatment on Impact Behavior of Areca Fibers Reinforced Polymer Composites
Authors: Srinivasa C. V., Bharath K. N.
Abstract:
Natural fibers are considered to have potential use as reinforcing agents in polymer composite materials because of their principal benefits: moderate strength and stiffness, low cost, and being an environmental friendly, degradable, and renewable material. A study has been carried out to evaluate impact properties of composites made by areca fibers reinforced urea formaldehyde, melamine urea formaldehyde and epoxy resins. The extracted areca fibers from the areca husk were alkali treated with potassium hydroxide (KOH) to obtain better interfacial bonding between fiber and matrix. Then composites were produced by means of compression molding technique with varying process parameters, such as fiber condition (untreated and alkali treated), and fiber loading percentages (50% and 60% by weight). The developed areca fiber reinforced composites were then characterized by impact test. The results show that, impact strength increase with increase in the loading percentage. It is observed that, treated areca fiber reinforcement increases impact strength when compared to untreated areca fiber reinforcement.
Keywords: Lignocellulosic Fibers Composites, Areca Fibers, Alkali Treatment, Impact Strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3783620 A Two Level Load Balancing Approach for Cloud Environment
Authors: Anurag Jain, Rajneesh Kumar
Abstract:
Cloud computing is the outcome of rapid growth of internet. Due to elastic nature of cloud computing and unpredictable behavior of user, load balancing is the major issue in cloud computing paradigm. An efficient load balancing technique can improve the performance in terms of efficient resource utilization and higher customer satisfaction. Load balancing can be implemented through task scheduling, resource allocation and task migration. Various parameters to analyze the performance of load balancing approach are response time, cost, data processing time and throughput. This paper demonstrates a two level load balancer approach by combining join idle queue and join shortest queue approach. Authors have used cloud analyst simulator to test proposed two level load balancer approach. The results are analyzed and compared with the existing algorithms and as observed, proposed work is one step ahead of existing techniques.
Keywords: Cloud Analyst, Cloud Computing, Join Idle Queue, Join Shortest Queue, Load balancing, Task Scheduling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1992619 A Consistency Protocol Multi-Layer for Replicas Management in Large Scale Systems
Authors: Ghalem Belalem, Yahya Slimani
Abstract:
Large scale systems such as computational Grid is a distributed computing infrastructure that can provide globally available network resources. The evolution of information processing systems in Data Grid is characterized by a strong decentralization of data in several fields whose objective is to ensure the availability and the reliability of the data in the reason to provide a fault tolerance and scalability, which cannot be possible only with the use of the techniques of replication. Unfortunately the use of these techniques has a height cost, because it is necessary to maintain consistency between the distributed data. Nevertheless, to agree to live with certain imperfections can improve the performance of the system by improving competition. In this paper, we propose a multi-layer protocol combining the pessimistic and optimistic approaches conceived for the data consistency maintenance in large scale systems. Our approach is based on a hierarchical representation model with tree layers, whose objective is with double vocation, because it initially makes it possible to reduce response times compared to completely pessimistic approach and it the second time to improve the quality of service compared to an optimistic approach.Keywords: Data Grid, replication, consistency, optimistic approach, pessimistic approach.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575618 An Improved Particle Swarm Optimization Technique for Combined Economic and Environmental Power Dispatch Including Valve Point Loading Effects
Authors: Badr M. Alshammari, T. Guesmi
Abstract:
In recent years, the combined economic and emission power dispatch is one of the main problems of electrical power system. It aims to schedule the power generation of generators in order to minimize cost production and emission of harmful gases caused by fossil-fueled thermal units such as CO, CO2, NOx, and SO2. To solve this complicated multi-objective problem, an improved version of the particle swarm optimization technique that includes non-dominated sorting concept has been proposed. Valve point loading effects and system losses have been considered. The three-unit and ten-unit benchmark systems have been used to show the effectiveness of the suggested optimization technique for solving this kind of nonconvex problem. The simulation results have been compared with those obtained using genetic algorithm based method. Comparison results show that the proposed approach can provide a higher quality solution with better performance.
Keywords: Power dispatch, valve point loading effects, multiobjective optimization, Pareto solutions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 770617 Assesing Extension of Meeting System Performance in Information Technology in Defense and Aerospace Project
Authors: Hakan Gürkan, Ahmet Denker
Abstract:
The Ministry of Defense (MoD) spends hundreds of millions of dollars on software to support its infrastructure, operate its weapons and provide command, control, communications, computing, intelligence, surveillance, and reconnaissance (C4ISR) functions. These and other all new advanced systems have a common critical component is information technology. Defense and Aerospace environment is continuously striving to keep up with increasingly sophisticated Information Technology (IT) in order to remain effective in today-s dynamic and unpredictable threat environment. This makes it one of the largest and fastest growing expenses of Defense. Hundreds of millions of dollars spent a year on IT projects. But, too many of those millions are wasted on costly mistakes. Systems that do not work properly, new components that are not compatible with old once, trendily new applications that do not really satisfy defense needs or lost though poorly managed contracts. This paper investigates and compiles the effective strategies that aim to end exasperation with low returns and high cost of Information Technology Acquisition for defense; it tries to show how to maximize value while reducing time and expenditure.Keywords: Iterative Process, Acquisition Management, Project management, Software Economics, Requirement analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1243616 Estimating Marine Tidal Power Potential in Kenya
Authors: Lucy Patricia Onundo, Wilfred Njoroge Mwema
Abstract:
The rapidly diminishing fossil fuel reserves, their exorbitant cost and the increasingly apparent negative effect of fossil fuels to climate changes is a wake-up call to explore renewable energy. Wind, bio-fuel and solar power have already become staples of Kenyan electricity mix. The potential of electric power generation from marine tidal currents is enormous, with oceans covering more than 70% of the earth. However, attempts to harness marine tidal energy in Kenya, has yet to be studied thoroughly due to its promising, cyclic, reliable and predictable nature and the vast energy contained within it. The high load factors resulting from the fluid properties and the predictable resource characteristics make marine currents particularly attractive for power generation and advantageous when compared to others. Global-level resource assessments and oceanographic literature and data have been compiled in an analysis of the technology-specific requirements for tidal energy technologies and the physical resources. Temporal variations in resource intensity as well as the differences between small-scale applications are considered.
Keywords: Energy data assessment, environmental legislation, renewable energy, tidal-in-stream turbines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1368615 Optimization of Surface Roughness and Vibration in Turning of Aluminum Alloy AA2024 Using Taguchi Technique
Authors: Vladimir Aleksandrovich Rogov, Ghorbani Siamak
Abstract:
Determination of optimal conditions of machining parameters is important to reduce the production cost and achieve the desired surface quality. This paper investigates the influence of cutting parameters on surface roughness and natural frequency in turning of aluminum alloy AA2024. The experiments were performed at the lathe machine using two different cutting tools made of AISI 5140 and carbide cutting insert coated with TiC. Turning experiments were planned by Taguchi method L9 orthogonal array.Three levels for spindle speed, feed rate, depth of cut and tool overhang were chosen as cutting variables. The obtained experimental data has been analyzed using signal to noise ratio and analysis of variance. The main effects have been discussed and percentage contributions of various parameters affecting surface roughness and natural frequency, and optimal cutting conditions have been determined. Finally, optimization of the cutting parameters using Taguchi method was verified by confirmation experiments.
Keywords: Turning, Cutting conditions, Surface roughness, Natural frequency, Taguchi method, ANOVA, S/N ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4616614 Decoy-pulse Protocol for Frequency-coded Quantum Key Distribution
Authors: Sudeshna Bhattacharya, Pratyush Pandey, Pradeep Kumar K
Abstract:
We propose a decoy-pulse protocol for frequency-coded implementation of B92 quantum key distribution protocol. A direct extension of decoy-pulse method to frequency-coding scheme results in security loss as an eavesdropper can distinguish between signal and decoy pulses by measuring the carrier photon number without affecting other statistics. We overcome this problem by optimizing the ratio of carrier photon number of decoy-to-signal pulse to be as close to unity as possible. In our method the switching between signal and decoy pulses is achieved by changing the amplitude of RF signal as opposed to modulating the intensity of optical signal thus reducing system cost. We find an improvement by a factor of 100 approximately in the key generation rate using decoy-state protocol. We also study the effect of source fluctuation on key rate. Our simulation results show a key generation rate of 1.5×10-4/pulse for link lengths up to 70km. Finally, we discuss the optimum value of average photon number of signal pulse for a given key rate while also optimizing the carrier ratio.
Keywords: B92, decoy-pulse, frequency-coding, quantum key distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1727613 Risk Assessment in Durations and Costs for Construction of Industrial Facilities in Egypt Using Equations and Computer
Authors: M. Kamal Elbokl, Negadi Kheira
Abstract:
Risk Evaluation is an important step in protecting your workers and your business, as well as complying with the law. It helps you focus on the risks that really matter in your workplace – the ones with the potential to cause real harm. We are in this paper introduce basics of risk assessment then we mention some of ways to risk evaluation by computer especially Monte Carlo simulation and Microsoft project.
We use Program Evaluation and Review Technique (PERT) to deal with Risks in Industrial Facilities in Evaluation and Assessment for this risk. Using PERT Technique in Microsoft Project by the PERT toolbar and using PERTMASTER Program with Primavera Program we evaluate many hazards and make calculations for that by mathematical equation to make right decisions. We define and calculate risk factor and risk severity to ranking the type of the risk then dealing with it using in that many ways like probability computation, curves, and tables. By introducing variables in the equation of functions in computer programs we calculate the risk in the time and the cost in general case and then mention some examples in industrial facilities field.
Keywords: Risk, Industrial Facilities, PERT, Monte Carlo Simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1954612 Minimization Entropic Applied to Rotary Dryers to Reduce the Energy Consumption
Authors: I. O. Nascimento, J. T. Manzi
Abstract:
The drying process is an important operation in the chemical industry and it is widely used in the food, grain industry and fertilizer industry. However, for demanding a considerable consumption of energy, such a process requires a deep energetic analysis in order to reduce operating costs. This paper deals with thermodynamic optimization applied to rotary dryers based on the entropy production minimization, aiming at to reduce the energy consumption. To do this, the mass, energy and entropy balance was used for developing a relationship that represents the rate of entropy production. The use of the Second Law of Thermodynamics is essential because it takes into account constraints of nature. Since the entropy production rate is minimized, optimals conditions of operations can be established and the process can obtain a substantial gain in energy saving. The minimization strategy had been led using classical methods such as Lagrange multipliers and implemented in the MATLAB platform. As expected, the preliminary results reveal a significant energy saving by the application of the optimal parameters found by the procedure of the entropy minimization It is important to say that this method has shown easy implementation and low cost.Keywords: Drying, entropy minimization, modeling dryers, thermodynamic optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1430611 Secure Hashing Algorithm and Advance Encryption Algorithm in Cloud Computing
Authors: Jaimin Patel
Abstract:
Cloud computing is one of the most sharp and important movement in various computing technologies. It provides flexibility to users, cost effectiveness, location independence, easy maintenance, enables multitenancy, drastic performance improvements, and increased productivity. On the other hand, there are also major issues like security. Being a common server, security for a cloud is a major issue; it is important to provide security to protect user’s private data, and it is especially important in e-commerce and social networks. In this paper, encryption algorithms such as Advanced Encryption Standard algorithms, their vulnerabilities, risk of attacks, optimal time and complexity management and comparison with other algorithms based on software implementation is proposed. Encryption techniques to improve the performance of AES algorithms and to reduce risk management are given. Secure Hash Algorithms, their vulnerabilities, software implementations, risk of attacks and comparison with other hashing algorithms as well as the advantages and disadvantages between hashing techniques and encryption are given.
Keywords: Cloud computing, encryption algorithm, secure hashing algorithm, brute force attack, birthday attack, plaintext attack, man-in-the-middle attack.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1748610 Optimization of Tolerance Grades of a Bearing and Shaft Assembly in a Washing Machine with Regard to Fatigue Life
Authors: M. Cangi, T. Dolar, C. Ersoy, Y. E. Aydogdu, A. I. Aydeniz, A. Mugan
Abstract:
The drum is one of the critical parts in a washing machine in which the clothes are washed and spin by the rotational movement. It is activated by the drum shaft which is attached to an electric motor and subjected to dynamic loading. Being one of the critical components, failures of the drum require costly repairs of dynamic components. In this study, tolerance bands between the drum shaft and its two bearings were examined to develop a relationship between the fatigue life of the shaft and the interaction tolerances. Optimization of tolerance bands was completed in consideration of the fatigue life of the shaft as the cost function. The following methodology is followed: multibody dynamic model of a washing machine was constructed and used to calculate dynamic loading on the components. Then, these forces were used in finite element analyses to calculate the stress field in critical components which was used for fatigue life predictions. The factors affecting the fatigue life were examined to find optimum tolerance grade for a given test condition. Numerical results were verified by experimental observations.
Keywords: Fatigue life, finite element analysis, tolerance analysis, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 933609 The Impact of Semantic Web on E-Commerce
Authors: Karim Heidari
Abstract:
Semantic Web Technologies enable machines to interpret data published in a machine-interpretable form on the web. At the present time, only human beings are able to understand the product information published online. The emerging semantic Web technologies have the potential to deeply influence the further development of the Internet Economy. In this paper we propose a scenario based research approach to predict the effects of these new technologies on electronic markets and business models of traders and intermediaries and customers. Over 300 million searches are conducted everyday on the Internet by people trying to find what they need. A majority of these searches are in the domain of consumer ecommerce, where a web user is looking for something to buy. This represents a huge cost in terms of people hours and an enormous drain of resources. Agent enabled semantic search will have a dramatic impact on the precision of these searches. It will reduce and possibly eliminate information asymmetry where a better informed buyer gets the best value. By impacting this key determinant of market prices semantic web will foster the evolution of different business and economic models. We submit that there is a need for developing these futuristic models based on our current understanding of e-commerce models and nascent semantic web technologies. We believe these business models will encourage mainstream web developers and businesses to join the “semantic web revolution."Keywords: E-Commerce, E-Business, Semantic Web, XML.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3461608 Functionalization of Polypropylene with Chiral Monomer for Improving Hemocompatibility
Authors: Xiaodong Xu, Dan Zhao, Xiujuan Chang, Chunming Li, Huiyun Zhou, Xin Li, Qiang Shi, Shifang Luan, Jinghua Yin
Abstract:
Polypropylene (PP) is one of the most commonly used plastics because of its low density, outstanding mechanical properties, and low cost. However, its drawbacks such as low surface energy, poor dyeability, lack of chemical functionalities, and poor compatibility with polar polymers and inorganic materials, have restricted the application of PP. To expand its application in biomedical materials, functionalization is considered to be the most effective way. In this study, PP was functionalized with a chiral monomer, (S)-1-acryloylpyrrolidine-2-carboxylic acid ((S)-APCA), by free-radical grafting in the solid phase. The grafting degree of PP-g-APCA was determined by chemical titration method, and the chemical structure of functionalized PP was characterized by FTIR spectroscopy, which confirmed that the chiral monomer (S)-APCA was successfully grafted onto PP. Static water contact angle results suggested that the surface hydrophilicity of PP was significantly improved by solid phase grafting and assistance of surface water treatment. Protein adsorption and platelet adhesion results showed that hemocompatibility of PP was greatly improved by grafting the chiral monomer.
Keywords: Functionalization, polypropylene, chiral monomer, hemocompatibility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1224607 Estimation of Buffer Size of Internet Gateway Server via G/M/1 Queuing Model
Authors: Dr. L.K. Singh, Dr. R. M. L, Riktesh Srivastava
Abstract:
How to efficiently assign system resource to route the Client demand by Gateway servers is a tricky predicament. In this paper, we tender an enhanced proposal for autonomous recital of Gateway servers under highly vibrant traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.Keywords: Gateway Server, G/M/1 Queuing Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1595606 Automatic Reusability Appraisal of Software Components using Neuro-fuzzy Approach
Authors: Parvinder S. Sandhu, Hardeep Singh
Abstract:
Automatic reusability appraisal could be helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable components from existing legacy systems; that can save cost of developing the software from scratch. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. In this paper, we have mentioned two-tier approach by studying the structural attributes as well as usability or relevancy of the component to a particular domain. Latent semantic analysis is used for the feature vector representation of various software domains. It exploits the fact that FeatureVector codes can be seen as documents containing terms -the idenifiers present in the components- and so text modeling methods that capture co-occurrence information in low-dimensional spaces can be used. Further, we devised Neuro- Fuzzy hybrid Inference System, which takes structural metric values as input and calculates the reusability of the software component. Decision tree algorithm is used to decide initial set of fuzzy rules for the Neuro-fuzzy system. The results obtained are convincing enough to propose the system for economical identification and retrieval of reusable software components.Keywords: Clustering, ID3, LSA, Neuro-fuzzy System, SVD
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662605 Effects of Human Factors on Workforce Scheduling
Authors: M. Othman, N. Bhuiyan, G. J. Gouw
Abstract:
In today-s competitive market, most companies develop manufacturing systems that can help in cost reduction and maximum quality. Human issues are an important part of manufacturing systems, yet most companies ignore their effects on production performance. This paper aims to developing an integrated workforce planning system that incorporates the human being. Therefore, a multi-objective mixed integer nonlinear programming model is developed to determine the amount of hiring, firing, training, overtime for each worker type. This paper considers a workforce planning model including human aspects such as skills, training, workers- personalities, capacity, motivation, and learning rates. This model helps to minimize the hiring, firing, training and overtime costs, and maximize the workers- performance. The results indicate that the workers- differences should be considered in workforce scheduling to generate realistic plans with minimum costs. This paper also investigates the effects of human learning rates on the performance of the production systems.Keywords: Human Factors, Learning Curves, Workers' Differences, Workforce Scheduling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1862