Search results for: first mover advantage
246 Self-evolving Artificial Immune System via Developing T and B Cell for Permutation Flow-shop Scheduling Problems
Authors: Pei-Chann Chang, Wei-Hsiu Huang, Ching-Jung Ting, Hwei-Wen Luo, Yu-Peng Yu
Abstract:
Artificial Immune System is applied as a Heuristic Algorithm for decades. Nevertheless, many of these applications took advantage of the benefit of this algorithm but seldom proposed approaches for enhancing the efficiency. In this paper, a Self-evolving Artificial Immune System is proposed via developing the T and B cell in Immune System and built a self-evolving mechanism for the complexities of different problems. In this research, it focuses on enhancing the efficiency of Clonal selection which is responsible for producing Affinities to resist the invading of Antigens. T and B cell are the main mechanisms for Clonal Selection to produce different combinations of Antibodies. Therefore, the development of T and B cell will influence the efficiency of Clonal Selection for searching better solution. Furthermore, for better cooperation of the two cells, a co-evolutional strategy is applied to coordinate for more effective productions of Antibodies. This work finally adopts Flow-shop scheduling instances in OR-library to validate the proposed algorithm.Keywords: Artificial Immune System, Clonal Selection, Flow-shop Scheduling Problems, Co-evolutional strategy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1748245 Smart Cane Assisted Mobility for the Visually Impaired
Authors: Jayant Sakhardande, Pratik Pattanayak, Mita Bhowmick
Abstract:
An efficient reintegration of the disabled people in the family and society should be fulfilled; hence it is strongly needful to assist their diminished functions or to replace the totally lost functions. Assistive technology helps in neutralizing the impairment. Recent advancements in embedded systems have opened up a vast area of research and development for affordable and portable assistive devices for the visually impaired. Granted there are many assistive devices on the market that are able to detect obstacles, and numerous research and development currently in process to alleviate the cause, unfortunately the cost of devices, size of devices, intrusiveness and higher learning curve prevents the visually impaired from taking advantage of available devices. This project aims at the design and implementation of a detachable unit which is robust, low cost and user friendly, thus, trying to aggrandize the functionality of the existing white cane, to concede above-knee obstacle detection. The designed obstruction detector uses ultrasound sensors for detecting the obstructions before direct contact. It bestows haptic feedback to the user in accordance with the position of the obstacle.Keywords: Visually impaired, Ultrasonic sensors, Obstruction detector, Mobility aid
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6072244 Elastic and Plastic Collision Comparison Using Finite Element Method
Authors: Gustavo Rodrigues, Hans Weber, Larissa Driemeier
Abstract:
The prevision of post-impact conditions and the behavior of the bodies during the impact have been object of several collision models. The formulation from Hertz’s theory is generally used dated from the 19th century. These models consider the repulsive force as proportional to the deformation of the bodies under contact and may consider it proportional to the rate of deformation. The objective of the present work is to analyze the behavior of the bodies during impact using the Finite Element Method (FEM) with elastic and plastic material models. The main parameters to evaluate are, the contact force, the time of contact and the deformation of the bodies. An advantage of using the FEM approach is the possibility to apply a plastic deformation to the model according to the material definition: there will be used Johnson–Cook plasticity model whose parameters are obtained through empirical tests of real materials. This model allows analyzing the permanent deformation caused by impact, phenomenon observed in real world depending on the forces applied to the body. These results are compared between them and with the model-based Hertz theory.
Keywords: Collision, finite element method, Hertz’s Theory, impact models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 778243 Ethics in the Technology Driven Enterprise
Authors: Bobbie Green, James A. Nelson
Abstract:
Innovations in technology have created new ethical challenges. Essential use of electronic communication in the workplace has escalated at an astronomical rate over the past decade. As such, legal and ethical dilemmas confronted by both the employer and the employee concerning managerial control and ownership of einformation have increased dramatically in the USA. From the employer-s perspective, ownership and control of all information created for the workplace is an undeniable source of economic advantage and must be monitored zealously. From the perspective of the employee, individual rights, such as privacy, freedom of speech, and freedom from unreasonable search and seizure, continue to be stalwart legal guarantees that employers are not legally or ethically entitled to abridge in the workplace. These issues have been the source of great debate and the catalyst for legal reform. The fine line between ethical and legal has been complicated by emerging technologies. This manuscript will identify and discuss a number of specific legal and ethical issues raised by the dynamic electronic workplace and conclude with suggestions that employers should follow to respect the delicate balance between employees- legal rights to privacy and the employer's right to protect its knowledge systems and infrastructure.Keywords: Information, ethics, legal, privacy
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2055242 Finite Element Application to Estimate Inservice Material Properties using Miniature Specimen
Authors: G. Partheepan, D.K. Sehgal, R.K. Pandey
Abstract:
This paper presents a method for determining the uniaxial tensile properties such as Young-s modulus, yield strength and the flow behaviour of a material in a virtually non-destructive manner. To achieve this, a new dumb-bell shaped miniature specimen has been designed. This helps in avoiding the removal of large size material samples from the in-service component for the evaluation of current material properties. The proposed miniature specimen has an advantage in finite element modelling with respect to computational time and memory space. Test fixtures have been developed to enable the tension tests on the miniature specimen in a testing machine. The studies have been conducted in a chromium (H11) steel and an aluminum alloy (AR66). The output from the miniature test viz. load-elongation diagram is obtained and the finite element simulation of the test is carried out using a 2D plane stress analysis. The results are compared with the experimental results. It is observed that the results from the finite element simulation corroborate well with the miniature test results. The approach seems to have potential to predict the mechanical properties of the materials, which could be used in remaining life estimation of the various in-service structures.Keywords: ABAQUS, finite element, miniature test, tensileproperties
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1730241 Multiple Job Shop-Scheduling using Hybrid Heuristic Algorithm
Authors: R.A.Mahdavinejad
Abstract:
In this paper, multi-processors job shop scheduling problems are solved by a heuristic algorithm based on the hybrid of priority dispatching rules according to an ant colony optimization algorithm. The objective function is to minimize the makespan, i.e. total completion time, in which a simultanous presence of various kinds of ferons is allowed. By using the suitable hybrid of priority dispatching rules, the process of finding the best solution will be improved. Ant colony optimization algorithm, not only promote the ability of this proposed algorithm, but also decreases the total working time because of decreasing in setup times and modifying the working production line. Thus, the similar work has the same production lines. Other advantage of this algorithm is that the similar machines (not the same) can be considered. So, these machines are able to process a job with different processing and setup times. According to this capability and from this algorithm evaluation point of view, a number of test problems are solved and the associated results are analyzed. The results show a significant decrease in throughput time. It also shows that, this algorithm is able to recognize the bottleneck machine and to schedule jobs in an efficient way.
Keywords: Job shops scheduling, Priority dispatching rules, Makespan, Hybrid heuristic algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1669240 Pattern Classification of Back-Propagation Algorithm Using Exclusive Connecting Network
Authors: Insung Jung, Gi-Nam Wang
Abstract:
The objective of this paper is to a design of pattern classification model based on the back-propagation (BP) algorithm for decision support system. Standard BP model has done full connection of each node in the layers from input to output layers. Therefore, it takes a lot of computing time and iteration computing for good performance and less accepted error rate when we are doing some pattern generation or training the network. However, this model is using exclusive connection in between hidden layer nodes and output nodes. The advantage of this model is less number of iteration and better performance compare with standard back-propagation model. We simulated some cases of classification data and different setting of network factors (e.g. hidden layer number and nodes, number of classification and iteration). During our simulation, we found that most of simulations cases were satisfied by BP based using exclusive connection network model compared to standard BP. We expect that this algorithm can be available to identification of user face, analysis of data, mapping data in between environment data and information.Keywords: Neural network, Back-propagation, classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1656239 Field Programmable Gate Array Based Infinite Impulse Response Filter Using Multipliers
Authors: Rajesh Mehra, Bharti Thakur
Abstract:
In this paper, an Infinite Impulse Response (IIR) filter has been designed and simulated on an Field Programmable Gate Arrays (FPGA). The implementation is based on Multiply Add and Accumulate (MAC) algorithm which uses multiply operations for design implementation. Parallel Pipelined structure is used to implement the proposed IIR Filter taking optimal advantage of the look up table of target device. The designed filter has been synthesized on Digital Signal Processor (DSP) slice based FPGA to perform multiplier function of MAC unit. The DSP slices are useful to enhance the speed performance. The proposed design is simulated with Matlab, synthesized with Xilinx Synthesis Tool, and implemented on FPGA devices. The Virtex 5 FPGA based design can operate at an estimated frequency of 81.5 MHz as compared to 40.5 MHz in case of Spartan 3 ADSP based design. The Virtex 5 based implementation also consumes less slices and slice flip flops of target FPGA in comparison to Spartan 3 ADSP based implementation to provide cost effective solution for signal processing applications.Keywords: Butterworth, DSP, IIR, MAC, FPGA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1871238 Improving Production Capacity through Efficient PPC System: Lesson from Leather Manufacturing
Authors: Mengist Hailemariam, Silma Yoseph
Abstract:
A well designed and executed Production Planning and Control (PPC) system is one of the key levers for superior performance in the current manufacturing set-up. Hence, measuring the PPC system performance has become a necessity for long term success. The present study examined PPC related issues which impact the production capacity and productivity of leather companies with special focus on Kombolcha Tannery Share Company (KTSC), Ethiopia. Physical observation, interview, and questionnaire were used to generate necessary information from the respondents and reach valid conclusions. Company annual reports were referred and analyzed to triangulate primary data. Consequently, the study revealed that KTSC runs below its capacity due to its inefficient PPC system being in use for which the root causes were identified. The study thereby conceptualizes a PPC system improvement framework comprising three pillars viz., management culture, internal capability and performance measurement together with key considerations in each case. The study findings enable the company to recognize the importance of efficient PPC system as a source of competitive advantage. It also aid managers in evaluating various PPC execution schemes to enhance productivity.
Keywords: Ethiopia, Leather manufacturing, Production planning and control, PPC improvement framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3734237 Hazardous Waste Generated in the Peruvian Textile Industry: Haute Couture, Alpaca Fiber and Tannery
Authors: Huiman C. Alberto
Abstract:
The research cites the various hazardous waste generated in the textile industry. The method used is descriptive and comparative, the process consisted of the search and evaluation of information, both nationally and internationally. The results indicate: (1) Waste is generated from the alpaca fiber industry in the various stages of camelid rearing, they stand out for their dangerousness: excreta, residual fiber and yarn scraps. (2) The main hazardous waste generated by the tannery industry are grease, hides, hair, plastic containers with traces of toxic substances, chips and pieces of leather with chrome. (3) Three companies' Solid Waste Management Plans were analyzed, randomly selected, and none of them detail waste treatment processes and warn of the lack of supervision by the authorities. It is concluded that the hazardous waste generated can affect human and environmental health. There is the possibility of taking advantage of certain hazardous waste such as manure and alpaca fiber, after treatment; while non-hazardous waste from the tannery such as yarn, panel weaving, cloth, scraps, and thread, can be used to produce new products, generating a production chain in favor of the entrepreneur himself.
Keywords: Alpaca fiber, excreta, Haute couture, hazardous waste tannery, hazardous waste treatment, textile waste,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 416236 Gabriel-constrained Parametric Surface Triangulation
Authors: Oscar E. Ruiz, Carlos Cadavid, Juan G. Lalinde, Ricardo Serrano, Guillermo Peris-Fajarnes
Abstract:
The Boundary Representation of a 3D manifold contains FACES (connected subsets of a parametric surface S : R2 -! R3). In many science and engineering applications it is cumbersome and algebraically difficult to deal with the polynomial set and constraints (LOOPs) representing the FACE. Because of this reason, a Piecewise Linear (PL) approximation of the FACE is needed, which is usually represented in terms of triangles (i.e. 2-simplices). Solving the problem of FACE triangulation requires producing quality triangles which are: (i) independent of the arguments of S, (ii) sensitive to the local curvatures, and (iii) compliant with the boundaries of the FACE and (iv) topologically compatible with the triangles of the neighboring FACEs. In the existing literature there are no guarantees for the point (iii). This article contributes to the topic of triangulations conforming to the boundaries of the FACE by applying the concept of parameterindependent Gabriel complex, which improves the correctness of the triangulation regarding aspects (iii) and (iv). In addition, the article applies the geometric concept of tangent ball to a surface at a point to address points (i) and (ii). Additional research is needed in algorithms that (i) take advantage of the concepts presented in the heuristic algorithm proposed and (ii) can be proved correct.Keywords: surface triangulation, conforming triangulation, surfacesampling, Gabriel complex.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662235 Adaptive PID Controller based on Reinforcement Learning for Wind Turbine Control
Authors: M. Sedighizadeh, A. Rezazadeh
Abstract:
A self tuning PID control strategy using reinforcement learning is proposed in this paper to deal with the control of wind energy conversion systems (WECS). Actor-Critic learning is used to tune PID parameters in an adaptive way by taking advantage of the model-free and on-line learning properties of reinforcement learning effectively. In order to reduce the demand of storage space and to improve the learning efficiency, a single RBF neural network is used to approximate the policy function of Actor and the value function of Critic simultaneously. The inputs of RBF network are the system error, as well as the first and the second-order differences of error. The Actor can realize the mapping from the system state to PID parameters, while the Critic evaluates the outputs of the Actor and produces TD error. Based on TD error performance index and gradient descent method, the updating rules of RBF kernel function and network weights were given. Simulation results show that the proposed controller is efficient for WECS and it is perfectly adaptable and strongly robust, which is better than that of a conventional PID controller.Keywords: Wind energy conversion systems, reinforcementlearning; Actor-Critic learning; adaptive PID control; RBF network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4937234 High Level Synthesis of Canny Edge Detection Algorithm on Zynq Platform
Authors: Hanaa M. Abdelgawad, Mona Safar, Ayman M. Wahba
Abstract:
Real time image and video processing is a demand in many computer vision applications, e.g. video surveillance, traffic management and medical imaging. The processing of those video applications requires high computational power. Thus, the optimal solution is the collaboration of CPU and hardware accelerators. In this paper, a Canny edge detection hardware accelerator is proposed. Edge detection is one of the basic building blocks of video and image processing applications. It is a common block in the pre-processing phase of image and video processing pipeline. Our presented approach targets offloading the Canny edge detection algorithm from processing system (PS) to programmable logic (PL) taking the advantage of High Level Synthesis (HLS) tool flow to accelerate the implementation on Zynq platform. The resulting implementation enables up to a 100x performance improvement through hardware acceleration. The CPU utilization drops down and the frame rate jumps to 60 fps of 1080p full HD input video stream.
Keywords: High Level Synthesis, Canny edge detection, Hardware accelerators, and Computer Vision.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5431233 Hybrid Neuro Fuzzy Approach for Automatic Generation Control of Two -Area Interconnected Power System
Authors: Gayadhar Panda, Sidhartha Panda, C. Ardil
Abstract:
The main objective of Automatic Generation Control (AGC) is to balance the total system generation against system load losses so that the desired frequency and power interchange with neighboring systems is maintained. Any mismatch between generation and demand causes the system frequency to deviate from its nominal value. Thus high frequency deviation may lead to system collapse. This necessitates a very fast and accurate controller to maintain the nominal system frequency. This paper deals with a novel approach of artificial intelligence (AI) technique called Hybrid Neuro-Fuzzy (HNF) approach for an (AGC). The advantage of this controller is that it can handle the non-linearities at the same time it is faster than other conventional controllers. The effectiveness of the proposed controller in increasing the damping of local and inter area modes of oscillation is demonstrated in a two area interconnected power system. The result shows that intelligent controller is having improved dynamic response and at the same time faster than conventional controller.
Keywords: Automatic Generation Control (AGC), Dynamic Model, Two-area Power System, Fuzzy Logic Controller, Neural Network, Hybrid Neuro-Fuzzy(HNF).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2461232 Comparative Study on Soil Tillage Using Rotary Tiller and Power Harrow
Authors: Watcharachan Sukcharoenvipharat, Prathuang Usaborisut, Sirisak Choedkiatphon
Abstract:
Farmers try to reduce steps of soil preparation by using subsoiler and then following by equipment for soil pulverization such as a rotary tiller and a power harrow which take advantage of using a power take-off of a tractor. Therefore, this study was conducted to compare the tilling performances of a rotary tiller and a power harrow applying after subsoiling. The results showed that both the rotary tiller and the power harrow had negative slip, indicating that they generated force to push a tractor. The rotary tiller created negative vertical force to lift up the tractor whereas opposite result was found when using the power harrow. Since working depths were different, vertical forces, torques and PTO powers for two equipment types were significantly different. However, no significant differences were found for the forward speeds, slips, drawbar pulls and drawbar powers. Comparative analysis showed that two equipment types had significant difference in PTO power to working depth, drawbar power to working depth, PTO power to working area, drawbar power to working area and soil pulverization.
Keywords: Rotary tiller, power harrow, drawbar pull, drawbar power, PTO power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2120231 Novel NMR-Technology to Assess Food Quality and Safety
Authors: Markus Link, Manfred Spraul, Hartmut Schaefer, Fang Fang, Birk Schuetz
Abstract:
High Resolution NMR Spectroscopy offers unique screening capabilities for food quality and safety by combining non-targeted and targeted screening in one analysis.
The objective is to demonstrate, that due to its extreme reproducibility NMR can detect smallest changes in concentrations of many components in a mixture, which is best monitored by statistical evaluation however also delivers reliable quantification results.
The methodology typically uses a 400 MHz high resolution instrument under full automation after minimized sample preparation.
For example one fruit juice analysis in a push button operation takes at maximum 15 minutes and delivers a multitude of results, which are automatically summarized in a PDF report.
The method has been proven on fruit juices, where so far unknown frauds could be detected. In addition conventional targeted parameters are obtained in the same analysis. This technology has the advantage that NMR is completely quantitative and concentration calibration only has to be done once for all compounds. Since NMR is so reproducible, it is also transferable between different instruments (with same field strength) and laboratories. Based on strict SOP`s, statistical models developed once can be used on multiple instruments and strategies for compound identification and quantification are applicable as well across labs.
Keywords: Automated solution, NMR, non-targeted screening, targeted screening.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2247230 Dynamic Modeling of Underplateform Damper used in Turbomachinery
Authors: Vikas Rastogi, Vipan Kumar, Loveleen Kumar Bhagi
Abstract:
The present work deals with the structural analysis of turbine blades and modeling of turbine blades. A common failure mode for turbine machines is high cycle of fatigue of compressor and turbine blades due to high dynamic stresses caused by blade vibration and resonance within the operation range of the machinery. In this work, proper damping system will be analyzed to reduce the vibrating blade. The main focus of the work is the modeling of under platform damper to evaluate the dynamic analysis of turbine-blade vibrations. The system is analyzed using Bond graph technique. Bond graph is one of the most convenient ways to represent a system from the physical aspect in foreground. It has advantage of putting together multi-energy domains of a system in a single representation in a unified manner. The bond graph model of dry friction damper is simulated on SYMBOLS-shakti® software. In this work, the blades are modeled as Timoshenko beam. Blade Vibrations under different working conditions are being analyzed numerically.Keywords: Turbine blade vibrations, Friction dampers, Timoshenko Beam, Bond graph modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2328229 Reactive Absorption of Hydrogen Sulfide in Aqueous Ferric Sulfate Solution
Authors: Z. Gholami, M. Torabi Angaji, F. Gholami, S. A. Razavi Alavi
Abstract:
Many commercial processes are available for the removal of H2S from gaseous streams. The desulfurization of gas streams using aqueous ferric sulfate solution as washing liquor is studied. Apart from sulfur, only H2O is generated in the process, and consequently, no waste treatment facilities are required. A distinct advantage of the process is that the reaction of H2S with is so rapid and complete that there remains no danger of discharging toxic waste gas. In this study, the reactive absorption of hydrogen sulfide into aqueous ferric sulfate solution has been studied and design calculations for equipments have been done and effective operation parameters on this process considered. Results show that high temperature and low pressure are suitable for absorption reaction. Variation of hydrogen sulfide concentration and Fe3+ concentration with time in absorption reaction shown that the reaction of ferric sulfate and hydrogen sulfide is first order with respect to the both reactant. At low Fe2(SO4)3 concentration the absorption rate of H2S increase with increasing the Fe2(SO4)3 concentration. At higher concentration a decrease in the absorption rate was found. At higher concentration of Fe2(SO4)3, the ionic strength and viscosity of solution increase remarkably resulting in a decrease of solubility, diffusivity and hence absorption rate.Keywords: Absorption, Fe2(SO4)3, H2S, Reactive Absorption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3943228 Effect of Polyvinyl Pyrrolidone and Ethyl Cellulose Concentration on Release Profile and Kinetics of Glibenclamide Extended Release Dosage Form System
Authors: Amit Kumar, Peeyush Sharma, Anil Bhandari
Abstract:
The aim of present work was to optimize the effect of Ethyl Cellulose (EC) and Polyvinyl Pyrrolidone (PVP) concentration in extended release solid dispersion of Glibenclamide using combination of hydrophilic and hydrophobic polymers such as Polyvinyl Pyrrolidone and Ethyl cellulose. The advantage of solid dispersion technique provides a unique approach to particle size reduction and increased rates of dissolution. The compatibility studies of the drug and polymers were studied by TLC and results suggested no interaction between drug and polymers. Solid dispersions of Glibenclamide were prepared by common solvent evaporation method using Polyvinyl Pyrrolidone and Ethyl cellulose. The results indicated that homogeneous or heterogeneous conditions during the preparation methods employed governed the internal structures of the polymer matrices while retaining the drug in an amorphous form. F2 formulation prepared by solid dispersion method, displayed extended drug release followed by Higuchi matrix model indicating diffusion release of GLB from polymer matrices.
Keywords: Ethyl Cellulose, Glibenclamide, Polyvinyl Pyrrolidone, Solid Dispersion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2526227 Impact of Modeling Different Fading Channels on Wireless MAN Fixed IEEE802.16d OFDM System with Diversity Transmission Technique
Authors: Shanar Askar, Shahzad Memon, LachhmanDas, MSKalhoro
Abstract:
Wimax (Worldwide Interoperability for Microwave Access) is a promising technology which can offer high speed data, voice and video service to the customer end, which is presently, dominated by the cable and digital subscriber line (DSL) technologies. The performance assessment of Wimax systems is dealt with. The biggest advantage of Broadband wireless application (BWA) over its wired competitors is its increased capacity and ease of deployment. The aims of this paper are to model and simulate the fixed OFDM IEEE 802.16d physical layer under variant combinations of digital modulation (BPSK, QPSK, and 16-QAM) over diverse combination of fading channels (AWGN, SUIs). Stanford University Interim (SUI) Channel serial was proposed to simulate the fixed broadband wireless access channel environments where IEEE 802.16d is to be deployed. It has six channel models that are grouped into three categories according to three typical different outdoor Terrains, in order to give a comprehensive effect of fading channels on the overall performance of the system.Keywords: WIMAX, OFDM, Additive White Gaussian Noise, Fading Channel, SUI, Doppler Effect.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2138226 A New Face Detection Technique using 2D DCT and Self Organizing Feature Map
Authors: Abdallah S. Abdallah, A. Lynn Abbott, Mohamad Abou El-Nasr
Abstract:
This paper presents a new technique for detection of human faces within color images. The approach relies on image segmentation based on skin color, features extracted from the two-dimensional discrete cosine transform (DCT), and self-organizing maps (SOM). After candidate skin regions are extracted, feature vectors are constructed using DCT coefficients computed from those regions. A supervised SOM training session is used to cluster feature vectors into groups, and to assign “face" or “non-face" labels to those clusters. Evaluation was performed using a new image database of 286 images, containing 1027 faces. After training, our detection technique achieved a detection rate of 77.94% during subsequent tests, with a false positive rate of 5.14%. To our knowledge, the proposed technique is the first to combine DCT-based feature extraction with a SOM for detecting human faces within color images. It is also one of a few attempts to combine a feature-invariant approach, such as color-based skin segmentation, together with appearance-based face detection. The main advantage of the new technique is its low computational requirements, in terms of both processing speed and memory utilization.Keywords: Face detection, skin color segmentation, self-organizingmap.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2543225 Data Envelopment Analysis with Partially Perfect Objects
Authors: Alexander Y. Vaninsky
Abstract:
This paper presents a simplified version of Data Envelopment Analysis (DEA) - a conventional approach to evaluating the performance and ranking of competitive objects characterized by two groups of factors acting in opposite directions: inputs and outputs. DEA with a Perfect Object (DEA PO) augments the group of actual objects with a virtual Perfect Object - the one having greatest outputs and smallest inputs. It allows for obtaining an explicit analytical solution and making a step to an absolute efficiency. This paper develops this approach further and introduces a DEA model with Partially Perfect Objects. DEA PPO consecutively eliminates the smallest relative inputs or greatest relative outputs, and applies DEA PO to the reduced collections of indicators. The partial efficiency scores are combined to get the weighted efficiency score. The computational scheme remains simple, like that of DEA PO, but the advantage of the DEA PPO is taking into account all of the inputs and outputs for each actual object. Firm evaluation is considered as an example.
Keywords: Data Envelopment Analysis, Perfect object, Partially perfect object, Partial efficiency, Explicit solution, Simplified algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1697224 A Framework for Scalable Autonomous P2P Resource Discovery for the Grid Implementation
Authors: Hesham A. Ali, Mofreh M. Salem, Ahmed A. Hamza
Abstract:
Recently, there have been considerable efforts towards the convergence between P2P and Grid computing in order to reach a solution that takes the best of both worlds by exploiting the advantages that each offers. Augmenting the peer-to-peer model to the services of the Grid promises to eliminate bottlenecks and ensure greater scalability, availability, and fault-tolerance. The Grid Information Service (GIS) directly influences quality of service for grid platforms. Most of the proposed solutions for decentralizing the GIS are based on completely flat overlays. The main contributions for this paper are: the investigation of a novel resource discovery framework for Grid implementations based on a hierarchy of structured peer-to-peer overlay networks, and introducing a discovery algorithm utilizing the proposed framework. Validation of the framework-s performance is done via simulation. Experimental results show that the proposed organization has the advantage of being scalable while providing fault-isolation, effective bandwidth utilization, and hierarchical access control. In addition, it will lead to a reliable, guaranteed sub-linear search which returns results within a bounded interval of time and with a smaller amount of generated traffic within each domain.
Keywords: Grid computing, grid information service, P2P, resource discovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1976223 Minimizing Risk Costs through Optimal Responses in NPD Projects
Authors: Chan-Sik Kim, Jong-Seong Kim, Se Won Lee, Hoo-Gon Choi
Abstract:
In rapidly changing market environment, firms are investing a lot of time and resources into new product development (NPD) projects to make profit and to obtain competitive advantage. However, failure rate of NPD projects is becoming high due to various internal and external risks which hinder successful NPD projects. To reduce the failure rate, it is critical that risks have to be managed effectively and efficiently through good strategy, and treated by optimal responses to minimize risk cost. Four strategies are adopted to handle the risks in this study. The optimal responses are characterized by high reduction of risk costs with high efficiency. This study suggests a framework to decide the optimal responses considering the core risks, risk costs, response efficiency and response costs for successful NPD projects. Both binary particles warm optimization (BPSO) and multi-objective particle swarm optimization (MOPSO) methods are mainly used in the framework. Although several limitations exist in use for real industries, the frame work shows good strength for handling the risks with highly scientific ways through an example.
Keywords: NPD projects, risk cost, strategy, optimal responses, Particle Swarm Optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1957222 Nonparametric Control Chart Using Density Weighted Support Vector Data Description
Authors: Myungraee Cha, Jun Seok Kim, Seung Hwan Park, Jun-Geol Baek
Abstract:
In manufacturing industries, development of measurement leads to increase the number of monitoring variables and eventually the importance of multivariate control comes to the fore. Statistical process control (SPC) is one of the most widely used as multivariate control chart. Nevertheless, SPC is restricted to apply in processes because its assumption of data as following specific distribution. Unfortunately, process data are composed by the mixture of several processes and it is hard to estimate as one certain distribution. To alternative conventional SPC, therefore, nonparametric control chart come into the picture because of the strength of nonparametric control chart, the absence of parameter estimation. SVDD based control chart is one of the nonparametric control charts having the advantage of flexible control boundary. However,basic concept of SVDD has been an oversight to the important of data characteristic, density distribution. Therefore, we proposed DW-SVDD (Density Weighted SVDD) to cover up the weakness of conventional SVDD. DW-SVDD makes a new attempt to consider dense of data as introducing the notion of density Weight. We extend as control chart using new proposed SVDD and a simulation study of various distributional data is conducted to demonstrate the improvement of performance.
Keywords: Density estimation, Multivariate control chart, Oneclass classification, Support vector data description (SVDD)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2121221 Evaluation of Energy Upgrade Measures and Connection of Renewable Energy Sources Using Software Tools: Case Study of an Academic Library Building in Larissa, Greece
Authors: Giwrgos S. Gkarmpounis, Aikaterini G. Rokkou, Marios N. Moschakis
Abstract:
Increased energy consumption in the academic buildings, creates the need to implement energy saving measures and to take advantage of the renewable energy sources to cover the electrical needs of those buildings. An Academic Library will be used as a case study. With the aid of RETScreen software that takes into account the energy consumptions and characteristics of the Library Building, it is proved that measures such as the replacement of fluorescent lights with led lights, the installation of outdoor shading, the replacement of the openings and Building Management System installation, provide a high level of energy savings. Moreover, given the available space of the building and the climatic data, the installation of a photovoltaic system of 100 kW can also cover a serious amount of the building energy consumption, unlike a wind system that seems uncompromising. Lastly, HOMER software is used to compare the use of a photovoltaic system against a wind system in order to verify the results that came up from the RETScreen software concerning the renewable energy sources.
Keywords: Energy saving measures, homer software, renewable energy sources, RETScreen software, energy efficiency and quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 992220 Adaptive Non-linear Filtering Technique for Image Restoration
Authors: S. K. Satpathy, S. Panda, K. K. Nagwanshi, S. K. Nayak, C. Ardil
Abstract:
Removing noise from the any processed images is very important. Noise should be removed in such a way that important information of image should be preserved. A decisionbased nonlinear algorithm for elimination of band lines, drop lines, mark, band lost and impulses in images is presented in this paper. The algorithm performs two simultaneous operations, namely, detection of corrupted pixels and evaluation of new pixels for replacing the corrupted pixels. Removal of these artifacts is achieved without damaging edges and details. However, the restricted window size renders median operation less effective whenever noise is excessive in that case the proposed algorithm automatically switches to mean filtering. The performance of the algorithm is analyzed in terms of Mean Square Error [MSE], Peak-Signal-to-Noise Ratio [PSNR], Signal-to-Noise Ratio Improved [SNRI], Percentage Of Noise Attenuated [PONA], and Percentage Of Spoiled Pixels [POSP]. This is compared with standard algorithms already in use and improved performance of the proposed algorithm is presented. The advantage of the proposed algorithm is that a single algorithm can replace several independent algorithms which are required for removal of different artifacts.
Keywords: Filtering, Decision Based Algorithm, noise, imagerestoration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2158219 Stock Market Integration Measurement: Investigation of Malaysia and Singapore Stock Markets
Authors: B. K. Yeoh, Z. Arsad, C. W. Hooy
Abstract:
This paper tests the level of market integration between Malaysia and Singapore stock markets with the world market. Kalman Filter (KF) methodology is used on the International Capital Asset Pricing Model (ICAPM) and the pricing errors estimated within the framework of ICAPM are used as a measure of market integration or segmentation. The advantage of the KF technique is that it allows for time-varying coefficients in estimating ICAPM and hence able to capture the varying degree of market integration. Empirical results show clear evidence of varying degree of market integration for both case of Malaysia and Singapore. Furthermore, the results show that the changes in the level of market integration are found to coincide with certain economic events that have taken placed. The findings certainly provide evidence on the practicability of the KF technique to estimate stock markets integration. In the comparison between Malaysia and Singapore stock market, the result shows that the trends of the market integration indices for Malaysia and Singapore look similar through time but the magnitude is notably different with the Malaysia stock market showing greater degree of market integration. Finally, significant evidence of varying degree of market integration shows the inappropriate use of OLS in estimating the level of market integration.
Keywords: ICAPM, Kalman filter, stock market integration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2173218 Dimensional Modeling of HIV Data Using Open Source
Authors: Charles D. Otine, Samuel B. Kucel, Lena Trojer
Abstract:
Selecting the data modeling technique for an information system is determined by the objective of the resultant data model. Dimensional modeling is the preferred modeling technique for data destined for data warehouses and data mining, presenting data models that ease analysis and queries which are in contrast with entity relationship modeling. The establishment of data warehouses as components of information system landscapes in many organizations has subsequently led to the development of dimensional modeling. This has been significantly more developed and reported for the commercial database management systems as compared to the open sources thereby making it less affordable for those in resource constrained settings. This paper presents dimensional modeling of HIV patient information using open source modeling tools. It aims to take advantage of the fact that the most affected regions by the HIV virus are also heavily resource constrained (sub-Saharan Africa) whereas having large quantities of HIV data. Two HIV data source systems were studied to identify appropriate dimensions and facts these were then modeled using two open source dimensional modeling tools. Use of open source would reduce the software costs for dimensional modeling and in turn make data warehousing and data mining more feasible even for those in resource constrained settings but with data available.Keywords: About Database, Data Mining, Data warehouse, Dimensional Modeling, Open Source.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959217 GPS and SMS-Based Child Tracking System Using Smart Phone
Authors: A. Al-Mazloum, E. Omer, M. F. A. Abdullah
Abstract:
Recently many cases of missing children between ages 14 and 17 years are reported. Parents always worry about the possibility of kidnapping of their children. This paper proposes an Android based solution to aid parents to track their children in real time. Nowadays, most mobile phones are equipped with location services capabilities allowing us to get the device’s geographic position in real time. The proposed solution takes the advantage of the location services provided by mobile phone since most of kids carry mobile phones. The mobile application use the GPS and SMS services found in Android mobile phones. It allows the parent to get their child’s location on a real time map. The system consists of two sides, child side and parent side. A parent’s device main duty is to send a request location SMS to the child’s device to get the location of the child. On the other hand, the child’s device main responsibility is to reply the GPS position to the parent’s device upon request.
Keywords: Child Tracking System, Global Positioning System (GPS), SMS-based Mobile Application.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15516