Search results for: poisson random measures
1035 An ICA Algorithm for Separation of Convolutive Mixture of Speech Signals
Authors: Rajkishore Prasad, Hiroshi Saruwatari, Kiyohiro Shikano
Abstract:
This paper describes Independent Component Analysis (ICA) based fixed-point algorithm for the blind separation of the convolutive mixture of speech, picked-up by a linear microphone array. The proposed algorithm extracts independent sources by non- Gaussianizing the Time-Frequency Series of Speech (TFSS) in a deflationary way. The degree of non-Gaussianization is measured by negentropy. The relative performances of algorithm under random initialization and Null beamformer (NBF) based initialization are studied. It has been found that an NBF based initial value gives speedy convergence as well as better separation performance
Keywords: Blind signal separation, independent component analysis, negentropy, convolutive mixture.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17781034 A Decision Support Tool for Evaluating Mobility Projects
Abstract:
Success is a European project that will implement several clean transport offers in three European cities and evaluate the environmental impacts. The goal of these measures is to improve urban mobility or the displacement of residents inside cities. For e.g. park and ride, electric vehicles, hybrid bus and bike sharing etc. A list of 28 criteria and 60 measures has been established for evaluation of these transport projects. The evaluation criteria can be grouped into: Transport, environment, social, economic and fuel consumption. This article proposes a decision support system based that encapsulates a hybrid approach based on fuzzy logic, multicriteria analysis and belief theory for the evaluation of impacts of urban mobility solutions. A web-based tool called DeSSIA (Decision Support System for Impacts Assessment) has been developed that treats complex data. The tool has several functionalities starting from data integration (import of data), evaluation of projects and finishes by graphical display of results. The tool development is based on the concept of MVC (Model, View, and Controller). The MVC is a conception model adapted to the creation of software's which impose separation between data, their treatment and presentation. Effort is laid on the ergonomic aspects of the application. It has codes compatible with the latest norms (XHTML, CSS) and has been validated by W3C (World Wide Web Consortium). The main ergonomic aspect focuses on the usability of the application, ease of learning and adoption. By the usage of technologies such as AJAX (XML and Java Script asynchrones), the application is more rapid and convivial. The positive points of our approach are that it treats heterogeneous data (qualitative, quantitative) from various information sources (human experts, survey, sensors, model etc.).
Keywords: Decision support tool, hybrid approach, urban mobility.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19941033 Thermodynamic Study of Hot Potassium Carbonate Solution Using Aspen Plus
Authors: O. Eisa, M. Shuhaimi
Abstract:
This paper presents a study on the thermodynamics and transport properties of hot potassium carbonate aqueous system (HPC) using electrolyte non-random two liquid, (ELECNRTL) model. The operation conditions are varied to determine the system liquid phase stability range at the standard and critical conditions. A case study involving 30 wt% K2CO3, H2O standard system at pressure of 1 bar and temperature range from 280.15 to 366.15 K has been studied. The estimated solubility index, viscosity, water activity, and density which obtained from the simulation showed a good agreement with the experimental work. Furthermore, the saturation temperature of the solution has been estimated.Keywords: Viscosity, Saturation index, Activity coefficient, Density.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 53901032 Resistive Switching Characteristics of Resistive Random Access Memory Devices after Furnace Annealing Processes
Authors: Chi-Yan Chu, Kai-Chi Chuang, Huang-Chung Cheng
Abstract:
In this study, the RRAM devices with the TiN/Ti/HfOx/TiN structure were fabricated, then the electrical characteristics of the devices without annealing and after 400 °C and 500 °C of the furnace annealing (FA) temperature processes were compared. The RRAM devices after the FA’s 400 °C showed the lower forming, set and reset voltages than the other devices without annealing. However, the RRAM devices after the FA’s 500 °C did not show any electrical characteristics because the TiN/Ti/HfOx/TiN device was oxidized, as shown in the XPS analysis. From these results, the RRAM devices after the FA’s 400 °C showed the best electrical characteristics.
Keywords: RRAM, furnace annealing, forming, set and reset voltages, XPS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10961031 A Spectral Decomposition Method for Ordinary Differential Equation Systems with Constant or Linear Right Hand Sides
Authors: R. B. Ogunrinde, C. C. Jibunoh
Abstract:
In this paper, a spectral decomposition method is developed for the direct integration of stiff and nonstiff homogeneous linear (ODE) systems with linear, constant, or zero right hand sides (RHSs). The method does not require iteration but obtains solutions at any random points of t, by direct evaluation, in the interval of integration. All the numerical solutions obtained for the class of systems coincide with the exact theoretical solutions. In particular, solutions of homogeneous linear systems, i.e. with zero RHS, conform to the exact analytical solutions of the systems in terms of t.Keywords: Spectral decomposition, eigenvalues of the Jacobian, linear RHS, homogeneous linear systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11501030 Data Placement in Heterogeneous Storage of Short Videos
Authors: W. Jaipahkdee, C. Srinilta
Abstract:
The overall service performance of I/O intensive system depends mainly on workload on its storage system. In heterogeneous storage environment where storage elements from different vendors with different capacity and performance are put together, workload should be distributed according to storage capability. This paper addresses data placement issue in short video sharing website. Workload contributed by a video is estimated by the number of views and life time span of existing videos in same category. Experiment was conducted on 42,000 video titles in six weeks. Result showed that the proposed algorithm distributed workload and maintained balance better than round robin and random algorithms.Keywords: data placement, heterogeneous storage system, YouTube, short videos
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14881029 Surface Topography Assessment Techniques based on an In-process Monitoring Approach of Tool Wear and Cutting Force Signature
Authors: A. M. Alaskari, S. E. Oraby
Abstract:
The quality of a machined surface is becoming more and more important to justify the increasing demands of sophisticated component performance, longevity, and reliability. Usually, any machining operation leaves its own characteristic evidence on the machined surface in the form of finely spaced micro irregularities (surface roughness) left by the associated indeterministic characteristics of the different elements of the system: tool-machineworkpart- cutting parameters. However, one of the most influential sources in machining affecting surface roughness is the instantaneous state of tool edge. The main objective of the current work is to relate the in-process immeasurable cutting edge deformation and surface roughness to a more reliable easy-to-measure force signals using a robust non-linear time-dependent modeling regression techniques. Time-dependent modeling is beneficial when modern machining systems, such as adaptive control techniques are considered, where the state of the machined surface and the health of the cutting edge are monitored, assessed and controlled online using realtime information provided by the variability encountered in the measured force signals. Correlation between wear propagation and roughness variation is developed throughout the different edge lifetimes. The surface roughness is further evaluated in the light of the variation in both the static and the dynamic force signals. Consistent correlation is found between surface roughness variation and tool wear progress within its initial and constant regions. At the first few seconds of cutting, expected and well known trend of the effect of the cutting parameters is observed. Surface roughness is positively influenced by the level of the feed rate and negatively by the cutting speed. As cutting continues, roughness is affected, to different extents, by the rather localized wear modes either on the tool nose or on its flank areas. Moreover, it seems that roughness varies as wear attitude transfers from one mode to another and, in general, it is shown that it is improved as wear increases but with possible corresponding workpart dimensional inaccuracy. The dynamic force signals are found reasonably sensitive to simulate either the progressive or the random modes of tool edge deformation. While the frictional force components, feeding and radial, are found informative regarding progressive wear modes, the vertical (power) components is found more representative carrier to system instability resulting from the edge-s random deformation.
Keywords: Dynamic force signals, surface roughness (finish), tool wear and deformation, tool wear modes (nose, flank)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13491028 dynr.mi: An R Program for Multiple Imputation in Dynamic Modeling
Authors: Yanling Li, Linying Ji, Zita Oravecz, Timothy R. Brick, Michael D. Hunter, Sy-Miin Chow
Abstract:
Assessing several individuals intensively over time yields intensive longitudinal data (ILD). Even though ILD provide rich information, they also bring other data analytic challenges. One of these is the increased occurrence of missingness with increased study length, possibly under non-ignorable missingness scenarios. Multiple imputation (MI) handles missing data by creating several imputed data sets, and pooling the estimation results across imputed data sets to yield final estimates for inferential purposes. In this article, we introduce dynr.mi(), a function in the R package, Dynamic Modeling in R (dynr). The package dynr provides a suite of fast and accessible functions for estimating and visualizing the results from fitting linear and nonlinear dynamic systems models in discrete as well as continuous time. By integrating the estimation functions in dynr and the MI procedures available from the R package, Multivariate Imputation by Chained Equations (MICE), the dynr.mi() routine is designed to handle possibly non-ignorable missingness in the dependent variables and/or covariates in a user-specified dynamic systems model via MI, with convergence diagnostic check. We utilized dynr.mi() to examine, in the context of a vector autoregressive model, the relationships among individuals’ ambulatory physiological measures, and self-report affect valence and arousal. The results from MI were compared to those from listwise deletion of entries with missingness in the covariates. When we determined the number of iterations based on the convergence diagnostics available from dynr.mi(), differences in the statistical significance of the covariate parameters were observed between the listwise deletion and MI approaches. These results underscore the importance of considering diagnostic information in the implementation of MI procedures.Keywords: Dynamic modeling, missing data, multiple imputation, physiological measures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8101027 Piecewise Interpolation Filter for Effective Processing of Large Signal Sets
Authors: Anatoli Torokhti, Stanley Miklavcic
Abstract:
Suppose KY and KX are large sets of observed and reference signals, respectively, each containing N signals. Is it possible to construct a filter F : KY → KX that requires a priori information only on few signals, p N, from KX but performs better than the known filters based on a priori information on every reference signal from KX? It is shown that the positive answer is achievable under quite unrestrictive assumptions. The device behind the proposed method is based on a special extension of the piecewise linear interpolation technique to the case of random signal sets. The proposed technique provides a single filter to process any signal from the arbitrarily large signal set. The filter is determined in terms of pseudo-inverse matrices so that it always exists.Keywords: Wiener filter, filtering of stochastic signals.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14121026 Modelling of Soil Erosion by Non Conventional Methods
Authors: Ganesh D. Kale, Sheela N. Vadsola
Abstract:
Soil erosion is the most serious problem faced at global and local level. So planning of soil conservation measures has become prominent agenda in the view of water basin managers. To plan for the soil conservation measures, the information on soil erosion is essential. Universal Soil Loss Equation (USLE), Revised Universal Soil Loss Equation 1 (RUSLE1or RUSLE) and Modified Universal Soil Loss Equation (MUSLE), RUSLE 1.06, RUSLE1.06c, RUSLE2 are most widely used conventional erosion estimation methods. The essential drawbacks of USLE, RUSLE1 equations are that they are based on average annual values of its parameters and so their applicability to small temporal scale is questionable. Also these equations do not estimate runoff generated soil erosion. So applicability of these equations to estimate runoff generated soil erosion is questionable. Data used in formation of USLE, RUSLE1 equations was plot data so its applicability at greater spatial scale needs some scale correction factors to be induced. On the other hand MUSLE is unsuitable for predicting sediment yield of small and large events. Although the new revised forms of USLE like RUSLE 1.06, RUSLE1.06c and RUSLE2 were land use independent and they have almost cleared all the drawbacks in earlier versions like USLE and RUSLE1, they are based on the regional data of specific area and their applicability to other areas having different climate, soil, land use is questionable. These conventional equations are applicable for sheet and rill erosion and unable to predict gully erosion and spatial pattern of rills. So the research was focused on development of nonconventional (other than conventional) methods of soil erosion estimation. When these non-conventional methods are combined with GIS and RS, gives spatial distribution of soil erosion. In the present paper the review of literature on non- conventional methods of soil erosion estimation supported by GIS and RS is presented.Keywords: Conventional methods, GIS, non-conventionalmethods, remote sensing, soil erosion modeling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42911025 Customer Satisfaction on Reliability Dimension of Service Quality in Indian Higher Education
Authors: Rajasekhar Mamilla, Janardhana G., Anjan Babu G.
Abstract:
The present research study analyses the students’ satisfaction with university performance regarding the reliability dimension, ability of professors and staff to perform the promised services with quality to students in the post-graduate courses offered by Sri Venkateswara University in India. The research is done with the notion that the student compares the perceived performance with prior expectations. Customer satisfaction is seen as the outcome of this comparison. The sample respondents were administered with schedule based on stratified random technique for this study. Statistical techniques such as factor analysis, t-test and correlation analysis were used to accomplish the respective objectives of the study.
Keywords: Satisfaction, Reliability, Service Quality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47261024 Enhancing Transit Trade, Facilitation System and Supply Chain Security for Local, Regional and an International Corridor
Authors: Moh’d A. AL-Shboul
Abstract:
Recently, and due to Arab spring and terrorism around the globe, pushing and driving most governments potentially to harmonize their border measures particularly the regional and an international transit trade within and among Customs Unions. The main purpose of this study is to investigate and provide an insight for monitoring and controlling the trade supply chain within and among different countries by using technological advancement (i.e. an electronic tracking system, etc.); furthermore, facilitate the local and intra-regional trade among countries through reviewing the recent trends and practical implementation of an electronic transit traffic and cargo that related to customs measures by introducing and supporting some case studies of several international and landlocked transit trade countries. The research methodology employed in this study was described as qualitative by conducting few interviews with managers, transit truck drivers, and traders and reviewing the related literature to collect qualitative data from secondary sources such as statistical reports, previous studies, etc. The results in this study show that Jordan and other countries around the globe that used an electronic tracking system for monitoring transit trade has led to a significant reduction in cost, effort and time in physical movement of goods internally and crossing through other countries. Therefore, there is no need to escort transit trucks by customs staff; hence, the rate of escort transit trucks is reduced by more than ninety percent, except the bulky and high duty goods. Electronic transit traffic has been increased; the average transit time journey has been reduced by more than seventy percent and has led to decrease in rates of smuggling up to fifty percent. The researcher recommends considering Jordan as regional and international office for tracking electronically and monitoring the transit trade for many considerations.Keywords: Electronic tracking system, facilitation system, regional and international corridor, supply chain security, transit trade.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14021023 Methods for Material and Process Monitoring by Characterization of (Second and Third Order) Elastic Properties with Lamb Waves
Abstract:
In accordance with the industry 4.0 concept, manufacturing process steps as well as the materials themselves are going to be more and more digitalized within the next years. The “digital twin” representing the simulated and measured dataset of the (semi-finished) product can be used to control and optimize the individual processing steps and help to reduce costs and expenditure of time in product development, manufacturing, and recycling. In the present work, two material characterization methods based on Lamb waves were evaluated and compared. For demonstration purpose, both methods were shown at a standard industrial product - copper ribbons, often used in photovoltaic modules as well as in high-current microelectronic devices. By numerical approximation of the Rayleigh-Lamb dispersion model on measured phase velocities second order elastic constants (Young’s modulus, Poisson’s ratio) were determined. Furthermore, the effective third order elastic constants were evaluated by applying elastic, “non-destructive”, mechanical stress on the samples. In this way, small microstructural variations due to mechanical preconditioning could be detected for the first time. Both methods were compared with respect to precision and inline application capabilities. Microstructure of the samples was systematically varied by mechanical loading and annealing. Changes in the elastic ultrasound transport properties were correlated with results from microstructural analysis and mechanical testing. In summary, monitoring the elastic material properties of plate-like structures using Lamb waves is valuable for inline and non-destructive material characterization and manufacturing process control. Second order elastic constants analysis is robust over wide environmental and sample conditions, whereas the effective third order elastic constants highly increase the sensitivity with respect to small microstructural changes. Both Lamb wave based characterization methods are fitting perfectly into the industry 4.0 concept.
Keywords: Lamb waves, industry 4.0, process control, elasticity, acoustoelasticity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10981022 Optimization of Communication Protocols by stochastic Delay Mechanisms
Authors: J. Levendovszky, I. Koncz, P. Boros
Abstract:
The paper is concerned with developing stochastic delay mechanisms for efficient multicast protocols and for smooth mobile handover processes which are capable of preserving a given Quality of Service (QoS). In both applications the participating entities (receiver nodes or subscribers) sample a stochastic timer and generate load after a random delay. In this way, the load on the networking resources is evenly distributed which helps to maintain QoS communication. The optimal timer distributions have been sought in different p.d.f. families (e.g. exponential, power law and radial basis function) and the optimal parameter have been found in a recursive manner. Detailed simulations have demonstrated the improvement in performance both in the case of multicast and mobile handover applications.
Keywords: Multicast communication, stochactic delay mechanisms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15491021 Blind Non-Minimum Phase Channel Identification Using 3rd and 4th Order Cumulants
Authors: S. Safi, A. Zeroual
Abstract:
In this paper we propose a family of algorithms based on 3rd and 4th order cumulants for blind single-input single-output (SISO) Non-Minimum Phase (NMP) Finite Impulse Response (FIR) channel estimation driven by non-Gaussian signal. The input signal represents the signal used in 10GBASE-T (or IEEE 802.3an-2006) as a Tomlinson-Harashima Precoded (THP) version of random Pulse-Amplitude Modulation with 16 discrete levels (PAM-16). The proposed algorithms are tested using three non-minimum phase channel for different Signal-to-Noise Ratios (SNR) and for different data input length. Numerical simulation results are presented to illustrate the performance of the proposed algorithms.Keywords: Higher Order Cumulants, Channel identification, Ethernet communication.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14481020 A New Effective Local Search Heuristic for the Maximum Clique Problem
Authors: S. Balaji
Abstract:
An edge based local search algorithm, called ELS, is proposed for the maximum clique problem (MCP), a well-known combinatorial optimization problem. ELS is a two phased local search method effectively £nds the near optimal solutions for the MCP. A parameter ’support’ of vertices de£ned in the ELS greatly reduces the more number of random selections among vertices and also the number of iterations and running times. Computational results on BHOSLIB and DIMACS benchmark graphs indicate that ELS is capable of achieving state-of-the-art-performance for the maximum clique with reasonable average running times.
Keywords: Maximum clique, local search, heuristic, NP-complete.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22561019 Overview of Adaptive Spline Interpolation
Authors: Rongli Gai, Zhiyuan Chang, Xiaohong Wang, Jingyu Liu
Abstract:
In view of various situations in the interpolation process, most researchers use self-adaptation to adjust the interpolation process, which is also one of the current and future research hotspots in the field of CNC (Computerized Numerical Control) machining. In the interpolation process, according to the overview of the spline curve interpolation algorithm, the adaptive analysis is carried out from the factors affecting the interpolation process. The adaptive operation is reflected in various aspects, such as speed, parameters, errors, nodes, feed rates, random period, sensitive point, step size, curvature, adaptive segmentation, adaptive optimization, etc. This paper will analyze and summarize the research of adaptive imputation in the direction of the above factors affecting imputation.
Keywords: Adaptive algorithm, CNC machining, interpolation constraints, spline curve interpolation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5521018 Performance Comparison of Real Time EDAC Systems for Applications On-Board Small Satellites
Authors: Y. Bentoutou
Abstract:
On-board Error Detection and Correction (EDAC) devices aim to secure data transmitted between the central processing unit (CPU) of a satellite onboard computer and its local memory. This paper presents a comparison of the performance of four low complexity EDAC techniques for application in Random Access Memories (RAMs) on-board small satellites. The performance of a newly proposed EDAC architecture is measured and compared with three different EDAC strategies, using the same FPGA technology. A statistical analysis of single-event upset (SEU) and multiple-bit upset (MBU) activity in commercial memories onboard Alsat-1 is given for a period of 8 yearsKeywords: Error Detection and Correction; On-board computer; small satellite missions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22621017 Generalized Maximum Entropy Method for Cosmic Source Localization
Authors: Youssef Khmou, Said Safi, Miloud Frikel
Abstract:
The Maximum entropy principle in spectral analysis was used as an estimator of Direction of Arrival (DoA) of electromagnetic or acoustic sources impinging on an array of sensors, indeed the maximum entropy operator is very efficient when the signals of the radiating sources are ergodic and complex zero mean random processes which is the case for cosmic sources. In this paper, we present basic review of the maximum entropy method (MEM) which consists of rank one operator but not a projector, and we elaborate a new operator which is full rank and sum of all possible projectors. Two dimensional Simulation results based on Monte Carlo trials prove the resolution power of the new operator where the MEM presents some erroneous fluctuations.
Keywords: Maximum entropy, Cosmic source, Localization, operator, projector, azimuth, elevation, DoA, circular array.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21331016 Effect of Trataka on Anxiety among Adolescents
Authors: Pushp Lata Rajpoot, Pushpa Vaishnav
Abstract:
Anxiety is a common psychological problem and also implicated as a contributor to many chronic diseases which decreased quality of life even with pharmacological treatment. At the present time several yogic practices- meditation, pranayama, and mantra, etcetera are playing important role in treating physiological and psychological problems. Hence, the present investigation is aimed to see the effect of Trataka on the level of anxiety among adolescents. For the present study, a sample of 30 adolescents belonging to the age range 20-30 years was selected from Devsanskriti Vishwa Vidyalaya Haridwar through random sampling. In this investigation, Sinha’s Comprehensive anxiety test has been used to measure the level of anxiety. Statistical analysis has been done by using t-test. Findings of this study reveal that Trataka significantly decreases the level of anxiety among adolescents.
Keywords: Adolescents, Anxiety, Trataka.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 57191015 Measures and Influence of a Baw Filter on Digital Radio-Communications Signals
Authors: A. Diet, M. Villegas, G. Baudoin
Abstract:
This work concerns the measurements of a Bulk Acoustic Waves (BAW) emission filter S parameters and compare with prototypes simulated types. Thanks to HP-ADS, a co-simulation of filters- characteristics in a digital radio-communication chain is performed. Four cases of modulation schemes are studied in order to illustrate the impact of the spectral occupation of the modulated signal. Results of simulations and co-simulation are given in terms of Error Vector Measurements to be useful for a general sensibility analysis of 4th/3rd Generation (G.) emitters (wideband QAM and OFDM signals)Keywords: RF architectures, BAW filters.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27731014 Volunteers’ Preparedness for Natural Disasters and EVANDE Project
Authors: A. Kourou, A. Ioakeimidou, E. Bafa, C. Fassoulas, M. Panoutsopoulou
Abstract:
The role of volunteers in disaster management is of decisive importance and the need of their involvement is well recognized, both for prevention measures and for disaster management. During major catastrophes, whereas professional personnel are outsourced, the role of volunteers is crucial. In Greece experience has shown that various groups operating in the civil protection mechanism like local administration staff or volunteers, in many cases do not have the necessary knowledge and information on best practices to act against natural disasters. One of the major problems is the lack of volunteers’ education and training. In the above given framework, this paper presents the results of a survey aimed to identify the level of education and preparedness of civil protection volunteers in Greece. Furthermore, the implementation of earthquake protection measures at individual, family and working level, are explored. More specifically, the survey questionnaire investigates issues regarding pre-earthquake protection actions, appropriate attitudes and behaviors during an earthquake and existence of contingency plans in the workplace. The questionnaires were administered to citizens from different regions of the country and who attend the civil protection training program: “Protect Myself and Others”. A closed-form questionnaire was developed for the survey, which contained questions regarding the following: a) knowledge of self-protective actions; b) existence of emergency planning at home; c) existence of emergency planning at workplace (hazard mitigation actions, evacuation plan, and performance of drills); and, d) respondents` perception about their level of earthquake preparedness. The results revealed a serious lack of knowledge and preparedness among respondents. Taking into consideration the aforementioned gap and in order to raise awareness and improve preparedness and effective response of volunteers acting in civil protection, the EVANDE project was submitted and approved by the European Commission (EC). The aim of that project is to educate and train civil protection volunteers on the most serious natural disasters, such as forest fires, floods, and earthquakes, and thus, increase their performance.
Keywords: Civil protection, earthquake preparedness, volunteers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12141013 Complex Energy Signal Model for Digital Human Fingerprint Matching
Authors: Jason Zalev, Reza Sedaghat
Abstract:
This paper describes a complex energy signal model that is isomorphic with digital human fingerprint images. By using signal models, the problem of fingerprint matching is transformed into the signal processing problem of finding a correlation between two complex signals that differ by phase-rotation and time-scaling. A technique for minutiae matching that is independent of image translation, rotation and linear-scaling, and is resistant to missing minutiae is proposed. The method was tested using random data points. The results show that for matching prints the scaling and rotation angles are closely estimated and a stronger match will have a higher correlation.Keywords: Affine Invariant, Fingerprint Recognition, Matching, Minutiae.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13181012 Alternative Computational Arrangements on g-Group (g > 2) Profile Analysis
Authors: Emmanuel U. Ohaegbulem, Felix N. Nwobi
Abstract:
Alternative and simple computational arrangements in carrying out multivariate profile analysis when more than two groups (populations) are involved are presented. These arrangements have been demonstrated to not only yield equivalent results for the test statistics (the Wilks lambdas), but they have less computational efforts relative to other arrangements so far presented in the literature; in addition to being quite simple and easy to apply.Keywords: Coincident profiles, g-group profile analysis, level profiles, parallel profiles, repeated measures MANOVA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12401011 Knowledge Continuity as a Part of Business Continuity Management
Authors: H. Urbancova, J. Urbanec
Abstract:
Today the intangible assets are the capital of knowledge and are the most important and the most valuable resource for organizations. All employees have knowledge independently of the kind of jobs they do. Knowledge is thus an asset, which influences business operations. The objective of this article is to identify knowledge continuity as an objective of business continuity management. The article has been prepared based on the analysis of secondary sources and the evaluation of primary sources of data by means of a quantitative survey conducted in the Czech Republic. The conclusion of the article is that organizations that apply business continuity management do not focus on the preservation of the knowledge of key employees. Organizations ensure knowledge continuity only intuitively, on a random basis, non-systematically and discontinuously. The non-ensuring of knowledge continuity represents a threat of loss of key knowledge for organizations and can also negatively affect business continuity.
Keywords: Business continuity, knowledge, organizations, survey.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35331010 A Development of a Simulation Tool for Production Planning with Capacity-Booking at Specialty Store Retailer of Private Label Apparel Firms
Authors: Erika Yamaguchi, Sirawadee Arunyanrt, Shunichi Ohmori, Kazuho Yoshimoto
Abstract:
In this paper, we suggest a simulation tool to make a decision of monthly production planning for maximizing a profit of Specialty store retailer of Private label Apparel (SPA) firms. Most of SPA firms are fabless and make outsourcing deals for productions with factories of their subcontractors. Every month, SPA firms make a booking for production lines and manpower in the factories. The booking is conducted a few months in advance based on a demand prediction and a monthly production planning at that time. However, the demand prediction is updated month by month, and the monthly production planning would change to meet the latest demand prediction. Then, SPA firms have to change the capacities initially booked within a certain range to suit to the monthly production planning. The booking system is called “capacity-booking”. These days, though it is an issue for SPA firms to make precise monthly production planning, many firms are still conducting the production planning by empirical rules. In addition, it is also a challenge for SPA firms to match their products and factories with considering their demand predictabilities and regulation abilities. In this paper, we suggest a model for considering these two issues. An objective is to maximize a total profit of certain periods, which is sales minus costs of production, inventory, and capacity-booking penalty. To make a better monthly production planning at SPA firms, these points should be considered: demand predictabilities by random trends, previous and next month’s production planning of the target month, and regulation abilities of the capacity-booking. To decide matching products and factories for outsourcing, it is important to consider seasonality, volume, and predictability of each product, production possibility, size, and regulation ability of each factory. SPA firms have to consider these constructions and decide orders with several factories per one product. We modeled these issues as a linear programming. To validate the model, an example of several computational experiments with a SPA firm is presented. We suppose four typical product groups: basic, seasonal (Spring / Summer), seasonal (Fall / Winter), and spot product. As a result of the experiments, a monthly production planning was provided. In the planning, demand predictabilities from random trend are reduced by producing products which are different product types. Moreover, priorities to produce are given to high-margin products. In conclusion, we developed a simulation tool to make a decision of monthly production planning which is useful when the production planning is set every month. We considered the features of capacity-booking, and matching of products and factories which have different features and conditions.
Keywords: Capacity-booking, SPA, monthly production planning, linear programming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13301009 Evolutionary Design of Polynomial Controller
Authors: R. Matousek, S. Lang, P. Minar, P. Pivonka
Abstract:
In the control theory one attempts to find a controller that provides the best possible performance with respect to some given measures of performance. There are many sorts of controllers e.g. a typical PID controller, LQR controller, Fuzzy controller etc. In the paper will be introduced polynomial controller with novel tuning method which is based on the special pole placement encoding scheme and optimization by Genetic Algorithms (GA). The examples will show the performance of the novel designed polynomial controller with comparison to common PID controller.Keywords: Evolutionary design, Genetic algorithms, PID controller, Pole placement, Polynomial controller
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21571008 Measurements of Radial Velocity in Fixed Fluidized Bed for Fischer-Tropsch Synthesis Using LDV
Authors: Xiaolai Zhang, Haitao Zhang, Qiwen Sun, Weixin Qian, Weiyong Ying
Abstract:
High temperature Fischer-Tropsch synthesis process use fixed fluidized bed as a reactor. In order to understand the flow behavior in the fluidized bed better, the research of how the radial velocity affects the entire flow field is necessary. Laser Doppler Velocimetry (LDV) was used to study the radial velocity distribution along the diameter direction of the cross-section of the particle in a fixed fluidized bed. The velocity in the cross-section is fluctuating within a small range. The direction of the speed is a random phenomenon. In addition to r/R is 1, the axial velocity are more than 6 times of the radial velocity, the radial velocity has little impact on the axial velocity in a fixed fluidized bed.Keywords: LDV, fixed fluidized bed, velocity, Fischer-Tropsch synthesis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16281007 Different Teaching Methods for Program Design and Algorithmic Language
Authors: Yue Zhao, Jianping Li
Abstract:
This paper covers the present situation and problem of experimental teaching of mathematics specialty in recent years, puts forward and demonstrates experimental teaching methods for different education. From the aspects of content and experimental teaching approach, uses as an example the course “Experiment for Program Designing & Algorithmic Language" and discusses teaching practice and laboratory course work. In addition a series of successful methods and measures are introduced in experimental teaching.Keywords: Differentiated teaching, experimental teaching, program design and algorithmic language, teaching method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16321006 Classification of Prostate Cell Nuclei using Artificial Neural Network Methods
Authors: M. Sinecen, M. Makinacı
Abstract:
The purpose of this paper is to assess the value of neural networks for classification of cancer and noncancer prostate cells. Gauss Markov Random Fields, Fourier entropy and wavelet average deviation features are calculated from 80 noncancer and 80 cancer prostate cell nuclei. For classification, artificial neural network techniques which are multilayer perceptron, radial basis function and learning vector quantization are used. Two methods are utilized for multilayer perceptron. First method has single hidden layer and between 3-15 nodes, second method has two hidden layer and each layer has between 3-15 nodes. Overall classification rate of 86.88% is achieved.
Keywords: Artificial neural networks, texture classification, cancer diagnosis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1591