Search results for: phase error accumulation methodology
3042 Automatic Classification of the Stand-to-Sit Phase in the TUG Test Using Machine Learning
Authors: Y. A. Adla, R. Soubra, M. Kasab, M. O. Diab, A. Chkeir
Abstract:
Over the past several years, researchers have shown a great interest in assessing the mobility of elderly people to measure their functional status. Usually, such an assessment is done by conducting tests that require the subject to walk a certain distance, turn around, and finally sit back down. Consequently, this study aims to provide an at home monitoring system to assess the patient’s status continuously. Thus, we proposed a technique to automatically detect when a subject sits down while walking at home. In this study, we utilized a Doppler radar system to capture the motion of the subjects. More than 20 features were extracted from the radar signals out of which 11 were chosen based on their Intraclass Correlation Coefficient (ICC > 0.75). Accordingly, the sequential floating forward selection wrapper was applied to further narrow down the final feature vector. Finally, five features were introduced to the Linear Discriminant Analysis classifier and an accuracy of 93.75% was achieved as well as a precision and recall of 95% and 90% respectively.
Keywords: Doppler radar system, stand-to-sit phase, TUG test, machine learning, classification
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4533041 The Optimal Equilibrium Capacity of Information Hiding Based on Game Theory
Authors: Ziquan Hu, Kun She, Shahzad Ali, Kai Yan
Abstract:
Game theory could be used to analyze the conflicted issues in the field of information hiding. In this paper, 2-phase game can be used to build the embedder-attacker system to analyze the limits of hiding capacity of embedding algorithms: the embedder minimizes the expected damage and the attacker maximizes it. In the system, the embedder first consumes its resource to build embedded units (EU) and insert the secret information into EU. Then the attacker distributes its resource evenly to the attacked EU. The expected equilibrium damage, which is maximum damage in value from the point of view of the attacker and minimum from the embedder against the attacker, is evaluated by the case when the attacker attacks a subset from all the EU. Furthermore, the optimal equilibrium capacity of hiding information is calculated through the optimal number of EU with the embedded secret information. Finally, illustrative examples of the optimal equilibrium capacity are presented.Keywords: 2-Phase Game, Expected Equilibrium damage, InformationHiding, Optimal Equilibrium Capacity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16213040 A Framework for Evaluation of Enterprise Architecture Implementation Methodologies
Authors: Babak Darvish Rouhani, Mohd Naz’ri Mahrin, Fatemeh Nikpay, Maryam Khanian Najafabadi, Pourya Nikfard
Abstract:
Enterprise Architecture (EA) Implementation Methodologies have become an important part of EA projects. Several implementation methodologies have been proposed, as a theoretical and practical approach, to facilitate and support the development of EA within an enterprise. A significant question when facing the starting of EA implementation is deciding which methodology to utilize. In order to answer this question, a framework with several criteria is applied in this paper for the comparative analysis of existing EA implementation methodologies. Five EA implementation methodologies including: EAP, TOGAF, DODAF, Gartner, and FEA are selected in order to compare with proposed framework. The results of the comparison indicate that those methodologies have not reached a sufficient maturity as whole due to lack of consideration on requirement management, maintenance, continuum, and complexities in their process. The framework has also ability for the evaluation of any kind of EA implementation methodologies.
Keywords: Enterprise Architecture, Enterprise Architecture Implementation Methodology. EAIM, Evaluating EAIM, Framework for evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55853039 Development of Researcher Knowledge in Mathematics Education: Towards a Confluence Framework
Authors: I. Kontorovich, R. Zazkis
Abstract:
We present a framework of researcher knowledge development in conducting a study in mathematics education. The key components of the framework are: knowledge germane to conducting a particular study, processes of knowledge accumulation, and catalyzing filters that influence a researcher decision making. The components of the framework originated from a confluence between constructs and theories in Mathematics Education, Higher Education and Sociology. Drawing on a self-reflective interview with a leading researcher in mathematics education, Professor Michèle Artigue, we illustrate how the framework can be utilized in data analysis. Criteria for framework evaluation are discussed.
Keywords: Community of practice, knowledge development, mathematics education research, researcher knowledge.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18503038 Quadrotor Black-Box System Identification
Authors: Ionel Stanculeanu, Theodor Borangiu
Abstract:
This paper presents a new approach in the identification of the quadrotor dynamic model using a black-box system for identification. Also the paper considers the problems which appear during the identification in the closed-loop and offers a technical solution for overcoming the correlation between the input noise present in the output
Keywords: System identification, UAV, prediction error method, quadrotor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34593037 Bail-in Capital: The New Box
Authors: Manu Krishnan, Phil Jacoby
Abstract:
In this paper, we discuss the paradigm shift in bank capital from the “gone concern" to the “going concern" mindset. We then propose a methodology for pricing a product of this shift called Contingent Capital Notes (“CoCos"). The Merton Model can determine a price for credit risk by using the firm-s equity value as a call option on those assets. Our pricing methodology for CoCos also uses the credit spread implied by the Merton Model in a subsequent derivative form created by John Hull et al . Here, a market implied asset volatility is calculated by using observed market CDS spreads. This implied asset volatility is then used to estimate the probability of triggering a predetermined “contingency event" given the distanceto- trigger (DTT). The paper then investigates the effect of varying DTTs and recovery assumptions on the CoCo yield. We conclude with an investment rationale.Keywords: CoCo, Contingent capital, Bank Capital, Tier1 Capital
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15443036 Development and Validation of a UPLC Method for the Determination of Albendazole Residues on Pharmaceutical Manufacturing Equipment Surfaces
Authors: R. S. Chandan, M. Vasudevan, Deecaraman, B. M. Gurupadayya
Abstract:
In Pharmaceutical industries, it is very important to remove drug residues from the equipment and areas used. The cleaning procedure must be validated, so special attention must be devoted to the methods used for analysis of trace amounts of drugs. A rapid, sensitive and specific reverse phase ultra performance liquid chromatographic (UPLC) method was developed for the quantitative determination of Albendazole in cleaning validation swab samples. The method was validated using an ACQUITY HSS C18, 50 x 2.1mm, 1.8μ column with a isocratic mobile phase containing a mixture of 1.36g of Potassium dihydrogenphosphate in 1000mL MilliQ water, 2mL of triethylamine and pH adjusted to 2.3 ± 0.05 with ortho-phosphoric acid, Acetonitrile and Methanol (50:40:10 v/v). The flow rate of the mobile phase was 0.5 mL min-1 with a column temperature of 350C and detection wavelength at 254nm using PDA detector. The injection volume was 2µl. Cotton swabs, moisten with acetonitrile were used to remove any residue of drug from stainless steel, teflon, rubber and silicon plates which mimic the production equipment surface and the mean extraction-recovery was found to be 91.8. The selected chromatographic condition was found to effectively elute Albendazole with retention time of 0.67min. The proposed method was found to be linear over the range of 0.2 to 150µg/mL and correlation coefficient obtained is 0.9992. The proposed method was found to be accurate, precise, reproducible and specific and it can also be used for routine quality control analysis of these drugs in biological samples either alone or in combined pharmaceutical dosage forms.
Keywords: Cleaning validation, Albendazole, residues, swab analysis, UPLC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31053035 Optimum Control Strategy of Three-Phase Shunt Active Filter System
Authors: Mihaela Popescu, Alexandru Bitoleanu, Mircea Dobriceanu, Vlad Suru
Abstract:
The aim of this paper is to identify an optimum control strategy of three-phase shunt active filters to minimize the total harmonic distortion factor of the supply current. A classical PIPI cascade control solution of the output current of the active filterand the voltage across the DC capacitor based on Modulus–Optimum criterion is taken into consideration. The control system operation has been simulated using Matlab-Simulink environment and the results agree with the theoretical expectation. It is shown that there is an optimum value of the DC-bus voltage which minimizes the supply current harmonic distortion factor. It corresponds to the equality of the apparent power at the output of the active filter and the apparent power across the capacitor. Finally, predicted results are verified experimentally on a MaxSine active power filter.Keywords: Active filtering, Controller tuning, Modulus Optimum criterion, Optimum control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21563034 Outsourcing the Front End of Innovation
Abstract:
The paper presents a new method for efficient innovation process management. Even though the innovation management methods, tools and knowledge are well established and documented in literature, most of the companies still do not manage it efficiently. Especially in SMEs the front end of innovation - problem identification, idea creation and selection - is often not optimally performed. Our eMIPS methodology represents a sort of "umbrella methodology" - a well-defined set of procedures, which can be dynamically adapted to the concrete case in a company. In daily practice, various methods (e.g. for problem identification and idea creation) can be applied, depending on the company's needs. It is based on the proactive involvement of the company's employees supported by the appropriate methodology and external experts. The presented phases are performed via a mixture of face-to-face activities (workshops) and online (eLearning) activities taking place in eLearning Moodle environment and using other e-communication channels. One part of the outcomes is an identified set of opportunities and concrete solutions ready for implementation. The other also very important result is connected to innovation competences for the participating employees related with concrete tools and methods for idea management. In addition, the employees get a strong experience for dynamic, efficient and solution oriented managing of the invention process. The eMIPS also represents a way of establishing or improving the innovation culture in the organization. The first results in a pilot company showed excellent results regarding the motivation of participants and also as to the results achieved.
Keywords: Creativity, distance learning, front end, innovation, problem.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22093033 Persuasive Communication on Social Egg Freezing in California from a Framing Theory Perspective
Authors: Leila Mohammadi
Abstract:
This paper presents the impact of persuasive communication implemented by fertility clinics websites, and how this information influences women at their decision-making for undertaking this procedure. The influential factors for women decisions to do social egg freezing (SEF) are analyzed from a framing theory perspective, with a specific focus on the impact of persuasive information on women’s decision making. This study follows a quantitative approach. A two-phase survey has been conducted to examine the interest rate to undertake SEF. In the first phase, a questionnaire was available during a month (May 2015) to women to answer whether or not they knew enough information of this process, with a total of 230 answers. The second phase took place in the two last weeks of July 2015. All the respondents were invited to a seminars called ‘All about egg freezing’ and afretwards they were requested to answer the second questionnaire. After the seminar, in which they were given an extensive amount of information about egg freezing, a total of 115 women replied the questionnaire. The collected data during this process were analyzed using descriptive statistics. Most of the respondents changed their opinion in the second questionaire which was after receiving information. Although in the first questionnaire their self-evaluation of having knowledge about this process and the implemented technologies was very high, they realized that they still need to access more information from different sources in order to be able to make a decision. The study reached the conclusion that persuasive and framed information by clinics would affect the decisions of these women. Despite the reasons women have to do egg freezing and their motivations behind it, providing people necessary information and unprejudiced data about this process (such as its positive and negative aspects, requirements, suppositions, possibilities and consequences) would help them to make a more precise and reasonable decision about what they are buying.
Keywords: Decision making, fertility clinics, framing theory, persuasive information, social egg freezing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9703032 A Propagator Method like Algorithm for Estimation of Multiple Real-Valued Sinusoidal Signal Frequencies
Authors: Sambit Prasad Kar, P.Palanisamy
Abstract:
In this paper a novel method for multiple one dimensional real valued sinusoidal signal frequency estimation in the presence of additive Gaussian noise is postulated. A computationally simple frequency estimation method with efficient statistical performance is attractive in many array signal processing applications. The prime focus of this paper is to combine the subspace-based technique and a simple peak search approach. This paper presents a variant of the Propagator Method (PM), where a collaborative approach of SUMWE and Propagator method is applied in order to estimate the multiple real valued sine wave frequencies. A new data model is proposed, which gives the dimension of the signal subspace is equal to the number of frequencies present in the observation. But, the signal subspace dimension is twice the number of frequencies in the conventional MUSIC method for estimating frequencies of real-valued sinusoidal signal. The statistical analysis of the proposed method is studied, and the explicit expression of asymptotic (large-sample) mean-squared-error (MSE) or variance of the estimation error is derived. The performance of the method is demonstrated, and the theoretical analysis is substantiated through numerical examples. The proposed method can achieve sustainable high estimation accuracy and frequency resolution at a lower SNR, which is verified by simulation by comparing with conventional MUSIC, ESPRIT and Propagator Method.
Keywords: Frequency estimation, peak search, subspace-based method without eigen decomposition, quadratic convex function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17313031 Geochemical Assessment of Heavy Metals Concentration in Surface Sediment of West Port, Malaysia
Authors: B.Tavakoly Sany, A. Salleh, A.H .Sulaiman, A. Mehdinia, GH. Monazami
Abstract:
One year (November 2009-October 2010) sediment monitoring was used to evaluate pollution status, concentration and distribution of heavy metals (As, Cu, Cd, Cr, Hg, Ni, Pb and Zn) in West Port of Malaysia. Sediment sample were collected from nine stations every four months. Geo-accumulation factor and Pollution Load Index (PLI) were estimated to better understand the pollution level in study area. The heavy metal concentration (Mg/g dry weight) were ranged from 20.2 to 162 for As, 7.4 to 27.6 for Cu, 0.244 to 3.53 for Cd, 11.5 to 61.5 for Cr, 0.11 to 0.409 for Hg, 7.2 to 22.2 for Ni, 22.3 to 80 for Pb and 23 to 98.3 for Zn. In general, concentration some metals (As,Cd, Hg and Pb) was higher than background values that are considered as serious concern for aquatic life and the human health.
Keywords: Heavy metals, Sediment Quality, geo-accumulationindex, Pollution Load Index
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25293030 Investigation of Combined use of MFCC and LPC Features in Speech Recognition Systems
Authors: К. R. Aida–Zade, C. Ardil, S. S. Rustamov
Abstract:
Statement of the automatic speech recognition problem, the assignment of speech recognition and the application fields are shown in the paper. At the same time as Azerbaijan speech, the establishment principles of speech recognition system and the problems arising in the system are investigated. The computing algorithms of speech features, being the main part of speech recognition system, are analyzed. From this point of view, the determination algorithms of Mel Frequency Cepstral Coefficients (MFCC) and Linear Predictive Coding (LPC) coefficients expressing the basic speech features are developed. Combined use of cepstrals of MFCC and LPC in speech recognition system is suggested to improve the reliability of speech recognition system. To this end, the recognition system is divided into MFCC and LPC-based recognition subsystems. The training and recognition processes are realized in both subsystems separately, and recognition system gets the decision being the same results of each subsystems. This results in decrease of error rate during recognition. The training and recognition processes are realized by artificial neural networks in the automatic speech recognition system. The neural networks are trained by the conjugate gradient method. In the paper the problems observed by the number of speech features at training the neural networks of MFCC and LPC-based speech recognition subsystems are investigated. The variety of results of neural networks trained from different initial points in training process is analyzed. Methodology of combined use of neural networks trained from different initial points in speech recognition system is suggested to improve the reliability of recognition system and increase the recognition quality, and obtained practical results are shown.Keywords: Speech recognition, cepstral analysis, Voice activation detection algorithm, Mel Frequency Cepstral Coefficients, features of speech, Cepstral Mean Subtraction, neural networks, Linear Predictive Coding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9133029 Ultrasound Assisted Method to Increase the Aluminum Dissolve Rate from Acidified Water
Authors: Wen Po Cheng, Chi Hua Fu, Ping Hung Chen, Ruey Fang Yu
Abstract:
Aluminum salt that is generally presents as a solid phase in the water purification sludge (WPS) can be dissolved, recovering a liquid phase, by adding strong acid to the sludge solution. According to the reaction kinetics, when reactant is in the form of small particles with a large specific surface area, or when the reaction temperature is high, the quantity of dissolved aluminum salt or reaction rate, respectively are high. Therefore, in this investigation, water purification sludge (WPS) solution was treated with ultrasonic waves to break down the sludge, and different acids (1 N HCl and 1 N H2SO4) were used to acidify it. Acid dosages that yielded the solution pH of less than two were used. The results thus obtained indicate that the quantity of dissolved aluminum in H2SO4-acidified solution exceeded that in HCl-acidified solution. Additionally, ultrasonic treatment increased the rate of dissolution of aluminum and the amount dissolved. The quantity of aluminum dissolved at 60℃ was 1.5 to 2.0 times higher than that at 25℃.Keywords: Coagulant, Aluminum, Ultrasonic, Acidification, Temperature, Sludge.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22763028 Effect of Alkaline Activator, Water, Superplasticiser and Slag Contents on the Compressive Strength and Workability of Slag-Fly Ash Based Geopolymer Mortar Cured under Ambient Temperature
Authors: M. Al-Majidi, A. Lampropoulos, A. Cundy
Abstract:
Geopolymer (cement-free) concrete is the most promising green alternative to ordinary Portland cement concrete and other cementitious materials. While a range of different geopolymer concretes have been produced, a common feature of these concretes is heat curing treatment which is essential in order to provide sufficient mechanical properties in the early age. However, there are several practical issues with the application of heat curing in large-scale structures. The purpose of this study is to develop cement-free concrete without heat curing treatment. Experimental investigations were carried out in two phases. In the first phase (Phase A), the optimum content of water, polycarboxylate based superplasticizer contents and potassium silicate activator in the mix was determined. In the second stage (Phase B), the effect of ground granulated blast furnace slag (GGBFS) incorporation on the compressive strength of fly ash (FA) and Slag based geopolymer mixtures was evaluated. Setting time and workability were also conducted alongside with compressive tests. The results showed that as the slag content was increased the setting time was reduced while the compressive strength was improved. The obtained compressive strength was in the range of 40-50 MPa for 50% slag replacement mixtures. Furthermore, the results indicated that increment of water and superplasticizer content resulted to retarding of the setting time and slight reduction of the compressive strength. The compressive strength of the examined mixes was considerably increased as potassium silicate content was increased.
Keywords: Fly ash, geopolymer, potassium silicate, room temperature treatment, slag.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26113027 Highly Conductive Polycrystalline Metallic Ring in a Magnetic Field
Authors: Isao Tomita
Abstract:
Electrical conduction in a quasi-one-dimensional polycrystalline metallic ring with a long electron phase coherence length realized at low temperature is investigated. In this situation, the wave nature of electrons is important in the ring, where the electrical current I can be induced by a vector potential that arises from a static magnetic field applied perpendicularly to the ring’s area. It is shown that if the average grain size of the polycrystalline ring becomes large (or comparable to the Fermi wavelength), the electrical current I increases to ~I0, where I0 is a current in a disorder-free ring. The cause of this increasing effect is examined, and this takes place if the electron localization length in the polycrystalline potential increases with increasing grain size, which gives rise to coherent connection of tails of a localized electron wave function in the ring and thus provides highly coherent electrical conduction.Keywords: Electrical Conduction, Electron Phase Coherence, Polycrystalline Metal, Magnetic Field.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16303026 Effect of Muscle Loss on Hip Muscular Effort during the Swing Phase of Transfemoral Amputee Gait: A Simulation Study
Authors: Dabiri Y, Najarian S, Eslami M R., Zahedi S, Moser D, Shirzad E, Allami M
Abstract:
The effect of muscle loss due to transfemoral amputation, on energy expenditure of hip joint and individual residual muscles was simulated. During swing phase of gait, with each muscle as an ideal force generator, the lower extremity was modeled as a two-degree of freedom linkage, for which hip and knee were joints. According to results, muscle loss will not lead to higher energy expenditure of hip joint, as long as other parameters of limb remain unaffected. This finding maybe due to the role of biarticular muscles in hip and knee joints motion. Moreover, if hip flexors are removed from the residual limb, residual flexors, and if hip extensors are removed, residual extensors will do more work. In line with the common practice in transfemoral amputation, this result demonstrates during transfemoral amputation, it is important to maintain the length of residual limb as much as possible.Keywords: Amputation Level, Simulation, Transfemoral Amputee.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17843025 A VR Cybersecurity Training Knowledge-Based Ontology
Authors: Shaila Rana, Wasim Alhamdani
Abstract:
Effective cybersecurity learning relies on an engaging, interactive, and entertaining activity that fosters positive learning outcomes. VR cybersecurity training may provide a training format that is engaging, interactive, and entertaining. A methodological approach and framework are needed to allow trainers and educators to employ VR cybersecurity training methods to promote positive learning outcomes. Thus, this paper aims to create an approach that cybersecurity trainers can follow to create a VR cybersecurity training module. This methodology utilizes concepts from other cybersecurity training frameworks, such as NICE and CyTrONE. Other cybersecurity training frameworks do not incorporate the use of VR. VR training proposes unique challenges that cannot be addressed in current cybersecurity training frameworks. Subsequently, this ontology utilizes concepts to develop VR training to create a relevant methodology for creating VR cybersecurity training modules.
Keywords: Virtual reality cybersecurity training, VR cybersecurity training, traditional cybersecurity training, ontology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5853024 Using Artificial Neural Network to Forecast Groundwater Depth in Union County Well
Authors: Zahra Ghadampour, Gholamreza Rakhshandehroo
Abstract:
A concern that researchers usually face in different applications of Artificial Neural Network (ANN) is determination of the size of effective domain in time series. In this paper, trial and error method was used on groundwater depth time series to determine the size of effective domain in the series in an observation well in Union County, New Jersey, U.S. different domains of 20, 40, 60, 80, 100, and 120 preceding day were examined and the 80 days was considered as effective length of the domain. Data sets in different domains were fed to a Feed Forward Back Propagation ANN with one hidden layer and the groundwater depths were forecasted. Root Mean Square Error (RMSE) and the correlation factor (R2) of estimated and observed groundwater depths for all domains were determined. In general, groundwater depth forecast improved, as evidenced by lower RMSEs and higher R2s, when the domain length increased from 20 to 120. However, 80 days was selected as the effective domain because the improvement was less than 1% beyond that. Forecasted ground water depths utilizing measured daily data (set #1) and data averaged over the effective domain (set #2) were compared. It was postulated that more accurate nature of measured daily data was the reason for a better forecast with lower RMSE (0.1027 m compared to 0.255 m) in set #1. However, the size of input data in this set was 80 times the size of input data in set #2; a factor that may increase the computational effort unpredictably. It was concluded that 80 daily data may be successfully utilized to lower the size of input data sets considerably, while maintaining the effective information in the data set.Keywords: Neural networks, groundwater depth, forecast.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25163023 Big Data: Concepts, Technologies and Applications in the Public Sector
Authors: A. Alexandru, C. A. Alexandru, D. Coardos, E. Tudora
Abstract:
Big Data (BD) is associated with a new generation of technologies and architectures which can harness the value of extremely large volumes of very varied data through real time processing and analysis. It involves changes in (1) data types, (2) accumulation speed, and (3) data volume. This paper presents the main concepts related to the BD paradigm, and introduces architectures and technologies for BD and BD sets. The integration of BD with the Hadoop Framework is also underlined. BD has attracted a lot of attention in the public sector due to the newly emerging technologies that allow the availability of network access. The volume of different types of data has exponentially increased. Some applications of BD in the public sector in Romania are briefly presented.
Keywords: Big data, big data Analytics, Hadoop framework, cloud computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23213022 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling
Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra
Abstract:
Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.
Keywords: Multi-temporal satellite image, urban growth, Non-stationarity, stochastic modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15043021 An Evaluation Method for Two-Dimensional Position Errors and Assembly Errors of a Rotational Table on a 4 Axis Machine Tool
Authors: Jooho Hwang, Chang-Kyu Song, Chun-Hong Park
Abstract:
This paper describes a method to measure and compensate a 4 axes ultra-precision machine tool that generates micro patterns on the large surfaces. The grooving machine is usually used for making a micro mold for many electrical parts such as a light guide plate for LCD and fuel cells. The ultra precision machine tool has three linear axes and one rotational table. Shaping is usually used to generate micro patterns. In the case of 50 μm pitch and 25 μm height pyramid pattern machining with a 90° wedge angle bite, one of linear axis is used for long stroke motion for high cutting speed and other linear axis are used for feeding. The triangular patterns can be generated with many times of long stroke of one axis. Then 90° rotation of work piece is needed to make pyramid patterns with superposition of machined two triangular patterns. To make a two dimensional positioning error, straightness of two axes in out of plane, squareness between the each axis are important. Positioning errors, straightness and squarness were measured by laser interferometer system. Those were compensated and confirmed by ISO230-6. One of difficult problem to measure the error motions is squareness or parallelism of axis between the rotational table and linear axis. It was investigated by simultaneous moving of rotary table and XY axes. This compensation method is introduced in this paper.Keywords: Ultra-precision machine tool, muti-axis errors, squraness, positioning errors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15813020 An Approach for Integration of Industrial Robot with Vision System and Simulation Software
Authors: Ahmed Sh. Khusheef, Ganesh Kothapalli, Majid Tolouei-Rad
Abstract:
Utilization of various sensors has made it possible to extend capabilities of industrial robots. Among these are vision sensors that are used for providing visual information to assist robot controllers. This paper presents a method of integrating a vision system and a simulation program with an industrial robot. The vision system is employed to detect a target object and compute its location in the robot environment. Then, the target object-s information is sent to the robot controller via parallel communication port. The robot controller uses the extracted object information and the simulation program to control the robot arm for approaching, grasping and relocating the object. This paper presents technical details of system components and describes the methodology used for this integration. It also provides a case study to prove the validity of the methodology developed.Keywords: industrial robot, integration, simulation, vision system
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22243019 Generative Adversarial Network Based Fingerprint Anti-Spoofing Limitations
Authors: Yehjune Heo
Abstract:
Fingerprint Anti-Spoofing approaches have been actively developed and applied in real-world applications. One of the main problems for Fingerprint Anti-Spoofing is not robust to unseen samples, especially in real-world scenarios. A possible solution will be to generate artificial, but realistic fingerprint samples and use them for training in order to achieve good generalization. This paper contains experimental and comparative results with currently popular GAN based methods and uses realistic synthesis of fingerprints in training in order to increase the performance. Among various GAN models, the most popular StyleGAN is used for the experiments. The CNN models were first trained with the dataset that did not contain generated fake images and the accuracy along with the mean average error rate were recorded. Then, the fake generated images (fake images of live fingerprints and fake images of spoof fingerprints) were each combined with the original images (real images of live fingerprints and real images of spoof fingerprints), and various CNN models were trained. The best performances for each CNN model, trained with the dataset of generated fake images and each time the accuracy and the mean average error rate, were recorded. We observe that current GAN based approaches need significant improvements for the Anti-Spoofing performance, although the overall quality of the synthesized fingerprints seems to be reasonable. We include the analysis of this performance degradation, especially with a small number of samples. In addition, we suggest several approaches towards improved generalization with a small number of samples, by focusing on what GAN based approaches should learn and should not learn.
Keywords: Anti-spoofing, CNN, fingerprint recognition, GAN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5943018 Sustainable Development in Construction
Authors: Ali Hemmati, Ali Kheyroddin
Abstract:
Semnan is a city in semnan province, northern Iran with a population estimated at 119,778 inhabitants. It is the provincial capital of semnan province. Iran is a developing country and construction is a basic factor of developing too. Hence, Semnan city needs to a special programming for construction of buildings, structures and infrastructures. Semnan municipality tries to begin this program. In addition to, city has some historical monuments which can be interesting for tourists. Hence, Semnan inhabitants can benefit from tourist industry. Optimization of Energy in construction industry is another activity of this municipality and the inhabitants who execute these regulations receive some discounts. Many parts of Iran such as semnan are located in highly seismic zones and structures must be constructed safe e.g., according to recent seismic codes. In this paper opportunities of IT in construction industry of Iran are investigated in three categories. Pre-construction phase, construction phase and earthquake disaster mitigation are studied. Studies show that information technology can be used in these items for reducing the losses and increasing the benefits. Both government and private sectors must contribute to this strategic project for obtaining the best result.Keywords: approval, building, construction, document, industry, IT, Semnan
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15463017 Surface Elevation Dynamics Assessment Using Digital Elevation Models, Light Detection and Ranging, GPS and Geospatial Information Science Analysis: Ecosystem Modelling Approach
Authors: Ali K. M. Al-Nasrawi, Uday A. Al-Hamdany, Sarah M. Hamylton, Brian G. Jones, Yasir M. Alyazichi
Abstract:
Surface elevation dynamics have always responded to disturbance regimes. Creating Digital Elevation Models (DEMs) to detect surface dynamics has led to the development of several methods, devices and data clouds. DEMs can provide accurate and quick results with cost efficiency, in comparison to the inherited geomatics survey techniques. Nowadays, remote sensing datasets have become a primary source to create DEMs, including LiDAR point clouds with GIS analytic tools. However, these data need to be tested for error detection and correction. This paper evaluates various DEMs from different data sources over time for Apple Orchard Island, a coastal site in southeastern Australia, in order to detect surface dynamics. Subsequently, 30 chosen locations were examined in the field to test the error of the DEMs surface detection using high resolution global positioning systems (GPSs). Results show significant surface elevation changes on Apple Orchard Island. Accretion occurred on most of the island while surface elevation loss due to erosion is limited to the northern and southern parts. Concurrently, the projected differential correction and validation method aimed to identify errors in the dataset. The resultant DEMs demonstrated a small error ratio (≤ 3%) from the gathered datasets when compared with the fieldwork survey using RTK-GPS. As modern modelling approaches need to become more effective and accurate, applying several tools to create different DEMs on a multi-temporal scale would allow easy predictions in time-cost-frames with more comprehensive coverage and greater accuracy. With a DEM technique for the eco-geomorphic context, such insights about the ecosystem dynamic detection, at such a coastal intertidal system, would be valuable to assess the accuracy of the predicted eco-geomorphic risk for the conservation management sustainability. Demonstrating this framework to evaluate the historical and current anthropogenic and environmental stressors on coastal surface elevation dynamism could be profitably applied worldwide.
Keywords: DEMs, eco-geomorphic-dynamic processes, geospatial information science. Remote sensing, surface elevation changes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11583016 Hydrogen Permeability of BSCY Proton-Conducting Perovskite Membrane
Authors: M. Heidari, A. Safekordi, A. Zamaniyan, E. Ganji Babakhani, M. Amanipour
Abstract:
Perovskite-type membrane Ba0.5Sr0.5Ce0.9Y0.1O3-δ (BSCY) was successfully synthesized by liquid citrate method. The hydrogen permeation and stability of BSCY perovskite-type membranes were studied at high temperatures. The phase structure of the powder was characterized by X-ray diffraction (XRD). Scanning electron microscopy (SEM) was used to characterize microstructures of the membrane sintered under various conditions. SEM results showed that increasing in sintering temperature, formed dense membrane with clear grains. XRD results for BSCY membrane that sintered in 1150 °C indicated single phase perovskite structure with orthorhombic configuration, and SEM results showed dense structure with clear grain size which is suitable for permeation tests. Partial substitution of Sr with Ba in SCY structure improved the hydrogen permeation flux through the membrane due to the larger ionic radius of Ba2+. BSCY membrane shows high hydrogen permeation flux of 1.6 ml/min.cm2 at 900 °C and partial pressure of 0.6.
Keywords: Hydrogen separation, perovskite, proton conducting membrane.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10613015 A Study on Algorithm Fusion for Recognition and Tracking of Moving Robot
Authors: Jungho Choi, Youngwan Cho
Abstract:
This paper presents an algorithm for the recognition and tracking of moving objects, 1/10 scale model car is used to verify performance of the algorithm. Presented algorithm for the recognition and tracking of moving objects in the paper is as follows. SURF algorithm is merged with Lucas-Kanade algorithm. SURF algorithm has strong performance on contrast, size, rotation changes and it recognizes objects but it is slow due to many computational complexities. Processing speed of Lucas-Kanade algorithm is fast but the recognition of objects is impossible. Its optical flow compares the previous and current frames so that can track the movement of a pixel. The fusion algorithm is created in order to solve problems which occurred using the Kalman Filter to estimate the position and the accumulated error compensation algorithm was implemented. Kalman filter is used to create presented algorithm to complement problems that is occurred when fusion two algorithms. Kalman filter is used to estimate next location, compensate for the accumulated error. The resolution of the camera (Vision Sensor) is fixed to be 640x480. To verify the performance of the fusion algorithm, test is compared to SURF algorithm under three situations, driving straight, curve, and recognizing cars behind the obstacles. Situation similar to the actual is possible using a model vehicle. Proposed fusion algorithm showed superior performance and accuracy than the existing object recognition and tracking algorithms. We will improve the performance of the algorithm, so that you can experiment with the images of the actual road environment.Keywords: SURF, Optical Flow Lucas-Kanade, Kalman Filter, object recognition, object tracking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22923014 Identification of the Best Blend Composition of Natural Rubber-High Density Polyethylene Blends for Roofing Applications
Authors: W. V. W. H. Wickramaarachchi, S. Walpalage, S. M. Egodage
Abstract:
Thermoplastic elastomer (TPE) is a multifunctional polymeric material which possesses a combination of excellent properties of parent materials. Basically, TPE has a rubber phase and a thermoplastic phase which gives processability as thermoplastics. When the rubber phase is partially or fully crosslinked in the thermoplastic matrix, TPE is called as thermoplastic elastomer vulcanizate (TPV). If the rubber phase is non-crosslinked, it is called as thermoplastic elastomer olefin (TPO). Nowadays TPEs are introduced into the commercial market with different products. However, the application of TPE as a roofing material is limited. Out of the commercially available roofing products from different materials, only single ply roofing membranes and plastic roofing sheets are produced from rubbers and plastics. Natural rubber (NR) and high density polyethylene (HDPE) are used in various industrial applications individually with some drawbacks. Therefore, this study was focused to develop both TPO and TPV blends from NR and HDPE at different compositions and then to identify the best blend composition to use as a roofing material. A series of blends by varying NR loading from 10 wt% to 50 wt%, at 10 wt% intervals, were prepared using a twin screw extruder. Dicumyl peroxide was used as a crosslinker for TPV. The standard properties for a roofing material like tensile properties tear strength, hardness, impact strength, water absorption, swell/gel analysis and thermal characteristics of the blends were investigated. Change of tensile strength after exposing to UV radiation was also studied. Tensile strength, hardness, tear strength, melting temperature and gel content of TPVs show higher values compared to TPOs at every loading studied, while water absorption and swelling index show lower values, suggesting TPVs are more suitable than TPOs for roofing applications. Most of the optimum properties were shown at 10/90 (NR/HDPE) composition. However, high impact strength and gel content were shown at 20/80 (NR/HDPE) composition. Impact strength, as being an energy absorbing property, is the most important for a roofing material in order to resist impact loads. Therefore, 20/80 (NR/HDPE) is identified as the best blend composition. UV resistance and other properties required for a roofing material could be achieved by incorporating suitable additives to TPVs.
Keywords: Thermoplastic elastomer, natural rubber, high density polyethylene, roofing material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9593013 Parametric Study of a Vapor Compression Refrigeration Cycle Using a Two-Phase Constant Area Ejector
Authors: E. Elgendy
Abstract:
There are several ways of improving the performance of a vapor compression refrigeration cycle. Use of an ejector as expansion device is one of the alternative ways. The present paper aims at evaluate the performance improvement of a vapor compression refrigeration cycle under a wide range of operating conditions. A numerical model is developed and a parametric study of important parameters such as condensation (30-50°C) and evaporation temperatures (-20-5°C), nozzle and diffuser efficiencies (0.75-0.95), subcooling and superheating degrees (0-15K) are investigated. The model verification gives a good agreement with the literature data. The simulation results revealed that condensation temperature has the highest effect (129%) on the performance improvement ratio while superheating has the lowest one (6.2%). Among ejector efficiencies, the diffuser efficiency has a significant effect on the COP of ejector expansion refrigeration cycle. The COP improvement percentage decreases from 10.9% to 4.6% as subcooling degrees increases by 15K.
Keywords: Numerical modeling, R134a, Two phase ejector, Vapor compression refrigeration system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5809