Search results for: multiple simultaneous requests
383 Optimal Channel Equalization for MIMO Time-Varying Channels
Authors: Ehab F. Badran, Guoxiang Gu
Abstract:
We consider optimal channel equalization for MIMO (multi-input/multi-output) time-varying channels in the sense of MMSE (minimum mean-squared-error), where the observation noise can be non-stationary. We show that all ZF (zero-forcing) receivers can be parameterized in an affine form which eliminates completely the ISI (inter-symbol-interference), and optimal channel equalizers can be designed through minimization of the MSE (mean-squarederror) between the detected signals and the transmitted signals, among all ZF receivers. We demonstrate that the optimal channel equalizer is a modified Kalman filter, and show that under the AWGN (additive white Gaussian noise) assumption, the proposed optimal channel equalizer minimizes the BER (bit error rate) among all possible ZF receivers. Our results are applicable to optimal channel equalization for DWMT (discrete wavelet multitone), multirate transmultiplexers, OFDM (orthogonal frequency division multiplexing), and DS (direct sequence) CDMA (code division multiple access) wireless data communication systems. A design algorithm for optimal channel equalization is developed, and several simulation examples are worked out to illustrate the proposed design algorithm.Keywords: Channel equalization, Kalman filtering, Time-varying systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1834382 Medical Image Fusion Based On Redundant Wavelet Transform and Morphological Processing
Authors: P. S. Gomathi, B. Kalaavathi
Abstract:
The process in which the complementary information from multiple images is integrated to provide composite image that contains more information than the original input images is called image fusion. Medical image fusion provides useful information from multimodality medical images that provides additional information to the doctor for diagnosis of diseases in a better way. This paper represents the wavelet based medical image fusion algorithm on different multimodality medical images. In order to fuse the medical images, images are decomposed using Redundant Wavelet Transform (RWT). The high frequency coefficients are convolved with morphological operator followed by the maximum-selection (MS) rule. The low frequency coefficients are processed by MS rule. The reconstructed image is obtained by inverse RWT. The quantitative measures which includes Mean, Standard Deviation, Average Gradient, Spatial frequency, Edge based Similarity Measures are considered for evaluating the fused images. The performance of this proposed method is compared with Pixel averaging, PCA, and DWT fusion methods. When compared with conventional methods, the proposed framework provides better performance for analysis of multimodality medical images.
Keywords: Discrete Wavelet Transform (DWT), Image Fusion, Morphological Processing, Redundant Wavelet Transform (RWT).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2158381 Verification of K-ω SST Turbulence Model for Supersonic Internal Flows
Abstract:
In this work, we try to find the best setting of Computational Fluid Dynamic solver available for the problems in the field of supersonic internal flows. We used the supersonic air-toair ejector to represent the typical problem in focus. There are multiple oblique shock waves, shear layers, boundary layers and normal shock interacting in the supersonic ejector making this device typical in field of supersonic inner flows. Modeling of shocks in general is demanding on the physical model of fluid, because ordinary conservation equation does not conform to real conditions in the near-shock region as found in many works. From these reasons, we decided to take special care about solver setting in this article by means of experimental approach of color Schlieren pictures and pneumatic measurement. Fast pressure transducers were used to measure unsteady static pressure in regimes with normal shock in mixing chamber. Physical behavior of ejector in several regimes is discussed. Best choice of eddy-viscosity setting is discussed on the theoretical base. The final verification of the k-ω SST is done on the base of comparison between experiment and numerical results.Keywords: CFD simulations, color Schlieren, k-ω SST, supersonic flows, shock waves.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6472380 Geospatial Network Analysis Using Particle Swarm Optimization
Authors: Varun Singh, Mainak Bandyopadhyay, Maharana Pratap Singh
Abstract:
The shortest path (SP) problem concerns with finding the shortest path from a specific origin to a specified destination in a given network while minimizing the total cost associated with the path. This problem has widespread applications. Important applications of the SP problem include vehicle routing in transportation systems particularly in the field of in-vehicle Route Guidance System (RGS) and traffic assignment problem (in transportation planning). Well known applications of evolutionary methods like Genetic Algorithms (GA), Ant Colony Optimization, Particle Swarm Optimization (PSO) have come up to solve complex optimization problems to overcome the shortcomings of existing shortest path analysis methods. It has been reported by various researchers that PSO performs better than other evolutionary optimization algorithms in terms of success rate and solution quality. Further Geographic Information Systems (GIS) have emerged as key information systems for geospatial data analysis and visualization. This research paper is focused towards the application of PSO for solving the shortest path problem between multiple points of interest (POI) based on spatial data of Allahabad City and traffic speed data collected using GPS. Geovisualization of results of analysis is carried out in GIS.
Keywords: GIS, Outliers, PSO, Traffic Data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2893379 Fast 3D Collision Detection Algorithm using 2D Intersection Area
Authors: Taehyun Yoon, Keechul Jung
Abstract:
There are many researches to detect collision between real object and virtual object in 3D space. In general, these techniques are need to huge computing power. So, many research and study are constructed by using cloud computing, network computing, and distribute computing. As a reason of these, this paper proposed a novel fast 3D collision detection algorithm between real and virtual object using 2D intersection area. Proposed algorithm uses 4 multiple cameras and coarse-and-fine method to improve accuracy and speed performance of collision detection. In the coarse step, this system examines the intersection area between real and virtual object silhouettes from all camera views. The result of this step is the index of virtual sensors which has a possibility of collision in 3D space. To decide collision accurately, at the fine step, this system examines the collision detection in 3D space by using the visual hull algorithm. Performance of the algorithm is verified by comparing with existing algorithm. We believe proposed algorithm help many other research, study and application fields such as HCI, augmented reality, intelligent space, and so on.
Keywords: Collision Detection, Computer Vision, Human Computer Interaction, Visual Hull
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2406378 Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data
Authors: J. D. D. Daniel, K. N. Goh, S. M. Yusop
Abstract:
Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.Keywords: Data Transformation Services (DTS), ObjectLinking and Embedding Database (OLEDB), Data Mart, OnlineAnalytical Processing (OLAP), Online Transactional Processing(OLTP).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2039377 Automatic Map Simplification for Visualization on Mobile Devices
Authors: Hang Yu
Abstract:
The visualization of geographic information on mobile devices has become popular as the widespread use of mobile Internet. The mobility of these devices brings about much convenience to people-s life. By the add-on location-based services of the devices, people can have an access to timely information relevant to their tasks. However, visual analysis of geographic data on mobile devices presents several challenges due to the small display and restricted computing resources. These limitations on the screen size and resources may impair the usability aspects of the visualization applications. In this paper, a variable-scale visualization method is proposed to handle the challenge of small mobile display. By merging multiple scales of information into a single image, the viewer is able to focus on the interesting region, while having a good grasp of the surrounding context. This is essentially visualizing the map through a fisheye lens. However, the fisheye lens induces undesirable geometric distortion in the peripheral, which renders the information meaningless. The proposed solution is to apply map generalization that removes excessive information around the peripheral and an automatic smoothing process to correct the distortion while keeping the local topology consistent. The proposed method is applied on both artificial and real geographical data for evaluation.
Keywords: Map simplification, visualization, mobile devices.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1436376 Satellite Imagery Classification Based on Deep Convolution Network
Authors: Zhong Ma, Zhuping Wang, Congxin Liu, Xiangzeng Liu
Abstract:
Satellite imagery classification is a challenging problem with many practical applications. In this paper, we designed a deep convolution neural network (DCNN) to classify the satellite imagery. The contributions of this paper are twofold — First, to cope with the large-scale variance in the satellite image, we introduced the inception module, which has multiple filters with different size at the same level, as the building block to build our DCNN model. Second, we proposed a genetic algorithm based method to efficiently search the best hyper-parameters of the DCNN in a large search space. The proposed method is evaluated on the benchmark database. The results of the proposed hyper-parameters search method show it will guide the search towards better regions of the parameter space. Based on the found hyper-parameters, we built our DCNN models, and evaluated its performance on satellite imagery classification, the results show the classification accuracy of proposed models outperform the state of the art method.
Keywords: Satellite imagery classification, deep convolution network, genetic algorithm, hyper-parameter optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2346375 Incorporating Multiple Supervised Learning Algorithms for Effective Intrusion Detection
Authors: Umar Albalawi, Sang C. Suh, Jinoh Kim
Abstract:
As internet continues to expand its usage with an enormous number of applications, cyber-threats have significantly increased accordingly. Thus, accurate detection of malicious traffic in a timely manner is a critical concern in today’s Internet for security. One approach for intrusion detection is to use Machine Learning (ML) techniques. Several methods based on ML algorithms have been introduced over the past years, but they are largely limited in terms of detection accuracy and/or time and space complexity to run. In this work, we present a novel method for intrusion detection that incorporates a set of supervised learning algorithms. The proposed technique provides high accuracy and outperforms existing techniques that simply utilizes a single learning method. In addition, our technique relies on partial flow information (rather than full information) for detection, and thus, it is light-weight and desirable for online operations with the property of early identification. With the mid-Atlantic CCDC intrusion dataset publicly available, we show that our proposed technique yields a high degree of detection rate over 99% with a very low false alarm rate (0.4%).
Keywords: Intrusion Detection, Supervised Learning, Traffic Classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2035374 Performance Evaluation of Para-virtualization on Modern Mobile Phone Platform
Authors: Yang Xu, Felix Bruns, Elizabeth Gonzalez, Shadi Traboulsi, Klaus Mott, Attila Bilgic
Abstract:
Emergence of smartphones brings to live the concept of converged devices with the availability of web amenities. Such trend also challenges the mobile devices manufactures and service providers in many aspects, such as security on mobile phones, complex and long time design flow, as well as higher development cost. Among these aspects, security on mobile phones is getting more and more attention. Microkernel based virtualization technology will play a critical role in addressing these challenges and meeting mobile market needs and preferences, since virtualization provides essential isolation for security reasons and it allows multiple operating systems to run on one processor accelerating development and cutting development cost. However, virtualization benefits do not come for free. As an additional software layer, it adds some inevitable virtualization overhead to the system, which may decrease the system performance. In this paper we evaluate and analyze the virtualization performance cost of L4 microkernel based virtualization on a competitive mobile phone by comparing the L4Linux, a para-virtualized Linux on top of L4 microkernel, with the native Linux performance using lmbench and a set of typical mobile phone applications.Keywords: L4 microkernel, virtualization overhead, mobilephone.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1976373 Online Partial Discharge Source Localization and Characterization Using Non-Conventional Method
Authors: Ammar Anwar Khan, Nissar R. Wani, Nazar Malik, Abdulrehman Al-Arainy, and Saad Alghuwainem
Abstract:
Power cables are vulnerable to failure due to aging or defects that occur with the passage of time under continuous operation and loading stresses. PD detection and characterization provide information on the location, nature, form and extent of the degradation. As a result, PD monitoring has become an important part of condition based maintenance (CBM) program among power utilities. Online partial discharge (PD) localization of defect sources in power cable system is possible using the time of flight method. The information regarding the time difference between the main and reflected pulses and cable length can help in locating the partial discharge source along the cable length. However, if the length of the cable is not known and the defect source is located at the extreme ends of the cable or in the middle of the cable, then double ended measurement is required to indicate the location of PD source. Use of multiple sensors can also help in discriminating the cable PD or local/ external PD. This paper presents the experience and results from online partial discharge measurements conducted in the laboratory and the challenges in partial discharge source localization.Keywords: Power cables, partial discharge localization, HFCT, condition based monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2826372 The Integrated Urban Strategies Based on Deep Urban History and Modern Technology Study: Tourism and Leisure Industries as Driving Force to Reactivate Historical Area
Authors: Cheng Li, Jie Shen, Yutian Tang
Abstract:
Embracing the upcoming era of urbanization with the challenges of limitation of resources, disappearing cultural identities and conflicts among different groups of stakeholders, new integrated approaches are offered in our urban practice to help decision-makers and stakeholders frame and develop well-conceived, practical strategies for urban developing trajectories to approach urban-level sustainability in multiple social, cultural, ecological dimensions. Through bottom-up participation, we take advantage of tourism and leisure industries as driving forces for urbanization in China to promote integrated sustainable systems, with the hope of approaching both historical and ecological aspects of urban sustainability; and also thanks to top-down participation, we have codes, standards and rules established by the governments to strengthen the implementation of ecological urban sustainability. The results are monitored and evaluated experimentally and multidimensionally and the sustainable systems we constructed with local stakeholder groups turned out to be effective. The presentation of our selected projects would indicate our different focuses on urban sustainability.Keywords: Urban sustainability, integrated urban strategy, tourism and leisure industries, history, modern technology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1587371 Aspect Oriented Software Architecture
Authors: Pradip Peter Dey, Ronald F. Gonzales, Gordon W. Romney, Mohammad Amin, Bhaskar Raj Sinha
Abstract:
Natural language processing systems pose a unique challenge for software architectural design as system complexity has increased continually and systems cannot be easily constructed from loosely coupled modules. Lexical, syntactic, semantic, and pragmatic aspects of linguistic information are tightly coupled in a manner that requires separation of concerns in a special way in design, implementation and maintenance. An aspect oriented software architecture is proposed in this paper after critically reviewing relevant architectural issues. For the purpose of this paper, the syntactic aspect is characterized by an augmented context-free grammar. The semantic aspect is composed of multiple perspectives including denotational, operational, axiomatic and case frame approaches. Case frame semantics matured in India from deep thematic analysis. It is argued that lexical, syntactic, semantic and pragmatic aspects work together in a mutually dependent way and their synergy is best represented in the aspect oriented approach. The software architecture is presented with an augmented Unified Modeling Language.Keywords: Language engineering, parsing, software design, user experience.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743370 Testing the Validity of Maturity Model for E-Government Implementation in Indonesia
Authors: Darmawan Napitupulu, Dana Indra Sensuse, Aniati Murni
Abstract:
The research was conducted to empirically validate the proposed maturity model of e-Government implementation, composed of four dimensions, further specified by 54 success factors as attributes. To do so, there are two steps were performed. First, expert’s judgment was conducted to test its content validity. The second, reliability study was performed to evaluate inter-rater agreement by using Fleiss Kappa approach. The kappa statistic (kappa coefficient) is the most commonly used method for testing the consistency among raters. Fleiss Kappa was a generalization of Kappa in extensions to the case of more than two raters (multiple raters) with multi-categorical ratings. Our findings show that most attributes of the proposed model were related to their corresponding dimensions. According to our results, The percentage of agree answers given by the experts was 73.69% in dimension A, 89.76% in B, 81.5% in C and 60.37% in D. This means that more than half of the attributes of each dimensions were appropriate or relevant to the dimensions they were supposed to measure, while 85% of attributes were relevant enough to their corresponding dimensions. Inter-rater reliability coefficient also showed satisfactory result and interpreted as substantial agreement among raters. Therefore, the proposed model in this paper was valid and reliable to measure the maturity of e-Government implementation.
Keywords: E-Government, Model, Maturity, Validity, Reliability Kappa.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2202369 A Propose of Personnel Assessment Method Including a Two-Way Assessment for Evaluating Evaluators and Employees
Authors: Shunsuke Saito, Kazuho Yoshimoto, Shunichi Ohmori, Sirawadee Arunyanart
Abstract:
In this paper, we suggest a mechanism of assessment that rater and Ratee (or employees) to convince. There are many problems exist in the personnel assessment. In particular, we were focusing on the three. (1) Raters are not sufficiently recognized assessment point. (2) Ratee are not convinced by the mechanism of assessment. (3) Raters (or Evaluators) and ratees have empathy. We suggest 1: Setting of "understanding of the assessment points." 2: Setting of "relative assessment ability." 3: Proposal of two-way assessment mechanism to solve these problems. As a prerequisite, it is assumed that there are multiple raters. This is because has been a growing importance of multi-faceted assessment. In this model, it determines the weight of each assessment point evaluators by the degree of understanding and assessment ability of raters and ratee. We used the ANP (Analytic Network Process) is a theory that an extension of the decision-making technique AHP (Analytic Hierarchy Process). ANP can be to address the problem of forming a network and assessment of Two-Way is possible. We apply this technique personnel assessment, the weights of rater of each point can be reasonably determined. We suggest absolute assessment for Two-Way assessment by ANP. We have verified that the consent of the two approaches is higher than conventional mechanism. Also, human resources consultant we got a comment about the application of the practice.
Keywords: Personnel assessment, ANP (analytic network process), two-way.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 789368 The Determination of Aflatoxins in Paddy and Milled Fractions of Rice in Guyana: Preliminary Results
Authors: Donna M. Morrison, Lambert Chester, Coretta A. N. Samuels, David R. Ledoux
Abstract:
A survey was conducted in the five rice-growing regions in Guyana to determine the presence of aflatoxins in multiple fractions of rice in June/October 2015 growing season. The fractions were paddy, steamed paddy, cargo rice, white rice and parboiled rice. Samples were analyzed by High Performance Liquid Chromatography. A subset of the samples was further analyzed by enzyme-linked immunosorbent assay (ELISA) for concurrence. All analyses were conducted at the University of Missouri, USA. Of the 186 samples tested, 16 had aflatoxin concentrations greater than 20 ppb the recommended limit for aflatoxins in food according to the United States Food and Drug Administration. An additional three samples had aflatoxin B1 concentrations greater than the European Union Commission maximum levels for aflatoxin B1 in rice at 5 µg/kg and total aflatoxins (B1, B2, G1 and G2) at 10 µg/kg. The survey indicates that there is no widespread aflatoxin problem in rice in Guyana. The incidence of aflatoxins appears to be localized.
Keywords: Aflatoxins, enzyme-linked immunosorbent assay, high-performance liquid chromatography, rice fractions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1580367 Nonlinear Impact Responses for a Damped Frame Supported by Nonlinear Springs with Hysteresis Using Fast FEA
Authors: T. Yamaguchi, M. Watanabe, M. Sasajima, C. Yuan, S. Maruyama, T. B. Ibrahim, H. Tomita
Abstract:
This paper deals with nonlinear vibration analysis using finite element method for frame structures consisting of elastic and viscoelastic damping layers supported by multiple nonlinear concentrated springs with hysteresis damping. The frame is supported by four nonlinear concentrated springs near the four corners. The restoring forces of the springs have cubic non-linearity and linear component of the nonlinear springs has complex quantity to represent linear hysteresis damping. The damping layer of the frame structures has complex modulus of elasticity. Further, the discretized equations in physical coordinate are transformed into the nonlinear ordinary coupled differential equations using normal coordinate corresponding to linear natural modes. Comparing shares of strain energy of the elastic frame, the damping layer and the springs, we evaluate the influences of the damping couplings on the linear and nonlinear impact responses. We also investigate influences of damping changed by stiffness of the elastic frame on the nonlinear coupling in the damped impact responses.Keywords: Dynamic response, Nonlinear impact response, Finite Element analysis, Numerical analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1720366 Impact of Out-of-Plane Stiffness of the Diaphragm on Deflection of Wood Light-Frame Shear Walls
Authors: M. M. Bagheri, G. Doudak, M. Gong
Abstract:
The in-plane rigidity of light frame diaphragms has been investigated by researchers due to the importance of this subsystem regarding lateral force distribution between the lateral force resisting system (LFRS). Where research has lacked is in evaluating the impact of out-of-plane raigidity of the diaphragm on the deflection of shear walls. This study aims at investigating the effect of the diaphragm on the behavior of wood light-frame shear walls, in particular its out-of-plane rigidity was simulated by modeling the floors as beam. The out of plane stiffness of the diaphragm was investigated for idealized (infinitely stiff or flexible) as well as “realistic”. The results showed reductions in the shear wall deflection in the magnitude of approximately 80% considering the out of plane rigidity of the diaphragm. It was also concluded that considering conservative estimates of out-of-plane stiffness might lead to a very significant reduction in deflection and that assuming the floor diaphragm to be infinitely rigid out of plan seems to be reasonable. For diaphragms supported on multiple panels, further reduction in the deflection was observed. More work, particularly at the experimental level, is needed to verify the finding obtained in the numerical investigation related to the effect of out of plane diaphragm stiffness.
Keywords: Deflection of light-frame wood shear walls, out-of-plane stiffness of the diaphragm, initial stiffness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 782365 A Cuckoo Search with Differential Evolution for Clustering Microarray Gene Expression Data
Authors: M. Pandi, K. Premalatha
Abstract:
A DNA microarray technology is a collection of microscopic DNA spots attached to a solid surface. Scientists use DNA microarrays to measure the expression levels of large numbers of genes simultaneously or to genotype multiple regions of a genome. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. It is handled by clustering which reveals the natural structures and identifying the interesting patterns in the underlying data. In this paper, gene based clustering in gene expression data is proposed using Cuckoo Search with Differential Evolution (CS-DE). The experiment results are analyzed with gene expression benchmark datasets. The results show that CS-DE outperforms CS in benchmark datasets. To find the validation of the clustering results, this work is tested with one internal and one external cluster validation indexes.
Keywords: DNA, Microarray, genomics, Cuckoo Search, Differential Evolution, Gene expression data, Clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1483364 Enhancing the Effectiveness of Air Defense Systems through Simulation Analysis
Authors: F. Felipe
Abstract:
Air Defense Systems contain high-value assets that are expected to fulfill their mission for several years - in many cases, even decades - while operating in a fast-changing, technology-driven environment. Thus, it is paramount that decision-makers can assess how effective an Air Defense System is in the face of new developing threats, as well as to identify the bottlenecks that could jeopardize the security of the airspace of a country. Given the broad extent of activities and the great variety of assets necessary to achieve the strategic objectives, a systems approach was taken in order to delineate the core requirements and the physical architecture of an Air Defense System. Then, value-focused thinking helped in the definition of the measures of effectiveness. Furthermore, analytical methods were applied to create a formal structure that preliminarily assesses such measures. To validate the proposed methodology, a powerful simulation was also used to determine the measures of effectiveness, now in more complex environments that incorporate both uncertainty and multiple interactions of the entities. The results regarding the validity of this methodology suggest that the approach can support decisions aimed at enhancing the capabilities of Air Defense Systems. In conclusion, this paper sheds some light on how consolidated approaches of Systems Engineering and Operations Research can be used as valid techniques for solving problems regarding a complex and yet vital matter.Keywords: Air defense, effectiveness, system, simulation, decision-support.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 408363 Coding based Synchronization Algorithm for Secondary Synchronization Channel in WCDMA
Authors: Deng Liao, Dongyu Qiu, Ahmed K. Elhakeem
Abstract:
A new code synchronization algorithm is proposed in this paper for the secondary cell-search stage in wideband CDMA systems. Rather than using the Cyclically Permutable (CP) code in the Secondary Synchronization Channel (S-SCH) to simultaneously determine the frame boundary and scrambling code group, the new synchronization algorithm implements the same function with less system complexity and less Mean Acquisition Time (MAT). The Secondary Synchronization Code (SSC) is redesigned by splitting into two sub-sequences. We treat the information of scrambling code group as data bits and use simple time diversity BCH coding for further reliability. It avoids involved and time-costly Reed-Solomon (RS) code computations and comparisons. Analysis and simulation results show that the Synchronization Error Rate (SER) yielded by the new algorithm in Rayleigh fading channels is close to that of the conventional algorithm in the standard. This new synchronization algorithm reduces system complexities, shortens the average cell-search time and can be implemented in the slot-based cell-search pipeline. By taking antenna diversity and pipelining correlation processes, the new algorithm also shows its flexible application in multiple antenna systems.Keywords: WCDMA cell-search, synchronization algorithm, secondary synchronization channel, antenna diversity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2391362 Educational Data Mining: The Case of Department of Mathematics and Computing in the Period 2009-2018
Authors: M. Sitoe, O. Zacarias
Abstract:
University education is influenced by several factors that range from the adoption of strategies to strengthen the whole process to the academic performance improvement of the students themselves. This work uses data mining techniques to develop a predictive model to identify students with a tendency to evasion and retention. To this end, a database of real students’ data from the Department of University Admission (DAU) and the Department of Mathematics and Informatics (DMI) was used. The data comprised 388 undergraduate students admitted in the years 2009 to 2014. The Weka tool was used for model building, using three different techniques, namely: K-nearest neighbor, random forest, and logistic regression. To allow for training on multiple train-test splits, a cross-validation approach was employed with a varying number of folds. To reduce bias variance and improve the performance of the models, ensemble methods of Bagging and Stacking were used. After comparing the results obtained by the three classifiers, Logistic Regression using Bagging with seven folds obtained the best performance, showing results above 90% in all evaluated metrics: accuracy, rate of true positives, and precision. Retention is the most common tendency.
Keywords: Evasion and retention, cross validation, bagging, stacking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 121361 Effects of a Methanol Fraction of the Leaves of Leonotis leonurus on the Blood Pressure and Heart Rate of Normotensive Male Wistar Rats
Authors: K. Obikeze, P. Mugabo, I. Green, D. Dietrich, A. Burger
Abstract:
Leonotisleonurus a shrub indigenous to Southern Africa is widely used in traditional medicine to treat a variety of conditions ranging from skin diseases and cough to epileptic fits and ‘heart problems’. Studies on the aqueous extract of the leaves have indicated cycloxegenase enzyme inhibitory activity and an antihypertensive effect. Five methanol leaf extract fractions (MLEa - MLEe) of L. leonurus were tested on anaesthetized normotensive male Wistar rats (AWR) and isolated perfused working rat hearts (IWH). Fraction MLEc (0.01mg/kg – 0.05mg/kg) induced significant increases in BP and HR in AWR and positive chronotropic and inotropic effects in IWH (1.0mg/ml – 5.0mg/ml). Pre-administration of atenolol (2.0mg/kg) and prazosin (60μg/kg) significantly inhibited MLEc effect on HR and MAP respectively in vivo, while atenolol (7.0mg/ml) pre-perfusion significantly inhibited MLEc effect in vitro. The hypertensive effect of MLEc is probably via β1agonism. Results also indicate the presence of multiple cardioactive compounds in L. leonurus.
Keywords: Cardiovascular effect, in vitro, in vivo, isolated perfused working heart, Leonotis leonurus, rat.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2092360 Clustering Mixed Data Using Non-normal Regression Tree for Process Monitoring
Authors: Youngji Yoo, Cheong-Sool Park, Jun Seok Kim, Young-Hak Lee, Sung-Shick Kim, Jun-Geol Baek
Abstract:
In the semiconductor manufacturing process, large amounts of data are collected from various sensors of multiple facilities. The collected data from sensors have several different characteristics due to variables such as types of products, former processes and recipes. In general, Statistical Quality Control (SQC) methods assume the normality of the data to detect out-of-control states of processes. Although the collected data have different characteristics, using the data as inputs of SQC will increase variations of data, require wide control limits, and decrease performance to detect outof- control. Therefore, it is necessary to separate similar data groups from mixed data for more accurate process control. In the paper, we propose a regression tree using split algorithm based on Pearson distribution to handle non-normal distribution in parametric method. The regression tree finds similar properties of data from different variables. The experiments using real semiconductor manufacturing process data show improved performance in fault detecting ability.Keywords: Semiconductor, non-normal mixed process data, clustering, Statistical Quality Control (SQC), regression tree, Pearson distribution system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1780359 On the Computation of a Common n-finger Robotic Grasp for a Set of Objects
Authors: Avishai Sintov, Roland Menassa, Amir Shapiro
Abstract:
Industrial robotic arms utilize multiple end-effectors, each for a specific part and for a specific task. We propose a novel algorithm which will define a single end-effector’s configuration able to grasp a given set of objects with different geometries. The algorithm will have great benefit in production lines allowing a single robot to grasp various parts. Hence, reducing the number of endeffectors needed. Moreover, the algorithm will reduce end-effector design and manufacturing time and final product cost. The algorithm searches for a common grasp over the set of objects. The search algorithm maps all possible grasps for each object which satisfy a quality criterion and takes into account possible external wrenches (forces and torques) applied to the object. The mapped grasps are- represented by high-dimensional feature vectors which describes the shape of the gripper. We generate a database of all possible grasps for each object in the feature space. Then we use a search and classification algorithm for intersecting all possible grasps over all parts and finding a single common grasp suitable for all objects. We present simulations of planar and spatial objects to validate the feasibility of the approach.
Keywords: Common Grasping, Search Algorithm, Robotic End-Effector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1677358 New Features for Specific JPEG Steganalysis
Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura
Abstract:
We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.
Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2162357 A Multi-Feature Deep Learning Algorithm for Urban Traffic Classification with Limited Labeled Data
Authors: Rohan Putatunda, Aryya Gangopadhyay
Abstract:
Acoustic sensors, if embedded in smart street lights, can help in capturing the activities (car honking, sirens, events, traffic, etc.) in cities. Needless to say, the acoustic data from such scenarios are complex due to multiple audio streams originating from different events, and when decomposed to independent signals, the amount of retrieved data volume is small in quantity which is inadequate to train deep neural networks. So, in this paper, we address the two challenges: a) separating the mixed signals, and b) developing an efficient acoustic classifier under data paucity. So, to address these challenges, we propose an architecture with supervised deep learning, where the initial captured mixed acoustics data are analyzed with Fast Fourier Transformation (FFT), followed by filtering the noise from the signal, and then decomposed to independent signals by fast independent component analysis (Fast ICA). To address the challenge of data paucity, we propose a multi feature-based deep neural network with high performance that is reflected in our experiments when compared to the conventional convolutional neural network (CNN) and multi-layer perceptron (MLP).
Keywords: FFT, ICA, vehicle classification, multi-feature DNN, CNN, MLP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 432356 Performance Determinants for Convenience Store Suppliers
Authors: Zainah Abdullah, Aznur Hajar Abdullah
Abstract:
This paper examines the impact of information and communication technology (ICT) usage, internal relationship, supplier-retailer relationship, logistics services and inventory management on convenience store suppliers- performance. Data was collected from 275 convenience store managers in Malaysia using a set of questionnaire. The multiple linear regression results indicate that inventory management, supplier-retailer relationship, logistics services and internal relationship are predictors of supplier performance as perceived by convenience store managers. However, ICT usage is not a predictor of supplier performance. The study focuses only on convenience stores and petrol station convenience stores and concentrates only on managers. The results provide insights to suppliers who serve convenience stores and possibly similar retail format on factors to consider in improving their service to retailers. The results also provide insights to government in its aspiration to improve business operations of convenience store to consider ways to enhance the adoption of ICT by retailers and suppliers.Keywords: Information and communication technology (ICT), internal relationship, inventory management, logistics services, supplier performance, supplier-retailer relationship.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4009355 Application of Single Subject Experimental Designs in Adapted Physical Activity Research: A Descriptive Analysis
Authors: Jiabei Zhang, Ying Qi
Abstract:
The purpose of this study was to develop a descriptive profile of the adapted physical activity research using single subject experimental designs. All research articles using single subject experimental designs published in the journal of Adapted Physical Activity Quarterly from 1984 to 2013 were employed as the data source. Each of the articles was coded in a subcategory of seven categories: (a) the size of sample; (b) the age of participants; (c) the type of disabilities; (d) the type of data analysis; (e) the type of designs, (f) the independent variable, and (g) the dependent variable. Frequencies, percentages, and trend inspection were used to analyze the data and develop a profile. The profile developed characterizes a small portion of research articles used single subject designs, in which most researchers used a small sample size, recruited children as subjects, emphasized learning and behavior impairments, selected visual inspection with descriptive statistics, preferred a multiple baseline design, focused on effects of therapy, inclusion, and strategy, and measured desired behaviors more often, with a decreasing trend over years.Keywords: Adapted physical activity research, single subject experimental designs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1842354 Evaluation of Fitts’ Law Index of Difficulty Formulation for Screen Size Variations
Authors: Hidehiko Okada, Takayuki Akiba
Abstract:
It is well-known as Fitts’ law that the time for a user to point a target on a GUI screen can be modeled as a linear function of “index of difficulty (ID).” In this paper, the authors investigate whether the traditional ID formulation is appropriate independently of device screen sizes. Result of our experiment reveals that the ID formulation may not consistently capture actual difficulty: users’ pointing performances are not consistent among pointing target variations of which index of difficulty are consistent. The term A/W may not be appropriate because the term causes the observed inconsistency. Based on this finding, the authors then evaluate the applicability of possible models other than Fitts’ one. Multiple regression models are found to be able to appropriately represent the effects of target design variations. The authors next make an attempt to improve the definition of ID in Fitts’ model. Our idea is to raise the size or the distance values depending on the screen size. The modified model is found to fit well to the users’ pointing data, which supports the idea.
Keywords: Fitts’ law, pointing device, small screen, touch user interface, usability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1628