Search results for: video image processing
1019 Recognition of Tifinagh Characters with Missing Parts Using Neural Network
Authors: El Mahdi Barrah, Said Safi, Abdessamad Malaoui
Abstract:
In this paper, we present an algorithm for reconstruction from incomplete 2D scans for tifinagh characters. This algorithm is based on using correlation between the lost block and its neighbors. This system proposed contains three main parts: pre-processing, features extraction and recognition. In the first step, we construct a database of tifinagh characters. In the second step, we will apply “shape analysis algorithm”. In classification part, we will use Neural Network. The simulation results demonstrate that the proposed method give good results.
Keywords: Tifinagh character recognition, Neural networks, Local cost computation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12851018 Coupled Dynamics in Host-Guest Complex Systems Duplicates Emergent Behavior in the Brain
Authors: Sergio Pissanetzky
Abstract:
The ability of the brain to organize information and generate the functional structures we use to act, think and communicate, is a common and easily observable natural phenomenon. In object-oriented analysis, these structures are represented by objects. Objects have been extensively studied and documented, but the process that creates them is not understood. In this work, a new class of discrete, deterministic, dissipative, host-guest dynamical systems is introduced. The new systems have extraordinary self-organizing properties. They can host information representing other physical systems and generate the same functional structures as the brain does. A simple mathematical model is proposed. The new systems are easy to simulate by computer, and measurements needed to confirm the assumptions are abundant and readily available. Experimental results presented here confirm the findings. Applications are many, but among the most immediate are object-oriented engineering, image and voice recognition, search engines, and Neuroscience.
Keywords: AI, artificial intelligence, complex system, object oriented, OO, refactoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20881017 Processing Web-Cam Images by a Neuro-Fuzzy Approach for Vehicular Traffic Monitoring
Authors: A. Faro, D. Giordano, C. Spampinato
Abstract:
Traffic management in an urban area is highly facilitated by the knowledge of the traffic conditions in every street or highway involved in the vehicular mobility system. Aim of the paper is to propose a neuro-fuzzy approach able to compute the main parameters of a traffic system, i.e., car density, velocity and flow, by using the images collected by the web-cams located at the crossroads of the traffic network. The performances of this approach encourage its application when the traffic system is far from the saturation. A fuzzy model is also outlined to evaluate when it is suitable to use more accurate, even if more time consuming, algorithms for measuring traffic conditions near to saturation.
Keywords: Neuro-fuzzy networks, computer vision, Fuzzy systems, intelligent transportation system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15911016 Performance Evaluation of Wavelet Based Coders on Brain MRI Volumetric Medical Datasets for Storage and Wireless Transmission
Authors: D. Dhouib, A. Naït-Ali, C. Olivier, M. S. Naceur
Abstract:
In this paper, we evaluate the performance of some wavelet based coding algorithms such as 3D QT-L, 3D SPIHT and JPEG2K. In the first step we achieve an objective comparison between three coders, namely 3D SPIHT, 3D QT-L and JPEG2K. For this purpose, eight MRI head scan test sets of 256 x 256x124 voxels have been used. Results show superior performance of 3D SPIHT algorithm, whereas 3D QT-L outperforms JPEG2K. The second step consists of evaluating the robustness of 3D SPIHT and JPEG2K coding algorithm over wireless transmission. Compressed dataset images are then transmitted over AWGN wireless channel or over Rayleigh wireless channel. Results show the superiority of JPEG2K over these two models. In fact, it has been deduced that JPEG2K is more robust regarding coding errors. Thus we may conclude the necessity of using corrector codes in order to protect the transmitted medical information.
Keywords: Image coding, medical imaging, wavelet basedcoder, wireless transmission.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19401015 Semi-Supervised Outlier Detection Using a Generative and Adversary Framework
Authors: Jindong Gu, Matthias Schubert, Volker Tresp
Abstract:
In many outlier detection tasks, only training data belonging to one class, i.e., the positive class, is available. The task is then to predict a new data point as belonging either to the positive class or to the negative class, in which case the data point is considered an outlier. For this task, we propose a novel corrupted Generative Adversarial Network (CorGAN). In the adversarial process of training CorGAN, the Generator generates outlier samples for the negative class, and the Discriminator is trained to distinguish the positive training data from the generated negative data. The proposed framework is evaluated using an image dataset and a real-world network intrusion dataset. Our outlier-detection method achieves state-of-the-art performance on both tasks.Keywords: Outlier detection, generative adversary networks, semi-supervised learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10721014 AI-Based Techniques for Online Social Media Network Sentiment Analysis: A Methodical Review
Authors: A. M. John-Otumu, M. M. Rahman, O. C. Nwokonkwo, M. C. Onuoha
Abstract:
Online social media networks have long served as a primary arena for group conversations, gossip, text-based information sharing and distribution. The use of natural language processing techniques for text classification and unbiased decision making has not been far-fetched. Proper classification of these textual information in a given context has also been very difficult. As a result, a systematic review was conducted from previous literature on sentiment classification and AI-based techniques. The study was done in order to gain a better understanding of the process of designing and developing a robust and more accurate sentiment classifier that could correctly classify social media textual information of a given context between hate speech and inverted compliments with a high level of accuracy using the knowledge gain from the evaluation of different artificial intelligence techniques reviewed. The study evaluated over 250 articles from digital sources like ACM digital library, Google Scholar, and IEEE Xplore; and whittled down the number of research to 52 articles. Findings revealed that deep learning approaches such as Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), Bidirectional Encoder Representations from Transformer (BERT), and Long Short-Term Memory (LSTM) outperformed various machine learning techniques in terms of performance accuracy. A large dataset is also required to develop a robust sentiment classifier. Results also revealed that data can be obtained from places like Twitter, movie reviews, Kaggle, Stanford Sentiment Treebank (SST), and SemEval Task4 based on the required domain. The hybrid deep learning techniques like CNN+LSTM, CNN+ Gated Recurrent Unit (GRU), CNN+BERT outperformed single deep learning techniques and machine learning techniques. Python programming language outperformed Java programming language in terms of development simplicity and AI-based library functionalities. Finally, the study recommended the findings obtained for building robust sentiment classifier in the future.
Keywords: Artificial Intelligence, Natural Language Processing, Sentiment Analysis, Social Network, Text.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5921013 An Advanced Technology for Renovation of Extruding Shafts
Authors: Dimitar Karastoyanov, Vladimir Monov
Abstract:
The paper is concerned with the technological process of renovation of shafts used in industrial manufacturing for extruding of sheet material. In the classical renovation technologies, a chrome based coating is applied to the working surface of the shaft in galvanic baths. The process, however, is known to be exclusively harmful due to the waste cyanide products. In this work, we present an advanced nanotechnology based on nonelectric chemical laying of a nickel coating with included nanoparticles. The technology is environmentally harmless and the new coating features an increased hardness and wear resistance. Results from experimental tests of the nanostructured nickel coating are presented and discussed.Keywords: Materials processing, nanoparticles, nickel coating, shafts renovation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18701012 Homogeneity of Microstructure and Mechanical Properties in Horizontal Continuous Cast Billet
Authors: V. Arbabi , I. Ebrahimzadeh, H. Ghanbari, M.M. Kaykha
Abstract:
Horizontal continuous casting is widely used to produce semi-finished non-Ferrous products. Homogeneity in the metallurgical characteristics and mechanical properties for this product is vital for industrial application. In the present work, the microstructure and mechanical properties of a horizontal continuous cast two-phase brass billet have been studied. Impact strength and hardness variations were examined and the phase composition and porosity studied with image analysis software. Distinct differences in mechanical properties were observed between the upper, middle and lower parts of the billet, which are explained in terms of the morphology and size of the phase in the microstructure. Hardness variation in the length of billet is higher in upper area but impact strength is higher in lower areas.Keywords: Horizontal Continuous Casting, Two-phase brasses, CuZn40Al1 alloy, Microstructure, Impact Strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21811011 Rheological Modeling for Shape-Memory Thermoplastic Polymers
Authors: H. Hosseini, B. V. Berdyshev, I. Iskopintsev
Abstract:
This paper presents a rheological model for producing shape-memory thermoplastic polymers. Shape-memory occurs as a result of internal rearrangement of the structural elements of a polymer. A non-linear viscoelastic model was developed that allows qualitative and quantitative prediction of the stress-strain behavior of shape-memory polymers during heating. This research was done to develop a technique to determine the maximum possible change in size of shape-memory products during heating. The rheological model used in this work was particularly suitable for defining process parameters and constructive parameters of the processing equipment.Keywords: Elastic deformation, heating, shape-memory polymers, stress-strain behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17661010 Brainwave Classification for Brain Balancing Index (BBI) via 3D EEG Model Using k-NN Technique
Authors: N. Fuad, M. N. Taib, R. Jailani, M. E. Marwan
Abstract:
In this paper, the comparison between k-Nearest Neighbor (kNN) algorithms for classifying the 3D EEG model in brain balancing is presented. The EEG signal recording was conducted on 51 healthy subjects. Development of 3D EEG models involves pre-processing of raw EEG signals and construction of spectrogram images. Then, maximum PSD values were extracted as features from the model. There are three indexes for balanced brain; index 3, index 4 and index 5. There are significant different of the EEG signals due to the brain balancing index (BBI). Alpha-α (8–13 Hz) and beta-β (13–30 Hz) were used as input signals for the classification model. The k-NN classification result is 88.46% accuracy. These results proved that k-NN can be used in order to predict the brain balancing application.
Keywords: Brain balancing, kNN, power spectral density, 3D EEG model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26241009 Cryptanalysis of Yang-Li-Liao’s Simple Three-Party Key Exchange (S-3PAKE) Protocol
Authors: Hae-Soon Ahn, Eun-Jun Yoon
Abstract:
Three-party password authenticated key exchange (3PAKE) protocols are widely deployed on lots of remote user authentication system due to its simplicity and convenience of maintaining a human-memorable password at client side to achieve secure communication within a hostile network. Recently, an improvement of 3PAKE protocol by processing a built-in data attached to other party for identity authentication to individual data was proposed by some researchers. However, this paper points out that the improved 3PAKE protocol is still vulnerable to undetectable on-line dictionary attack and off-line dictionary attack.
Keywords: Three-party key exchange, 3PAKE, Passwordauthenticated key exchange, Network security, Dictionary attack
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21211008 Characterization and Development of Anthropomorphic Phantoms Liver for Use in Nuclear Medicine
Authors: Ferreira F. C. L., Souza D. N., Rodrigues T. M. A., Cunha C. J., Dullius M. A., Andrade J. E., Sousa A. H., Vieira J. P. C., Carvalho Júnior A. B., Santos L. P. B., Passos R. O.
Abstract:
The objective this study was to characterize and develop anthropomorphic liver phantoms in tomography hepatic procedures for quality control and improvement professionals in nuclear medicine. For the conformation of the anthropomorphic phantom was used in plaster and acrylic. We constructed three phantoms representing processes with liver cirrhosis. The phantoms were filled with 99mTc diluted with water to obtain the scintigraphic images. Tomography images were analyzed anterior and posterior phantom representing a body with a greater degree cirrhotic. It was noted that the phantoms allow the acquisition of images similar to real liver with cirrhosis. Simulations of hemangiomas may contribute to continued professional education of nuclear medicine, on the question of image acquisition, allowing of the study parameters such of the matrix, energy window and count statistics.Keywords: Nuclear medicine, liver phantom, control quality
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16771007 The Nature of the Complicated Fabric Textures: How to Represent in Primary Visual Cortex
Authors: J. L. Liu, L. Wang, B. Zhu, J. Zhou, W. D. Gao
Abstract:
Fabric textures are very common in our daily life. However, the representation of fabric textures has never been explored from neuroscience view. Theoretical studies suggest that primary visual cortex (V1) uses a sparse code to efficiently represent natural images. However, how the simple cells in V1 encode the artificial textures is still a mystery. So, here we will take fabric texture as stimulus to study the response of independent component analysis that is established to model the receptive field of simple cells in V1. We choose 140 types of fabrics to get the classical fabric textures as materials. Experiment results indicate that the receptive fields of simple cells have obvious selectivity in orientation, frequency and phase when drifting gratings are used to determine their tuning properties. Additionally, the distribution of optimal orientation and frequency shows that the patch size selected from each original fabric image has a significant effect on the frequency selectivity.Keywords: Fabric Texture, Receptive Filed, Simple Cell, Spare Coding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14731006 Smart and Connected Aircraft Cabin: A Balancing Act between Operational Cabin Management, Airline Business and Passenger Expectations
Authors: Ralf God, Lothar Kerschgens, Leonardo Goratti, Steven Lemaire
Abstract:
Ubiquitous connectivity is a reality and a basic need for users on ground. Air travel connectivity in the cabin is also becoming increasingly important for passengers during cabin use. Wireless sensor networks that provide information to cabin management systems are being used by airlines to optimize cabin crew workload. In networked cabin systems, communications and digitally transmitted data must be managed by airlines in every direction. Security and privacy, information processing and knowledge management are the current and future requirements for a smart and connected cabin.
Keywords: Smart and connected cabin management, Internet of Things, power management, airline business.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4351005 Automatic Detection of Mass Type Breast Cancer using Texture Analysis in Korean Digital Mammography
Authors: E. B. Jo, J. H. Lee, J. Y. Park, S. M. Kim
Abstract:
In this study, we present an advanced detection technique for mass type breast cancer based on texture information of organs. The proposed method detects the cancer areas in three stages. In the first stage, the midpoints of mass area are determined based on AHE (Adaptive Histogram Equalization). In the second stage, we set the threshold coefficient of homogeneity by using MLE (Maximum Likelihood Estimation) to compute the uniformity of texture. Finally, mass type cancer tissues are extracted from the original image. As a result, it was observed that the proposed method shows an improved detection performance on dense breast tissues of Korean women compared with the existing methods. It is expected that the proposed method may provide additional diagnostic information for detection of mass-type breast cancer.Keywords: Mass Type Breast Cancer, Mammography, Maximum Likelihood Estimation (MLE), Ranklets, SVM
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19881004 Approach Based on Fuzzy C-Means for Band Selection in Hyperspectral Images
Authors: Diego Saqui, José H. Saito, José R. Campos, Lúcio A. de C. Jorge
Abstract:
Hyperspectral images and remote sensing are important for many applications. A problem in the use of these images is the high volume of data to be processed, stored and transferred. Dimensionality reduction techniques can be used to reduce the volume of data. In this paper, an approach to band selection based on clustering algorithms is presented. This approach allows to reduce the volume of data. The proposed structure is based on Fuzzy C-Means (or K-Means) and NWHFC algorithms. New attributes in relation to other studies in the literature, such as kurtosis and low correlation, are also considered. A comparison of the results of the approach using the Fuzzy C-Means and K-Means with different attributes is performed. The use of both algorithms show similar good results but, particularly when used attributes variance and kurtosis in the clustering process, however applicable in hyperspectral images.
Keywords: Band selection, fuzzy C-means, K-means, hyperspectral image.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18141003 Detection Characteristics of the Random and Deterministic Signals in Antenna Arrays
Authors: Olesya Bolkhovskaya, Alexey Davydov, Alexander Maltsev
Abstract:
In this paper, approach to incoherent signal detection in multi-element antenna array are researched and modeled. Two types of useful signals with unknown wavefront were considered: first one, deterministic (Barker code), and second one, random (Gaussian distribution). The derivation of the sufficient statistics took into account the linearity of the antenna array. The performance characteristics and detecting curves are modeled and compared for different useful signals parameters and for different number of elements of the antenna array. Results of researches in case of some additional conditions can be applied to a digital communications systems.Keywords: Antenna array, detection curves, performance characteristics, quadrature processing, signal detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17721002 Designing and Implementing a Novel Scheduler for Multiprocessor System using Genetic Algorithm
Authors: Iman Zangeneh, Mostafa Moradi, Mazyar Baranpouyan
Abstract:
System is using multiple processors for computing and information processing, is increasing rapidly speed operation of these systems compared with single processor systems, very significant impact on system performance is increased .important differences to yield a single multi-processor cpu, the scheduling policies, to reduce the implementation time of all processes. Notwithstanding the famous algorithms such as SPT, LPT, LSPT and RLPT for scheduling and there, but none led to the answer are not optimal.In this paper scheduling using genetic algorithms and innovative way to finish the whole process faster that we do and the result compared with three algorithms we mentioned.
Keywords: Multiprocessor system, genetic algorithms, time implementation process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15591001 A Resistorless High Input Impedance First Order All-Pass Filter Using CCCIIs
Authors: Kapil Dev Sharma, Kirat Pal, Costas Psychalinos
Abstract:
A new first order all-pass filter topology realized using current controlled current conveyors (CCCIIs) is introduced in this paper. Offered benefits are the high-impedance of the input node, the absence of external resistors because of the usage of CCCIIs with positive and negative intrinsic resistances, the presence of only grounded capacitors, and the capability of electronic adjustment of the phase shift through a single bias current. The correct operation of the introduced topology is conformed through simulation results, while its behavior is evaluated through comparison results.
Keywords: Active filters, All-pass filters, Analog signal processing, Current conveyors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17091000 Big Data: Concepts, Technologies and Applications in the Public Sector
Authors: A. Alexandru, C. A. Alexandru, D. Coardos, E. Tudora
Abstract:
Big Data (BD) is associated with a new generation of technologies and architectures which can harness the value of extremely large volumes of very varied data through real time processing and analysis. It involves changes in (1) data types, (2) accumulation speed, and (3) data volume. This paper presents the main concepts related to the BD paradigm, and introduces architectures and technologies for BD and BD sets. The integration of BD with the Hadoop Framework is also underlined. BD has attracted a lot of attention in the public sector due to the newly emerging technologies that allow the availability of network access. The volume of different types of data has exponentially increased. Some applications of BD in the public sector in Romania are briefly presented.
Keywords: Big data, big data Analytics, Hadoop framework, cloud computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2320999 A Method of Planar-Template- Based Camera Self-Calibration for Single-View
Abstract:
Camera calibration is an important step in 3D reconstruction. Camera calibration may be classified into two major types: traditional calibration and self-calibration. However, a calibration method in using a checkerboard is intermediate between traditional calibration and self-calibration. A self is proposed based on a square in this paper. Only a square in the planar template, the camera self-calibration can be completed through the single view. The proposed algorithm is that the virtual circle and straight line are established by a square on planar template, and circular points, vanishing points in straight lines and the relation between them are be used, in order to obtain the image of the absolute conic (IAC) and establish the camera intrinsic parameters. To make the calibration template is simpler, as compared with the Zhang Zhengyou-s method. Through real experiments and experiments, the experimental results show that this algorithm is feasible and available, and has a certain precision and robustness.Keywords: Absolute conic, camera calibration, circle point, vanishing point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1894998 Texture Characterization Based on a Chandrasekhar Fast Adaptive Filter
Authors: Mounir Sayadi, Farhat Fnaiech
Abstract:
In the framework of adaptive parametric modelling of images, we propose in this paper a new technique based on the Chandrasekhar fast adaptive filter for texture characterization. An Auto-Regressive (AR) linear model of texture is obtained by scanning the image row by row and modelling this data with an adaptive Chandrasekhar linear filter. The characterization efficiency of the obtained model is compared with the model adapted with the Least Mean Square (LMS) 2-D adaptive algorithm and with the cooccurrence method features. The comparison criteria is based on the computation of a characterization degree using the ratio of "betweenclass" variances with respect to "within-class" variances of the estimated coefficients. Extensive experiments show that the coefficients estimated by the use of Chandrasekhar adaptive filter give better results in texture discrimination than those estimated by other algorithms, even in a noisy context.
Keywords: Texture analysis, statistical features, adaptive filters, Chandrasekhar algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614997 Multiple Object Tracking using Particle Swarm Optimization
Authors: Chen-Chien Hsu, Guo-Tang Dai
Abstract:
This paper presents a particle swarm optimization (PSO) based approach for multiple object tracking based on histogram matching. To start with, gray-level histograms are calculated to establish a feature model for each of the target object. The difference between the gray-level histogram corresponding to each particle in the search space and the target object is used as the fitness value. Multiple swarms are created depending on the number of the target objects under tracking. Because of the efficiency and simplicity of the PSO algorithm for global optimization, target objects can be tracked as iterations continue. Experimental results confirm that the proposed PSO algorithm can rapidly converge, allowing real-time tracking of each target object. When the objects being tracked move outside the tracking range, global search capability of the PSO resumes to re-trace the target objects.Keywords: multiple object tracking, particle swarm optimization, gray-level histogram, image
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4099996 A Multistage Sulphidisation Flotation Procedure for a Low Grade Malachite Copper Ore
Authors: Tebogo P. Phetla, Edison Muzenda
Abstract:
This study was carried out to develop a flotation procedure for an oxide copper ore from a Region in Central Africa for producing an 18% copper concentrate for downstream processing at maximum recovery from a 4% copper feed grade. The copper recoveries achieved from the test work were less than 50% despite changes in reagent conditions (multistage sulphidisation, use of RCA emulsion and mixture, use of AM 2, etc). The poor recoveries were attributed to the mineralogy of the ore from which copper silicates accounted for approximately 70% (mass) of the copper minerals in the ore. These can be complex and difficult to float using conventional flotation methods. Best results were obtained using basic sulphidisation procedures, a high flotation temperature and extended flotation residence time.Keywords: Froth flotation, Sulphidisation, Copper oxide ore, Mineralogy, Recovery
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5827995 Face Detection using Variance based Haar-Like feature and SVM
Authors: Cuong Nguyen Khac, Ju H. Park, Ho-Youl Jung
Abstract:
This paper proposes a new approach to perform the problem of real-time face detection. The proposed method combines primitive Haar-Like feature and variance value to construct a new feature, so-called Variance based Haar-Like feature. Face in image can be represented with a small quantity of features using this new feature. We used SVM instead of AdaBoost for training and classification. We made a database containing 5,000 face samples and 10,000 non-face samples extracted from real images for learning purposed. The 5,000 face samples contain many images which have many differences of light conditions. And experiments showed that face detection system using Variance based Haar-Like feature and SVM can be much more efficient than face detection system using primitive Haar-Like feature and AdaBoost. We tested our method on two Face databases and one Non-Face database. We have obtained 96.17% of correct detection rate on YaleB face database, which is higher 4.21% than that of using primitive Haar-Like feature and AdaBoost.Keywords: AdaBoost, Haar-Like feature, SVM, variance, Variance based Haar-Like feature.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3733994 Factors Having Impact on Marketing and Improvement Measures in the Real Estate Sector of Turkey
Authors: Ali Ihtiyar, Serdar Durdyev, Syuhaida Ismail
Abstract:
Marketing is an essential issue to the survival of any real estate company in Turkey. There are some factors which are constraining the achievements of the marketing and sales strategies in the Turkey real estate industry. This study aims to identify and prioritise the most significant constraints to marketing in real estate sector and new strategies based on those constraints. This study is based on survey method, where the respondents such as credit counsellors, real estate investors, consultants, academicians and marketing representatives in Turkey were asked to rank forty seven sub-factors according to their levels of impact. The results of Multiattribute analytical technique indicated that the main subcomponents having impact on marketing in real estate sector are interest rates, real estate credit availability, accessibility, company image and consumer real income, respectively. The identified constraints are expected to guide the marketing team in a sales-effective way.Keywords: Marketing, marketing constraints, Real estate marketing, Turkey real estate sector
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579993 Comparing Hilditch, Rosenfeld, Zhang-Suen,and Nagendraprasad -Wang-Gupta Thinning
Authors: Anastasia Rita Widiarti
Abstract:
This paper compares Hilditch, Rosenfeld, Zhang- Suen, dan Nagendraprasad Wang Gupta (NWG) thinning algorithms for Javanese character image recognition. Thinning is an effective process when the focus in not on the size of the pattern, but rather on the relative position of the strokes in the pattern. The research analyzes the thinning of 60 Javanese characters. Time-wise, Zhang-Suen algorithm gives the best results with the average process time being 0.00455188 seconds. But if we look at the percentage of pixels that meet one-pixel thickness, Rosenfelt algorithm gives the best results, with a 99.98% success rate. From the number of pixels that are erased, NWG algorithm gives the best results with the average number of pixels erased being 84.12%. It can be concluded that the Hilditch algorithm performs least successfully compared to the other three algorithms.Keywords: Hilditch algorithm, Nagendraprasad-Wang-Guptaalgorithm, Rosenfeld algorithm, Thinning, Zhang-suen algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3917992 Mean Shift-based Preprocessing Methodology for Improved 3D Buildings Reconstruction
Authors: Nikolaos Vassilas, Theocharis Tsenoglou, Djamchid Ghazanfarpour
Abstract:
In this work, we explore the capability of the mean shift algorithm as a powerful preprocessing tool for improving the quality of spatial data, acquired from airborne scanners, from densely built urban areas. On one hand, high resolution image data corrupted by noise caused by lossy compression techniques are appropriately smoothed while at the same time preserving the optical edges and, on the other, low resolution LiDAR data in the form of normalized Digital Surface Map (nDSM) is upsampled through the joint mean shift algorithm. Experiments on both the edge-preserving smoothing and upsampling capabilities using synthetic RGB-z data show that the mean shift algorithm is superior to bilateral filtering as well as to other classical smoothing and upsampling algorithms. Application of the proposed methodology for 3D reconstruction of buildings of a pilot region of Athens, Greece results in a significant visual improvement of the 3D building block model.Keywords: 3D buildings reconstruction, data fusion, data upsampling, mean shift.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2004991 Computer Aided Detection on Mammography
Authors: Giovanni Luca Masala
Abstract:
A typical definition of the Computer Aided Diagnosis (CAD), found in literature, can be: A diagnosis made by a radiologist using the output of a computerized scheme for automated image analysis as a diagnostic aid. Often it is possible to find the expression Computer Aided Detection (CAD or CADe): this definition emphasizes the intent of CAD to support rather than substitute the human observer in the analysis of radiographic images. In this article we will illustrate the application of CAD systems and the aim of these definitions. Commercially available CAD systems use computerized algorithms for identifying suspicious regions of interest. In this paper are described the general CAD systems as an expert system constituted of the following components: segmentation / detection, feature extraction, and classification / decision making. As example, in this work is shown the realization of a Computer- Aided Detection system that is able to assist the radiologist in identifying types of mammary tumor lesions. Furthermore this prototype of station uses a GRID configuration to work on a large distributed database of digitized mammographic images.Keywords: Computer Aided Detection, Computer Aided Diagnosis, mammography, GRID.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1926990 Optimal ECG Sampling Frequency for Multiscale Entropy-Based HRV
Authors: Manjit Singh
Abstract:
Multiscale entropy (MSE) is an extensively used index to provide a general understanding of multiple complexity of physiologic mechanism of heart rate variability (HRV) that operates on a wide range of time scales. Accurate selection of electrocardiogram (ECG) sampling frequency is an essential concern for clinically significant HRV quantification; high ECG sampling rate increase memory requirements and processing time, whereas low sampling rate degrade signal quality and results in clinically misinterpreted HRV. In this work, the impact of ECG sampling frequency on MSE based HRV have been quantified. MSE measures are found to be sensitive to ECG sampling frequency and effect of sampling frequency will be a function of time scale.Keywords: ECG, heart rate variability, HRV, multiscale entropy, sampling frequency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1351