Search results for: content filtering
6350 Open Source Knowledge Management Approach to Manage and Disseminate Distributed Content in a Global Enterprise
Authors: Rahul Thakur, Onkar Chandel
Abstract:
Red Hat is the world leader in providing open source software and solutions. A global enterprise, like Red Hat, has unique issues of connecting employees with content because of distributed offices, multiple teams spread across geographies, multiple languages, and different cultures. Employees, of a global company, create content that is distributed across departments, teams, regions, and countries. This makes finding the best content difficult since owners keep iterating on the existing content. When employees are unable to find the content, they end up creating it once again and in the process duplicating existing material and effort. Also, employees may not find the relevant content and spend time reviewing obsolete duplicate, or irrelevant content. On an average, a person spends 15 minutes/day in failed searches that might result in missed business opportunities, employee frustration, and substandard deliverables. Red Hat Knowledge Management Office (KMO) applied 'open source strategy' to solve the above problems. Under the Open Source Strategy, decisions are taken collectively. The strategy aims at accomplishing common goals with the help of communities. The objectives of this initiative were to save employees' time, get them authentic content, improve their content search experience, avoid duplicate content creation, provide context based search, improve analytics, improve content management workflows, automate content classification, and automate content upload. This session will describe open source strategy, its applicability in content management, challenges, recommended solutions, and outcome.Keywords: content classification, content management, knowledge management, open source
Procedia PDF Downloads 2106349 The Effect of Technology on Advanced Automotive Electronics
Authors: Abanob Nady Wasef Moawed
Abstract:
In more complicated systems, inclusive of automotive gearboxes, a rigorous remedy of the data is essential because there are several transferring elements (gears, bearings, shafts, and many others.), and in this way, there are numerous viable sources of mistakes and also noise. The fundamental goal of these elements are the detection of damage in car gearbox. The detection strategies used are the wavelet technique, the bispectrum, advanced filtering techniques (selective filtering) of vibrational alerts and mathematical morphology. Gearbox vibration assessments were achieved (gearboxes in proper circumstance and with defects) of a manufacturing line of a huge car assembler. The vibration indicators have acquired the use of five accelerometers in distinct positions of the sample. The effects acquired using the kurtosis, bispectrum, wavelet and mathematical morphology confirmed that it's far possible to identify the lifestyles of defects in automobile gearboxes.Keywords: 3D-shaped electronics, electronic components, thermoforming, component positioningautomotive gearbox, mathematical morphology, wavelet, bispectrum
Procedia PDF Downloads 226348 Three-Dimensional Jet Refraction Simulation Using a Gradient Term Suppression and Filtering Method
Authors: Lican Wang, Rongqian Chen, Yancheng You, Ruofan Qiu
Abstract:
In the applications of jet engine, open-jet wind tunnel and airframe, there wildly exists a shear layer formed by the velocity and temperature gradients between jet flow and surrounded medium. The presence of shear layer will refract and reflect the sound path that consequently influences the measurement results in far-field. To investigate and evaluate the shear layer effect, a gradient term suppression and filtering method is adopted to simulate sound propagation through a steady sheared flow in three dimensions. Two typical configurations are considered: one is an incompressible and cold jet flow in wind tunnel and the other is a compressible and hot jet flow in turbofan engine. A numerically linear microphone array is used to localize the position of given sound source. The localization error is presented and linearly fitted.Keywords: aeroacoustic, linearized Euler equation, acoustic propagation, source localization
Procedia PDF Downloads 2006347 Online Prediction of Nonlinear Signal Processing Problems Based Kernel Adaptive Filtering
Authors: Hamza Nejib, Okba Taouali
Abstract:
This paper presents two of the most knowing kernel adaptive filtering (KAF) approaches, the kernel least mean squares and the kernel recursive least squares, in order to predict a new output of nonlinear signal processing. Both of these methods implement a nonlinear transfer function using kernel methods in a particular space named reproducing kernel Hilbert space (RKHS) where the model is a linear combination of kernel functions applied to transform the observed data from the input space to a high dimensional feature space of vectors, this idea known as the kernel trick. Then KAF is the developing filters in RKHS. We use two nonlinear signal processing problems, Mackey Glass chaotic time series prediction and nonlinear channel equalization to figure the performance of the approaches presented and finally to result which of them is the adapted one.Keywords: online prediction, KAF, signal processing, RKHS, Kernel methods, KRLS, KLMS
Procedia PDF Downloads 3976346 Evaluation of Sensor Pattern Noise Estimators for Source Camera Identification
Authors: Benjamin Anderson-Sackaney, Amr Abdel-Dayem
Abstract:
This paper presents a comprehensive survey of recent source camera identification (SCI) systems. Then, the performance of various sensor pattern noise (SPN) estimators was experimentally assessed, under common photo response non-uniformity (PRNU) frameworks. The experiments used 1350 natural and 900 flat-field images, captured by 18 individual cameras. 12 different experiments, grouped into three sets, were conducted. The results were analyzed using the receiver operator characteristic (ROC) curves. The experimental results demonstrated that combining the basic SPN estimator with a wavelet-based filtering scheme provides promising results. However, the phase SPN estimator fits better with both patch-based (BM3D) and anisotropic diffusion (AD) filtering schemes.Keywords: sensor pattern noise, source camera identification, photo response non-uniformity, anisotropic diffusion, peak to correlation energy ratio
Procedia PDF Downloads 4386345 PDDA: Priority-Based, Dynamic Data Aggregation Approach for Sensor-Based Big Data Framework
Authors: Lutful Karim, Mohammed S. Al-kahtani
Abstract:
Sensors are being used in various applications such as agriculture, health monitoring, air and water pollution monitoring, traffic monitoring and control and hence, play the vital role in the growth of big data. However, sensors collect redundant data. Thus, aggregating and filtering sensors data are significantly important to design an efficient big data framework. Current researches do not focus on aggregating and filtering data at multiple layers of sensor-based big data framework. Thus, this paper introduces (i) three layers data aggregation and framework for big data and (ii) a priority-based, dynamic data aggregation scheme (PDDA) for the lowest layer at sensors. Simulation results show that the PDDA outperforms existing tree and cluster-based data aggregation scheme in terms of overall network energy consumptions and end-to-end data transmission delay.Keywords: big data, clustering, tree topology, data aggregation, sensor networks
Procedia PDF Downloads 3436344 Eucalyptus camendulensis and Its Drying Effect on Water and Essential Oil Content
Authors: Mehani Mouna, Segni Ladjel
Abstract:
Medicinal and aromatic plants are promising and are characterized by the biosynthesis of odorous molecules that make up the so-called essential oils (EO), which have long been known for their antiseptic and therapeutic activity in folk medicine. The objective of this study was to evaluate the influence of drying in the shade on the water content and on the content of essential oils extracted from leaves of Eucalyptus camendulensis for better quality control of medicinal and aromatic plants. The water content of the Eucalyptus camendulensis plant material decreases during the drying process. It increased from 100 % to 0.006 % for the drying in the shade after ten days. The moisture content is practically constant at the end of the drying period. The drying in the shade increases the concentration of essential oils of Eucalyptus camendulensis. When the leaves of Eucalyptus camendulensis plant are in the shade, the maximum of the essential oil content was obtained on the eighth days; the recorded value was 1.43% ± 0.01%. Beyond these periods, the content continuously drops in before stabilizing. The optimum drying time is between 6 and 9 days.Keywords: Eucalyptus camendulensis, drying, essential oils, water, content
Procedia PDF Downloads 3566343 Bipolar Impulse Noise Removal and Edge Preservation in Color Images and Video Using Improved Kuwahara Filter
Authors: Reji Thankachan, Varsha PS
Abstract:
Both image capturing devices and human visual systems are nonlinear. Hence nonlinear filtering methods outperforms its linear counterpart in many applications. Linear methods are unable to remove impulsive noise in images by preserving its edges and fine details. In addition, linear algorithms are unable to remove signal dependent or multiplicative noise in images. This paper presents an approach to denoise and smoothen the Bipolar impulse noised images and videos using improved Kuwahara filter. It involves a 2 stage algorithm which includes a noise detection followed by filtering. Numerous simulation demonstrate that proposed method outperforms the existing method by eliminating the painting like flattening effect along the local feature direction while preserving edge with improvement in PSNR and MSE.Keywords: bipolar impulse noise, Kuwahara, PSNR MSE, PDF
Procedia PDF Downloads 4976342 Teaching Science Content Area Literacy to 21st Century Learners
Authors: Melissa C. Ingram
Abstract:
The use of new literacies within science classrooms needs to be balanced by teachers to both teach different forms of communication while assessing content area proficiency. Using new literacies such as Twitter and Facebook needs to be incorporated into science content area literacy studies in addition to continuing to use generally-accepted forms of scientific content area presentation, which include scientific papers and textbooks. The research question this literature review seeks to answer is “What are some ways in which new forms of literacy are better suited to teach scientific content area literacy to 21st Century learners?” The research question is addressed through a literature review that highlights methods currently being used to educate the next wave of learners in the world of science content area literacy. Both temporal discourse analysis (TDA) and critical discourse analysis (CDA) were used to determine the need to use new literacies to teach science content area literacy. Increased use of digital technologies and a change in science content area pedagogy were explored.Keywords: science content area literacy, new literacies, critical discourse analysis, temporal discourse analysis
Procedia PDF Downloads 2206341 The Impact of Content Familiarity of Receptive Skills on Language Learning
Authors: Sara Fallahi
Abstract:
This paper reviews the importance of content familiarity of receptive skills and offers solutions to the issue of content unfamiliarity in language learning materials. Presently, language learning materials are mainly comprised of global issues and target language speakers’ culture(s) in receptive skills. This might leadlearners to focus on content rather than the language. As a solution, materials on receptive skills can be developed with a focus on learners’culture and social concerns, especially in the beginner levels of learning. Language learners often learn their target language through the receptive skills of listening and reading before language production ensues through speaking and writing. Students’ journey from receptive skills to productive skills is mainly concentrated on by teachers. There are barriers to language learning, such as time and energy, that can hinder learners’ understanding and ability to build the required background knowledge of the content. This is generated due to learners’ unfamiliarity with the skill’s content. Therefore, materials that improve content familiarity will help learners improve their language comprehension, learning, and usage. This presentation will conclude with practical solutions to help teachers and learners more authentically integrate language and culture to elevate language learning.Keywords: language learning, listening content, reading content, content familiarity, ESL books, language learning books, cultural familiarity
Procedia PDF Downloads 1176340 Noise Removal Techniques in Medical Images
Authors: Amhimmid Mohammed Saffour, Abdelkader Salama
Abstract:
Filtering is a part of image enhancement techniques, it is used to enhance certain details such as edges in the image that are relevant to the application. Additionally, filtering can even be used to eliminate unwanted components of noise. Medical images typically contain salt and pepper noise and Poisson noise. This noise appears to the presence of minute grey scale variations within the image. In this paper, different filters techniques namely (Median, Wiener, Rank order3, Rank order5, and Average) were applied on CT medical images (Brain and chest). We using all these filters to remove salt and pepper noise from these images. This type of noise consists of random pixels being set to black or white. Peak Signal to Noise Ratio (PSNR), Mean Square Error r(MSE) and Histogram were used to evaluated the quality of filtered images. The results, which we have achieved shows that, these filters, are more useful and they prove to be helpful for general medical practitioners to analyze the symptoms of the patients with no difficulty.Keywords: CT imaging, median filter, adaptive filter and average filter, MATLAB
Procedia PDF Downloads 3126339 Characterization of Biodiesel Produced from Cow-Tallow
Authors: Nwadike Emmanuel Chinagoron, Achebe Chukwunonso, Ezeliora Chukwuemeka Daniel, Azaka Onyemazuwa Andrew
Abstract:
In this research work, the process of biodiesel production in a pilot plant was studied using cow tallow as raw material, methanol as the solvent and potassium hydroxide as catalysts. The biodiesel quality was determined by characterization. The tallow used in the production had a molecular weight of 860g. Its oil had a density value of 0.8g/ml, iodine value of 63.45, viscosity at 300C was 9.83pas, acid value was 1.96, free fatty acid (FFA) of 0.98%, saponification value of 82.75mleq/kg, specific gravity of 0.898, flash point of 1100C, cloud point of 950C and Calorific value also called Higher Heating Value (HHV) of 38.365MJ/Kg. The produced biodiesel had a density of 0.82g/ml, iodine value of 126.9, viscosity of 4.32pas at 300C, acid value of 0.561, FFA of 0.2805%, saponification value of 137.45 mleq/kg.Flash point, cloud point and centane number of the biodiesel produced are 1390C, 980C and 57.5 respectively, with fat content, protein content, ash content, moisture content, fiber content and carbohydrate content values of 10%, 2.8%, 5%, 5%, 20%, and 37.2% respectively. The biodiesel higher heating values (calorific values) when estimated from viscosity, density and flash points were 41.4MJ/Kg, 63.8MJ/Kg, and 34.6MJ/Kg respectively. The biodiesel was blended with conventional diesel. The blend B-10 had values of 1320C and 960C for flash and cloud points, with Calorific value (or HHV) of 34.6 MJ/Kg (when estimated from its Flash point) and fat content, protein content, ash content, moisture content, fiber content and carbohydrate content values of 5%, 2.1%,10%, 5%, 15%, and 62.9% respectively.Keywords: biodiesel, characterization, cow-tallow, cetane rating
Procedia PDF Downloads 5346338 The Use of Image Processing Responses Tools Applied to Analysing Bouguer Gravity Anomaly Map (Tangier-Tetuan's Area-Morocco)
Authors: Saad Bakkali
Abstract:
Image processing is a powerful tool for the enhancement of edges in images used in the interpretation of geophysical potential field data. Arial and terrestrial gravimetric surveys were carried out in the region of Tangier-Tetuan. From the observed and measured data of gravity Bouguer gravity anomalies map was prepared. This paper reports the results and interpretations of the transformed maps of Bouguer gravity anomaly of the Tangier-Tetuan area using image processing. Filtering analysis based on classical image process was applied. Operator image process like logarithmic and gamma correction are used. This paper also present the results obtained from this image processing analysis of the enhancement edges of the Bouguer gravity anomaly map of the Tangier-Tetuan zone.Keywords: bouguer, tangier, filtering, gamma correction, logarithmic enhancement edges
Procedia PDF Downloads 4206337 A Background Subtraction Based Moving Object Detection Around the Host Vehicle
Authors: Hyojin Lim, Cuong Nguyen Khac, Ho-Youl Jung
Abstract:
In this paper, we propose moving object detection method which is helpful for driver to safely take his/her car out of parking lot. When moving objects such as motorbikes, pedestrians, the other cars and some obstacles are detected at the rear-side of host vehicle, the proposed algorithm can provide to driver warning. We assume that the host vehicle is just before departure. Gaussian Mixture Model (GMM) based background subtraction is basically applied. Pre-processing such as smoothing and post-processing as morphological filtering are added.We examine “which color space has better performance for detection of moving objects?” Three color spaces including RGB, YCbCr, and Y are applied and compared, in terms of detection rate. Through simulation, we prove that RGB space is more suitable for moving object detection based on background subtraction.Keywords: gaussian mixture model, background subtraction, moving object detection, color space, morphological filtering
Procedia PDF Downloads 6126336 Impact of Foliar Application of Zinc on Micro and Macro Elements Distribution in Phyllanthus amarus
Authors: Nguyen Cao Nguyen, Krasimir I. Ivanov, Penka S. Zapryanova
Abstract:
The present study was carried out to investigate the interaction of foliar applied zinc with other elements in Phyllanthus amarus plants. The plant samples for our experiment were collected from Lam Dong province, Vietnam. Seven suspension solutions of nanosized zinc hydroxide nitrate (Zn5(OH)8(NO3)2·2H2O) with different Zn concentration were used. Fertilization and irrigation were the same for all variants. The Zn content and the content of selected micro (Cu, Fe, Mn) and macro (Ca, Mg, P and K) nutrients in plant roots, and stems and leaves were determined. It was concluded that the zinc content of plant roots varies narrowly, with no significant impact of ZnHN fertilization. The same trend can be seen in the content of Cu, Mn, and macronutrients. The zinc content of plant stems and leaves varies within wide limits, with the significant impact of ZnHN fertilization. The trends in the content of Cu, Mn, and macronutrients are kept the same as in the root, whereas the iron trends to increase its content at increasing the zinc content.Keywords: Phyllanthus amarus, Zinc, Micro and macro elements, foliar fertilizer
Procedia PDF Downloads 1486335 Critical Review of Web Content Mining Extraction Mechanisms
Authors: Rabia Bashir, Sajjad Akbar
Abstract:
There is an inevitable demand of web mining due to rapid increase of huge information on the Internet, but the striking variety of web structures has made required content retrieval a difficult task. To counter this issue, Web Content Mining (WCM) emerges as a potential candidate which extracts and integrates suitable resources of data to users. In past few years, research has been done on several extraction techniques for WCM i.e. agent-based, template-based, assumption-based, statistic-based, wrapper-based and machine learning. However, it is still unclear that either these approaches are efficiently tackling the significant challenges of WCM or not. To answer this question, this paper identifies these challenges such as language independency, structure flexibility, performance, automation, dynamicity, redundancy handling, intelligence, relevant content retrieval, and privacy. Further, mapping of these challenges is done with existing extraction mechanisms which helps to adopt the most suitable WCM approach, given some conditions and characteristics at hand.Keywords: content mining challenges, web content mining, web content extraction approaches, web information retrieval
Procedia PDF Downloads 5456334 Network Mobility Support in Content-Centric Internet
Authors: Zhiwei Yan, Jong-Hyouk Lee, Yong-Jin Park, Xiaodong Lee
Abstract:
In this paper, we analyze NEtwork MObility (NEMO) supporting problems in Content-Centric Networking (CCN), and propose the CCN-NEMO which can well support the deployment of the content-centric paradigm in large-scale mobile Internet. The CCN-NEMO extends the signaling message of the basic CCN protocol, to support the mobility discovery and fast trigger of Interest re-issuing during the network mobility. Besides, the Mobile Router (MR) is extended to optimize the content searching and relaying in the local subnet. These features can be employed by the nested NEMO to maximize the advantages of content retrieving with CCN. Based on the analysis, we compare the performance on handover latency between the basic CCN and our proposed CCN-NEMO. The results show that our scheme can facilitate the content-retrieving in the NEMO scenario with improved performance.Keywords: NEMO, CCN, mobility, handover latency
Procedia PDF Downloads 4686333 Filtering and Reconstruction System for Grey-Level Forensic Images
Authors: Ahd Aljarf, Saad Amin
Abstract:
Images are important source of information used as evidence during any investigation process. Their clarity and accuracy is essential and of the utmost importance for any investigation. Images are vulnerable to losing blocks and having noise added to them either after alteration or when the image was taken initially, therefore, having a high performance image processing system and it is implementation is very important in a forensic point of view. This paper focuses on improving the quality of the forensic images. For different reasons packets that store data can be affected, harmed or even lost because of noise. For example, sending the image through a wireless channel can cause loss of bits. These types of errors might give difficulties generally for the visual display quality of the forensic images. Two of the images problems: noise and losing blocks are covered. However, information which gets transmitted through any way of communication may suffer alteration from its original state or even lose important data due to the channel noise. Therefore, a developed system is introduced to improve the quality and clarity of the forensic images.Keywords: image filtering, image reconstruction, image processing, forensic images
Procedia PDF Downloads 3616332 The Impact of Different Rhizobium leguminosarum Strains on the Protein Content of Peas and Broad Beans
Authors: Alise Senberga, Laila Dubova, Liene Strauta, Ina Alsina, Ieva Erdberga
Abstract:
Legume symbiotic relationship with nitrogen fixating bacteria Rhizobim leguminosarum is an important factor used to improve the productivity of legumes, due to the fact that rhizobia can supply plant with the necessary amount of nitrogen. R. leguminosarum strains have shown different activity in fixing nitrogen. Depending on the chosen R. leguminosarum strain, host plant biochemical content can be altered. In this study we focused particularly on the changes in protein content in beans (using two different varieties) and peas (five different varieties) due to the use of several different R. leguminosarum strains (four strains for both beans and peas). Overall, the protein content increase was observed after seed inoculation with R. leguminosarum. Strain and plant cultivar interaction specification was observed. The effect of R. leguminosarum inoculation on the content of protein was dependent on the R. leguminosarum strain used. Plant cultivar also appeared to have a decisive role in protein content formation with the help of R. leguminosaru.Keywords: legumes, protein content, rhizobia strains, soil
Procedia PDF Downloads 5196331 Evaluation of Diagnosis Performance Based on Pairwise Model Construction and Filtered Data
Authors: Hyun-Woo Cho
Abstract:
It is quite important to utilize right time and intelligent production monitoring and diagnosis of industrial processes in terms of quality and safety issues. When compared with monitoring task, fault diagnosis represents the task of finding process variables responsible causing a specific fault in the process. It can be helpful to process operators who should investigate and eliminate root causes more effectively and efficiently. This work focused on the active use of combining a nonlinear statistical technique with a preprocessing method in order to implement practical real-time fault identification schemes for data-rich cases. To compare its performance to existing identification schemes, a case study on a benchmark process was performed in several scenarios. The results showed that the proposed fault identification scheme produced more reliable diagnosis results than linear methods. In addition, the use of the filtering step improved the identification results for the complicated processes with massive data sets.Keywords: diagnosis, filtering, nonlinear statistical techniques, process monitoring
Procedia PDF Downloads 2416330 Digital Content Strategy (DCS) Detailed Review of the Key Content Components
Authors: Oksana Razina, Shakeel Ahmad, Jessie Qun Ren, Olufemi Isiaq
Abstract:
The modern life of businesses is categorically reliant on their established position online, where digital (and particularly website) content plays a significant role as the first point of information. Digital content, therefore, becomes essential – from making the first impression to the building and development of client relationships. Despite a number of valuable papers suggesting a strategic approach when dealing with digital data, other sources often do not view or accept the approach to digital content as a holistic or continuous process. Associations are frequently made with merely a one-off marketing campaign or similar. The challenge is to establish an agreed definition for the notion of Digital Content Strategy, which currently does not exist, as DCS is viewed from an excessive number of different angles. A strategic approach to content, nonetheless, is required, both practically and contextually. The researchers, therefore, aimed at attempting to identify the key content components comprising a digital content strategy to ensure all the aspects were covered and strategically applied – from the company’s understanding of the content value to the ability to display flexibility of content and advances in technology. This conceptual project evaluated existing literature on the topic of Digital Content Strategy (DCS) and related aspects, using the PRISMA Systematic Review Method, Document Analysis, Inclusion and Exclusion Criteria, Scoping Review, Snow-Balling Technique and Thematic Analysis. The data was collected from academic and statistical sources, government and relevant trade publications. Based on the suggestions from academics and trading sources related to the issues discussed, the researchers revealed the key actions for content creation and attempted to define the notion of DCS. The major finding of the study presented Key Content Components of Digital Content Strategy and can be considered for implementation in a business retail setting.Keywords: digital content strategy, key content components, websites, digital marketing strategy
Procedia PDF Downloads 1456329 Semantic-Based Collaborative Filtering to Improve Visitor Cold Start in Recommender Systems
Authors: Baba Mbaye
Abstract:
In collaborative filtering recommendation systems, a user receives suggested items based on the opinions and evaluations of a community of users. This type of recommendation system uses only the information (notes in numerical values) contained in a usage matrix as input data. This matrix can be constructed based on users' behaviors or by offering users to declare their opinions on the items they know. The cold start problem leads to very poor performance for new users. It is a phenomenon that occurs at the beginning of use, in the situation where the system lacks data to make recommendations. There are three types of cold start problems: cold start for a new item, a new system, and a new user. We are interested in this article at the cold start for a new user. When the system welcomes a new user, the profile exists but does not have enough data, and its communities with other users profiles are still unknown. This leads to recommendations not adapted to the profile of the new user. In this paper, we propose an approach that improves cold start by using the notions of similarity and semantic proximity between users profiles during cold start. We will use the cold-metadata available (metadata extracted from the new user's data) useful in positioning the new user within a community. The aim is to look for similarities and semantic proximities with the old and current user profiles of the system. Proximity is represented by close concepts considered to belong to the same group, while similarity groups together elements that appear similar. Similarity and proximity are two close but not similar concepts. This similarity leads us to the construction of similarity which is based on: a) the concepts (properties, terms, instances) independent of ontology structure and, b) the simultaneous representation of the two concepts (relations, presence of terms in a document, simultaneous presence of the authorities). We propose an ontology, OIVCSRS (Ontology of Improvement Visitor Cold Start in Recommender Systems), in order to structure the terms and concepts representing the meaning of an information field, whether by the metadata of a namespace, or the elements of a knowledge domain. This approach allows us to automatically attach the new user to a user community, partially compensate for the data that was not initially provided and ultimately to associate a better first profile with the cold start. Thus, the aim of this paper is to propose an approach to improving cold start using semantic technologies.Keywords: visitor cold start, recommender systems, collaborative filtering, semantic filtering
Procedia PDF Downloads 2176328 Analyzing On-Line Process Data for Industrial Production Quality Control
Authors: Hyun-Woo Cho
Abstract:
The monitoring of industrial production quality has to be implemented to alarm early warning for unusual operating conditions. Furthermore, identification of their assignable causes is necessary for a quality control purpose. For such tasks many multivariate statistical techniques have been applied and shown to be quite effective tools. This work presents a process data-based monitoring scheme for production processes. For more reliable results some additional steps of noise filtering and preprocessing are considered. It may lead to enhanced performance by eliminating unwanted variation of the data. The performance evaluation is executed using data sets from test processes. The proposed method is shown to provide reliable quality control results, and thus is more effective in quality monitoring in the example. For practical implementation of the method, an on-line data system must be available to gather historical and on-line data. Recently large amounts of data are collected on-line in most processes and implementation of the current scheme is feasible and does not give additional burdens to users.Keywords: detection, filtering, monitoring, process data
Procedia PDF Downloads 5576327 Dry Matter, Moisture, Ash and Crude Fibre Content in Distinct Segments of ‘Durian Kampung’ Husk
Authors: Norhanim Nordin, Rosnah Shamsudin, Azrina Azlan, Mohammad Effendy Ya’acob
Abstract:
An environmental friendly approach for disposal of voluminous durian husk waste could be implemented by substituting them into various valuable commodities, such as healthcare and biofuel products. Thus, the study of composition value in each segment of durian husk was very crucial to determine the suitable proportions of nutrients that need to be added and mixed in the product. A total of 12 ‘Durian Kampung’ fruits from Sg Ruan, Pahang were selected and each fruit husk was divided into four segments and labelled as P-L (thin neck area of white inner husk), P-B (thick bottom area of white inner husk), H (green and thorny outer husk) and W (whole combination of P-B and H). Four experiments have been carried out to determine the dry matter, moisture, ash and crude fibre content. The results show that the H segment has the highest dry matter content (30.47%), while the P-B segment has the highest percentage in moisture (81.83%) and ash (6.95%) content. It was calculated that the ash content of the P-B segment has a higher rate of moisture level which causes the ash content to increase about 2.89% from the P-L segment. These data have proven that each segment of durian husk has a significant difference in terms of composition value, which might be useful information to fully utilize every part of the durian husk in the future.Keywords: durian husk, crude fibre content, dry matter content, moisture content
Procedia PDF Downloads 2986326 Email Phishing Detection Using Natural Language Processing and Convolutional Neural Network
Abstract:
Phishing is one of the oldest and best known scams on the Internet. It can be defined as any type of telecommunications fraud that uses social engineering tricks to obtain confidential data from its victims. It’s a cybercrime aimed at stealing your sensitive information. Phishing is generally done via private email, so scammers impersonate large companies or other trusted entities to encourage victims to voluntarily provide information such as login credentials or, worse yet, credit card numbers. The COVID-19 theme is used by cybercriminals in multiple malicious campaigns like phishing. In this environment, messaging filtering solutions have become essential to protect devices that will now be used outside of the secure perimeter. Despite constantly updating methods to avoid these cyberattacks, the end result is currently insufficient. Many researchers are looking for optimal solutions to filter phishing emails, but we still need good results. In this work, we concentrated on solving the problem of detecting phishing emails using the different steps of NLP preprocessing, and we proposed and trained a model using one-dimensional CNN. Our study results show that our model obtained an accuracy of 99.99%, which demonstrates how well our model is working.Keywords: phishing, e-mail, NLP preprocessing, CNN, e-mail filtering
Procedia PDF Downloads 1246325 Frequency Decomposition Approach for Sub-Band Common Spatial Pattern Methods for Motor Imagery Based Brain-Computer Interface
Authors: Vitor M. Vilas Boas, Cleison D. Silva, Gustavo S. Mafra, Alexandre Trofino Neto
Abstract:
Motor imagery (MI) based brain-computer interfaces (BCI) uses event-related (de)synchronization (ERS/ ERD), typically recorded using electroencephalography (EEG), to translate brain electrical activity into control commands. To mitigate undesirable artifacts and noise measurements on EEG signals, methods based on band-pass filters defined by a specific frequency band (i.e., 8 – 30Hz), such as the Infinity Impulse Response (IIR) filters, are typically used. Spatial techniques, such as Common Spatial Patterns (CSP), are also used to estimate the variations of the filtered signal and extract features that define the imagined motion. The CSP effectiveness depends on the subject's discriminative frequency, and approaches based on the decomposition of the band of interest into sub-bands with smaller frequency ranges (SBCSP) have been suggested to EEG signals classification. However, despite providing good results, the SBCSP approach generally increases the computational cost of the filtering step in IM-based BCI systems. This paper proposes the use of the Fast Fourier Transform (FFT) algorithm in the IM-based BCI filtering stage that implements SBCSP. The goal is to apply the FFT algorithm to reduce the computational cost of the processing step of these systems and to make them more efficient without compromising classification accuracy. The proposal is based on the representation of EEG signals in a matrix of coefficients resulting from the frequency decomposition performed by the FFT, which is then submitted to the SBCSP process. The structure of the SBCSP contemplates dividing the band of interest, initially defined between 0 and 40Hz, into a set of 33 sub-bands spanning specific frequency bands which are processed in parallel each by a CSP filter and an LDA classifier. A Bayesian meta-classifier is then used to represent the LDA outputs of each sub-band as scores and organize them into a single vector, and then used as a training vector of an SVM global classifier. Initially, the public EEG data set IIa of the BCI Competition IV is used to validate the approach. The first contribution of the proposed method is that, in addition to being more compact, because it has a 68% smaller dimension than the original signal, the resulting FFT matrix maintains the signal information relevant to class discrimination. In addition, the results showed an average reduction of 31.6% in the computational cost in relation to the application of filtering methods based on IIR filters, suggesting FFT efficiency when applied in the filtering step. Finally, the frequency decomposition approach improves the overall system classification rate significantly compared to the commonly used filtering, going from 73.7% using IIR to 84.2% using FFT. The accuracy improvement above 10% and the computational cost reduction denote the potential of FFT in EEG signal filtering applied to the context of IM-based BCI implementing SBCSP. Tests with other data sets are currently being performed to reinforce such conclusions.Keywords: brain-computer interfaces, fast Fourier transform algorithm, motor imagery, sub-band common spatial patterns
Procedia PDF Downloads 1286324 Geomatic Techniques to Filter Vegetation from Point Clouds
Authors: M. Amparo Núñez-Andrés, Felipe Buill, Albert Prades
Abstract:
More and more frequently, geomatics techniques such as terrestrial laser scanning or digital photogrammetry, either terrestrial or from drones, are being used to obtain digital terrain models (DTM) used for the monitoring of geological phenomena that cause natural disasters, such as landslides, rockfalls, debris-flow. One of the main multitemporal analyses developed from these models is the quantification of volume changes in the slopes and hillsides, either caused by erosion, fall, or land movement in the source area or sedimentation in the deposition zone. To carry out this task, it is necessary to filter the point clouds of all those elements that do not belong to the slopes. Among these elements, vegetation stands out as it is the one we find with the greatest presence and its constant change, both seasonal and daily, as it is affected by factors such as wind. One of the best-known indexes to detect vegetation on the image is the NVDI (Normalized Difference Vegetation Index), which is obtained from the combination of the infrared and red channels. Therefore it is necessary to have a multispectral camera. These cameras are generally of lower resolution than conventional RGB cameras, while their cost is much higher. Therefore we have to look for alternative indices based on RGB. In this communication, we present the results obtained in Georisk project (PID2019‐103974RB‐I00/MCIN/AEI/10.13039/501100011033) by using the GLI (Green Leaf Index) and ExG (Excessive Greenness), as well as the change to the Hue-Saturation-Value (HSV) color space being the H coordinate the one that gives us the most information for vegetation filtering. These filters are applied both to the images, creating binary masks to be used when applying the SfM algorithms, and to the point cloud obtained directly by the photogrammetric process without any previous filter or the one obtained by TLS (Terrestrial Laser Scanning). In this last case, we have also tried to work with a Riegl VZ400i sensor that allows the reception, as in the aerial LiDAR, of several returns of the signal. Information to be used for the classification on the point cloud. After applying all the techniques in different locations, the results show that the color-based filters allow correct filtering in those areas where the presence of shadows is not excessive and there is a contrast between the color of the slope lithology and the vegetation. As we have advanced in the case of using the HSV color space, it is the H coordinate that responds best for this filtering. Finally, the use of the various returns of the TLS signal allows filtering with some limitations.Keywords: RGB index, TLS, photogrammetry, multispectral camera, point cloud
Procedia PDF Downloads 1516323 Fat-Tail Test of Regulatory DNA Sequences
Authors: Jian-Jun Shu
Abstract:
The statistical properties of CRMs are explored by estimating similar-word set occurrence distribution. It is observed that CRMs tend to have a fat-tail distribution for similar-word set occurrence. Thus, the fat-tail test with two fatness coefficients is proposed to distinguish CRMs from non-CRMs, especially from exons. For the first fatness coefficient, the separation accuracy between CRMs and exons is increased as compared with the existing content-based CRM prediction method – fluffy-tail test. For the second fatness coefficient, the computing time is reduced as compared with fluffy-tail test, making it very suitable for long sequences and large data-base analysis in the post-genome time. Moreover, these indexes may be used to predict the CRMs which have not yet been observed experimentally. This can serve as a valuable filtering process for experiment.Keywords: statistical approach, transcription factor binding sites, cis-regulatory modules, DNA sequences
Procedia PDF Downloads 2896322 Power Quality Modeling Using Recognition Learning Methods for Waveform Disturbances
Authors: Sang-Keun Moon, Hong-Rok Lim, Jin-O Kim
Abstract:
This paper presents a Power Quality (PQ) modeling and filtering processes for the distribution system disturbances using recognition learning methods. Typical PQ waveforms with mathematical applications and gathered field data are applied to the proposed models. The objective of this paper is analyzing PQ data with respect to monitoring, discriminating, and evaluating the waveform of power disturbances to ensure the system preventative system failure protections and complex system problem estimations. Examined signal filtering techniques are used for the field waveform noises and feature extractions. Using extraction and learning classification techniques, the efficiency was verified for the recognition of the PQ disturbances with focusing on interactive modeling methods in this paper. The waveform of selected 8 disturbances is modeled with randomized parameters of IEEE 1159 PQ ranges. The range, parameters, and weights are updated regarding field waveform obtained. Along with voltages, currents have same process to obtain the waveform features as the voltage apart from some of ratings and filters. Changing loads are causing the distortion in the voltage waveform due to the drawing of the different patterns of current variation. In the conclusion, PQ disturbances in the voltage and current waveforms indicate different types of patterns of variations and disturbance, and a modified technique based on the symmetrical components in time domain was proposed in this paper for the PQ disturbances detection and then classification. Our method is based on the fact that obtained waveforms from suggested trigger conditions contain potential information for abnormality detections. The extracted features are sequentially applied to estimation and recognition learning modules for further studies.Keywords: power quality recognition, PQ modeling, waveform feature extraction, disturbance trigger condition, PQ signal filtering
Procedia PDF Downloads 1866321 The Content-Based Classroom: Perspectives on Integrating Language and Content
Authors: Mourad Ben Bennani
Abstract:
Views of language and language learning have undergone a tremendous change over the last decades. Language is no longer seen as a set of structured rules. It is rather viewed as a tool of interaction and communication. This shift in views has resulted in change in viewing language learning, which gave birth to various approaches and methodologies of language teaching. Two of these approaches are content-based instruction and content and language integrated learning (CLIL). These are similar approaches which integrate content and foreign/second language learning through various methodologies and models as a result of different implementations around the world. This presentation deals with sociocultural view of CBI and CLIL. It also defines language and content as vital components of CBI and CLIL. Next it reviews the origins of CBI and the continuum perspectives and CLIL definitions and models featured in the literature. Finally it summarizes current aspects around research in program evaluation with a focus on the benefits and challenges of these innovative approaches for second language teaching.Keywords: CBI, CLIL, CBI continuum, CLIL models
Procedia PDF Downloads 432