Search results for: Extraction and data integration
7387 Data Mining on the Router Logs for Statistical Application Classification
Authors: M. Rahmati, S.M. Mirzababaei
Abstract:
With the advance of information technology in the new era the applications of Internet to access data resources has steadily increased and huge amount of data have become accessible in various forms. Obviously, the network providers and agencies, look after to prevent electronic attacks that may be harmful or may be related to terrorist applications. Thus, these have facilitated the authorities to under take a variety of methods to protect the special regions from harmful data. One of the most important approaches is to use firewall in the network facilities. The main objectives of firewalls are to stop the transfer of suspicious packets in several ways. However because of its blind packet stopping, high process power requirements and expensive prices some of the providers are reluctant to use the firewall. In this paper we proposed a method to find a discriminate function to distinguish between usual packets and harmful ones by the statistical processing on the network router logs. By discriminating these data, an administrator may take an approach action against the user. This method is very fast and can be used simply in adjacent with the Internet routers.Keywords: Data Mining, Firewall, Optimization, Packetclassification, Statistical Pattern Recognition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16297386 Use of GIS for the Performance Evaluation of Canal Irrigation System in Rice Wheat Cropping Zone
Authors: Umm-e- Kalsoom, M. Arshad, Sadia Iqbal, M. Usman, M. Adnan
Abstract:
The research study evaluated the performance of irrigation system by using special scientific tools like Remote Sensing and GIS technology, so that proper measurements could be taken for the sustainable agriculture and water management. Different performance evaluation parameters had been calculated for the purposed data was gathered from field investigation and different government and private organizations. According to the calculations, organic matter ranges from 0.19% (low value) to 0.76% (high value). In flat irrigation system for wheat yield ranges from 3347.16 to 5260.39 kg/ha, while the total water applied to wheat crop ranges from 252.94 to 279.19 mm and WUE ranges from 13.07 to 18.37 kg/ha/mm. For rice yield ranges from 3347.47 to 5433.07 kg/ha with total water supplied to rice crop ranges from 764.71 to 978.15 mm and WUE ranges from 3.49 to 5.71 kg/ha/mm. Similarly, in raised bed system wheat yield ranges from 4569.13 to 6008.60 kg/ha, total water supplied ranges from 158.87 to 185.09 mm and WUE ranges from 27.20 to 33.54 kg/ha/mm while in rice crop, yield ranges from 5285.04 to 6716.69 kg/ha, total water supplied ranges from 600.72 to 755.06 mm and WUE ranges from 6.41 to 10.05 kg/ha/mm. Almost 51.3% water saving is observed in bed irrigation system as compared to flat system. Less water supplied to beds is more affective as its WUE value is higher than flat system where more water is supplied in both the seasons. Similarly, RWS values show that maximum water deficit while minimum area is getting adequate water supply. Greater yield is recorded in bed system as plant per square meter is more in bed system in comparison of flat system Thus, the integration of GIS tools to regularly compute performance indices could provide irrigation managers with the means for managing efficiently the irrigation system.Keywords: Field survey, Relative Water Supply (RWS), Remote sensing maps, Water Use Efficiency (WUE).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23997385 A Sliding Mesh Technique and Compressibility Correction Effects of Two-equation Turbulence Models for a Pintle-Perturbed Flow Analysis
Authors: J. Y. Heo, H. G. Sung
Abstract:
Numerical simulations have been performed for assessment of compressibility correction of two-equation turbulence models suitable for large scale separation flows perturbed by pintle strokes. In order to take into account pintle movement, a sliding mesh method was applied. The chamber pressure, mass flow rate, and thrust have been analyzed, and the response lag and sensitivity at the chamber and nozzle were estimated for a movable pintle. The nozzle performance for pintle reciprocating as its insertion and extraction processes, were analyzed to better understand the dynamic performance of the pintle nozzle.
Keywords: Pintle, sliding mesh, turbulent model, compressibility correction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22407384 Improvement of Data Transfer over Simple Object Access Protocol (SOAP)
Authors: Khaled Ahmed Kadouh, Kamal Ali Albashiri
Abstract:
This paper presents a designed algorithm involves improvement of transferring data over Simple Object Access Protocol (SOAP). The aim of this work is to establish whether using SOAP in exchanging XML messages has any added advantages or not. The results showed that XML messages without SOAP take longer time and consume more memory, especially with binary data.
Keywords: JAX-WS, SMTP, SOAP, Web service, XML.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21087383 Comparison of Machine Learning and Deep Learning Algorithms for Automatic Classification of 80 Different Pollen Species
Authors: Endrick Barnacin, Jean-Luc Henry, Jimmy Nagau, Jack Molinié
Abstract:
Palynology is a field of interest in many disciplines due to its multiple applications: chronological dating, climatology, allergy treatment, and honey characterization. Unfortunately, the analysis of a pollen slide is a complicated and time consuming task that requires the intervention of experts in the field, which are becoming increasingly rare due to economic and social conditions. In this context, the automation of this task is urgent. In this work, we compare classical feature extraction methods (Shape, GLCM, LBP, and others) and Deep Learning (CNN and Transfer Learning) to perform a recognition task over 80 regional pollen species. It has been found that the use of Transfer Learning seems to be more precise than the other approaches.
Keywords: Image segmentation, stuck particles separation, Sobel operator, thresholding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1687382 Numerical Simulations of Flood and Inundation in Jobaru River Basin Using Laser Profiler Data
Authors: Hiroto Nakashima, Toshihiro Morita, Koichiro Ohgushi
Abstract:
Laser Profiler (LP) data from aerial laser surveys have been increasingly used as topographical inputs to numerical simulations of flooding and inundation in river basins. LP data has great potential for reproducing topography, but its effective usage has not yet been fully established. In this study, flooding and inundation are simulated numerically using LP data for the Jobaru River basin of Japan’s Saga Plain. The analysis shows that the topography is reproduced satisfactorily in the computational domain with urban and agricultural areas requiring different grid sizes. A 2-D numerical simulation shows that flood flow behavior changes as grid size is varied.
Keywords: LP data, numerical simulation, topological analysis, mesh size.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15127381 Channels Splitting Strategy for Optical Local Area Networks of Passive Star Topology
Authors: Peristera Baziana
Abstract:
In this paper, we present a network configuration for a WDM LANs of passive star topology that assume that the set of data WDM channels is split into two separate sets of channels, with different access rights over them. Especially, a synchronous transmission WDMA access algorithm is adopted in order to increase the probability of successful transmission over the data channels and consequently to reduce the probability of data packets transmission cancellation in order to avoid the data channels collisions. Thus, a control pre-transmission access scheme is followed over a separate control channel. An analytical Markovian model is studied and the average throughput is mathematically derived. The performance is studied for several numbers of data channels and various values of control phase duration.Keywords: Access algorithm, channels division, collisions avoidance, wavelength division multiplexing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9937380 A Testbed for the Experiments Performed in Missing Value Treatments
Authors: Dias de J. C. Lilian, Lobato M. F. Fábio, de Santana L. Ádamo
Abstract:
The occurrence of missing values in database is a serious problem for Data Mining tasks, responsible for degrading data quality and accuracy of analyses. In this context, the area has shown a lack of standardization for experiments to treat missing values, introducing difficulties to the evaluation process among different researches due to the absence in the use of common parameters. This paper proposes a testbed intended to facilitate the experiments implementation and provide unbiased parameters using available datasets and suited performance metrics in order to optimize the evaluation and comparison between the state of art missing values treatments.
Keywords: Data imputation, data mining, missing values treatment, testbed.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14977379 Data-Reusing Adaptive Filtering Algorithms with Adaptive Error Constraint
Authors: Young-Seok Choi
Abstract:
We present a family of data-reusing and affine projection algorithms. For identification of a noisy linear finite impulse response channel, a partial knowledge of a channel, especially noise, can be used to improve the performance of the adaptive filter. Motivated by this fact, the proposed scheme incorporates an estimate of a knowledge of noise. A constraint, called the adaptive noise constraint, estimates an unknown information of noise. By imposing this constraint on a cost function of data-reusing and affine projection algorithms, a cost function based on the adaptive noise constraint and Lagrange multiplier is defined. Minimizing the new cost function leads to the adaptive noise constrained (ANC) data-reusing and affine projection algorithms. Experimental results comparing the proposed schemes to standard data-reusing and affine projection algorithms clearly indicate their superior performance.Keywords: Data-reusing, affine projection algorithm, error constraint, system identification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15967378 Balancing Strategies for Parallel Content-based Data Retrieval Algorithms in a k-tree Structured Database
Authors: Radu Dobrescu, Matei Dobrescu, Daniela Hossu
Abstract:
The paper proposes a unified model for multimedia data retrieval which includes data representatives, content representatives, index structure, and search algorithms. The multimedia data are defined as k-dimensional signals indexed in a multidimensional k-tree structure. The benefits of using the k-tree unified model were demonstrated by running the data retrieval application on a six networked nodes test bed cluster. The tests were performed with two retrieval algorithms, one that allows parallel searching using a single feature, the second that performs a weighted cascade search for multiple features querying. The experiments show a significant reduction of retrieval time while maintaining the quality of results.
Keywords: balancing strategies, multimedia databases, parallelprocessing, retrieval algorithms
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14067377 Improved Plasmonic Demultiplexer Based on Tapered and Rectangular Slot MIM Waveguide
Authors: Aso Rahimzadegan, Seyyed Poorya Hosseini, Kamran Qaderi
Abstract:
In this paper, we have proposed two novel plasmonic demultiplexing structures based on metal-insulator-metal surfaces which, beside their compact size, have a very good transmission spectrum. The impact of the key internal parameters on the transmission spectrum is numerically analyzed by using the twodimensional (2D) finite difference time domain (FDTD) method. The proposed structures could be used to develop ultra-compact photonic wavelength demultiplexing devices for large-scale photonic integration.
Keywords: Photonic integrated devices, Plasmonics, Metalinsulator- metal (MIM) waveguide, Demultiplexers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20547376 The HDH Model for the Development of Creative Structural Thinking and Its Applications to Other Systems
Authors: Mosseri Avraham
Abstract:
Teaching structures and structural design in architectural studies is considered a difficult mission due to complex reasons and circumstances. This article proposes a new conceptual model (HDH) for teaching structures and structural design in architectural studies. Because of its systems-thinking orientation it is also relevant and applicable to other fields and systems. The HDH model was developed in order to encourage the integration of science and art, especially in relation to structures, in architectural studies.
Keywords: Structural Thinking, Conceptual Design, Teaching Structures, Systems Thinking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13927375 Handling Mobility using Virtual Grid in Static Wireless Sensor Networks
Authors: T.P. Sharma
Abstract:
Querying a data source and routing data towards sink becomes a serious challenge in static wireless sensor networks if sink and/or data source are mobile. Many a times the event to be observed either moves or spreads across wide area making maintenance of continuous path between source and sink a challenge. Also, sink can move while query is being issued or data is on its way towards sink. In this paper, we extend our already proposed Grid Based Data Dissemination (GBDD) scheme which is a virtual grid based topology management scheme restricting impact of movement of sink(s) and event(s) to some specific cells of a grid. This obviates the need for frequent path modifications and hence maintains continuous flow of data while minimizing the network energy consumptions. Simulation experiments show significant improvements in network energy savings and average packet delay for a packet to reach at sink.Keywords: Mobility in WSNs, virtual grid, GBDD, clustering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15337374 Experimental Modal Analysis and Model Validation of Antenna Structures
Authors: B.R. Potgieter, G. Venter
Abstract:
Numerical design optimization is a powerful tool that can be used by engineers during any stage of the design process. There are many different applications for structural optimization. A specific application that will be discussed in the following paper is experimental data matching. Data obtained through tests on a physical structure will be matched with data from a numerical model of that same structure. The data of interest will be the dynamic characteristics of an antenna structure focusing on the mode shapes and modal frequencies. The structure used was a scaled and simplified model of the Karoo Array Telescope-7 (KAT-7) antenna structure. This kind of data matching is a complex and difficult task. This paper discusses how optimization can assist an engineer during the process of correlating a finite element model with vibration test data.Keywords: Finite Element Model (FEM), Karoo Array Telescope(KAT-7), modal frequencies, mode shapes, optimization, shape optimization, size optimization, vibration tests
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18247373 Compressed Suffix Arrays to Self-Indexes Based on Partitioned Elias-Fano
Abstract:
A practical and simple self-indexing data structure, Partitioned Elias-Fano (PEF) - Compressed Suffix Arrays (CSA), is built in linear time for the CSA based on PEF indexes. Moreover, the PEF-CSA is compared with two classical compressed indexing methods, Ferragina and Manzini implementation (FMI) and Sad-CSA on different type and size files in Pizza & Chili. The PEF-CSA performs better on the existing data in terms of the compression ratio, count, and locates time except for the evenly distributed data such as proteins data. The observations of the experiments are that the distribution of the φ is more important than the alphabet size on the compression ratio. Unevenly distributed data φ makes better compression effect, and the larger the size of the hit counts, the longer the count and locate time.
Keywords: Compressed suffix array, self-indexing, partitioned Elias-Fano, PEF-CSA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10597372 Real-time Performance Study of EPA Periodic Data Transmission
Authors: Liu Ning, Zhong Chongquan, Teng Hongfei
Abstract:
EPA (Ethernet for Plant Automation) resolves the nondeterministic problem of standard Ethernet and accomplishes real-time communication by means of micro-segment topology and deterministic scheduling mechanism. This paper studies the real-time performance of EPA periodic data transmission from theoretical and experimental perspective. By analyzing information transmission characteristics and EPA deterministic scheduling mechanism, 5 indicators including delivery time, time synchronization accuracy, data-sending time offset accuracy, utilization percentage of configured timeslice and non-RTE bandwidth that can be used to specify the real-time performance of EPA periodic data transmission are presented and investigated. On this basis, the test principles and test methods of the indicators are respectively studied and some formulas for real-time performance of EPA system are derived. Furthermore, an experiment platform is developed to test the indicators of EPA periodic data transmission in a micro-segment. According to the analysis and the experiment, the methods to improve the real-time performance of EPA periodic data transmission including optimizing network structure, studying self-adaptive adjustment method of timeslice and providing data-sending time offset accuracy for configuration are proposed.
Keywords: EPA system, Industrial Ethernet, Periodic data, Real-time performance
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14527371 Numerical Study on the Effect of Spudcan Penetration on the Jacket Platform
Authors: Xiangming Ge, Bing Pan, Wei He, Hao Chen, Yong Zhou, Jiayao Wu, Weijiang Chu
Abstract:
How the extraction and penetration of spudcan affect the performance of the adjacent pile foundation supporting the jacket platform was studied in the program FLAC3D depending on a wind farm project in Bohai sea. The simulations were conducted at the end of the spudcan penetration, which induced a pockmark in the seabed. The effects of the distance between the pile foundation and the pockmark were studied. The displacement at the mudline arose when the pockmark was closer. The bearing capacity of this jacket platform with deep pile foundations has been less influenced by the process of spudcan penetration, which can induce severe stresses on the pile foundation. The induced rotation was also satisfied with the serviceability constraints.
Keywords: Offshore foundation, pile-soil interaction, spudcan penetration, FLAC3D.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3207370 Data Quality Enhancement with String Length Distribution
Authors: Qi Xiu, Hiromu Hota, Yohsuke Ishii, Takuya Oda
Abstract:
Recently, collectable manufacturing data are rapidly increasing. On the other hand, mega recall is getting serious as a social problem. Under such circumstances, there are increasing needs for preventing mega recalls by defect analysis such as root cause analysis and abnormal detection utilizing manufacturing data. However, the time to classify strings in manufacturing data by traditional method is too long to meet requirement of quick defect analysis. Therefore, we present String Length Distribution Classification method (SLDC) to correctly classify strings in a short time. This method learns character features, especially string length distribution from Product ID, Machine ID in BOM and asset list. By applying the proposal to strings in actual manufacturing data, we verified that the classification time of strings can be reduced by 80%. As a result, it can be estimated that the requirement of quick defect analysis can be fulfilled.Keywords: Data quality, feature selection, probability distribution, string classification, string length.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13067369 Modern Kazakhstan in Global World After Independence
Authors: Dmitri Men, Byong-soon Chun, Soon-ok Myong
Abstract:
The article deals with the problems of political and economic processes in Kazakhstan since independence in the context of globalization. It analyzes the geopolitical situation and selfpositioning processes in the world after the end of the "cold war". It examines the problems of internal economization of the Republic for 20 years of independence. The authors argue that the reforms proceeded in the economic sphere have brought ambiguous and tangible results. Despite the difficult economic and political conditions facing a world economical crisis the country has undergone fundamental and radical transformations in the whole socio-economic systemKeywords: Globalization, Kazakhstan, integration, economic processes, financial crisis, multi-vector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 46587368 Efficient and Extensible Data Processing Framework in Ubiquitious Sensor Networks
Authors: Junghoon Lee, Gyung-Leen Park, Ho-Young Kwak, Cheol Min Kim
Abstract:
This paper presents the design and implements the prototype of an intelligent data processing framework in ubiquitous sensor networks. Much focus is put on how to handle the sensor data stream as well as the interoperability between the low-level sensor data and application clients. Our framework first addresses systematic middleware which mitigates the interaction between the application layer and low-level sensors, for the sake of analyzing a great volume of sensor data by filtering and integrating to create value-added context information. Then, an agent-based architecture is proposed for real-time data distribution to efficiently forward a specific event to the appropriate application registered in the directory service via the open interface. The prototype implementation demonstrates that our framework can host a sophisticated application on the ubiquitous sensor network and it can autonomously evolve to new middleware, taking advantages of promising technologies such as software agents, XML, cloud computing, and the like.
Keywords: sensor network, intelligent farm, middleware, event detection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13367367 An Approach for the Prediction of Cardiovascular Diseases
Authors: Nebi Gedik
Abstract:
Regardless of age or gender, cardiovascular illnesses are a serious health concern because of things like poor eating habits, stress, a sedentary lifestyle, hard work schedules, alcohol use, and weight. It tends to happen suddenly and has a high rate of recurrence. Machine learning models can be implemented to assist healthcare systems in the accurate detection and diagnosis of cardiovascular disease (CVD) in patients. Improved heart failure prediction is one of the primary goals of researchers using the heart disease dataset. The purpose of this study is to identify the feature or features that offer the best classification prediction for CVD detection. The support vector machine classifier is used to compare each feature's performance. It has been determined which feature produces the best results.
Keywords: Cardiovascular disease, feature extraction, supervised learning, support vector machine.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1097366 Eliciting and Confirming Data, Information, Knowledge and Wisdom in a Specialist Health Care Setting: The WICKED Method
Authors: S. Impey, D. Berry, S. Furtado, M. Galvin, L. Grogan, O. Hardiman, L. Hederman, M. Heverin, V. Wade, L. Douris, D. O'Sullivan, G. Stephens
Abstract:
Healthcare is a knowledge-rich environment. This knowledge, while valuable, is not always accessible outside the borders of individual clinics. This research aims to address part of this problem (at a study site) by constructing a maximal data set (knowledge artefact) for motor neurone disease (MND). This data set is proposed as an initial knowledge base for a concurrent project to develop an MND patient data platform. It represents the domain knowledge at the study site for the duration of the research (12 months). A knowledge elicitation method was also developed from the lessons learned during this process - the WICKED method. WICKED is an anagram of the words: eliciting and confirming data, information, knowledge, wisdom. But it is also a reference to the concept of wicked problems, which are complex and challenging, as is eliciting expert knowledge. The method was evaluated at a second site, and benefits and limitations were noted. Benefits include that the method provided a systematic way to manage data, information, knowledge and wisdom (DIKW) from various sources, including healthcare specialists and existing data sets. Limitations surrounded the time required and how the data set produced only represents DIKW known during the research period. Future work is underway to address these limitations.
Keywords: Healthcare, knowledge acquisition, maximal data sets, action design science.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4567365 Temporally Coherent 3D Animation Reconstruction from RGB-D Video Data
Authors: Salam Khalifa, Naveed Ahmed
Abstract:
We present a new method to reconstruct a temporally coherent 3D animation from single or multi-view RGB-D video data using unbiased feature point sampling. Given RGB-D video data, in form of a 3D point cloud sequence, our method first extracts feature points using both color and depth information. In the subsequent steps, these feature points are used to match two 3D point clouds in consecutive frames independent of their resolution. Our new motion vectors based dynamic alignement method then fully reconstruct a spatio-temporally coherent 3D animation. We perform extensive quantitative validation using novel error functions to analyze the results. We show that despite the limiting factors of temporal and spatial noise associated to RGB-D data, it is possible to extract temporal coherence to faithfully reconstruct a temporally coherent 3D animation from RGB-D video data.
Keywords: 3D video, 3D animation, RGB-D video, Temporally Coherent 3D Animation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20527364 Framework for the Modeling of the Supply Chain Collaborative Planning Process
Authors: D. Pérez, M. M. E. Alemany
Abstract:
In this work, a framework to model the Supply Chain (SC) Collaborative Planning (CP) process is proposed. The main contributions of this framework concern 1) the presentation of the decision view, the most important one due to the characteristics of the process, jointly within the physical, organisation and information views, and 2) the simultaneous consideration of the spatial and temporal integration among the different supply chain decision centres. This framework provides the basis for a realistic and integrated perspective of the supply chain collaborative planning process and also the analytical modeling of each of its decisional activities.Keywords: Collaborative Planning, Decision View, Distributed Decision-Making, Framework.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13057363 ECG Analysis using Nature Inspired Algorithm
Authors: A.Sankara Subramanian, G.Gurusamy, G.Selvakumar, P.Gnanasekar, A.Nagappan
Abstract:
This paper presents an algorithm based on the wavelet decomposition, for feature extraction from the ECG signal and recognition of three types of Ventricular Arrhythmias using neural networks. A set of Discrete Wavelet Transform (DWT) coefficients, which contain the maximum information about the arrhythmias, is selected from the wavelet decomposition. After that a novel clustering algorithm based on nature inspired algorithm (Ant Colony Optimization) is developed for classifying arrhythmia types. The algorithm is applied on the ECG registrations from the MIT-BIH arrhythmia and malignant ventricular arrhythmia databases. We applied Daubechies 4 wavelet in our algorithm. The wavelet decomposition enabled us to perform the task efficiently and produced reliable results.Keywords: Daubechies 4 Wavelet, ECG, Nature inspired algorithm, Ventricular Arrhythmias, Wavelet Decomposition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22887362 Application of Multi-Dimensional Principal Component Analysis to Medical Data
Authors: Naoki Yamamoto, Jun Murakami, Chiharu Okuma, Yutaro Shigeto, Satoko Saito, Takashi Izumi, Nozomi Hayashida
Abstract:
Multi-dimensional principal component analysis (PCA) is the extension of the PCA, which is used widely as the dimensionality reduction technique in multivariate data analysis, to handle multi-dimensional data. To calculate the PCA the singular value decomposition (SVD) is commonly employed by the reason of its numerical stability. The multi-dimensional PCA can be calculated by using the higher-order SVD (HOSVD), which is proposed by Lathauwer et al., similarly with the case of ordinary PCA. In this paper, we apply the multi-dimensional PCA to the multi-dimensional medical data including the functional independence measure (FIM) score, and describe the results of experimental analysis.Keywords: multi-dimensional principal component analysis, higher-order SVD (HOSVD), functional independence measure (FIM), medical data, tensor decomposition
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24757361 Management Prospects of Winery By-Products Based on Phenolic Compounds and Antioxidant Activity of Grape Skins: The Case of Greek Ionian Islands
Authors: Marinos Xagoraris, Iliada K. Lappa, Charalambos Kanakis, Dimitra Daferera, Christina Papadopoulou, Georgios Sourounis, Charilaos Giotis, Pavlos Bouchagier, Christos S. Pappas, Petros A. Tarantilis, Efstathia Skotti
Abstract:
The aim of this work was to recover phenolic compounds from grape skins produced in Greek varieties of the Ionian Islands in order to form the basis of calculations for their further utilization in the context of the circular economy. Isolation and further utilization of phenolic compounds is an important issue in winery by-products. For this purpose, 37 samples were collected, extracted, and analyzed in an attempt to provide the appropriate basis for their sustainable exploitation. Extraction of the bioactive compounds was held using an eco-friendly, non-toxic, and highly effective water-glycerol solvent system. Then, extracts were analyzed using UV-Vis, liquid chromatography-mass spectrometry (LC-MS), FTIR, and Raman spectroscopy. Also, total phenolic content and antioxidant activity were measured. LC-MS chromatography showed qualitative differences between different varieties. Peaks were attributed to monomeric 3-flavanols as well as monomeric, dimeric, and trimeric proanthocyanidins. The FT-IR and Raman spectra agreed with the chromatographic data and contributed to identifying phenolic compounds. Grape skins exhibited high total phenolic content (TPC), and it was proved that during vinification, a large number of polyphenols remained in the pomace. This study confirmed that grape skins from Ionian Islands are a promising source of bioactive compounds, suggesting their utilization under a bio-economic and environmental strategic framework.
Keywords: Antioxidant activity, grape skin, phenolic compounds, waste recovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4787360 A Geographical Spatial Analysis on the Benefits of Using Wind Energy in Kuwait
Authors: Obaid AlOtaibi, Salman Hussain
Abstract:
Wind energy is associated with many geographical factors including wind speed, climate change, surface topography, environmental impacts, and several economic factors, most notably the advancement of wind technology and energy prices. It is the fastest-growing and least economically expensive method for generating electricity. Wind energy generation is directly related to the characteristics of spatial wind. Therefore, the feasibility study for the wind energy conversion system is based on the value of the energy obtained relative to the initial investment and the cost of operation and maintenance. In Kuwait, wind energy is an appropriate choice as a source of energy generation. It can be used in groundwater extraction in agricultural areas such as Al-Abdali in the north and Al-Wafra in the south, or in fresh and brackish groundwater fields or remote and isolated locations such as border areas and projects away from conventional power electricity services, to take advantage of alternative energy, reduce pollutants, and reduce energy production costs. The study covers the State of Kuwait with an exception of metropolitan area. Climatic data were attained through the readings of eight distributed monitoring stations affiliated with Kuwait Institute for Scientific Research (KISR). The data were used to assess the daily, monthly, quarterly, and annual available wind energy accessible for utilization. The researchers applied the Suitability Model to analyze the study by using the ArcGIS program. It is a model of spatial analysis that compares more than one location based on grading weights to choose the most suitable one. The study criteria are: the average annual wind speed, land use, topography of land, distance from the main road networks, urban areas. According to the previous criteria, the four proposed locations to establish wind farm projects are selected based on the weights of the degree of suitability (excellent, good, average, and poor). The percentage of areas that represents the most suitable locations with an excellent rank (4) is 8% of Kuwait’s area. It is relatively distributed as follows: Al-Shqaya, Al-Dabdeba, Al-Salmi (5.22%), Al-Abdali (1.22%), Umm al-Hayman (0.70%), North Wafra and Al-Shaqeeq (0.86%). The study recommends to decision-makers to consider the proposed location (No.1), (Al-Shqaya, Al-Dabdaba, and Al-Salmi) as the most suitable location for future development of wind farms in Kuwait, this location is economically feasible.Keywords: Kuwait, renewable energy, spatial analysis, wind energy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8677359 Dynamic Data Partition Algorithm for a Parallel H.264 Encoder
Authors: Juntae Kim, Jaeyoung Park, Kyoungkun Lee, Jong Tae Kim
Abstract:
The H.264/AVC standard is a highly efficient video codec providing high-quality videos at low bit-rates. As employing advanced techniques, the computational complexity has been increased. The complexity brings about the major problem in the implementation of a real-time encoder and decoder. Parallelism is the one of approaches which can be implemented by multi-core system. We analyze macroblock-level parallelism which ensures the same bit rate with high concurrency of processors. In order to reduce the encoding time, dynamic data partition based on macroblock region is proposed. The data partition has the advantages in load balancing and data communication overhead. Using the data partition, the encoder obtains more than 3.59x speed-up on a four-processor system. This work can be applied to other multimedia processing applications.Keywords: H.264/AVC, video coding, thread-level parallelism, OpenMP, multimedia
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17807358 Developing Structured Sizing Systems for Manufacturing Ready-Made Garments of Indian Females Using Decision Tree-Based Data Mining
Authors: Hina Kausher, Sangita Srivastava
Abstract:
In India, there is a lack of standard, systematic sizing approach for producing readymade garments. Garments manufacturing companies use their own created size tables by modifying international sizing charts of ready-made garments. The purpose of this study is to tabulate the anthropometric data which cover the variety of figure proportions in both height and girth. 3,000 data have been collected by an anthropometric survey undertaken over females between the ages of 16 to 80 years from the some states of India to produce the sizing system suitable for clothing manufacture and retailing. The data are used for the statistical analysis of body measurements, the formulation of sizing systems and body measurements tables. Factor analysis technique is used to filter the control body dimensions from the large number of variables. Decision tree-based data mining is used to cluster the data. The standard and structured sizing system can facilitate pattern grading and garment production. Moreover, it can exceed buying ratios and upgrade size allocations to retail segments.Keywords: Anthropometric data, data mining, decision tree, garments manufacturing, ready-made garments, sizing systems.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 923