Search results for: patent data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24313

Search results for: patent data

24223 Jagiellonian-PET: A Novel TOF-PET Detector Based on Plastic Scintillators

Authors: P. Moskal, T. Bednarski, P. Bialas, E. Czerwinski, A. Gajos, A. Gruntowski, D. Kaminska, L. Kaplon, G. Korcyl, P. Kowalski, T. Kozik, W. Krzemien, E. Kubicz, Sz. Niedzwiecki, M. Palka, L. Raczynski, Z. Rudy, P. Salabura, N. G. Sharma, M. Silarski, A. Slomski, J. Smyrski, A. Strzelecki, A. Wieczorek, W. Wislicki, M. Zielinski, N. Zon

Abstract:

A new concept and results of the performance tests of the TOF-PET detection system developed at the Jagiellonian University will be presented. The novelty of the concept lies in employing long strips of polymer scintillators instead of crystals as detectors of annihilation quanta, and in using predominantly the timing of signals instead of their amplitudes for the reconstruction of Lines-of-Response. The diagnostic chamber consists of plastic scintillator strips readout by pairs of photo multipliers arranged axially around a cylindrical surface. To take advantage of the superior timing properties of plastic scintillators the signals are probed in the voltage domain with the accuracy of 20 ps by a newly developed electronics, and the data are collected by the novel trigger-less and reconfigurable data acquisition system. The hit-position and hit-time are reconstructed by the dedicated reconstruction methods based on the compressing sensing theory and the library of synchronized model signals. The solutions are subject to twelve patent applications. So far a time-of-flight resolution of ~120 ps (sigma) was achieved for a double-strip prototype with 30 cm field-of-view (FOV). It is by more than a factor of two better than TOF resolution achievable in current TOF-PET modalities and at the same time the FOV of 30 cm long prototype is significantly larger with respect to typical commercial PET devices. The Jagiellonian PET (J-PET) detector with plastic scintillators arranged axially possesses also another advantage. Its diagnostic chamber is free of any electronic devices and magnetic materials thus giving unique possibilities of combining J-PET with CT and J-PET with MRI for scanning the same part of a patient at the same time with both methods.

Keywords: PET-CT, PET-MRI, TOF-PET, scintillator

Procedia PDF Downloads 463
24222 Analysis of Big Data

Authors: Sandeep Sharma, Sarabjit Singh

Abstract:

As per the user demand and growth trends of large free data the storage solutions are now becoming more challenge-able to protect, store and to retrieve data. The days are not so far when the storage companies and organizations are start saying 'no' to store our valuable data or they will start charging a huge amount for its storage and protection. On the other hand as per the environmental conditions it becomes challenge-able to maintain and establish new data warehouses and data centers to protect global warming threats. A challenge of small data is over now, the challenges are big that how to manage the exponential growth of data. In this paper we have analyzed the growth trend of big data and its future implications. We have also focused on the impact of the unstructured data on various concerns and we have also suggested some possible remedies to streamline big data.

Keywords: big data, unstructured data, volume, variety, velocity

Procedia PDF Downloads 517
24221 Ultrasonic Densitometry of Alveolar Bone Jaw during Retention Period of Orthodontic Treatment

Authors: Margarita A. Belousova, Sergey N. Ermoliev, Nina K. Loginova

Abstract:

The method of intraoral ultrasound densitometry developed to diagnose mineral density of alveolar bone jaws during retention period of orthodontic treatment (Patent of Russian Federation № 2541038). It was revealed significant decrease of the ultrasonic wave speed and bone mineral density in patients with relapses dentition anomalies during retention period of orthodontic treatment.

Keywords: intraoral ultrasonic densitometry, speed of sound, alveolar jaw bone, relapses of dentition anomalies, retention period of orthodontic treatment

Procedia PDF Downloads 352
24220 The Nature and the Structure of Scientific and Innovative Collaboration Networks

Authors: Afshin Moazami, Andrea Schiffauerova

Abstract:

The objective of this work is to investigate the development and the role of collaboration networks in the creation of knowledge and innovations in the US and Canada, with a special focus on Quebec. In order to create scientific networks, the data on journal articles were extracted from SCOPUS, and the networks were built based on the co-authorship of the journal papers. For innovation networks, the USPTO database was used, and the networks were built on the patent co-inventorship. Various indicators characterizing the evolution of the network structure and the positions of the researchers and inventors in the networks were calculated. The comparison between the United States, Canada, and Quebec was then carried out. The preliminary results show that the nature of scientific collaboration networks differs from the one seen in innovation networks. Scientists work in bigger teams and are mostly interconnected within one giant network component, whereas the innovation network is much more clustered and fragmented, the inventors work more repetitively with the same partners, often in smaller isolated groups. In both Canada and the US, an increasing tendency towards collaboration was observed, and it was found that networks are getting bigger and more centralized with time. Moreover, a declining share of knowledge transfers per scientist was detected, suggesting an increasing specialization of science. The US collaboration networks tend to be more centralized than the Canadian ones. Quebec shares a lot of features with the Canadian network, but some differences were observed, for example, Quebec inventors rely more on the knowledge transmission through intermediaries.

Keywords: Canada, collaboration, innovation network, scientific network, Quebec, United States

Procedia PDF Downloads 172
24219 Research of Data Cleaning Methods Based on Dependency Rules

Authors: Yang Bao, Shi Wei Deng, WangQun Lin

Abstract:

This paper introduces the concept and principle of data cleaning, analyzes the types and causes of dirty data, and proposes several key steps of typical cleaning process, puts forward a well scalability and versatility data cleaning framework, in view of data with attribute dependency relation, designs several of violation data discovery algorithms by formal formula, which can obtain inconsistent data to all target columns with condition attribute dependent no matter data is structured (SQL) or unstructured (NoSQL), and gives 6 data cleaning methods based on these algorithms.

Keywords: data cleaning, dependency rules, violation data discovery, data repair

Procedia PDF Downloads 536
24218 Mining Big Data in Telecommunications Industry: Challenges, Techniques, and Revenue Opportunity

Authors: Hoda A. Abdel Hafez

Abstract:

Mining big data represents a big challenge nowadays. Many types of research are concerned with mining massive amounts of data and big data streams. Mining big data faces a lot of challenges including scalability, speed, heterogeneity, accuracy, provenance and privacy. In telecommunication industry, mining big data is like a mining for gold; it represents a big opportunity and maximizing the revenue streams in this industry. This paper discusses the characteristics of big data (volume, variety, velocity and veracity), data mining techniques and tools for handling very large data sets, mining big data in telecommunication and the benefits and opportunities gained from them.

Keywords: mining big data, big data, machine learning, telecommunication

Procedia PDF Downloads 372
24217 Producing Fertilizers of Increased Environmental and Agrochemical Efficiency via Application of Plant-available Inorganic Coatings

Authors: Andrey Norov

Abstract:

Reduction of inefficient losses of nutrients when using mineral fertilizers is a very important and urgent challenge, which is of both economic and environmental significance. The loss of nutrients to the environment leads to the release of greenhouse gases, eutrophication of water bodies, soil salinization and degradation, and other undesirable phenomena. This report focuses on slow and controlled release fertilizers produced through the application of inorganic coatings, which make the released nutrients plant-available. There are shown the advantages of these fertilizers their improved physical and chemical properties, as well as the effect of the coatings on yield growth and on the degree of nutrient efficiency. This type of fertilizers is an alternative to other polymer-coated fertilizers and is more ecofriendly. The production method is protected by the Russian patent.

Keywords: coatings, controlled release, fertilizer, nutrients, nutrient efficiency, yield increase

Procedia PDF Downloads 66
24216 Process of Research, Development and Application of New Pelletizer

Authors: Ľubomír Šooš, Peter Križan, Juraj Beniak, Miloš Matúš

Abstract:

The success of introducing a new product on the market is the new principle of production, or progressive design, improved efficiency or high quality of manufactured products. Proportionally with the growth of interest in press-biofuels - pellets or briquettes, is also growing interest in the new design better, more efficiently machines produce pellets, briquettes or granules completely new shapes. Our department has for years dedicated to the development of new highly productive designs pressing machines and new optimized press-biofuels. In this field, we have more than 40 national and international patents. The aim of paper is description of the introduction of a new principle pelleting mill and the description of his process of research, development, manufacturing and testing to deployment into production.

Keywords: compacting process, pellets mill, design, new conception, press-biofuels, patent, waste

Procedia PDF Downloads 357
24215 JavaScript Object Notation Data against eXtensible Markup Language Data in Software Applications a Software Testing Approach

Authors: Theertha Chandroth

Abstract:

This paper presents a comparative study on how to check JSON (JavaScript Object Notation) data against XML (eXtensible Markup Language) data from a software testing point of view. JSON and XML are widely used data interchange formats, each with its unique syntax and structure. The objective is to explore various techniques and methodologies for validating comparison and integration between JSON data to XML and vice versa. By understanding the process of checking JSON data against XML data, testers, developers and data practitioners can ensure accurate data representation, seamless data interchange, and effective data validation.

Keywords: XML, JSON, data comparison, integration testing, Python, SQL

Procedia PDF Downloads 94
24214 Multi-Source Data Fusion for Urban Comprehensive Management

Authors: Bolin Hua

Abstract:

In city governance, various data are involved, including city component data, demographic data, housing data and all kinds of business data. These data reflects different aspects of people, events and activities. Data generated from various systems are different in form and data source are different because they may come from different sectors. In order to reflect one or several facets of an event or rule, data from multiple sources need fusion together. Data from different sources using different ways of collection raised several issues which need to be resolved. Problem of data fusion include data update and synchronization, data exchange and sharing, file parsing and entry, duplicate data and its comparison, resource catalogue construction. Governments adopt statistical analysis, time series analysis, extrapolation, monitoring analysis, value mining, scenario prediction in order to achieve pattern discovery, law verification, root cause analysis and public opinion monitoring. The result of Multi-source data fusion is to form a uniform central database, which includes people data, location data, object data, and institution data, business data and space data. We need to use meta data to be referred to and read when application needs to access, manipulate and display the data. A uniform meta data management ensures effectiveness and consistency of data in the process of data exchange, data modeling, data cleansing, data loading, data storing, data analysis, data search and data delivery.

Keywords: multi-source data fusion, urban comprehensive management, information fusion, government data

Procedia PDF Downloads 353
24213 Reviewing Privacy Preserving Distributed Data Mining

Authors: Sajjad Baghernezhad, Saeideh Baghernezhad

Abstract:

Nowadays considering human involved in increasing data development some methods such as data mining to extract science are unavoidable. One of the discussions of data mining is inherent distribution of the data usually the bases creating or receiving such data belong to corporate or non-corporate persons and do not give their information freely to others. Yet there is no guarantee to enable someone to mine special data without entering in the owner’s privacy. Sending data and then gathering them by each vertical or horizontal software depends on the type of their preserving type and also executed to improve data privacy. In this study it was attempted to compare comprehensively preserving data methods; also general methods such as random data, coding and strong and weak points of each one are examined.

Keywords: data mining, distributed data mining, privacy protection, privacy preserving

Procedia PDF Downloads 492
24212 The Right to Data Portability and Its Influence on the Development of Digital Services

Authors: Roman Bieda

Abstract:

The General Data Protection Regulation (GDPR) will come into force on 25 May 2018 which will create a new legal framework for the protection of personal data in the European Union. Article 20 of GDPR introduces a right to data portability. This right allows for data subjects to receive the personal data which they have provided to a data controller, in a structured, commonly used and machine-readable format, and to transmit this data to another data controller. The right to data portability, by facilitating transferring personal data between IT environments (e.g.: applications), will also facilitate changing the provider of services (e.g. changing a bank or a cloud computing service provider). Therefore, it will contribute to the development of competition and the digital market. The aim of this paper is to discuss the right to data portability and its influence on the development of new digital services.

Keywords: data portability, digital market, GDPR, personal data

Procedia PDF Downloads 445
24211 The Romero-System Clarinet: A Milestone in the 19th Century Clarinet Manufacture

Authors: Pedro Rubio

Abstract:

Antonio Romero y Andía, was one of the most active and interesting figures in 19th century Spanish music. He was not only an exceptional clarinetist, he was also a publisher, a brilliant oboist, a music critic, and he revitalized Madrid’s musical scene by promoting orchestras and a national opera. In 1849, Romero was appointed Professor of Clarinet at the Conservatory of Madrid. Shortly after, Romero introduced to Spain the Boehm-System clarinet recently appeared in France. However, when initial interest in that system waned, he conceived his own system in 1853. The clarinet was manufactured in Paris by Lefêvre, who registered its first patent in 1862. In 1867 a second version was patented, and a year earlier, in 1866, the Romero clarinet was adopted as an official instrument for teaching the clarinet at the Conservatory of Madrid. The Romero-System clarinet mechanism has incorporated numerous additional devices and several extra keys, its skillful combination in a single instrument represents not only one of the pinnacles in the manufacture of musical instruments of the 19th century, but also an authentic synthesis of knowledge and practice in an era in which woodwind instruments were shaped as we know them today. Through the description and analysis of the data related to the aforementioned historical period, this lecture will try to show a crucial time in the history of all woodwind instruments, a period of technological effervescence in which the Romero-System clarinet emerged. The different stages of conception of the clarinet will be described, as well as its manufacturing and marketing process. Romero played with his clarinet system over twenty-five years. The research has identified the repertoire associated with this instrument whose conclusions will be presented in its case in the Congress.

Keywords: Antonio Romero, clarinet, keywork, 19th century

Procedia PDF Downloads 91
24210 Ultrasonic Densitometry of Bone Tissue of Jaws and Phalanges of Fingers in Patients after Orthodontic Treatment

Authors: Margarita Belousova

Abstract:

The ultrasonic densitometry (RU patent № 2541038) was used to assess the density of the bone tissue in the jaws of patients after orthodontic treatment. In addition, by ultrasonic densitometry assessed the state of the bone tissue in the region III phalanges of middle fingers in above mentioned patients. A comparative study was carried out in healthy volunteers of same age. It was established a significant decrease of the ultrasound wave speed and bone mineral density after active period of orthodontic treatment. Statistically, significant differences in bone mineral density of the fingers by ultrasonic densitometry in both groups of patients were not detected.

Keywords: intraoral ultrasonic densitometry, bone tissue density of jaws, bone tissue density of phalanges of fingers, orthodontic treatment

Procedia PDF Downloads 246
24209 Recent Advances in Data Warehouse

Authors: Fahad Hanash Alzahrani

Abstract:

This paper describes some recent advances in a quickly developing area of data storing and processing based on Data Warehouses and Data Mining techniques, which are associated with software, hardware, data mining algorithms and visualisation techniques having common features for any specific problems and tasks of their implementation.

Keywords: data warehouse, data mining, knowledge discovery in databases, on-line analytical processing

Procedia PDF Downloads 370
24208 How to Use Big Data in Logistics Issues

Authors: Mehmet Akif Aslan, Mehmet Simsek, Eyup Sensoy

Abstract:

Big Data stands for today’s cutting-edge technology. As the technology becomes widespread, so does Data. Utilizing massive data sets enable companies to get competitive advantages over their adversaries. Out of many area of Big Data usage, logistics has significance role in both commercial sector and military. This paper lays out what big data is and how it is used in both military and commercial logistics.

Keywords: big data, logistics, operational efficiency, risk management

Procedia PDF Downloads 613
24207 Polymorphism of HMW-GS in Collection of Wheat Genotypes

Authors: M. Chňapek, M. Tomka, R. Peroutková, Z. Gálová

Abstract:

Processes of plant breeding, testing and licensing of new varieties, patent protection in seed production, relations in trade and protection of copyright are dependent on identification, differentiation and characterization of plant genotypes. Therefore, we focused our research on utilization of wheat storage proteins as genetic markers suitable not only for differentiation of individual genotypes, but also for identification and characterization of their considerable properties. We analyzed a collection of 102 genotypes of bread wheat (Triticum aestivum L.), 41 genotypes of spelt wheat (Triticum spelta L.), and 35 genotypes of durum wheat (Triticum durum Desf.), in this study. Our results show, that genotypes of bread wheat and durum wheat were homogenous and single line, but spelt wheat genotypes were heterogenous. We observed variability of HMW-GS composition according to environmental factors and level of breeding and predict technological quality on the basis of Glu-score calculation.

Keywords: genotype identification, HMW-GS, wheat quality, polymorphism

Procedia PDF Downloads 435
24206 An Early Detection Type 2 Diabetes Using K - Nearest Neighbor Algorithm

Authors: Ng Liang Shen, Ngahzaifa Abdul Ghani

Abstract:

This research aimed at developing an early warning system for pre-diabetic and diabetics by analyzing simple and easily determinable signs and symptoms of diabetes among the people living in Malaysia using Particle Swarm Optimized Artificial. With the skyrocketing prevalence of Type 2 diabetes in Malaysia, the system can be used to encourage affected people to seek further medical attention to prevent the onset of diabetes or start managing it early enough to avoid the associated complications. The study sought to find out the best predictive variables of Type 2 Diabetes Mellitus, developed a system to diagnose diabetes from the variables using Artificial Neural Networks and tested the system on accuracy to find out the patent generated from diabetes diagnosis result in machine learning algorithms even at primary or advanced stages.

Keywords: diabetes diagnosis, Artificial Neural Networks, artificial intelligence, soft computing, medical diagnosis

Procedia PDF Downloads 304
24205 Implementation of an IoT Sensor Data Collection and Analysis Library

Authors: Jihyun Song, Kyeongjoo Kim, Minsoo Lee

Abstract:

Due to the development of information technology and wireless Internet technology, various data are being generated in various fields. These data are advantageous in that they provide real-time information to the users themselves. However, when the data are accumulated and analyzed, more various information can be extracted. In addition, development and dissemination of boards such as Arduino and Raspberry Pie have made it possible to easily test various sensors, and it is possible to collect sensor data directly by using database application tools such as MySQL. These directly collected data can be used for various research and can be useful as data for data mining. However, there are many difficulties in using the board to collect data, and there are many difficulties in using it when the user is not a computer programmer, or when using it for the first time. Even if data are collected, lack of expert knowledge or experience may cause difficulties in data analysis and visualization. In this paper, we aim to construct a library for sensor data collection and analysis to overcome these problems.

Keywords: clustering, data mining, DBSCAN, k-means, k-medoids, sensor data

Procedia PDF Downloads 349
24204 Government (Big) Data Ecosystem: Definition, Classification of Actors, and Their Roles

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Organizations, including governments, generate (big) data that are high in volume, velocity, veracity, and come from a variety of sources. Public Administrations are using (big) data, implementing base registries, and enforcing data sharing within the entire government to deliver (big) data related integrated services, provision of insights to users, and for good governance. Government (Big) data ecosystem actors represent distinct entities that provide data, consume data, manipulate data to offer paid services, and extend data services like data storage, hosting services to other actors. In this research work, we perform a systematic literature review. The key objectives of this paper are to propose a robust definition of government (big) data ecosystem and a classification of government (big) data ecosystem actors and their roles. We showcase a graphical view of actors, roles, and their relationship in the government (big) data ecosystem. We also discuss our research findings. We did not find too much published research articles about the government (big) data ecosystem, including its definition and classification of actors and their roles. Therefore, we lent ideas for the government (big) data ecosystem from numerous areas that include scientific research data, humanitarian data, open government data, industry data, in the literature.

Keywords: big data, big data ecosystem, classification of big data actors, big data actors roles, definition of government (big) data ecosystem, data-driven government, eGovernment, gaps in data ecosystems, government (big) data, public administration, systematic literature review

Procedia PDF Downloads 133
24203 Government Big Data Ecosystem: A Systematic Literature Review

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Data that is high in volume, velocity, veracity and comes from a variety of sources is usually generated in all sectors including the government sector. Globally public administrations are pursuing (big) data as new technology and trying to adopt a data-centric architecture for hosting and sharing data. Properly executed, big data and data analytics in the government (big) data ecosystem can be led to data-driven government and have a direct impact on the way policymakers work and citizens interact with governments. In this research paper, we conduct a systematic literature review. The main aims of this paper are to highlight essential aspects of the government (big) data ecosystem and to explore the most critical socio-technical factors that contribute to the successful implementation of government (big) data ecosystem. The essential aspects of government (big) data ecosystem include definition, data types, data lifecycle models, and actors and their roles. We also discuss the potential impact of (big) data in public administration and gaps in the government data ecosystems literature. As this is a new topic, we did not find specific articles on government (big) data ecosystem and therefore focused our research on various relevant areas like humanitarian data, open government data, scientific research data, industry data, etc.

Keywords: applications of big data, big data, big data types. big data ecosystem, critical success factors, data-driven government, egovernment, gaps in data ecosystems, government (big) data, literature review, public administration, systematic review

Procedia PDF Downloads 187
24202 A Machine Learning Decision Support Framework for Industrial Engineering Purposes

Authors: Anli Du Preez, James Bekker

Abstract:

Data is currently one of the most critical and influential emerging technologies. However, the true potential of data is yet to be exploited since, currently, about 1% of generated data are ever actually analyzed for value creation. There is a data gap where data is not explored due to the lack of data analytics infrastructure and the required data analytics skills. This study developed a decision support framework for data analytics by following Jabareen’s framework development methodology. The study focused on machine learning algorithms, which is a subset of data analytics. The developed framework is designed to assist data analysts with little experience, in choosing the appropriate machine learning algorithm given the purpose of their application.

Keywords: Data analytics, Industrial engineering, Machine learning, Value creation

Procedia PDF Downloads 142
24201 Providing Security to Private Cloud Using Advanced Encryption Standard Algorithm

Authors: Annapureddy Srikant Reddy, Atthanti Mahendra, Samala Chinni Krishna, N. Neelima

Abstract:

In our present world, we are generating a lot of data and we, need a specific device to store all these data. Generally, we store data in pen drives, hard drives, etc. Sometimes we may loss the data due to the corruption of devices. To overcome all these issues, we implemented a cloud space for storing the data, and it provides more security to the data. We can access the data with just using the internet from anywhere in the world. We implemented all these with the java using Net beans IDE. Once user uploads the data, he does not have any rights to change the data. Users uploaded files are stored in the cloud with the file name as system time and the directory will be created with some random words. Cloud accepts the data only if the size of the file is less than 2MB.

Keywords: cloud space, AES, FTP, NetBeans IDE

Procedia PDF Downloads 177
24200 Business Intelligence for Profiling of Telecommunication Customer

Authors: Rokhmatul Insani, Hira Laksmiwati Soemitro

Abstract:

Business Intelligence is a methodology that exploits the data to produce information and knowledge systematically, business intelligence can support the decision-making process. Some methods in business intelligence are data warehouse and data mining. A data warehouse can store historical data from transactional data. For data modelling in data warehouse, we apply dimensional modelling by Kimball. While data mining is used to extracting patterns from the data and get insight from the data. Data mining has many techniques, one of which is segmentation. For profiling of telecommunication customer, we use customer segmentation according to customer’s usage of services, customer invoice and customer payment. Customers can be grouped according to their characteristics and can be identified the profitable customers. We apply K-Means Clustering Algorithm for segmentation. The input variable for that algorithm we use RFM (Recency, Frequency and Monetary) model. All process in data mining, we use tools IBM SPSS modeller.

Keywords: business intelligence, customer segmentation, data warehouse, data mining

Procedia PDF Downloads 450
24199 Offshore Power Transition Project

Authors: Kashmir Johal

Abstract:

Within a wider context of improving whole-life effectiveness of gas and oil fields, we have been researching how to generate power local to the wellhead. (Provision of external power to a subsea wellhead can be prohibitively expensive and results in uneconomic fields. This has been an oil/gas industry challenge for many years.) We have been developing a possible approach to “local” power generation and have been conducting technical, environmental, (and economic) research to develop a viable approach. We sought to create a workable design for a new type of power generation system that makes use of differential pressure that can exist between the sea surface and a gas (or oil reservoir). The challenge has not just been to design a system capable of generating power from potential energy but also to design it in such a way that it anticipates and deals with the wide range of technological, environmental, and chemical constraints faced in such environments. We believe this project shows the enormous opportunity in deriving clean, economic, and zero emissions renewable energy from offshore sources. Since this technology is not currently available, a patent has been filed to protect the advancement of this technology.

Keywords: renewable, energy, power, offshore

Procedia PDF Downloads 48
24198 Permanent Magnet Machine Can Be a Vibration Sensor for Itself

Authors: M. Barański

Abstract:

The article presents a new vibration diagnostic method designed to (PM) machines with permanent magnets. Those devices are commonly used in small wind and water systems or vehicles drives. The author’s method is very innovative and unique. Specific structural properties of PM machines are used in this method - electromotive force (EMF) generated due to vibrations. There was analysed number of publications which describe vibration diagnostic methods and tests of electrical PM machines and there was no method found to determine the technical condition of such machine basing on their own signals. In this article, the method genesis, the similarity of machines with permanent magnet to vibration sensor and simulation and laboratory tests results will be discussed. The method of determination the technical condition of electrical machine with permanent magnets basing on its own signals is the subject of patent application No P.405669, and it is the main thesis of author’s doctoral dissertation.

Keywords: vibrations, generator, permanent magnet, traction drive, electrical vehicle

Procedia PDF Downloads 346
24197 Imputation Technique for Feature Selection in Microarray Data Set

Authors: Younies Saeed Hassan Mahmoud, Mai Mabrouk, Elsayed Sallam

Abstract:

Analysing DNA microarray data sets is a great challenge, which faces the bioinformaticians due to the complication of using statistical and machine learning techniques. The challenge will be doubled if the microarray data sets contain missing data, which happens regularly because these techniques cannot deal with missing data. One of the most important data analysis process on the microarray data set is feature selection. This process finds the most important genes that affect certain disease. In this paper, we introduce a technique for imputing the missing data in microarray data sets while performing feature selection.

Keywords: DNA microarray, feature selection, missing data, bioinformatics

Procedia PDF Downloads 536
24196 PDDA: Priority-Based, Dynamic Data Aggregation Approach for Sensor-Based Big Data Framework

Authors: Lutful Karim, Mohammed S. Al-kahtani

Abstract:

Sensors are being used in various applications such as agriculture, health monitoring, air and water pollution monitoring, traffic monitoring and control and hence, play the vital role in the growth of big data. However, sensors collect redundant data. Thus, aggregating and filtering sensors data are significantly important to design an efficient big data framework. Current researches do not focus on aggregating and filtering data at multiple layers of sensor-based big data framework. Thus, this paper introduces (i) three layers data aggregation and framework for big data and (ii) a priority-based, dynamic data aggregation scheme (PDDA) for the lowest layer at sensors. Simulation results show that the PDDA outperforms existing tree and cluster-based data aggregation scheme in terms of overall network energy consumptions and end-to-end data transmission delay.

Keywords: big data, clustering, tree topology, data aggregation, sensor networks

Procedia PDF Downloads 306
24195 Influence of Geometrical Parameters of a Wind Turbine on the Optimal Tip-Speed Ratio

Authors: Zdzislaw Piotr Kaminski, Miroslaw Wendeker, Zbigniew Czyz

Abstract:

The paper describes the geometric model, calculation algorithm and results of the CFD simulation of the airflow around a rotor in the vertical axis wind turbine (VAWT) with the ANSYS Fluent computational solver. The CFD method enables creating aerodynamic characteristics of forces acting on rotor working surfaces and determining parameters such as torque or power generated by the rotor assembly. The object of the research was a rotor whose construction is based on patent no.PL219985. The conducted tests enabled a mathematical model with a description of the generation of aerodynamic forces acting on each rotor blade. Additionally, this model was compared to the results of the wind tunnel tests. The analysis also focused on the influence of the blade angle on turbine power and the TSR. The research has shown that the turbine blade angle has a significant impact on the optimal value of the TSR.

Keywords: computational fluid dynamics, numerical analysis, renewable energy, wind turbine

Procedia PDF Downloads 121
24194 Control the Flow of Big Data

Authors: Shizra Waris, Saleem Akhtar

Abstract:

Big data is a research area receiving attention from academia and IT communities. In the digital world, the amounts of data produced and stored have within a short period of time. Consequently this fast increasing rate of data has created many challenges. In this paper, we use functionalism and structuralism paradigms to analyze the genesis of big data applications and its current trends. This paper presents a complete discussion on state-of-the-art big data technologies based on group and stream data processing. Moreover, strengths and weaknesses of these technologies are analyzed. This study also covers big data analytics techniques, processing methods, some reported case studies from different vendor, several open research challenges and the chances brought about by big data. The similarities and differences of these techniques and technologies based on important limitations are also investigated. Emerging technologies are suggested as a solution for big data problems.

Keywords: computer, it community, industry, big data

Procedia PDF Downloads 162