Search results for: data reproducibility
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25210

Search results for: data reproducibility

25150 How to Use Big Data in Logistics Issues

Authors: Mehmet Akif Aslan, Mehmet Simsek, Eyup Sensoy

Abstract:

Big Data stands for today’s cutting-edge technology. As the technology becomes widespread, so does Data. Utilizing massive data sets enable companies to get competitive advantages over their adversaries. Out of many area of Big Data usage, logistics has significance role in both commercial sector and military. This paper lays out what big data is and how it is used in both military and commercial logistics.

Keywords: big data, logistics, operational efficiency, risk management

Procedia PDF Downloads 641
25149 Detection of Telomerase Activity as Cancer Biomarker Using Nanogap-Rich Au Nanowire SERS Sensor

Authors: G. Eom, H. Kim, A. Hwang, T. Kang, B. Kim

Abstract:

Telomerase activity is overexpressed in over 85% of human cancers while suppressed in normal somatic cells. Telomerase has been attracted as a universal cancer biomarker. Therefore, the development of effective telomerase activity detection methods is urgently demanded in cancer diagnosis and therapy. Herein, we report a nanogap-rich Au nanowire (NW) surface-enhanced Raman scattering (SERS) sensor for detection of human telomerase activity. The nanogap-rich Au NW SERS sensors were prepared simply by uniformly depositing nanoparticles (NPs) on single-crystalline Au NWs. We measured SERS spectra of methylene blue (MB) from 60 different nanogap-rich Au NWs and obtained the relative standard deviation (RSD) of 4.80%, confirming the superb reproducibility of nanogap-rich Au NW SERS sensors. The nanogap-rich Au NW SERS sensors enable us to detect telomerase activity in 0.2 cancer cells/mL. Furthermore, telomerase activity is detectable in 7 different cancer cell lines whereas undetectable in normal cell lines, which suggest the potential applicability of nanogap-rich Au NW SERS sensor in cancer diagnosis. We expect that the present nanogap-rich Au NW SERS sensor can be useful in biomedical applications including a diverse biomarker sensing.

Keywords: cancer biomarker, nanowires, surface-enhanced Raman scattering, telomerase

Procedia PDF Downloads 349
25148 Implementation of an IoT Sensor Data Collection and Analysis Library

Authors: Jihyun Song, Kyeongjoo Kim, Minsoo Lee

Abstract:

Due to the development of information technology and wireless Internet technology, various data are being generated in various fields. These data are advantageous in that they provide real-time information to the users themselves. However, when the data are accumulated and analyzed, more various information can be extracted. In addition, development and dissemination of boards such as Arduino and Raspberry Pie have made it possible to easily test various sensors, and it is possible to collect sensor data directly by using database application tools such as MySQL. These directly collected data can be used for various research and can be useful as data for data mining. However, there are many difficulties in using the board to collect data, and there are many difficulties in using it when the user is not a computer programmer, or when using it for the first time. Even if data are collected, lack of expert knowledge or experience may cause difficulties in data analysis and visualization. In this paper, we aim to construct a library for sensor data collection and analysis to overcome these problems.

Keywords: clustering, data mining, DBSCAN, k-means, k-medoids, sensor data

Procedia PDF Downloads 378
25147 Government (Big) Data Ecosystem: Definition, Classification of Actors, and Their Roles

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Organizations, including governments, generate (big) data that are high in volume, velocity, veracity, and come from a variety of sources. Public Administrations are using (big) data, implementing base registries, and enforcing data sharing within the entire government to deliver (big) data related integrated services, provision of insights to users, and for good governance. Government (Big) data ecosystem actors represent distinct entities that provide data, consume data, manipulate data to offer paid services, and extend data services like data storage, hosting services to other actors. In this research work, we perform a systematic literature review. The key objectives of this paper are to propose a robust definition of government (big) data ecosystem and a classification of government (big) data ecosystem actors and their roles. We showcase a graphical view of actors, roles, and their relationship in the government (big) data ecosystem. We also discuss our research findings. We did not find too much published research articles about the government (big) data ecosystem, including its definition and classification of actors and their roles. Therefore, we lent ideas for the government (big) data ecosystem from numerous areas that include scientific research data, humanitarian data, open government data, industry data, in the literature.

Keywords: big data, big data ecosystem, classification of big data actors, big data actors roles, definition of government (big) data ecosystem, data-driven government, eGovernment, gaps in data ecosystems, government (big) data, public administration, systematic literature review

Procedia PDF Downloads 162
25146 Electromagnetic Interference Shielding Characteristics for Stainless Wire Mesh and Number of Plies of Carbon Fiber Reinforced Plastic

Authors: Min Sang Lee, Hee Jae Shin, In Pyo Cha, Hyun Kyung Yoon, Seong Woo Hong, Min Jae Yu, Hong Gun Kim, Lee Ku Kwac

Abstract:

In this paper, the electromagnetic shielding characteristics of an up-to-date typical carbon filler material, carbon fiber used with a metal mesh were investigated. Carbon fiber 12k-prepregs, where carbon fibers were impregnated with epoxy, were laminated with wire meshes, vacuum bag-molded and hardened to manufacture hybrid-type specimens, with which an electromagnetic shield test was performed in accordance with ASTM D4935-10, through which was known as the most excellent reproducibility is obtainable among electromagnetic shield tests. In addition, glass fiber prepress whose electromagnetic shielding effect were known as insignificant were laminated and formed with wire meshes to verify the validity of the electromagnetic shield effect of wire meshes in order to confirm the electromagnetic shielding effect of metal meshes corresponding existing carbon fiber 12k-prepregs. By grafting carbon fibers, on which studies are being actively underway in the environmental aspects and electromagnetic shielding effect, with hybrid-type wire meshes that were analyzed through the tests, in this study, the applicability and possibility are proposed.

Keywords: Carbon Fiber Reinforced Plastic(CFRP), Glass Fiber Reinforced Plastic(GFRP), stainless wire mesh, electromagnetic shielding

Procedia PDF Downloads 415
25145 A Statistical Model for the Geotechnical Parameters of Cement-Stabilised Hightown’s Soft Soil: A Case Stufy of Liverpool, UK

Authors: Hassnen M. Jafer, Khalid S. Hashim, W. Atherton, Ali W. Alattabi

Abstract:

This study investigates the effect of two important parameters (length of curing period and percentage of the added binder) on the strength of soil treated with OPC. An intermediate plasticity silty clayey soil with medium organic content was used in this study. This soft soil was treated with different percentages of a commercially available cement type 32.5-N. laboratory experiments were carried out on the soil treated with 0, 1.5, 3, 6, 9, and 12% OPC by the dry weight to determine the effect of OPC on the compaction parameters, consistency limits, and the compressive strength. Unconfined compressive strength (UCS) test was carried out on cement-treated specimens after exposing them to different curing periods (1, 3, 7, 14, 28, and 90 days). The results of UCS test were used to develop a non-linear multi-regression model to find the relationship between the predicted and the measured maximum compressive strength of the treated soil (qu). The results indicated that there was a significant improvement in the index of plasticity (IP) by treating with OPC; IP was decreased from 20.2 to 14.1 by using 12% of OPC; this percentage was enough to increase the UCS of the treated soil up to 1362 kPa after 90 days of curing. With respect to the statistical model of the predicted qu, the results showed that the regression coefficients (R2) was equal to 0.8534 which indicates a good reproducibility for the constructed model.

Keywords: cement admixtures, soft soil stabilisation, geotechnical parameters, multi-regression model

Procedia PDF Downloads 366
25144 Government Big Data Ecosystem: A Systematic Literature Review

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Data that is high in volume, velocity, veracity and comes from a variety of sources is usually generated in all sectors including the government sector. Globally public administrations are pursuing (big) data as new technology and trying to adopt a data-centric architecture for hosting and sharing data. Properly executed, big data and data analytics in the government (big) data ecosystem can be led to data-driven government and have a direct impact on the way policymakers work and citizens interact with governments. In this research paper, we conduct a systematic literature review. The main aims of this paper are to highlight essential aspects of the government (big) data ecosystem and to explore the most critical socio-technical factors that contribute to the successful implementation of government (big) data ecosystem. The essential aspects of government (big) data ecosystem include definition, data types, data lifecycle models, and actors and their roles. We also discuss the potential impact of (big) data in public administration and gaps in the government data ecosystems literature. As this is a new topic, we did not find specific articles on government (big) data ecosystem and therefore focused our research on various relevant areas like humanitarian data, open government data, scientific research data, industry data, etc.

Keywords: applications of big data, big data, big data types. big data ecosystem, critical success factors, data-driven government, egovernment, gaps in data ecosystems, government (big) data, literature review, public administration, systematic review

Procedia PDF Downloads 228
25143 A Machine Learning Decision Support Framework for Industrial Engineering Purposes

Authors: Anli Du Preez, James Bekker

Abstract:

Data is currently one of the most critical and influential emerging technologies. However, the true potential of data is yet to be exploited since, currently, about 1% of generated data are ever actually analyzed for value creation. There is a data gap where data is not explored due to the lack of data analytics infrastructure and the required data analytics skills. This study developed a decision support framework for data analytics by following Jabareen’s framework development methodology. The study focused on machine learning algorithms, which is a subset of data analytics. The developed framework is designed to assist data analysts with little experience, in choosing the appropriate machine learning algorithm given the purpose of their application.

Keywords: Data analytics, Industrial engineering, Machine learning, Value creation

Procedia PDF Downloads 168
25142 The Perceptions of Parents Regarding the Appropriateness of the Early Childhood Financial Literacy Program for Children 3 to 6 Years of Age Presented at an Early Childhood Facility in South Africa: A Case Study

Authors: M. Naude, R. Joubert, A. du Plessis, S. Pelser, M. Trollip

Abstract:

Context: The study focuses on the perceptions of South African parents and teachers regarding a play-based financial literacy program for children aged 3 to 6 years at an early childhood facility. It emphasizes the importance of early interventions in financial education to reduce poverty and inequality. Research Aim: To explore how parental involvement in teaching money management concepts to young children can support financial literacy education both at school and at home. Methodology: A qualitative deductive case study was conducted at a South African early childhood facility involving 90 children, their teachers and their families. Thematic content analysis of online survey responses and focus group discussions with teachers were used to identify patterns and themes related to participants’ perceptions of the financial literacy program. Validity: The study's validity and reproducibility are ensured by the depth and honesty of the data, participant involvement, and the inquirer's objectivity. Reliability aligns with the interpretive paradigm of this study, while transparency in data gathering and analysis enhances its trustworthiness. Credibility is further supported by using two triangulation methods: focus group interviews with teachers and open-ended questionnaires from parents. Findings: Parents reported overall satisfaction with the program and highlighted the development of essential money management skills in their children. They emphasized the collaborative role of home and school environments in fostering financial literacy in early childhood. Teachers reported that communication and interaction with the parents increased and grew. Healthy and positive relationships were established between the teachers and the parents which contributed to the success of the classroom financial literacy program. Theoretical Importance: The study underscores the significance of play-based financial literacy education in early childhood and the critical role of parental involvement in reinforcing money management concepts. It contributes to laying a solid foundation for children's future financial well-being. Data Collection: Data was collected through an online survey administered to parents of children participating in the financial literacy program over a period of 10 weeks. Focus group discussions were utilized with the teachers of each class after the conclusion of the program. Analysis Procedures: Thematic content analysis was applied to the survey responses to identify patterns, themes, and insights related to the participants’ perceptions of the program's effectiveness in teaching money management concepts to young children. Question Addressed: How does parental involvement in teaching money management concepts to young children support financial literacy education in early childhood? Conclusion: The study highlights the positive impact of a play-based financial literacy program for children aged 3 to 6 years and underscores the importance of collaboration between home and school environments in fostering financial literacy skills.

Keywords: early childhood, financial literacy, money management, parent involvement, play-based learning, South Africa

Procedia PDF Downloads 14
25141 Total-Reflection X-Ray Spectroscopy as a Tool for Element Screening in Food Samples

Authors: Hagen Stosnach

Abstract:

The analytical demands on modern instruments for element analysis in food samples include the analysis of major, trace and ultra-trace essential elements as well as potentially toxic trace elements. In this study total reflection, X-ray fluorescence analysis (TXRF) is presented as an analytical technique, which meets the requirements, defined by the Association of Official Agricultural Chemists (AOAC) regarding the limit of quantification, repeatability, reproducibility and recovery for most of the target elements. The advantages of TXRF are the small sample mass required, the broad linear range from µg/kg up to wt.-% values, no consumption of gases or cooling water, and the flexible and easy sample preparation. Liquid samples like alcoholic or non-alcoholic beverages can be analyzed without any preparation. For solid food samples, the most common sample pre-treatment methods are mineralization, direct deposition of the sample onto the reflector without/with minimal treatment, mainly as solid suspensions or after extraction. The main disadvantages are due to the possible peaks overlapping, which may lower the accuracy of quantitative analysis and the limit in the element identification. This analytical technique will be presented by several application examples, covering a broad range of liquid and solid food types.

Keywords: essential elements, toxic metals, XRF, spectroscopy

Procedia PDF Downloads 133
25140 Determination of Bisphenol A and Uric Acid by Modified Single-Walled Carbon Nanotube with Magnesium Layered Hydroxide 3-(4-Methoxyphenyl)Propionic Acid Nanocomposite

Authors: Illyas Md Isa, Maryam Musfirah Che Sobry, Mohamad Syahrizal Ahmad, Nurashikin Abd Azis

Abstract:

A single-walled carbon nanotube (SWCNT) that has been modified with magnesium layered hydroxide 3-(4-methoxyphenyl)propionic acid nanocomposite was proposed for the determination of uric acid and bisphenol A by square wave voltammetry. The results obtained denote that MLH-MPP nanocomposites enhance the sensitivity of the voltammetry detection responses. The best performance is shown by the modified carbon nanotube paste electrode (CNTPE) with the composition of single-walled carbon nanotube: magnesium layered hydroxide 3-(4-methoxyphenyl)propionic acid nanocomposite at 100:15 (% w/w). The linear range where the sensor works well is within the concentration 1.0 10-7 – 1.0 10-4 and 3.0 10-7 – 1.0 10-4 for uric acid and bisphenol A respectively with the limit of detection of 1.0 10-7 M for both organics. The interferences of uric acid and bisphenol A with other organic were studied and most of them did not interfere. The results shown for each experimental parameter on the proposed CNTPE showed that it has high sensitivity, good selectivity, repeatability and reproducibility. Therefore, the modified CNTPE can be used for the determination of uric acid and bisphenol A in real samples such as blood, plastic bottles and foods.

Keywords: bisphenol A, magnesium layered hydroxide 3-(4-methoxyphenyl)propionic acid nanocomposite, Nanocomposite, uric acid

Procedia PDF Downloads 212
25139 Providing Security to Private Cloud Using Advanced Encryption Standard Algorithm

Authors: Annapureddy Srikant Reddy, Atthanti Mahendra, Samala Chinni Krishna, N. Neelima

Abstract:

In our present world, we are generating a lot of data and we, need a specific device to store all these data. Generally, we store data in pen drives, hard drives, etc. Sometimes we may loss the data due to the corruption of devices. To overcome all these issues, we implemented a cloud space for storing the data, and it provides more security to the data. We can access the data with just using the internet from anywhere in the world. We implemented all these with the java using Net beans IDE. Once user uploads the data, he does not have any rights to change the data. Users uploaded files are stored in the cloud with the file name as system time and the directory will be created with some random words. Cloud accepts the data only if the size of the file is less than 2MB.

Keywords: cloud space, AES, FTP, NetBeans IDE

Procedia PDF Downloads 206
25138 Investigating Nanocrystalline CaF2:Tm for Carbon Beam and Gamma Radiation Dosimetry

Authors: Kanika Sharma, Shaila Bahl, Birendra Singh, Pratik Kumar, S. P. Lochab, A. Pandey

Abstract:

In the present investigation, initially nano-particles of CaF2 were prepared by the chemical co-precipitation method and later the prepared salt was activated by thulium (0.1 mol%) using the combustion technique. The final product was characterized and confirmed by X-Ray diffraction (XRD) and transmission electron microscopy (TEM). Further, the thermoluminescence (TL) properties of the nanophosphor were studied by irradiating it with 1.25 MeV of gamma radiation and 65 MeV of carbon (C6+) ion beam. For gamma rays, two prominent TL peaks were observed with a low temperature peak at around 1070C and a high temperature peak at around 1570C. Furthermore, the nanophosphor maintained a linear TL response for the entire range of studied doses i.e. 10 Gy to 2000 Gy for both the temperature peaks. Moreover, when the nanophosphor was irradiated with 65 MeV of C6+ ion beam the shape and structure of the glow curves remained spectacularly similar and the nanophosphor displayed a linear TL response for the full range of studied fluences i.e. 5*1010 ions/cm2 to 1 *1012 ions/ cm2. Finally, various tests like reproducibility test and batch homogeneity were also carried out to define the final product. Thus, co-precipitation method followed by combustion technique was successful in effectively producing dosimetric grade CaF2:Tm for dosimetry of gamma as well as carbon (C6+) beam.

Keywords: gamma radiation, ion beam, nanocrystalline, radiation dosimetry

Procedia PDF Downloads 184
25137 Business Intelligence for Profiling of Telecommunication Customer

Authors: Rokhmatul Insani, Hira Laksmiwati Soemitro

Abstract:

Business Intelligence is a methodology that exploits the data to produce information and knowledge systematically, business intelligence can support the decision-making process. Some methods in business intelligence are data warehouse and data mining. A data warehouse can store historical data from transactional data. For data modelling in data warehouse, we apply dimensional modelling by Kimball. While data mining is used to extracting patterns from the data and get insight from the data. Data mining has many techniques, one of which is segmentation. For profiling of telecommunication customer, we use customer segmentation according to customer’s usage of services, customer invoice and customer payment. Customers can be grouped according to their characteristics and can be identified the profitable customers. We apply K-Means Clustering Algorithm for segmentation. The input variable for that algorithm we use RFM (Recency, Frequency and Monetary) model. All process in data mining, we use tools IBM SPSS modeller.

Keywords: business intelligence, customer segmentation, data warehouse, data mining

Procedia PDF Downloads 483
25136 Imputation Technique for Feature Selection in Microarray Data Set

Authors: Younies Saeed Hassan Mahmoud, Mai Mabrouk, Elsayed Sallam

Abstract:

Analysing DNA microarray data sets is a great challenge, which faces the bioinformaticians due to the complication of using statistical and machine learning techniques. The challenge will be doubled if the microarray data sets contain missing data, which happens regularly because these techniques cannot deal with missing data. One of the most important data analysis process on the microarray data set is feature selection. This process finds the most important genes that affect certain disease. In this paper, we introduce a technique for imputing the missing data in microarray data sets while performing feature selection.

Keywords: DNA microarray, feature selection, missing data, bioinformatics

Procedia PDF Downloads 574
25135 PDDA: Priority-Based, Dynamic Data Aggregation Approach for Sensor-Based Big Data Framework

Authors: Lutful Karim, Mohammed S. Al-kahtani

Abstract:

Sensors are being used in various applications such as agriculture, health monitoring, air and water pollution monitoring, traffic monitoring and control and hence, play the vital role in the growth of big data. However, sensors collect redundant data. Thus, aggregating and filtering sensors data are significantly important to design an efficient big data framework. Current researches do not focus on aggregating and filtering data at multiple layers of sensor-based big data framework. Thus, this paper introduces (i) three layers data aggregation and framework for big data and (ii) a priority-based, dynamic data aggregation scheme (PDDA) for the lowest layer at sensors. Simulation results show that the PDDA outperforms existing tree and cluster-based data aggregation scheme in terms of overall network energy consumptions and end-to-end data transmission delay.

Keywords: big data, clustering, tree topology, data aggregation, sensor networks

Procedia PDF Downloads 345
25134 Noninvasive Disease Diagnosis through Breath Analysis Using DNA-functionalized SWNT Sensor Array

Authors: W. J. Zhang, Y. Q. Du, M. L. Wang

Abstract:

Noninvasive diagnostics of diseases via breath analysis has attracted considerable scientific and clinical interest for many years and become more and more promising with the rapid advancement in nanotechnology and biotechnology. The volatile organic compounds (VOCs) in exhaled breath, which are mainly blood borne, particularly provide highly valuable information about individuals’ physiological and pathophysiological conditions. Additionally, breath analysis is noninvasive, real-time, painless and agreeable to patients. We have developed a wireless sensor array based on single-stranded DNA (ssDNA)-decorated single-walled carbon nanotubes (SWNT) for the detection of a number of physiological indicators in breath. Eight DNA sequences were used to functionalize SWNT sensors to detect trace amount of methanol, benzene, dimethyl sulfide, hydrogen sulfide, acetone and ethanol, which are indicators of heavy smoking, excessive drinking, and diseases such as lung cancer, breast cancer, cirrhosis and diabetes. Our tests indicated that DNA functionalized SWNT sensors exhibit great selectivity, sensitivity, reproducibility, and repeatability. Furthermore, different molecules can be distinguished through pattern recognition enabled by this sensor array. Thus, the DNA-SWNT sensor array has great potential to be applied in chemical or bimolecular detection for the noninvasive diagnostics of diseases and health monitoring.

Keywords: breath analysis, diagnosis, DNA-SWNT sensor array, noninvasive

Procedia PDF Downloads 348
25133 Control the Flow of Big Data

Authors: Shizra Waris, Saleem Akhtar

Abstract:

Big data is a research area receiving attention from academia and IT communities. In the digital world, the amounts of data produced and stored have within a short period of time. Consequently this fast increasing rate of data has created many challenges. In this paper, we use functionalism and structuralism paradigms to analyze the genesis of big data applications and its current trends. This paper presents a complete discussion on state-of-the-art big data technologies based on group and stream data processing. Moreover, strengths and weaknesses of these technologies are analyzed. This study also covers big data analytics techniques, processing methods, some reported case studies from different vendor, several open research challenges and the chances brought about by big data. The similarities and differences of these techniques and technologies based on important limitations are also investigated. Emerging technologies are suggested as a solution for big data problems.

Keywords: computer, it community, industry, big data

Procedia PDF Downloads 194
25132 Glyco-Biosensing as a Novel Tool for Prostate Cancer Early-Stage Diagnosis

Authors: Pavel Damborsky, Martina Zamorova, Jaroslav Katrlik

Abstract:

Prostate cancer is annually the most common newly diagnosed cancer among men. An extensive number of evidence suggests that traditional serum Prostate-specific antigen (PSA) assay still suffers from a lack of sufficient specificity and sensitivity resulting in vast over-diagnosis and overtreatment. Thus, the early-stage detection of prostate cancer (PCa) plays undisputedly a critical role for successful treatment and improved quality of life. Over the last decade, particular altered glycans have been described that are associated with a range of chronic diseases, including cancer and inflammation. These glycans differences enable a distinction to be made between physiological and pathological state and suggest a valuable biosensing tool for diagnosis and follow-up purposes. Aberrant glycosylation is one of the major characteristics of disease progression. Consequently, the aim of this study was to develop a more reliable tool for early-stage PCa diagnosis employing lectins as glyco-recognition elements. Biosensor and biochip technology putting to use lectin-based glyco-profiling is one of the most promising strategies aimed at providing fast and efficient analysis of glycoproteins. The proof-of-concept experiments based on sandwich assay employing anti-PSA antibody and an aptamer as a capture molecules followed by lectin glycoprofiling were performed. We present a lectin-based biosensing assay for glycoprofiling of serum biomarker PSA using different biosensor and biochip platforms such as label-free surface plasmon resonance (SPR) and microarray with fluorescent label. The results suggest significant differences in interaction of particular lectins with PSA. The antibody-based assay is frequently associated with the sensitivity, reproducibility, and cross-reactivity issues. Aptamers provide remarkable advantages over antibodies due to the nucleic acid origin, stability and no glycosylation. All these data are further step for construction of highly selective, sensitive and reliable sensors for early-stage diagnosis. The experimental set-up also holds promise for the development of comparable assays with other glycosylated disease biomarkers.

Keywords: biomarker, glycosylation, lectin, prostate cancer

Procedia PDF Downloads 406
25131 High Performance Computing and Big Data Analytics

Authors: Branci Sarra, Branci Saadia

Abstract:

Because of the multiplied data growth, many computer science tools have been developed to process and analyze these Big Data. High-performance computing architectures have been designed to meet the treatment needs of Big Data (view transaction processing standpoint, strategic, and tactical analytics). The purpose of this article is to provide a historical and global perspective on the recent trend of high-performance computing architectures especially what has a relation with Analytics and Data Mining.

Keywords: high performance computing, HPC, big data, data analysis

Procedia PDF Downloads 520
25130 A Landscape of Research Data Repositories in Re3data.org Registry: A Case Study of Indian Repositories

Authors: Prashant Shrivastava

Abstract:

The purpose of this study is to explore re3dat.org registry to identify research data repositories registration workflow process. Further objective is to depict a graph for present development of research data repositories in India. Preliminarily with an approach to understand re3data.org registry framework and schema design then further proceed to explore the status of research data repositories of India in re3data.org registry. Research data repositories are getting wider relevance due to e-research concepts. Now available registry re3data.org is a good tool for users and researchers to identify appropriate research data repositories as per their research requirements. In Indian environment, a compatible National Research Data Policy is the need of the time to boost the management of research data. Registry for Research Data Repositories is a crucial tool to discover specific information in specific domain. Also, Research Data Repositories in India have not been studied. Re3data.org registry and status of Indian research data repositories both discussed in this study.

Keywords: research data, research data repositories, research data registry, re3data.org

Procedia PDF Downloads 324
25129 A Study of Cloud Computing Solution for Transportation Big Data Processing

Authors: Ilgin Gökaşar, Saman Ghaffarian

Abstract:

The need for fast processed big data of transportation ridership (eg., smartcard data) and traffic operation (e.g., traffic detectors data) which requires a lot of computational power is incontrovertible in Intelligent Transportation Systems. Nowadays cloud computing is one of the important subjects and popular information technology solution for data processing. It enables users to process enormous measure of data without having their own particular computing power. Thus, it can also be a good selection for transportation big data processing as well. This paper intends to examine how the cloud computing can enhance transportation big data process with contrasting its advantages and disadvantages, and discussing cloud computing features.

Keywords: big data, cloud computing, Intelligent Transportation Systems, ITS, traffic data processing

Procedia PDF Downloads 467
25128 Harmonic Data Preparation for Clustering and Classification

Authors: Ali Asheibi

Abstract:

The rapid increase in the size of databases required to store power quality monitoring data has demanded new techniques for analysing and understanding the data. One suggested technique to assist in analysis is data mining. Preparing raw data to be ready for data mining exploration take up most of the effort and time spent in the whole data mining process. Clustering is an important technique in data mining and machine learning in which underlying and meaningful groups of data are discovered. Large amounts of harmonic data have been collected from an actual harmonic monitoring system in a distribution system in Australia for three years. This amount of acquired data makes it difficult to identify operational events that significantly impact the harmonics generated on the system. In this paper, harmonic data preparation processes to better understanding of the data have been presented. Underlying classes in this data has then been identified using clustering technique based on the Minimum Message Length (MML) method. The underlying operational information contained within the clusters can be rapidly visualised by the engineers. The C5.0 algorithm was used for classification and interpretation of the generated clusters.

Keywords: data mining, harmonic data, clustering, classification

Procedia PDF Downloads 247
25127 Linguistic Summarization of Structured Patent Data

Authors: E. Y. Igde, S. Aydogan, F. E. Boran, D. Akay

Abstract:

Patent data have an increasingly important role in economic growth, innovation, technical advantages and business strategies and even in countries competitions. Analyzing of patent data is crucial since patents cover large part of all technological information of the world. In this paper, we have used the linguistic summarization technique to prove the validity of the hypotheses related to patent data stated in the literature.

Keywords: data mining, fuzzy sets, linguistic summarization, patent data

Procedia PDF Downloads 272
25126 Proposal of Data Collection from Probes

Authors: M. Kebisek, L. Spendla, M. Kopcek, T. Skulavik

Abstract:

In our paper we describe the security capabilities of data collection. Data are collected with probes located in the near and distant surroundings of the company. Considering the numerous obstacles e.g. forests, hills, urban areas, the data collection is realized in several ways. The collection of data uses connection via wireless communication, LAN network, GSM network and in certain areas data are collected by using vehicles. In order to ensure the connection to the server most of the probes have ability to communicate in several ways. Collected data are archived and subsequently used in supervisory applications. To ensure the collection of the required data, it is necessary to propose algorithms that will allow the probes to select suitable communication channel.

Keywords: communication, computer network, data collection, probe

Procedia PDF Downloads 360
25125 Investigation of Mode II Fracture Toughness in Orthotropic Materials

Authors: Mahdi Fakoor, Nabi Mehri Khansari, Ahmadreza Farokhi

Abstract:

Evaluation of mode II fracture toughness (KIIC) in composite materials is very hard problem to be solved, since it can be affected by many mechanisms of dissipation. Furthermore, non-linearity in its behavior can offer an extra difficulty to obtain accuracy in the results. Different reported values for KIIC in various references can prove the mentioned assertion. In this research, some solutions proposed based on the form of necessary corrections that should be executed on the common test fixtures. Due to the fact that the common test fixtures are not able to active toughening mechanisms in pure Mode II correctly, we have employed some structural modifications on common fixtures. Particularly, the Iosipescu test is used as start point. The tests are applied on graphite/epoxy; PMMA and Western White Pine Wood. Also, mixed mode I/II fracture limit curves are used to indicate the scattering in test results are really relevant to the creation of Fracture Process Zone (FPZ). In the present paper, shear load consideration applied at the predicted shear zone by considering some significant structural amendments that can active mode II toughening mechanisms. Indeed, the employed empirical method causes significant developing in repeatability and reproducibility as well. Moreover, a 3D Finite Element (FE) is performed for verification of the obtained results. Eventually, it is figured out that, a remarkable precision can be obtained in common test fixture in comparison with the previous one.

Keywords: FPZ, shear test fixture, mode II fracture toughness, composite material, FEM

Procedia PDF Downloads 361
25124 Open Science Philosophy, Research and Innovation

Authors: C.Ardil

Abstract:

Open Science translates the understanding and application of various theories and practices in open science philosophy, systems, paradigms and epistemology. Open Science originates with the premise that universal scientific knowledge is a product of a collective scholarly and social collaboration involving all stakeholders and knowledge belongs to the global society. Scientific outputs generated by public research are a public good that should be available to all at no cost and without barriers or restrictions. Open Science has the potential to increase the quality, impact and benefits of science and to accelerate advancement of knowledge by making it more reliable, more efficient and accurate, better understandable by society and responsive to societal challenges, and has the potential to enable growth and innovation through reuse of scientific results by all stakeholders at all levels of society, and ultimately contribute to growth and competitiveness of global society. Open Science is a global movement to improve accessibility to and reusability of research practices and outputs. In its broadest definition, it encompasses open access to publications, open research data and methods, open source, open educational resources, open evaluation, and citizen science. The implementation of open science provides an excellent opportunity to renegotiate the social roles and responsibilities of publicly funded research and to rethink the science system as a whole. Open Science is the practice of science in such a way that others can collaborate and contribute, where research data, lab notes and other research processes are freely available, under terms that enable reuse, redistribution and reproduction of the research and its underlying data and methods. Open Science represents a novel systematic approach to the scientific process, shifting from the standard practices of publishing research results in scientific publications towards sharing and using all available knowledge at an earlier stage in the research process, based on cooperative work and diffusing scholarly knowledge with no barriers and restrictions. Open Science refers to efforts to make the primary outputs of publicly funded research results (publications and the research data) publicly accessible in digital format with no limitations. Open Science is about extending the principles of openness to the whole research cycle, fostering, sharing and collaboration as early as possible, thus entailing a systemic change to the way science and research is done. Open Science is the ongoing transition in how open research is carried out, disseminated, deployed, and transformed to make scholarly research more open, global, collaborative, creative and closer to society. Open Science involves various movements aiming to remove the barriers for sharing any kind of output, resources, methods or tools, at any stage of the research process. Open Science embraces open access to publications, research data, source software, collaboration, peer review, notebooks, educational resources, monographs, citizen science, or research crowdfunding. The recognition and adoption of open science practices, including open science policies that increase open access to scientific literature and encourage data and code sharing, is increasing in the open science philosophy. Revolutionary open science policies are motivated by ethical, moral or utilitarian arguments, such as the right to access digital research literature for open source research or science data accumulation, research indicators, transparency in the field of academic practice, and reproducibility. Open science philosophy is adopted primarily to demonstrate the benefits of open science practices. Researchers use open science applications for their own advantage in order to get more offers, increase citations, attract media attention, potential collaborators, career opportunities, donations and funding opportunities. In open science philosophy, open data findings are evidence that open science practices provide significant benefits to researchers in scientific research creation, collaboration, communication, and evaluation according to more traditional closed science practices. Open science considers concerns such as the rigor of peer review, common research facts such as financing and career development, and the sacrifice of author rights. Therefore, researchers are recommended to implement open science research within the framework of existing academic evaluation and incentives. As a result, open science research issues are addressed in the areas of publishing, financing, collaboration, resource management and sharing, career development, discussion of open science questions and conclusions.

Keywords: Open Science, Open Science Philosophy, Open Science Research, Open Science Data

Procedia PDF Downloads 131
25123 A Review on Big Data Movement with Different Approaches

Authors: Nay Myo Sandar

Abstract:

With the growth of technologies and applications, a large amount of data has been producing at increasing rate from various resources such as social media networks, sensor devices, and other information serving devices. This large collection of massive, complex and exponential growth of dataset is called big data. The traditional database systems cannot store and process such data due to large and complexity. Consequently, cloud computing is a potential solution for data storage and processing since it can provide a pool of resources for servers and storage. However, moving large amount of data to and from is a challenging issue since it can encounter a high latency due to large data size. With respect to big data movement problem, this paper reviews the literature of previous works, discusses about research issues, finds out approaches for dealing with big data movement problem.

Keywords: Big Data, Cloud Computing, Big Data Movement, Network Techniques

Procedia PDF Downloads 86
25122 Nanostructure of Gamma-Alumina Prepared by a Modified Sol-Gel Technique

Authors: Débora N. Zambrano, Marina O. Gosatti, Leandro M. Dufou, Daniel A. Serrano, M. Mónica Guraya, Soledad Perez-Catán

Abstract:

Nanoporous g-Al2O3 samples were synthesized via a sol-gel technique, introducing changes in the Yoldas´ method. The aim of the work was to achieve an effective control of the nanostructure properties and morphology of the final g-Al2O3. The influence of the reagent temperature during the hydrolysis was evaluated in case of water at 5 ºC and 98 ºC, and alkoxide at -18 ºC and room temperature. Sol-gel transitions were performed at 120 ºC and room temperature. All g-Al2O3 samples were characterized by X-ray diffraction, nitrogen adsorption and thermal analysis. Our results showed that temperature of both water and alkoxide has not much influence on the nanostructure of the final g-Al2O3, thus giving a structure very similar to that of samples obtained by the reference method as long as the reaction temperature above 75 ºC is reached soon enough. XRD characterization showed diffraction patterns corresponding to g-Al2O3 for all samples. Also BET specific area values (253-280 m2/g) were similar to those obtained by Yoldas’s original method. The temperature of the sol-gel transition does not affect the resulting sample structure, and crystalline boehmite particles were identified in all dried gels. We analyzed the reproducibility of the samples’ structure by preparing different samples under identical conditions; we found that performing the sol-gel transition at 120 ºC favors the production of more reproducible samples and also reduces significantly the time of the sol-gel reaction.

Keywords: nanostructure alumina, boehmite, sol-gel technique, N2 adsorption/desorption isotherm, pore size distribution, BET area.

Procedia PDF Downloads 323
25121 Optimized Approach for Secure Data Sharing in Distributed Database

Authors: Ahmed Mateen, Zhu Qingsheng, Ahmad Bilal

Abstract:

In the current age of technology, information is the most precious asset of a company. Today, companies have a large amount of data. As the data become larger, access to data for some particular information is becoming slower day by day. Faster data processing to shape it in the form of information is the biggest issue. The major problems in distributed databases are the efficiency of data distribution and response time of data distribution. The security of data distribution is also a big issue. For these problems, we proposed a strategy that can maximize the efficiency of data distribution and also increase its response time. This technique gives better results for secure data distribution from multiple heterogeneous sources. The newly proposed technique facilitates the companies for secure data sharing efficiently and quickly.

Keywords: ER-schema, electronic record, P2P framework, API, query formulation

Procedia PDF Downloads 333