Search results for: climatological weather data measurement
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26521

Search results for: climatological weather data measurement

25291 Open Data for e-Governance: Case Study of Bangladesh

Authors: Sami Kabir, Sadek Hossain Khoka

Abstract:

Open Government Data (OGD) refers to all data produced by government which are accessible in reusable way by common people with access to Internet and at free of cost. In line with “Digital Bangladesh” vision of Bangladesh government, the concept of open data has been gaining momentum in the country. Opening all government data in digital and customizable format from single platform can enhance e-governance which will make government more transparent to the people. This paper presents a well-in-progress case study on OGD portal by Bangladesh Government in order to link decentralized data. The initiative is intended to facilitate e-service towards citizens through this one-stop web portal. The paper further discusses ways of collecting data in digital format from relevant agencies with a view to making it publicly available through this single point of access. Further, possible layout of this web portal is presented.

Keywords: e-governance, one-stop web portal, open government data, reusable data, web of data

Procedia PDF Downloads 332
25290 Adaptation Nature-Based Solutions: CBA of Woodlands for Flood Risk Management in the Aire Catchment, UK

Authors: Olivia R. Rendon

Abstract:

More than half of the world population lives in cities, in the UK, for example, 82% of the population was urban by 2013. Cities concentrate valuable and numerous infrastructure and sectors of the national economies. Cities are particularly vulnerable to climate change which will lead to higher damage costs in the future. There is thus a need to develop and invest in adaptation measures for cities to reduce the impact of flooding and other extreme weather events. Recent flood episodes present a significant and growing challenge to the UK and the estimated cost of urban flood damage is 270 million a year for England and Wales. This study aims to carry out cost-benefit analysis (CBA) of a nature-based approach for flood risk management in cities, focusing on the city of Leeds and the wider Aire catchment as a case study. Leeds was chosen as a case study due to its being one of the most flood vulnerable cities in the UK. In Leeds, over 4,500 properties are currently vulnerable to flooding and approximately £450 million of direct damage is estimated for a potential major flood from the River Aire. Leeds is also the second largest Metropolitan District in England with a projected population of 770,000 for 2014. So far the city council has mainly focused its flood risk management efforts on hard infrastructure solutions for the city centre. However, the wider Leeds district is at significant flood risk which could benefit from greener adaptation measures. This study presents estimates of a nature-based adaptation approach for flood risk management in Leeds. This land use management estimate is based on generating costings utilising primary and secondary data. This research contributes findings on the costs of different adaptation measures to flood risk management in a UK city, including the trade-offs and challenges of utilising nature-based solutions. Results also explore the potential implementation of the adaptation measures in the case study and the challenges of data collection and analysis for adaptation in flood risk management.

Keywords: green infrastructure, ecosystem services, woodland, adaptation, flood risk

Procedia PDF Downloads 265
25289 A Preliminary Study of the Effects of Abiotic Environmental Variables on Early Diptera Carrion Colonizers in Algiers, Algeria

Authors: M. Taleb, G. Tail, F. Z. Kara, B. Djedouani T. Moussa

Abstract:

Necrophagous insects usually colonize cadavers within a short time after death. However, they are influenced by weather conditions, and their distribution and activity vary according to different time scales, which can affect the post-mortem interval (PMI) estimation. As no data have been published in Algeria on necrophagous insects visiting corpses, two field surveys were conducted in July 2012 and March 2013 at the National Institute for Criminalistics and Criminology (INCC) using rabbit carcasses (Oryctolagus cuniculus L.). The trials were designed to identify the necrophagous Diptera fauna of Algiers, Algeria and examine their variations according to environmental variables. Four hundred and eighteen Diptera adults belonging to five families were captured during this study. The species which were identified on human corpses in different regions of Algeria were also observed on the rabbit carcasses. Although seasonal variations of the species were observed, their abundance did not significantly vary between the two seasons. In addition to seasonal effects, the ambient temperature, the wind speed, and precipitation affect the number of trapped flies. These conclusions highlight the necessity of considering the environmental factors at a scene to estimate the post-mortem interval accurately. It is hoped that these findings provide basic information regarding the necrophagous Diptera fauna of Algeria.

Keywords: forensic entomology, necrophagous diptera, post-mortem interval, abiotic factors, Algeria

Procedia PDF Downloads 371
25288 Ceramic Employees’ Occupational Health and Safety Training Expectations in Turkey

Authors: Erol Karaca

Abstract:

This study aims to analyze ceramic employees’ occupational health and safety training expectations. To that general objective, the study tries to examine whether occupational health and safety training expectations of ceramic employees meaningfully differentiate depending on demographic features and professional, social and economic conditions. For this purpose, the research data was collected through “Questionnaire of Occupational Health and Safety Training Expectation” (QSOHSTE) consisting of 25 open and close-ended questions developed by the researcher on the base of the literature review. QSOHSTE was applied to 125 ceramic employees working in Kutahya, Turkey. Data obtained from questionnaires were analyzed via SPSS 21. The findings, obtained from the study, revealed that employees’ agreement level to occupational health and safety training expectation statements is generally high-level. These findings also reveals that employees have various expectations about occupational health and safety training. These expectations are increasing sensitivity towards occupational health and safety training about the prevention of occupational accidents and diseases, contributing occupational health and safety training in establishing healthy and safe working environment, requiring occupational health and safety training before starting work, in case of changing working equipment and new technological applications, necessity of measurement and evaluation after occupational health and safety training. Besides these findings, employees’ agreement level to occupational health and safety training expectation statements also varies in terms of educational level, professional seniority, income level and perception of economic condition.

Keywords: occupational health and safety, occupational training, occupational expectation, professional seniority

Procedia PDF Downloads 430
25287 Application of Neural Networks to Predict Changing the Diameters of Bubbles in Pool Boiling Distilled Water

Authors: V. Nikkhah Rashidabad, M. Manteghian, M. Masoumi, S. Mousavian, D. Ashouri

Abstract:

In this research, the capability of neural networks in modeling and learning complicated and nonlinear relations has been used to develop a model for the prediction of changes in the diameter of bubbles in pool boiling distilled water. The input parameters used in the development of this network include element temperature, heat flux, and retention time of bubbles. The test data obtained from the experiment of the pool boiling of distilled water, and the measurement of the bubbles form on the cylindrical element. The model was developed based on training algorithm, which is typologically of back-propagation type. Considering the correlation coefficient obtained from this model is 0.9633. This shows that this model can be trusted for the simulation and modeling of the size of bubble and thermal transfer of boiling.

Keywords: bubble diameter, heat flux, neural network, training algorithm

Procedia PDF Downloads 430
25286 Geological Structure Identification in Semilir Formation: An Correlated Geological and Geophysical (Very Low Frequency) Data for Zonation Disaster with Current Density Parameters and Geological Surface Information

Authors: E. M. Rifqi Wilda Pradana, Bagus Bayu Prabowo, Meida Riski Pujiyati, Efraim Maykhel Hagana Ginting, Virgiawan Arya Hangga Reksa

Abstract:

The VLF (Very Low Frequency) method is an electromagnetic method that uses low frequencies between 10-30 KHz which results in a fairly deep penetration. In this study, the VLF method was used for zonation of disaster-prone areas by identifying geological structures in the form of faults. Data acquisition was carried out in Trimulyo Region, Jetis District, Bantul Regency, Special Region of Yogyakarta, Indonesia with 8 measurement paths. This study uses wave transmitters from Japan and Australia to obtain Tilt and Elipt values that can be used to create RAE (Rapat Arus Ekuivalen or Current Density) sections that can be used to identify areas that are easily crossed by electric current. This section will indicate the existence of a geological structure in the form of faults in the study area which is characterized by a high RAE value. In data processing of VLF method, it is obtained Tilt vs Elliptical graph and Moving Average (MA) Tilt vs Moving Average (MA) Elipt graph of each path that shows a fluctuating pattern and does not show any intersection at all. Data processing uses Matlab software and obtained areas with low RAE values that are 0%-6% which shows medium with low conductivity and high resistivity and can be interpreted as sandstone, claystone, and tuff lithology which is part of the Semilir Formation. Whereas a high RAE value of 10% -16% which shows a medium with high conductivity and low resistivity can be interpreted as a fault zone filled with fluid. The existence of the fault zone is strengthened by the discovery of a normal fault on the surface with strike N550W and dip 630E at coordinates X= 433256 and Y= 9127722 so that the activities of residents in the zone such as housing, mining activities and other activities can be avoided to reduce the risk of natural disasters.

Keywords: current density, faults, very low frequency, zonation

Procedia PDF Downloads 156
25285 An Investigation into the Use of Overset Mesh for a Vehicle Aerodynamics Case When Driving in Close Proximity

Authors: Kushal Kumar Chode, Remus Miahi Cirstea

Abstract:

In recent times, the drive towards more efficient vehicles and the increase in the number of vehicle on the roads has driven the aerodynamic researchers from studying the vehicle in isolation towards understanding the benefits of vehicle platooning. Vehicle platooning is defined as a series of vehicles traveling in close proximity. Due to the limitations in size and load measurement capabilities for the wind tunnels facilities, it is very difficult to perform this investigation experimentally. In this paper, the use of chimera or overset meshing technique is used within the STARCCM+ software to model the flow surrounding two identical vehicle models travelling in close proximity and also during an overtaking maneuver. The results are compared with data obtained from a polyhedral mesh and identical physics conditions. The benefits in terms of computational time and resources and the accuracy of the overset mesh approach are investigated.

Keywords: chimera mesh, computational accuracy, overset mesh, platooning vehicles

Procedia PDF Downloads 334
25284 Artificial Neural Network Model Based Setup Period Estimation for Polymer Cutting

Authors: Zsolt János Viharos, Krisztián Balázs Kis, Imre Paniti, Gábor Belső, Péter Németh, János Farkas

Abstract:

The paper presents the results and industrial applications in the production setup period estimation based on industrial data inherited from the field of polymer cutting. The literature of polymer cutting is very limited considering the number of publications. The first polymer cutting machine is known since the second half of the 20th century; however, the production of polymer parts with this kind of technology is still a challenging research topic. The products of the applying industrial partner must met high technical requirements, as they are used in medical, measurement instrumentation and painting industry branches. Typically, 20% of these parts are new work, which means every five years almost the entire product portfolio is replaced in their low series manufacturing environment. Consequently, it requires a flexible production system, where the estimation of the frequent setup periods' lengths is one of the key success factors. In the investigation, several (input) parameters have been studied and grouped to create an adequate training information set for an artificial neural network as a base for the estimation of the individual setup periods. In the first group, product information is collected such as the product name and number of items. The second group contains material data like material type and colour. In the third group, surface quality and tolerance information are collected including the finest surface and tightest (or narrowest) tolerance. The fourth group contains the setup data like machine type and work shift. One source of these parameters is the Manufacturing Execution System (MES) but some data were also collected from Computer Aided Design (CAD) drawings. The number of the applied tools is one of the key factors on which the industrial partners’ estimations were based previously. The artificial neural network model was trained on several thousands of real industrial data. The mean estimation accuracy of the setup periods' lengths was improved by 30%, and in the same time the deviation of the prognosis was also improved by 50%. Furthermore, an investigation on the mentioned parameter groups considering the manufacturing order was also researched. The paper also highlights the manufacturing introduction experiences and further improvements of the proposed methods, both on the shop floor and on the quotation preparation fields. Every week more than 100 real industrial setup events are given and the related data are collected.

Keywords: artificial neural network, low series manufacturing, polymer cutting, setup period estimation

Procedia PDF Downloads 233
25283 Resource Framework Descriptors for Interestingness in Data

Authors: C. B. Abhilash, Kavi Mahesh

Abstract:

Human beings are the most advanced species on earth; it's all because of the ability to communicate and share information via human language. In today's world, a huge amount of data is available on the web in text format. This has also resulted in the generation of big data in structured and unstructured formats. In general, the data is in the textual form, which is highly unstructured. To get insights and actionable content from this data, we need to incorporate the concepts of text mining and natural language processing. In our study, we mainly focus on Interesting data through which interesting facts are generated for the knowledge base. The approach is to derive the analytics from the text via the application of natural language processing. Using semantic web Resource framework descriptors (RDF), we generate the triple from the given data and derive the interesting patterns. The methodology also illustrates data integration using the RDF for reliable, interesting patterns.

Keywords: RDF, interestingness, knowledge base, semantic data

Procedia PDF Downloads 140
25282 Multi-Model Super Ensemble Based Advanced Approaches for Monsoon Rainfall Prediction

Authors: Swati Bhomia, C. M. Kishtawal, Neeru Jaiswal

Abstract:

Traditionally, monsoon forecasts have encountered many difficulties that stem from numerous issues such as lack of adequate upper air observations, mesoscale nature of convection, proper resolution, radiative interactions, planetary boundary layer physics, mesoscale air-sea fluxes, representation of orography, etc. Uncertainties in any of these areas lead to large systematic errors. Global circulation models (GCMs), which are developed independently at different institutes, each of which carries somewhat different representation of the above processes, can be combined to reduce the collective local biases in space, time, and for different variables from different models. This is the basic concept behind the multi-model superensemble and comprises of a training and a forecast phase. The training phase learns from the recent past performances of models and is used to determine statistical weights from a least square minimization via a simple multiple regression. These weights are then used in the forecast phase. The superensemble forecasts carry the highest skill compared to simple ensemble mean, bias corrected ensemble mean and the best model out of the participating member models. This approach is a powerful post-processing method for the estimation of weather forecast parameters reducing the direct model output errors. Although it can be applied successfully to the continuous parameters like temperature, humidity, wind speed, mean sea level pressure etc., in this paper, this approach is applied to rainfall, a parameter quite difficult to handle with standard post-processing methods, due to its high temporal and spatial variability. The present study aims at the development of advanced superensemble schemes comprising of 1-5 day daily precipitation forecasts from five state-of-the-art global circulation models (GCMs), i.e., European Centre for Medium Range Weather Forecasts (Europe), National Center for Environmental Prediction (USA), China Meteorological Administration (China), Canadian Meteorological Centre (Canada) and U.K. Meteorological Office (U.K.) obtained from THORPEX Interactive Grand Global Ensemble (TIGGE), which is one of the most complete data set available. The novel approaches include the dynamical model selection approach in which the selection of the superior models from the participating member models at each grid and for each forecast step in the training period is carried out. Multi-model superensemble based on the training using similar conditions is also discussed in the present study, which is based on the assumption that training with the similar type of conditions may provide the better forecasts in spite of the sequential training which is being used in the conventional multi-model ensemble (MME) approaches. Further, a variety of methods that incorporate a 'neighborhood' around each grid point which is available in literature to allow for spatial error or uncertainty, have also been experimented with the above mentioned approaches. The comparison of these schemes with respect to the observations verifies that the newly developed approaches provide more unified and skillful prediction of the summer monsoon (viz. June to September) rainfall compared to the conventional multi-model approach and the member models.

Keywords: multi-model superensemble, dynamical model selection, similarity criteria, neighborhood technique, rainfall prediction

Procedia PDF Downloads 120
25281 A Versatile Data Processing Package for Ground-Based Synthetic Aperture Radar Deformation Monitoring

Authors: Zheng Wang, Zhenhong Li, Jon Mills

Abstract:

Ground-based synthetic aperture radar (GBSAR) represents a powerful remote sensing tool for deformation monitoring towards various geohazards, e.g. landslides, mudflows, avalanches, infrastructure failures, and the subsidence of residential areas. Unlike spaceborne SAR with a fixed revisit period, GBSAR data can be acquired with an adjustable temporal resolution through either continuous or discontinuous operation. However, challenges arise from processing high temporal-resolution continuous GBSAR data, including the extreme cost of computational random-access-memory (RAM), the delay of displacement maps, and the loss of temporal evolution. Moreover, repositioning errors between discontinuous campaigns impede the accurate measurement of surface displacements. Therefore, a versatile package with two complete chains is developed in this study in order to process both continuous and discontinuous GBSAR data and address the aforementioned issues. The first chain is based on a small-baseline subset concept and it processes continuous GBSAR images unit by unit. Images within a window form a basic unit. By taking this strategy, the RAM requirement is reduced to only one unit of images and the chain can theoretically process an infinite number of images. The evolution of surface displacements can be detected as it keeps temporarily-coherent pixels which are present only in some certain units but not in the whole observation period. The chain supports real-time processing of the continuous data and the delay of creating displacement maps can be shortened without waiting for the entire dataset. The other chain aims to measure deformation between discontinuous campaigns. Temporal averaging is carried out on a stack of images in a single campaign in order to improve the signal-to-noise ratio of discontinuous data and minimise the loss of coherence. The temporal-averaged images are then processed by a particular interferometry procedure integrated with advanced interferometric SAR algorithms such as robust coherence estimation, non-local filtering, and selection of partially-coherent pixels. Experiments are conducted using both synthetic and real-world GBSAR data. Displacement time series at the level of a few sub-millimetres are achieved in several applications (e.g. a coastal cliff, a sand dune, a bridge, and a residential area), indicating the feasibility of the developed GBSAR data processing package for deformation monitoring of a wide range of scientific and practical applications.

Keywords: ground-based synthetic aperture radar, interferometry, small baseline subset algorithm, deformation monitoring

Procedia PDF Downloads 145
25280 Abilitest Battery: Presentation of Tests and Psychometric Properties

Authors: Sylwia Sumińska, Łukasz Kapica, Grzegorz Szczepański

Abstract:

Introduction: Cognitive skills are a crucial part of everyday functioning. Cognitive skills include perception, attention, language, memory, executive functions, and higher cognitive skills. With the aging of societies, there is an increasing percentage of people whose cognitive skills decline. Cognitive skills affect work performance. The appropriate diagnosis of a worker’s cognitive skills reduces the risk of errors and accidents at work which is also important for senior workers. The study aimed to prepare new cognitive tests for adults aged 20-60 and assess the psychometric properties of the tests. The project responds to the need for reliable and accurate methods of assessing cognitive performance. Computer tests were developed to assess psychomotor performance, attention, and working memory. Method: Two hundred eighty people aged 20-60 will participate in the study in 4 age groups. Inclusion criteria for the study were: no subjective cognitive impairment, no history of severe head injuries, chronic diseases, psychiatric and neurological diseases. The research will be conducted from February - to June 2022. Cognitive tests: 1) Measurement of psychomotor performance: Reaction time, Reaction time with selective attention component; 2) Measurement of sustained attention: Visual search (dots), Visual search (numbers); 3) Measurement of working memory: Remembering words, Remembering letters. To assess the validity and the reliability subjects will perform the Vienna Test System, i.e., “Reaction Test” (reaction time), “Signal Detection” (sustained attention), “Corsi Block-Tapping Test” (working memory), and Perception and Attention Test (TUS), Colour Trails Test (CTT), Digit Span – subtest from The Wechsler Adult Intelligence Scale. Eighty people will be invited to a session after three months aimed to assess the consistency over time. Results: Due to ongoing research, the detailed results from 280 people will be shown at the conference separately in each age group. The results of correlation analysis with the Vienna Test System will be demonstrated as well.

Keywords: aging, attention, cognitive skills, cognitive tests, psychomotor performance, working memory

Procedia PDF Downloads 91
25279 Evaluation of Intervention Effectiveness from the Client Perspective: Dimensions and Measurement of Wellbeing

Authors: Neşe Alkan

Abstract:

Purpose: The point that applied/clinical psychology, which is the practice and research discipline of the mental health field, has reached today can be summarized as the necessity of handling the psychological well-being of people from multiple perspectives and the goal of moving it to a higher level. Clients' subjective assessment of their own condition and wellbeing is an integral part of evidence-based interventions. There is a need for tools through which clients can evaluate the effectiveness of the psychotherapy/intervention performed with them and their contribution to the wellbeing and wellbeing of this process in a valid and reliable manner. The aim of this research is to meet this need, to test the reliability and validity of the index in Turkish, and explore its usability in the practices of both researchers and psychotherapists. Method: A total of 213 adults aged between 18-54, 69.5% working and 29.5% university students, were included in the study. Along with their demographic information, the participants were administered a set of scales: wellbeing, life satisfaction, spiritual satisfaction, shopping addiction, and loneliness, namely via an online platform. The construct validity of the wellbeing scale was tested with exploratory and confirmatory factor analyses, convergent and discriminant validity were tested with two-way full and partial correlation analyses and, measurement invariance was tested with one-way analysis of variance. Results: Factor analyzes showed that the scale consisted of six dimensions as it is in its original structure. The internal consistency of the scale was found to be Cronbach α = .82. Two-way correlation analyzes revealed that the wellbeing scale total score was positively correlated with general life satisfaction (r = .62) and spiritual satisfaction (r = .29), as expected. It was negatively correlated with loneliness (r = -.51) and shopping addiction (r = -.15). While the scale score did not vary by gender, previous illness, or nicotine addiction, it was found that the total wellbeing scale scores of the participants who had used antidepressant medication during the past year were lower than those who did not use antidepressant medication (F(1,204) = 7.713, p = .005). Conclusion: It has been concluded that the 12-item wellbeing scale consisting of six dimensions can be used in research and health sciences practices as a valid and reliable measurement tool. Further research which examines the reliability and validity of the scale in different widely used languages such as Spanish and Chinese is recommended.

Keywords: wellbeing, intervention effectiveness, reliability and validity, effectiveness

Procedia PDF Downloads 166
25278 Data Mining Practices: Practical Studies on the Telecommunication Companies in Jordan

Authors: Dina Ahmad Alkhodary

Abstract:

This study aimed to investigate the practices of Data Mining on the telecommunication companies in Jordan, from the viewpoint of the respondents. In order to achieve the goal of the study, and test the validity of hypotheses, the researcher has designed a questionnaire to collect data from managers and staff members from main department in the researched companies. The results shows improvements stages of the telecommunications companies towered Data Mining.

Keywords: data, mining, development, business

Procedia PDF Downloads 476
25277 Comparison of Corneal Curvature Measurements Conducted with Tomey AO-2000® and the Current Standard Biometer IOL Master®

Authors: Mohd Radzi Hilmi, Khairidzan Mohd Kamal, Che Azemin Mohd Zulfaezal, Ariffin Azrin Esmady

Abstract:

Purpose: Corneal curvature (CC) is an important anterior segment parameter. This study compared CC measurements conducted with two optical devices in phakic eyes. Methods: Sixty phakic eyes of 30 patients were enrolled in this study. CC was measured three times with the optical biometer and topography-keratometer Tomey AO-2000 (Tomey Corporation, Nagoya, Japan), then with the standard partial optical coherence interferometry (PCI) IOL Master (Carl Zeiss Meditec, Dublin, CA) and data were statistically analysed. Results: The measurements resulted in a mean CC of 43.86 ± 1.57 D with Tomey AO-2000 and 43.84 ± 1.55 D with IOL Master. Distribution of data is normal, and no significance difference in CC values was detected (P = 0.952) between the two devices. Correlation between CC measurements was highly significant (r = 0. 99; P < 0.0001). The mean difference of CC values between devices was 0.017 D and 95% limit of agreement was -0.088 to 0.12. Duration taken for measurements with the standard biometer IOL Master was longer (55.17 ± 2.24 seconds) than with Tomey AO-2000 (39.88 ± 2.38 seconds) in automatic mode. Duration of manual measurement with Tomey AO-2000 in manual mode was the shortest (28.57 ± 2.71 seconds). Conclusion: In phakic eyes, CC measured with Tomey AO-2000 and IOL Master showed similar values, and high correlation was observed between these two devices. This shows that both devices can be used interchangeably. Tomey AO-2000 is better in terms of faster to operate and has its own topography systems.

Keywords: corneal topography, corneal curvature, IOL Master, Tomey AO2000

Procedia PDF Downloads 365
25276 The Impact of System and Data Quality on Organizational Success in the Kingdom of Bahrain

Authors: Amal M. Alrayes

Abstract:

Data and system quality play a central role in organizational success, and the quality of any existing information system has a major influence on the effectiveness of overall system performance.Given the importance of system and data quality to an organization, it is relevant to highlight their importance on organizational performance in the Kingdom of Bahrain. This research aims to discover whether system quality and data quality are related, and to study the impact of system and data quality on organizational success. A theoretical model based on previous research is used to show the relationship between data and system quality, and organizational impact. We hypothesize, first, that system quality is positively associated with organizational impact, secondly that system quality is positively associated with data quality, and finally that data quality is positively associated with organizational impact. A questionnaire was conducted among public and private organizations in the Kingdom of Bahrain. The results show that there is a strong association between data and system quality, that affects organizational success.

Keywords: data quality, performance, system quality, Kingdom of Bahrain

Procedia PDF Downloads 473
25275 Cloud Computing in Data Mining: A Technical Survey

Authors: Ghaemi Reza, Abdollahi Hamid, Dashti Elham

Abstract:

Cloud computing poses a diversity of challenges in data mining operation arising out of the dynamic structure of data distribution as against the use of typical database scenarios in conventional architecture. Due to immense number of users seeking data on daily basis, there is a serious security concerns to cloud providers as well as data providers who put their data on the cloud computing environment. Big data analytics use compute intensive data mining algorithms (Hidden markov, MapReduce parallel programming, Mahot Project, Hadoop distributed file system, K-Means and KMediod, Apriori) that require efficient high performance processors to produce timely results. Data mining algorithms to solve or optimize the model parameters. The challenges that operation has to encounter is the successful transactions to be established with the existing virtual machine environment and the databases to be kept under the control. Several factors have led to the distributed data mining from normal or centralized mining. The approach is as a SaaS which uses multi-agent systems for implementing the different tasks of system. There are still some problems of data mining based on cloud computing, including design and selection of data mining algorithms.

Keywords: cloud computing, data mining, computing models, cloud services

Procedia PDF Downloads 458
25274 Evaluation of Methodologies for Measuring Harmonics and Inter-Harmonics in Photovoltaic Facilities

Authors: Anésio de Leles Ferreira Filho, Wesley Rodrigues de Oliveira, Jéssica Santoro Gonçalves, Jorge Andrés Cormane Angarita

Abstract:

The increase in electric power demand in face of environmental issues has intensified the participation of renewable energy sources such as photovoltaics, in the energy matrix of various countries. Due to their operational characteristics, they can generate time-varying harmonic and inter-harmonic distortions. For this reason, the application of methods of measurement based on traditional Fourier analysis, as proposed by IEC 61000-4-7, can provide inaccurate results. Considering the aspects mentioned herein, came the idea of the development of this work which aims to present the results of a comparative evaluation between a methodology arising from the combination of the Prony method with the Kalman filter and another method based on the IEC 61000-4-30 and IEC 61000-4-7 standards. Employed in this study were synthetic signals and data acquired through measurements in a 50kWp photovoltaic installation.

Keywords: harmonics, inter-harmonics, iec61000-4-7, parametric estimators, photovoltaic generation

Procedia PDF Downloads 466
25273 A Study of Cost and Revenue Earned from Tourist Walking Street Activities in Songkhla City Municipality, Thailand

Authors: Weerawan Marangkun

Abstract:

This study is a survey intended to investigate cost, revenue and factors affecting changes in revenue and to provide guidelines for improving factors affecting changes in revenue from tourist walking street activities in Songkhla City Municipality. Instruments used in this study were structured interviews, using Yaman table (1973) where the random sampling error was+ 10%. The sample consisting of 83 entrepreneurs were drawn from a total population of 272. The purposive sampling method was used. Data were collected during the 6-month period from December 2011 until May 2012. The findings indicate that the cost paid by an entrepreneur in connection with his/her services for tourists is mainly for travel, which stands at about 290 Baht per day. Each entrepreneur earns about 3,850 Baht per day, which means about 400,000 Baht per year. The combined total revenue from walking street tourist activities is about 108.8 million Baht per year. Such activities add economic value to tourist facilities due to expenditures by tourists and provide the entrepreneurs with considerable income. Factors affecting changes in revenue from tourist walking street activities are: the increase in the number of entrepreneurs; the holding of trade fairs, events or interesting shows in the vicinity; and weather conditions (e.g. abundant rainfall, which can contribute to a decrease in the number of tourists). Suggested measures to improve factors affecting changes in the income are: addition or creation of new activities; regulation of operations of the stalls and parking area; and generation of greater publicity through the social network.

Keywords: cost, revenue, tourist, walking street

Procedia PDF Downloads 344
25272 Cross-border Data Transfers to and from South Africa

Authors: Amy Gooden, Meshandren Naidoo

Abstract:

Genetic research and transfers of big data are not confined to a particular jurisdiction, but there is a lack of clarity regarding the legal requirements for importing and exporting such data. Using direct-to-consumer genetic testing (DTC-GT) as an example, this research assesses the status of data sharing into and out of South Africa (SA). While SA laws cover the sending of genetic data out of SA, prohibiting such transfer unless a legal ground exists, the position where genetic data comes into the country depends on the laws of the country from where it is sent – making the legal position less clear.

Keywords: cross-border, data, genetic testing, law, regulation, research, sharing, South Africa

Procedia PDF Downloads 115
25271 The Effectiveness of Solution-Focused Group Therapy on Improving Depressed Mothers of Child Abuser Families

Authors: Roya Maqami, Kaveh Qaderi Bagajan, Mohammad Mahdi Yousefi, Saeed Moradi

Abstract:

The purpose of this study is to investigate the efficacy of solution-focused group therapy on improving the depressed mothers of child abuser families. This study was carried out in the form of a semi-pilot, pre-test and post-test on two groups (experimental and control). Subjects include all mothers and their children that are the members of Shush and Naser Khosro child home. Beck Depression Inventory and Child Trauma Questionnaire were used to collect data. First, child abuse questionnaire was completed by children, Then Beck Depression Inventory was completed by their mothers that 22 of them were recognized as depressed and randomly divided in two groups of experimental and control. After applying pre-test for both of these groups, the intervention of solution- focused group therapy was performed in five sessions on experimental group. Finally, post-test was applied on both groups and subsequently in a month, follow-up test was performed. T-test, multivariate variance, and repeated measurement analysis of variance were used to analyze the data. According to the findings, it can be concluded that this therapy leads to the improvement of depressed mother's mood. As a result, the intervention of solution-focused group therapy is useful in order to improve the depressing mood of mothers of child abuser families.

Keywords: child abuse, depressed mothers, child abuser families, solution focused group therapy

Procedia PDF Downloads 346
25270 The Study of Security Techniques on Information System for Decision Making

Authors: Tejinder Singh

Abstract:

Information system is the flow of data from different levels to different directions for decision making and data operations in information system (IS). Data can be violated by different manner like manual or technical errors, data tampering or loss of integrity. Security system called firewall of IS is effected by such type of violations. The flow of data among various levels of Information System is done by networking system. The flow of data on network is in form of packets or frames. To protect these packets from unauthorized access, virus attacks, and to maintain the integrity level, network security is an important factor. To protect the data to get pirated, various security techniques are used. This paper represents the various security techniques and signifies different harmful attacks with the help of detailed data analysis. This paper will be beneficial for the organizations to make the system more secure, effective, and beneficial for future decisions making.

Keywords: information systems, data integrity, TCP/IP network, vulnerability, decision, data

Procedia PDF Downloads 286
25269 Data Integration with Geographic Information System Tools for Rural Environmental Monitoring

Authors: Tamas Jancso, Andrea Podor, Eva Nagyne Hajnal, Peter Udvardy, Gabor Nagy, Attila Varga, Meng Qingyan

Abstract:

The paper deals with the conditions and circumstances of integration of remotely sensed data for rural environmental monitoring purposes. The main task is to make decisions during the integration process when we have data sources with different resolution, location, spectral channels, and dimension. In order to have exact knowledge about the integration and data fusion possibilities, it is necessary to know the properties (metadata) that characterize the data. The paper explains the joining of these data sources using their attribute data through a sample project. The resulted product will be used for rural environmental analysis.

Keywords: remote sensing, GIS, metadata, integration, environmental analysis

Procedia PDF Downloads 105
25268 Guided Energy Theory of a Particle: Answered Questions Arise from Quantum Foundation

Authors: Desmond Agbolade Ademola

Abstract:

This work aimed to introduce a theory, called Guided Energy Theory of a particle that answered questions that arise from quantum foundation, quantum mechanics theory, and interpretation such as: what is nature of wavefunction? Is mathematical formalism of wavefunction correct? Does wavefunction collapse during measurement? Do quantum physical entanglement and many world interpretations really exist? In addition, is there uncertainty in the physical reality of our nature as being concluded in the Quantum theory? We have been able to show by the fundamental analysis presented in this work that the way quantum mechanics theory, and interpretation describes nature is not correlated with physical reality. Because, we discovered amongst others that, (1) Guided energy theory of a particle fundamentally provides complete physical observable series of quantized measurement of a particle momentum, force, energy e.t.c. in a given distance and time.In contrast, quantum mechanics wavefunction describes that nature has inherited probabilistic and indeterministic physical quantities, resulting in unobservable physical quantities that lead to many worldinterpretation.(2) Guided energy theory of a particle fundamentally predicts that it is mathematically possible to determine precise quantized measurementof position and momentum of a particle simultaneously. Because, there is no uncertainty in nature; nature however naturally guides itself against uncertainty. Contrary to the conclusion in quantum mechanics theory that, it is mathematically impossible to determine the position and the momentum of a particle simultaneously. Furthermore, we have been able to show by this theory that, it is mathematically possible to determine quantized measurement of force acting on a particle simultaneously, which is not possible on the premise of quantum mechanics theory. (3) It is evidently shown by our theory that, guided energy does not collapse, only describes the lopsided nature of a particle behavior in motion. This pretty offers us insight on gradual process of engagement - convergence and disengagement – divergence of guided energy holders which further highlight the picture how wave – like behavior return to particle-like behavior and how particle – like behavior return to wave – like behavior respectively. This further proves that the particles’ behavior in motion is oscillatory in nature. The mathematical formalism of Guided energy theory shows that nature is certainty whereas the mathematical formalism of Quantum mechanics theory shows that nature is absolutely probabilistics. In addition, the nature of wavefunction is the guided energy of the wave. In conclusion, the fundamental mathematical formalism of Quantum mechanics theory is wrong.

Keywords: momentum, physical entanglement, wavefunction, uncertainty

Procedia PDF Downloads 277
25267 Web Page Design Optimisation Based on Segment Analytics

Authors: Varsha V. Rohini, P. R. Shreya, B. Renukadevi

Abstract:

In the web analytics the information delivery and the web usage is optimized and the analysis of data is done. The analytics is the measurement, collection and analysis of webpage data. Page statistics and user metrics are the important factor in most of the web analytics tool. This is the limitation of the existing tools. It does not provide design inputs for the optimization of information. This paper aims at providing an extension for the scope of web analytics to provide analysis and statistics of each segment of a webpage. The number of click count is calculated and the concentration of links in a web page is obtained. Its user metrics are used to help in proper design of the displayed content in a webpage by Vision Based Page Segmentation (VIPS) algorithm. When the algorithm is applied on the web page it divides the entire web page into the visual block tree. The visual block tree generated will further divide the web page into visual blocks or segments which help us to understand the usage of each segment in a page and its content. The dynamic web pages and deep web pages are used to extend the scope of web page segment analytics. Space optimization concept is used with the help of the output obtained from the Vision Based Page Segmentation (VIPS) algorithm. This technique provides us the visibility of the user interaction with the WebPages and helps us to place the important links in the appropriate segments of the webpage and effectively manage space in a page and the concentration of links.

Keywords: analytics, design optimization, visual block trees, vision based technology

Procedia PDF Downloads 253
25266 Analysis of Genomics Big Data in Cloud Computing Using Fuzzy Logic

Authors: Mohammad Vahed, Ana Sadeghitohidi, Majid Vahed, Hiroki Takahashi

Abstract:

In the genomics field, the huge amounts of data have produced by the next-generation sequencers (NGS). Data volumes are very rapidly growing, as it is postulated that more than one billion bases will be produced per year in 2020. The growth rate of produced data is much faster than Moore's law in computer technology. This makes it more difficult to deal with genomics data, such as storing data, searching information, and finding the hidden information. It is required to develop the analysis platform for genomics big data. Cloud computing newly developed enables us to deal with big data more efficiently. Hadoop is one of the frameworks distributed computing and relies upon the core of a Big Data as a Service (BDaaS). Although many services have adopted this technology, e.g. amazon, there are a few applications in the biology field. Here, we propose a new algorithm to more efficiently deal with the genomics big data, e.g. sequencing data. Our algorithm consists of two parts: First is that BDaaS is applied for handling the data more efficiently. Second is that the hybrid method of MapReduce and Fuzzy logic is applied for data processing. This step can be parallelized in implementation. Our algorithm has great potential in computational analysis of genomics big data, e.g. de novo genome assembly and sequence similarity search. We will discuss our algorithm and its feasibility.

Keywords: big data, fuzzy logic, MapReduce, Hadoop, cloud computing

Procedia PDF Downloads 284
25265 Innovative Screening Tool Based on Physical Properties of Blood

Authors: Basant Singh Sikarwar, Mukesh Roy, Ayush Goyal, Priya Ranjan

Abstract:

This work combines two bodies of knowledge which includes biomedical basis of blood stain formation and fluid communities’ wisdom that such formation of blood stain depends heavily on physical properties. Moreover biomedical research tells that different patterns in stains of blood are robust indicator of blood donor’s health or lack thereof. Based on these valuable insights an innovative screening tool is proposed which can act as an aide in the diagnosis of diseases such Anemia, Hyperlipidaemia, Tuberculosis, Blood cancer, Leukemia, Malaria etc., with enhanced confidence in the proposed analysis. To realize this powerful technique, simple, robust and low-cost micro-fluidic devices, a micro-capillary viscometer and a pendant drop tensiometer are designed and proposed to be fabricated to measure the viscosity, surface tension and wettability of various blood samples. Once prognosis and diagnosis data has been generated, automated linear and nonlinear classifiers have been applied into the automated reasoning and presentation of results. A support vector machine (SVM) classifies data on a linear fashion. Discriminant analysis and nonlinear embedding’s are coupled with nonlinear manifold detection in data and detected decisions are made accordingly. In this way, physical properties can be used, using linear and non-linear classification techniques, for screening of various diseases in humans and cattle. Experiments are carried out to validate the physical properties measurement devices. This framework can be further developed towards a real life portable disease screening cum diagnostics tool. Small-scale production of screening cum diagnostic devices is proposed to carry out independent test.

Keywords: blood, physical properties, diagnostic, nonlinear, classifier, device, surface tension, viscosity, wettability

Procedia PDF Downloads 363
25264 Forthcoming Big Data on Smart Buildings and Cities: An Experimental Study on Correlations among Urban Data

Authors: Yu-Mi Song, Sung-Ah Kim, Dongyoun Shin

Abstract:

Cities are complex systems of diverse and inter-tangled activities. These activities and their complex interrelationships create diverse urban phenomena. And such urban phenomena have considerable influences on the lives of citizens. This research aimed to develop a method to reveal the causes and effects among diverse urban elements in order to enable better understanding of urban activities and, therefrom, to make better urban planning strategies. Specifically, this study was conducted to solve a data-recommendation problem found on a Korean public data homepage. First, a correlation analysis was conducted to find the correlations among random urban data. Then, based on the results of that correlation analysis, the weighted data network of each urban data was provided to people. It is expected that the weights of urban data thereby obtained will provide us with insights into cities and show us how diverse urban activities influence each other and induce feedback.

Keywords: big data, machine learning, ontology model, urban data model

Procedia PDF Downloads 396
25263 Utilization of Bauxite Residue in Construction Materials: An Experimental Study

Authors: Ryan Masoodi, Hossein Rostami

Abstract:

Aluminum has been credited for the massive advancement of many industrial products, from aerospace and automotive to electronics and even household appliances. These developments have come with a cost, which is a toxic by-product. The rise of aluminum production has been accompanied by the rise of a waste material called Bauxite Residue or Red Mud. This toxic material has been proved to be harmful to the environment, yet, there is no proper way to dispose or recycle it. Herewith, a new experimental method to utilize this waste in the building material is proposed. A method to mix red mud, fly ash, and some other ingredients is explored to create a new construction material that can satisfy the minimum required strength for bricks. It concludes that it is possible to produce bricks with enough strength that is suitable for constriction in environments with low to moderate weather conditions.

Keywords: bauxite residue, brick, red mud, recycling

Procedia PDF Downloads 152
25262 Text Similarity in Vector Space Models: A Comparative Study

Authors: Omid Shahmirzadi, Adam Lugowski, Kenneth Younge

Abstract:

Automatic measurement of semantic text similarity is an important task in natural language processing. In this paper, we evaluate the performance of different vector space models to perform this task. We address the real-world problem of modeling patent-to-patent similarity and compare TFIDF (and related extensions), topic models (e.g., latent semantic indexing), and neural models (e.g., paragraph vectors). Contrary to expectations, the added computational cost of text embedding methods is justified only when: 1) the target text is condensed; and 2) the similarity comparison is trivial. Otherwise, TFIDF performs surprisingly well in other cases: in particular for longer and more technical texts or for making finer-grained distinctions between nearest neighbors. Unexpectedly, extensions to the TFIDF method, such as adding noun phrases or calculating term weights incrementally, were not helpful in our context.

Keywords: big data, patent, text embedding, text similarity, vector space model

Procedia PDF Downloads 156