Search results for: Data Management in Cloud
7707 Strategic Decision Making Practice in Croatia – Which Decision Making Style is More Effective?
Authors: Ivana Bulog
Abstract:
Decision making is a vital part of the business world and any other field of human endeavor. Which way a business organization will take, and where that way will lead it, depends on broad range of decisions made by managers in the managerial structure. Strategic decisions are of the greatest importance for organizational success. Although much empirical research has been done trying to describe and explain its nature and effectiveness, knowledge about strategic decision making is still incomplete. This paper explores the nature of strategic decision making in particular setting - in Croatian companies. The main focus of this research is on the style that decision makers on strategic management level are following when making decisions of life importance for their companies. Two main decision making style that explain the way decision maker collects and processes available information and performs all the activities in strategic decision making process were empirical tested: rational and intuitive one. Besides analyzing their existence on strategic management level in Croatian companies, their effectiveness is analyzed as well. Results showed that decision makers at strategic management level are following both styles somewhat equally in order to function effectively, and that intuitive style is more effective when considering decisions outcomes.
Keywords: Decision making style, decision making effectiveness, strategic decisions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16967706 Water End-Use Classification with Contemporaneous Water-Energy Data and Deep Learning Network
Authors: Khoi A. Nguyen, Rodney A. Stewart, Hong Zhang
Abstract:
‘Water-related energy’ is energy use which is directly or indirectly influenced by changes to water use. Informatics applying a range of mathematical, statistical and rule-based approaches can be used to reveal important information on demand from the available data provided at second, minute or hourly intervals. This study aims to combine these two concepts to improve the current water end use disaggregation problem through applying a wide range of most advanced pattern recognition techniques to analyse the concurrent high-resolution water-energy consumption data. The obtained results have shown that recognition accuracies of all end-uses have significantly increased, especially for mechanised categories, including clothes washer, dishwasher and evaporative air cooler where over 95% of events were correctly classified.
Keywords: Deep learning network, smart metering, water end use, water-energy data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13647705 Role of GIS in Distribution Power Systems
Authors: N. Rezaee, M Nayeripour, A. Roosta, T. Niknam
Abstract:
With the prevalence of computer and development of information technology, Geographic Information Systems (GIS) have long used for a variety of applications in electrical engineering. GIS are designed to support the analysis, management, manipulation and mapping of spatial data. This paper presents several usages of GIS in power utilities such as automated route selection for the construction of new power lines which uses a dynamic programming model for route optimization, load forecasting and optimizing planning of substation-s location and capacity with comprehensive algorithm which involves an accurate small-area electric load forecasting procedure and simulates the different cost functions of substations.
Keywords: Geographic information systems (GIS), optimallocation and capacity, power distribution planning, route selection, spatial load forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 55057704 Watermark Bit Rate in Diverse Signal Domains
Authors: Nedeljko Cvejic, Tapio Sepp
Abstract:
A study of the obtainable watermark data rate for information hiding algorithms is presented in this paper. As the perceptual entropy for wideband monophonic audio signals is in the range of four to five bits per sample, a significant amount of additional information can be inserted into signal without causing any perceptual distortion. Experimental results showed that transform domain watermark embedding outperforms considerably watermark embedding in time domain and that signal decompositions with a high gain of transform coding, like the wavelet transform, are the most suitable for high data rate information hiding. Keywords?Digital watermarking, information hiding, audio watermarking, watermark data rate.
Keywords: Digital watermarking, information hiding, audio watermarking, watermark data rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16297703 Concurrent Access to Complex Entities
Authors: Cosmin Rablou
Abstract:
In this paper we present a way of controlling the concurrent access to data in a distributed application using the Pessimistic Offline Lock design pattern. In our case, the application processes a complex entity, which contains in a hierarchical structure different other entities (objects). It will be shown how the complex entity and the contained entities must be locked in order to control the concurrent access to data.Keywords: Object-oriented programming, Pessimistic Lock, Design pattern, Concurrent access to data, Processing complex entities
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13117702 Legal Regulation and Critical Analysis for an Effectively Treatment of Pharmaceutical Waste
Authors: Merita Dauti, Edita Alili-Idrizi, Sihana Ahmeti–Lika, Ledjan Malaj
Abstract:
The extermination and proper disposal of pharmaceutical wastes from expired and unused medications remains a disputable issue due to their specific nature and characteristics. Even though the hazards from these wastes are already well known in terms of environment and human health, people still treat them as usual wastes. At a national level, in many countries the management of pharmaceutical and medical wastes has been one of the main objectives in order to protect people’s health and the environment. Even though many legal regulations exist in this respect, there has not been a single law that would clearly explain the procedures of returning medicines, ways of selection, treatment and extermination of pharmaceutical wastes. This paper aims at analyzing the practices of pharmaceutical waste management and treatment in some European countries as well as a review of the legislation and official guidelines in managing these kinds of wastes and protecting the environment and human health. A suitable treatment and management of expired medications and other similar wastes would be in the interest of public health in the first place, as well as in the interest of healthcare institutions and other bodies engaged in environment protection.
Keywords: Pharmaceutical waste, legal regulation, proper disposal, environment pollution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28207701 Sparse Coding Based Classification of Electrocardiography Signals Using Data-Driven Complete Dictionary Learning
Authors: Fuad Noman, Sh-Hussain Salleh, Chee-Ming Ting, Hadri Hussain, Syed Rasul
Abstract:
In this paper, a data-driven dictionary approach is proposed for the automatic detection and classification of cardiovascular abnormalities. Electrocardiography (ECG) signal is represented by the trained complete dictionaries that contain prototypes or atoms to avoid the limitations of pre-defined dictionaries. The data-driven trained dictionaries simply take the ECG signal as input rather than extracting features to study the set of parameters that yield the most descriptive dictionary. The approach inherently learns the complicated morphological changes in ECG waveform, which is then used to improve the classification. The classification performance was evaluated with ECG data under two different preprocessing environments. In the first category, QT-database is baseline drift corrected with notch filter and it filters the 60 Hz power line noise. In the second category, the data are further filtered using fast moving average smoother. The experimental results on QT database confirm that our proposed algorithm shows a classification accuracy of 92%.Keywords: Electrocardiogram, dictionary learning, sparse coding, classification.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20957700 A Remote Sensing Approach to Calculate Population Using Roads Network Data in Lebanon
Authors: Kamel Allaw, Jocelyne Adjizian Gerard, Makram Chehayeb, Nada Badaro Saliba
Abstract:
In developing countries, such as Lebanon, the demographic data are hardly available due to the absence of the mechanization of population system. The aim of this study is to evaluate, using only remote sensing data, the correlations between the number of population and the characteristics of roads network (length of primary roads, length of secondary roads, total length of roads, density and percentage of roads and the number of intersections). In order to find the influence of the different factors on the demographic data, we studied the degree of correlation between each factor and the number of population. The results of this study have shown a strong correlation between the number of population and the density of roads and the number of intersections.
Keywords: Population, road network, statistical correlations, remote sensing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9967699 Wind Speed Data Analysis using Wavelet Transform
Authors: S. Avdakovic, A. Lukac, A. Nuhanovic, M. Music
Abstract:
Renewable energy systems are becoming a topic of great interest and investment in the world. In recent years wind power generation has experienced a very fast development in the whole world. For planning and successful implementations of good wind power plant projects, wind potential measurements are required. In these projects, of great importance is the effective choice of the micro location for wind potential measurements, installation of the measurement station with the appropriate measuring equipment, its maintenance and analysis of the gained data on wind potential characteristics. In this paper, a wavelet transform has been applied to analyze the wind speed data in the context of insight in the characteristics of the wind and the selection of suitable locations that could be the subject of a wind farm construction. This approach shows that it can be a useful tool in investigation of wind potential.Keywords: Wind potential, Wind speed data, Wavelettransform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26327698 A Biometric Template Security Approach to Fingerprints Based on Polynomial Transformations
Authors: Ramon Santana
Abstract:
The use of biometric identifiers in the field of information security, access control to resources, authentication in ATMs and banking among others, are of great concern because of the safety of biometric data. In the general architecture of a biometric system have been detected eight vulnerabilities, six of them allow obtaining minutiae template in plain text. The main consequence of obtaining minutia templates is the loss of biometric identifier for life. To mitigate these vulnerabilities several models to protect minutiae templates have been proposed. Several vulnerabilities in the cryptographic security of these models allow to obtain biometric data in plain text. In order to increase the cryptographic security and ease of reversibility, a minutiae templates protection model is proposed. The model aims to make the cryptographic protection and facilitate the reversibility of data using two levels of security. The first level of security is the data transformation level. In this level generates invariant data to rotation and translation, further transformation is irreversible. The second level of security is the evaluation level, where the encryption key is generated and data is evaluated using a defined evaluation function. The model is aimed at mitigating known vulnerabilities of the proposed models, basing its security on the impossibility of the polynomial reconstruction.Keywords: Fingerprint, template protection, bio-cryptography, minutiae protection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8437697 Accessible Business Process Modelling
Authors: D. D. Vaziri, D. DeOliveira
Abstract:
This article concerns with the accessibility of Business process modelling tools (BPMo tools) and business process modelling languages (BPMo languages). Therefore the reader will be introduced to business process management and the authors' motivation behind this inquiry. Afterwards, the paper will reflect problems when applying inaccessible BPMo tools. To illustrate these problems the authors distinguish between two different categories of issues and provide practical examples. Finally the article will present three approaches to improve the accessibility of BPMo tools and BPMo languages.Keywords: Accessibility, Business Process Management, BPM, Event Process Chains, Modelling Languages
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23477696 SIMGraph: Simplifying Contig Graph to Improve de Novo Genome Assembly Using Next-generation Sequencing Data
Authors: Chien-Ju Li, Chun-Hui Yu, Chi-Chuan Hwang, Tsunglin Liu , Darby Tien-Hao Chang
Abstract:
De novo genome assembly is always fragmented. Assembly fragmentation is more serious using the popular next generation sequencing (NGS) data because NGS sequences are shorter than the traditional Sanger sequences. As the data throughput of NGS is high, the fragmentations in assemblies are usually not the result of missing data. On the contrary, the assembled sequences, called contigs, are often connected to more than one other contigs in a complicated manner, leading to the fragmentations. False connections in such complicated connections between contigs, named a contig graph, are inevitable because of repeats and sequencing/assembly errors. Simplifying a contig graph by removing false connections directly improves genome assembly. In this work, we have developed a tool, SIMGraph, to resolve ambiguous connections between contigs using NGS data. Applying SIMGraph to the assembly of a fungus and a fish genome, we resolved 27.6% and 60.3% ambiguous contig connections, respectively. These results can reduce the experimental efforts in resolving contig connections.
Keywords: Contig graph, NGS, de novo assembly, scaffold.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17347695 Acute Coronary Syndrome Prediction Using Data Mining Techniques- An Application
Authors: Tahseen A. Jilani, Huda Yasin, Madiha Yasin, C. Ardil
Abstract:
In this paper we use data mining techniques to investigate factors that contribute significantly to enhancing the risk of acute coronary syndrome. We assume that the dependent variable is diagnosis – with dichotomous values showing presence or absence of disease. We have applied binary regression to the factors affecting the dependent variable. The data set has been taken from two different cardiac hospitals of Karachi, Pakistan. We have total sixteen variables out of which one is assumed dependent and other 15 are independent variables. For better performance of the regression model in predicting acute coronary syndrome, data reduction techniques like principle component analysis is applied. Based on results of data reduction, we have considered only 14 out of sixteen factors.
Keywords: Acute coronary syndrome (ACS), binary logistic regression analyses, myocardial ischemia (MI), principle component analysis, unstable angina (U.A.).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21147694 Improving the Analytical Power of Dynamic DEA Models, by the Consideration of the Shape of the Distribution of Inputs/Outputs Data: A Linear Piecewise Decomposition Approach
Authors: Elias K. Maragos, Petros E. Maravelakis
Abstract:
In Dynamic Data Envelopment Analysis (DDEA), which is a subfield of Data Envelopment Analysis (DEA), the productivity of Decision Making Units (DMUs) is considered in relation to time. In this case, as it is accepted by the most of the researchers, there are outputs, which are produced by a DMU to be used as inputs in a future time. Those outputs are known as intermediates. The common models, in DDEA, do not take into account the shape of the distribution of those inputs, outputs or intermediates data, assuming that the distribution of the virtual value of them does not deviate from linearity. This weakness causes the limitation of the accuracy of the analytical power of the traditional DDEA models. In this paper, the authors, using the concept of piecewise linear inputs and outputs, propose an extended DDEA model. The proposed model increases the flexibility of the traditional DDEA models and improves the measurement of the dynamic performance of DMUs.
Keywords: Data envelopment analysis, Dynamic DEA, Piecewise linear inputs, Piecewise linear outputs.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6567693 Enhancing Students’ Performance in Basic Science and Technology in Nigeria Using Moodle LMS
Authors: Olugbade Damola, Adekomi Adebimbo, Sofowora Olaniyi Alaba
Abstract:
One of the major problems facing education in Nigeria is the provision of quality Science and Technology education. Inadequate teaching facilities, non-usage of innovative teaching strategies, ineffective classroom management, lack of students’ motivation and poor integration of ICT has resulted in the increase in percentage of students who failed Basic Science and Technology in Junior Secondary Certification Examination for National Examination Council in Nigeria. To address these challenges, the Federal Government came up with a road map on education. This was with a view of enhancing quality education through integration of modern technology into teaching and learning, enhancing quality assurance through proper monitoring and introduction of innovative methods of teaching. This led the researcher to investigate how MOODLE LMS could be used to enhance students’ learning outcomes in BST. A sample of 120 students was purposively selected from four secondary schools in Ogbomoso. The experimental group was taught using MOODLE LMS, while the control group was taught using the conventional method. Data obtained were analyzed using mean, standard deviation and t-test. The result showed that MOODLE LMS was an effective learning platform in teaching BST in junior secondary schools (t=4.953, P<0.05). Students’ attitudes towards BST was also enhanced through MOODLE LMS (t=15.632, P<0.05). The use of MOODLE LMS significantly enhanced students’ retention (t=6.640, P<0.05). In conclusion, the Federal Government efforts at enhancing quality assurance through integration of modern technology and e-learning in Secondary schools proved to have yielded good result has students found MOODLE LMS to be motivating and interactive. Attendance was improved.
Keywords: MOODLE, learning management system, quality assurance, basic science and technology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35247692 Data Mining Determination of Sunlight Average Input for Solar Power Plant
Authors: Fl. Loury, P. Sablonière, C. Lamoureux, G. Magnier, Th. Gutierrez
Abstract:
A method is proposed to extract faithful representative patterns from data set of observations when they are suffering from non-negligible fluctuations. Supposing time interval between measurements to be extremely small compared to observation time, it consists in defining first a subset of intermediate time intervals characterizing coherent behavior. Data projection on these intervals gives a set of curves out of which an ideally “perfect” one is constructed by taking the sup limit of them. Then comparison with average real curve in corresponding interval gives an efficiency parameter expressing the degradation consecutive to fluctuation effect. The method is applied to sunlight data collected in a specific place, where ideal sunlight is the one resulting from direct exposure at location latitude over the year, and efficiency is resulting from action of meteorological parameters, mainly cloudiness, at different periods of the year. The extracted information already gives interesting element of decision, before being used for analysis of plant control.
Keywords: Base Input Reconstruction, Data Mining, Efficiency Factor, Information Pattern Operator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15287691 Job Satisfaction and Motivation as Predictors of Lecturers’ Effectiveness in Nigeria Police Academy
Authors: Abdulkareem Hussein Bibire
Abstract:
Job satisfaction and motivation have been given an important attention in psychology because they are seen as main instruments in maintaining organizational growth and development; they are also used to accomplish organizational aims and objectives. However, it has been observed that some institutions failed in motivating and stimulating their workers; in contrast, workers may be motivated but not satisfied with the job and failed to perform efficiently and effectively. It is hoped that the study of this nature would be of significance value to all stakeholders in education specifically, lecturers in higher institutions in Nigeria. Also, it is hoped that the findings of this study will enhance lecturers’ effectiveness and performance in discharging their duties. In the light of the above statements, this study investigated whether job satisfaction and motivation predict lecturers’ effectiveness in Nigeria Police Academy, Wudil, Kano State. Correlational research method was adopted for the study, while purposive sampling technique was used to choose the institution and the sampled lectures (70). Simple random sampling technique was used to select one hundred cadets across the academy. Two instruments were used to elicit information from both lecturers and cadets. These were job satisfaction and motivation; and lecturers’ effectiveness Questionnaires. The instruments were subjected to pilot testing and found to have reliability coefficient of 0.69 and 0.71 respectively. The results of the study revealed that there was a significance relationship among job satisfaction, motivation and lecturers effectiveness in Nigeria Police Academy. There was a significance relationship between job satisfaction and lecturers’ effectiveness in Nigeria Police Academy the cal r is 0.21 while the crt r is 0.19. at p<0.05 and; there was a significance relationship between job motivation and lecturers effectiveness in Nigeria Police Academy the cal r is 0.20 while the crt r is 0.19 at p<0.05This study therefore concluded that there was a significance relationship among job satisfaction, motivation and lecturers effectiveness in Nigeria Police Academy. Based on the data collected, collated and analyzed Recommendations were made for both the lecturers and the Academy management. It is also suggested that lecturers should be industrious in their primary assignment in other to make values to cadets lives and career. And management should also try to enhance lecturers performance by more motivational needs for the lecturers.Keywords: Satisfaction, motivation, lecturer effectiveness, academy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21357690 Transesterification of Waste Cooking Oil for Biodiesel Production Using Modified Clinoptilolite Zeolite as a Heterogeneous Catalyst
Authors: D. Mowla, N. Rasti, P. Keshavarz
Abstract:
Reduction of fossil fuels sources, increasing of pollution gases emission, and global warming effects increase the demand of renewable fuels. One of the main candidates of alternative fuels is biodiesel. Biodiesel limits greenhouse gas effects due to the closed CO2 cycle. Biodiesel has more biodegradability, lower combustion emissions such as CO, SOx, HC, PM and lower toxicity than petro diesel. However, biodiesel has high production cost due to high price of plant oils as raw material. So, the utilization of waste cooking oils (WCOs) as feedstock, due to their low price and disposal problems reduce biodiesel production cost. In this study, production of biodiesel by transesterification of methanol and WCO using modified sodic potassic (SP) clinoptilolite zeolite and sodic potassic calcic (SPC) clinoptilolite zeolite as heterogeneous catalysts have been investigated. These natural clinoptilolite zeolites were modified by KOH solution to increase the site activity. The optimum biodiesel yields for SP clinoptilolite and SPC clinoptilolite were 95.8% and 94.8%, respectively. Produced biodiesel were analyzed and compared with petro diesel and ASTM limits. The properties of produced biodiesel confirm well with ASTM limits. The density, kinematic viscosity, cetane index, flash point, cloud point, and pour point of produced biodiesel were all higher than petro diesel but its acid value was lower than petro diesel. Finally, the reusability and regeneration of catalysts were investigated. The results indicated that the spent zeolites cannot be reused directly for the transesterification, but they can be regenerated easily and can obtain high activity.Keywords: Biodiesel, renewable fuel, transesterification, waste cooking oil.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14817689 Social Work Practice to Labour Welfare: A Proposed Model of Field Work Practicum and Role of Social Worker in India
Authors: Naeem Ahmed
Abstract:
Social work is a professional activity based on the approach of “helping people to help themselves” (Stroup). Social work education and practice both are based on humanitarian philosophy in which social workers try to increase the happiness of the society and to reduce the problems of society. Labour welfare is a specialised field of social work which especially focuses on welfare of organised and unorganised labour. In India labour is facing numerous problems in both organised and unorganised sectors because of ignorance, illiteracy, high rate of unemployment etc. In most of the Indian social work institutions we have this specialization with different names like Human Resource Management or Industrial Relation and Personnel Management or Industrial Relations and Labour Welfare or Industrial Social Work etc. Field work practice is integrated part of social work education curriculum in all specialised field. In India we have different field work practice models being followed in different institutions. The main objective of this paper is to prepare a universal field work practicum model in the field of labour welfare. This paper is exploratory in nature, researcher used personal experience and secondary data (model of field work practice in different institutions like Aligarh Muslim University, Pondicherry University, Central University of Karnataka, University of Lucknow, MJP Rohilkhand University Bareilly etc.) Researcher found that there is an immediate need to upgrade the curriculum or field work practice in this particular field, as more than 40 percent of total population engaged in either unorganised or organised sector (NSSO 2011-12) and they are not aware about their rights. In this way a social worker can play an important role in existing labour welfare facilities by making them aware.
Keywords: Fieldwork, labour welfare, organised labour, social work practice, unorganised labour.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28667688 Reliability-Based Maintenance Management Methodology to Minimise Life Cycle Cost of Water Supply Networks
Authors: Mojtaba Mahmoodian, Joshua Phelan, Mehdi Shahparvari
Abstract:
With a large percentage of countries’ total infrastructure expenditure attributed to water network maintenance, it is essential to optimise maintenance strategies to rehabilitate or replace underground pipes before failure occurs. The aim of this paper is to provide water utility managers with a maintenance management approach for underground water pipes, subject to external loading and material corrosion, to give the lowest life cycle cost over a predetermined time period. This reliability-based maintenance management methodology details the optimal years for intervention, the ideal number of maintenance activities to perform before replacement and specifies feasible renewal options and intervention prioritisation to minimise the life cycle cost. The study was then extended to include feasible renewal methods by determining the structural condition index and potential for soil loss, then obtaining the failure impact rating to assist in prioritising pipe replacement. A case study on optimisation of maintenance plans for the Melbourne water pipe network is considered in this paper to evaluate the practicality of the proposed methodology. The results confirm that the suggested methodology can provide water utility managers with a reliable systematic approach to determining optimum maintenance plans for pipe networks.Keywords: Water pipe networks, maintenance management, reliability analysis, optimum maintenance plan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12567687 Optimization of Technical and Technological Solutions for the Development of Offshore Hydrocarbon Fields in the Kaliningrad Region
Authors: Pavel Shcherban, Viktoria Ivanova, Alexander Neprokin, Vladislav Golovanov
Abstract:
Currently, LLC «Lukoil-Kaliningradmorneft» is implementing a comprehensive program for the development of offshore fields of the Kaliningrad region. This is largely associated with the depletion of the resource base of land in the region, as well as the positive results of geological investigation surrounding the Baltic Sea area and the data on the volume of hydrocarbon recovery from a single offshore field are working on the Kaliningrad region – D-6 «Kravtsovskoye».The article analyzes the main stages of the LLC «Lukoil-Kaliningradmorneft»’s development program for the development of the hydrocarbon resources of the region's shelf and suggests an optimization algorithm that allows managing a multi-criteria process of development of shelf deposits. The algorithm is formed on the basis of the problem of sequential decision making, which is a section of dynamic programming. Application of the algorithm during the consolidation of the initial data, the elaboration of project documentation, the further exploration and development of offshore fields will allow to optimize the complex of technical and technological solutions and increase the economic efficiency of the field development project implemented by LLC «Lukoil-Kaliningradmorneft».
Keywords: Offshore fields of hydrocarbons of the Baltic Sea, Development of offshore oil and gas fields, Optimization of the field development scheme, Solution of multi-criteria tasks in the oil and gas complex, Quality management of technical and technological processes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8577686 Development of an Avionics System for Flight Data Collection of an UAV Helicopter
Authors: Nikhil Ramaswamy, S.N.Omkar, Kashyap.H.Nathwani, Anil.M.Vanjare
Abstract:
In this present work, the development of an avionics system for flight data collection of a Raptor 30 V2 is carried out. For the data acquisition both onground and onboard avionics systems are developed for testing of a small-scale Unmanned Aerial Vehicle (UAV) helicopter. The onboard avionics record the helicopter state outputs namely accelerations, angular rates and Euler angles, in real time, and the on ground avionics system record the inputs given to the radio controlled helicopter through a transmitter, in real time. The avionic systems are designed and developed taking into consideration low weight, small size, anti-vibration, low power consumption, and easy interfacing. To mitigate the medium frequency vibrations embedded on the UAV helicopter during flight, a damper is designed and its performance is evaluated. A number of flight tests are carried out and the data obtained is then analyzed for accuracy and repeatability and conclusions are inferred.Keywords: Data collection, Flight Testing, Onground and Onboard Avionics, UAV helicopter
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26427685 Equilibrium Modeling of Carbon Dioxide Adsorption on Zeolites
Authors: Alireza Behvandi, Somayeh Tourani
Abstract:
High pressure adsorption of carbon dioxide on zeolite 13X was investigated in the pressure range (0 to 4) Mpa and temperatures 298, 308 and 323K. The data fitting is accomplished with the Toth, UNILAN, Dubinin-Astakhov and virial adsorption models which are generally used for micro porous adsorbents such as zeolites. Comparison with experimental data from the literature indicated that the virial model would best determine results. These results may be partly attributed to the flexibility of the virial model which can accommodate as many constants as the data warrants.Keywords: adsorption models, zeolite, carbon dioxide
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28847684 Application of Java-based Pointcuts in Aspect Oriented Programming (AOP) for Data Race Detection
Authors: Sadaf Khalid, Fahim Arif
Abstract:
Wide applicability of concurrent programming practices in developing various software applications leads to different concurrency errors amongst which data race is the most important. Java provides greatest support for concurrent programming by introducing various concurrency packages. Aspect oriented programming (AOP) is modern programming paradigm facilitating the runtime interception of events of interest and can be effectively used to handle the concurrency problems. AspectJ being an aspect oriented extension to java facilitates the application of concepts of AOP for data race detection. Volatile variables are usually considered thread safe, but they can become the possible candidates of data races if non-atomic operations are performed concurrently upon them. Various data race detection algorithms have been proposed in the past but this issue of volatility and atomicity is still unaddressed. The aim of this research is to propose some suggestions for incorporating certain conditions for data race detection in java programs at the volatile fields by taking into account support for atomicity in java concurrency packages and making use of pointcuts. Two simple test programs will demonstrate the results of research. The results are verified on two different Java Development Kits (JDKs) for the purpose of comparison.Keywords: Aspect Bench Compiler (abc), Aspect OrientedProgramming (AOP), AspectJ, Aspects, Concurrency packages, Concurrent programming, Cross-cutting Concerns, Data race, Eclipse, Java, Java Development Kits (JDKs), Pointcuts
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19317683 Dissecting Big Trajectory Data to Analyse Road Network Travel Efficiency
Authors: Rania Alshikhe, Vinita Jindal
Abstract:
Digital innovation has played a crucial role in managing smart transportation. For this, big trajectory data collected from trav-eling vehicles, such as taxis through installed global positioning sys-tem (GPS)-enabled devices can be utilized. It offers an unprecedented opportunity to trace the movements of vehicles in fine spatiotemporal granularity. This paper aims to explore big trajectory data to measure the travel efficiency of road networks using the proposed statistical travel efficiency measure (STEM) across an entire city. Further, it identifies the cause of low travel efficiency by proposed least square approximation network-based causality exploration (LANCE). Finally, the resulting data analysis reveals the causes of low travel efficiency, along with the road segments that need to be optimized to improve the traffic conditions and thus minimize the average travel time from given point A to point B in the road network. Obtained results show that our proposed approach outperforms the baseline algorithms for measuring the travel efficiency of the road network.
Keywords: GPS trajectory, road network, taxi trips, digital map, big data, STEM, LANCE
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5107682 Actionable Rules: Issues and New Directions
Authors: Harleen Kaur
Abstract:
Knowledge Discovery in Databases (KDD) is the process of extracting previously unknown, hidden and interesting patterns from a huge amount of data stored in databases. Data mining is a stage of the KDD process that aims at selecting and applying a particular data mining algorithm to extract an interesting and useful knowledge. It is highly expected that data mining methods will find interesting patterns according to some measures, from databases. It is of vital importance to define good measures of interestingness that would allow the system to discover only the useful patterns. Measures of interestingness are divided into objective and subjective measures. Objective measures are those that depend only on the structure of a pattern and which can be quantified by using statistical methods. While, subjective measures depend only on the subjectivity and understandability of the user who examine the patterns. These subjective measures are further divided into actionable, unexpected and novel. The key issues that faces data mining community is how to make actions on the basis of discovered knowledge. For a pattern to be actionable, the user subjectivity is captured by providing his/her background knowledge about domain. Here, we consider the actionability of the discovered knowledge as a measure of interestingness and raise important issues which need to be addressed to discover actionable knowledge.
Keywords: Data Mining Community, Knowledge Discovery inDatabases (KDD), Interestingness, Subjective Measures, Actionability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19427681 Model Discovery and Validation for the Qsar Problem using Association Rule Mining
Authors: Luminita Dumitriu, Cristina Segal, Marian Craciun, Adina Cocu, Lucian P. Georgescu
Abstract:
There are several approaches in trying to solve the Quantitative 1Structure-Activity Relationship (QSAR) problem. These approaches are based either on statistical methods or on predictive data mining. Among the statistical methods, one should consider regression analysis, pattern recognition (such as cluster analysis, factor analysis and principal components analysis) or partial least squares. Predictive data mining techniques use either neural networks, or genetic programming, or neuro-fuzzy knowledge. These approaches have a low explanatory capability or non at all. This paper attempts to establish a new approach in solving QSAR problems using descriptive data mining. This way, the relationship between the chemical properties and the activity of a substance would be comprehensibly modeled.Keywords: association rules, classification, data mining, Quantitative Structure - Activity Relationship.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17887680 Managing of Work Risk in Small and Medium-Size Companies
Authors: Janusz K. Grabara, Bartłomiej Okwiet, Sebastian Kot
Abstract:
The purpose of the article is presentation and analysis of the aspect of job security in small and medium-size enterprises in Poland with reference to other EU countries. We show the theoretical aspects of the risk with reference to managing small and medium enterprises, next risk management in small and medium enterprises in Poland, which were subjected to a detailed analysis. We show in detail the risk associated with the operation of the mentioned above companies, as well as analyses its levels on various stages and for different kinds of conducted activity.Keywords: Job safety, small and medium-size companies, SME, work risk, risk management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13097679 Life Cycle-Based Analysis of Meat Production: Ecosystem Impacts
Authors: Michelle Zeyuan Ma, Hermann Heilmeier
Abstract:
Recently, meat production ecosystem impacts initiated many hot discussions and researchers, and it is a difficult implementation to reduce such impacts due to the demand of meat products. It calls for better management and control of ecosystem impacts from every aspects of meat production. This article analyzes the ecosystem impacts of meat production based on meat products life cycle. The analysis shows that considerable ecosystem impacts are caused by different meat production steps: initial establishment phase, animal raising, slaughterhouse processing, meat consumption, and wastes management. Based on this analysis, the impacts are summarized as: leading factor for biodiversity loss; water waste, land use waste and land degradation; greenhouse gases emissions; pollution to air, water, and soil; related major diseases. The article also provides a discussion on a solution-sustainable food system, which could help in reducing ecosystem impacts. The analysis method is based on the life cycle level, it provides a concept of the whole meat industry ecosystem impacts, and the analysis result could be useful to manage or control meat production ecosystem impacts from investor, producer and consumer sides.Keywords: Eutrophication, life cycle based analysis, sustainable food, waste management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12027678 Post Pandemic Mobility Analysis through Indexing and Sharding in MongoDB: Performance Optimization and Insights
Authors: Karan Vishavjit, Aakash Lakra, Shafaq Khan
Abstract:
The COVID-19 pandemic has pushed healthcare professionals to use big data analytics as a vital tool for tracking and evaluating the effects of contagious viruses. To effectively analyse huge datasets, efficient NoSQL databases are needed. The analysis of post-COVID-19 health and well-being outcomes and the evaluation of the effectiveness of government efforts during the pandemic is made possible by this research’s integration of several datasets, which cuts down on query processing time and creates predictive visual artifacts. We recommend applying sharding and indexing technologies to improve query effectiveness and scalability as the dataset expands. Effective data retrieval and analysis are made possible by spreading the datasets into a sharded database and doing indexing on individual shards. Analysis of connections between governmental activities, poverty levels, and post-pandemic wellbeing is the key goal. We want to evaluate the effectiveness of governmental initiatives to improve health and lower poverty levels. We will do this by utilising advanced data analysis and visualisations. The findings provide relevant data that support the advancement of UN sustainable objectives, future pandemic preparation, and evidence-based decision-making. This study shows how Big Data and NoSQL databases may be used to address problems with global health.
Keywords: COVID-19, big data, data analysis, indexing, NoSQL, sharding, scalability, poverty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 69