Search results for: big data ecosystem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25188

Search results for: big data ecosystem

24588 Internal Concept of Integrated Health by Agrarian Society in Malagasy Highlands for the Last Century

Authors: O. R. Razanakoto, L. Temple

Abstract:

Living in a least developed country, the Malagasy society has a weak capacity to internalize progress, including health concerns. Since the arrival in the fifteenth century of Arabic script, called Sorabe, that was mainly dedicated to the aristocracy, until the colonial era beginning at the end of the nineteenth century and that has popularized the current usual script of the occidental civilization, the upcoming manuscripts that deal with apparent scientific or at least academic issue have been slowly established. So that, the Malagasy communities’ way of life is not well documented yet to allow a precise understanding of the major concerns, reason, and purpose of the existence of the farmers that compose them. A question arises, according to literature, how does Malagasy community that is dominated by agrarian society conceive the conservation of its wellbeing? This study aims to emphasize the scope and the limits of the « One Health » concept or of the Health Integrated Approach (HIA) that evolves at global scale, with regard to the specific context of local Malagasy smallholder farms. It is expected to identify how this society represents linked risks and the mechanisms between human health, animal health, plant health, and ecosystem health within the last 100 years. To do so, the framework to conduct systematic review for agricultural research has been deployed to access available literature. This task has been coupled with the reading of articles that are not indexed by online scientific search engine but that mention part of a history of agriculture and of farmers in Madagascar. This literature review has informed the interactions between human illnesses and those affecting animals and plants (breeded or wild) with any unexpected event (ecological or economic) that has modified the equilibrium of the ecosystem, or that has disturbed the livelihoods of agrarian communities. Besides, drivers that may either accentuate or attenuate the devasting effects of these illnesses and changes were revealed. The study has established that the reasons of human worries are not only physiological. Among the factors that regulate global health, food system and contemporary medicine have helped to the improvement of life expectancy from 55 to 63 years in Madagascar during the last 50 years. However, threats to global health are still occurring. New human or animal illnesses and livestock / plant pathology or enemies may also appear, whereas ancient illnesses that are supposed to have disappeared may be back. This study has highlighted how much important are the risks associated to the impact of unmanaged externalities that weaken community’s life. Many risks, and also solutions, come from abroad and have long term effects even though those happen as punctual event. Thus, a constructivist strategy is suggested to the « One Health » global concept throughout the record of local facts. This approach should facilitate the exploration of methodological pathways and the identification of relevant indicators for research related to HIA.

Keywords: agrarian system, health integrated approach, history, madagascar, resilience, risk

Procedia PDF Downloads 99
24587 Influence of Esports Marketing Strategies on Consumer Behavior: A Case Study of Valorant

Authors: Alex Arghya Adhikari

Abstract:

Gaming and esports industry is one of the biggest and fastest growing industries in the world. Globally people have started investing more in this industry since now people believe just like traditional sports, esports can also sustain their future. Last year in the month of December, the Indian government also recognised esports as an official sport but there has not been any positive initiative by the government in encouraging people to enter esports. This is a problem which cannot be overlooked since we are already in the digital age and gaming and esports is the future industry. There is a need for multiple effective marketing strategies by the game publishers to stabilize the esports in the country. Purpose: To observe the marketing-communication strategies that are implemented by Riot Games’ Valorant and how those strategies influence the consumer behavior and the esports of the game. Methodology: Activities over the internet related to the game like livestreams, discord chats, Instagram posts and YouTube videos over a period of two months have been collected through the Digital Ethnography. To support and validate the observations of the data collected, in-depth online interviews have been conducted which includes streamers, journalists, LAN experienced players and casual players. Findings: The game publisher through its Dynamic Competitive Gaming Experience and Community-Engaged Ecosystem succeeded in making the game a Recreational activity and a Community which goes beyond the In-game experiences which helped in understanding the impact of audience engagement on esports and the loopholes and setbacks of Indian esports. Conclusion: The study provides a comprehensive analysis of how Valorant's successful marketing and community engagement strategies have contributed to its global popularity and competitive esports environment. It highlights the various strategies employed by Riot Games to keep players engaged and connected, and also the challenges in the Indian esports landscape which differentiates it from the global competition.

Keywords: esports, valorant, marketing, consumer behaviour

Procedia PDF Downloads 51
24586 Digital Female Entrepreneurs in South Africa: Drivers and Relationship to Economic Development

Authors: C. van den Berg, C. Pokpas

Abstract:

Popular discourse touts entrepreneurship as a universal solution for underdevelopment, unemployment, and poverty. Moreover, claims are made that women and other disadvantaged groups can achieve material and personal success through digital entrepreneurship. This paper examines the potential of digital technology in entrepreneurial ventures to stimulate economic growth for marginalized groups and communities. Although digital entrepreneurship is hailed as a means to empower under-resourced and socially marginalized people, these opportunities still exist within the confines of existing social and cultural practices. The perspectives of female digital entrepreneurs in developing countries are sorely understudied, particularly concerning an understanding of the complex underlying socio-cultural factors impeding women’s entrepreneurial behaviors. This qualitative study, guided by a feminist phenomenological perspective, focused on the experiences of digital female entrepreneurs in the Western Cape of South Africa. Data were collected via semi-structured interviews and analyzed through the interpretative phenomenological analysis (IPA) approach to determine the relationships between digital entrepreneurship and structural and agential enabling conditions. Findings show that digital entrepreneurship is not a panacea for economic growth in marginalized groups and communities and highlight the importance of addressing socio-cultural gender inequality to enable successful entrepreneurial activity. The paper concludes with recommendations for specialized training initiatives aimed at female entrepreneurs that address internalized constraints and barriers that keep women subservient and measures to shift gender and power beliefs. The outcome will benefit the stimulation of gender-specific public policies to develop a successful digital start-up ecosystem further.

Keywords: digital innovation, female digital entrepreneurs, feminist phenomenology, gender, marginalised communities

Procedia PDF Downloads 119
24585 Single-Cell Visualization with Minimum Volume Embedding

Authors: Zhenqiu Liu

Abstract:

Visualizing the heterogeneity within cell-populations for single-cell RNA-seq data is crucial for studying the functional diversity of a cell. However, because of the high level of noises, outlier, and dropouts, it is very challenging to measure the cell-to-cell similarity (distance), visualize and cluster the data in a low-dimension. Minimum volume embedding (MVE) projects the data into a lower-dimensional space and is a promising tool for data visualization. However, it is computationally inefficient to solve a semi-definite programming (SDP) when the sample size is large. Therefore, it is not applicable to single-cell RNA-seq data with thousands of samples. In this paper, we develop an efficient algorithm with an accelerated proximal gradient method and visualize the single-cell RNA-seq data efficiently. We demonstrate that the proposed approach separates known subpopulations more accurately in single-cell data sets than other existing dimension reduction methods.

Keywords: single-cell RNA-seq, minimum volume embedding, visualization, accelerated proximal gradient method

Procedia PDF Downloads 217
24584 Cloud Data Security Using Map/Reduce Implementation of Secret Sharing Schemes

Authors: Sara Ibn El Ahrache, Tajje-eddine Rachidi, Hassan Badir, Abderrahmane Sbihi

Abstract:

Recently, there has been increasing confidence for a favorable usage of big data drawn out from the huge amount of information deposited in a cloud computing system. Data kept on such systems can be retrieved through the network at the user’s convenience. However, the data that users send include private information, and therefore, information leakage from these data is now a major social problem. The usage of secret sharing schemes for cloud computing have lately been approved to be relevant in which users deal out their data to several servers. Notably, in a (k,n) threshold scheme, data security is assured if and only if all through the whole life of the secret the opponent cannot compromise more than k of the n servers. In fact, a number of secret sharing algorithms have been suggested to deal with these security issues. In this paper, we present a Mapreduce implementation of Shamir’s secret sharing scheme to increase its performance and to achieve optimal security for cloud data. Different tests were run and through it has been demonstrated the contributions of the proposed approach. These contributions are quite considerable in terms of both security and performance.

Keywords: cloud computing, data security, Mapreduce, Shamir's secret sharing

Procedia PDF Downloads 291
24583 Willingness to Pay for Environmental Conservation and Management of Nogas Island and Its Surrounding Waters Among the Residents of Anini-Y, Antique

Authors: Nichole Patricia Pedrina, Karl Jasper Sumande, Alice Joan Ferrer

Abstract:

Nogas Island situated in the municipality of Anini-y in the province of Antique is endowed with natural resources especially a thriving marine ecosystem that attracts tourists all year round. But despite its beauty and emerging popularity, the island and its surrounding waters remain vulnerable to degradation brought about by anthropocentric activities. An emphasis on the protection and conservation is paramount in order to ensure environmental sustainability over time. This study was conducted in order to determine the willingness-to-pay (WTP) of the local residents of Anini-y, Antique for the conservation of Nogas Island and its surrounding waters. The Contingent Valuation Method (CVM) was used to determine the WTP of the study participants. In addition, the study also described the socio-demographic and economic characteristics, the level of awareness, knowledge and attitude towards the conservation and the reasons for the willingness to pay off the residents for the conservation of the island and its surrounding waters. A pilot-tested interview schedule was used to collect data from 320 randomly selected study participants in 8 barangays in the municipality of Anini-y from January to December 2017. Binary logit regression was conducted in order to identify factors affecting the study participants’ WTP. The results revealed that 54.69 percent of the study participants were willing to pay (with adjustment to the level of certainty) for the conservation program. The sex, monthly household income, randomly assigned bid price and the knowledge index were the variables that affected the willingness-to-pay of the study participants for both with and without adjustment to the level of certainty. The monthly mean WTP of the study participants with and without adjustment to the level of certainty were P115 and P104.5, respectively. This study can serve as a guide for the municipality of Anini-y in creating a policy or program that aims to conserve and protect Nogas Island and its surrounding waters.

Keywords: economic valuation, environmental conservation, total economic value, willingness to pay

Procedia PDF Downloads 207
24582 A Modular Framework for Enabling Analysis for Educators with Different Levels of Data Mining Skills

Authors: Kyle De Freitas, Margaret Bernard

Abstract:

Enabling data mining analysis among a wider audience of educators is an active area of research within the educational data mining (EDM) community. The paper proposes a framework for developing an environment that caters for educators who have little technical data mining skills as well as for more advanced users with some data mining expertise. This framework architecture was developed through the review of the strengths and weaknesses of existing models in the literature. The proposed framework provides a modular architecture for future researchers to focus on the development of specific areas within the EDM process. Finally, the paper also highlights a strategy of enabling analysis through either the use of predefined questions or a guided data mining process and highlights how the developed questions and analysis conducted can be reused and extended over time.

Keywords: educational data mining, learning management system, learning analytics, EDM framework

Procedia PDF Downloads 313
24581 Using Audit Tools to Maintain Data Quality for ACC/NCDR PCI Registry Abstraction

Authors: Vikrum Malhotra, Manpreet Kaur, Ayesha Ghotto

Abstract:

Background: Cardiac registries such as ACC Percutaneous Coronary Intervention Registry require high quality data to be abstracted, including data elements such as nuclear cardiology, diagnostic coronary angiography, and PCI. Introduction: The audit tool created is used by data abstractors to provide data audits and assess the accuracy and inter-rater reliability of abstraction performed by the abstractors for a health system. This audit tool solution has been developed across 13 registries, including ACC/NCDR registries, PCI, STS, Get with the Guidelines. Methodology: The data audit tool was used to audit internal registry abstraction for all data elements, including stress test performed, type of stress test, data of stress test, results of stress test, risk/extent of ischemia, diagnostic catheterization detail, and PCI data elements for ACC/NCDR PCI registries. This is being used across 20 hospital systems internally and providing abstraction and audit services for them. Results: The data audit tool had inter-rater reliability and accuracy greater than 95% data accuracy and IRR score for the PCI registry in 50 PCI registry cases in 2021. Conclusion: The tool is being used internally for surgical societies and across hospital systems. The audit tool enables the abstractor to be assessed by an external abstractor and includes all of the data dictionary fields for each registry.

Keywords: abstraction, cardiac registry, cardiovascular registry, registry, data

Procedia PDF Downloads 94
24580 Artificial Intelligence Based Comparative Analysis for Supplier Selection in Multi-Echelon Automotive Supply Chains via GEP and ANN Models

Authors: Seyed Esmail Seyedi Bariran, Laysheng Ewe, Amy Ling

Abstract:

Since supplier selection appears as a vital decision, selecting supplier based on the best and most accurate ways has a lot of importance for enterprises. In this study, a new Artificial Intelligence approach is exerted to remove weaknesses of supplier selection. The paper has three parts. First part is choosing the appropriate criteria for assessing the suppliers’ performance. Next one is collecting the data set based on experts. Afterwards, the data set is divided into two parts, the training data set and the testing data set. By the training data set the best structure of GEP and ANN are selected and to evaluate the power of the mentioned methods the testing data set is used. The result obtained shows that the accuracy of GEP is more than ANN. Moreover, unlike ANN, a mathematical equation is presented by GEP for the supplier selection.

Keywords: supplier selection, automotive supply chains, ANN, GEP

Procedia PDF Downloads 615
24579 Effect of Mangrove Forests in Coastal Flood and Erosion

Authors: Majid Samiee Zenoozian

Abstract:

This paper studies the susceptibility of local settlements in the gulf of Oman mangrove forest zone to flooding and progressesconsiderate of acuities and reactions to historical and present coastal flooding.it is indirect thaterosionsproduced in coastal zones by the change of mangrove undergrowthsubsequent from the enduring influence of persons since the late 19th century. Confronted with the increasing impact of climate change on climate ambitiousalarms such as flooding and biodiversity damage, handling the relationship between mangroves and their atmosphere has become authoritative for their defense. Coastal flood dangers are increasing quickly. We offer high resolution approximations of the financial value of mangroves forests for flood risk discount. We progress a probabilistic, process-based estimate of the properties of mangroves on avoidanceharms to people and property. More significantly, it also establishes how the incessantsqualor of this significant ecosystem has the potential to unfavorably influence the future cyclone persuadeddangers in the area.

Keywords: mangrove forest, coastal, flood, erosion

Procedia PDF Downloads 101
24578 Increasing the Apparent Time Resolution of Tc-99m Diethylenetriamine Pentaacetic Acid Galactosyl Human Serum Albumin Dynamic SPECT by Use of an 180-Degree Interpolation Method

Authors: Yasuyuki Takahashi, Maya Yamashita, Kyoko Saito

Abstract:

In general, dynamic SPECT data acquisition needs a few minutes for one rotation. Thus, the time-activity curve (TAC) derived from the dynamic SPECT is relatively coarse. In order to effectively shorten the interval, between data points, we adopted a 180-degree interpolation method. This method is already used for reconstruction of the X-ray CT data. In this study, we applied this 180-degree interpolation method to SPECT and investigated its effectiveness.To briefly describe the 180-degree interpolation method: the 180-degree data in the second half of one rotation are combined with the 180-degree data in the first half of the next rotation to generate a 360-degree data set appropriate for the time halfway between the first and second rotations. In both a phantom and a patient study, the data points from the interpolated images fell in good agreement with the data points tracking the accumulation of 99mTc activity over time for appropriate region of interest. We conclude that data derived from interpolated images improves the apparent time resolution of dynamic SPECT.

Keywords: dynamic SPECT, time resolution, 180-degree interpolation method, 99mTc-GSA.

Procedia PDF Downloads 487
24577 Treatment of Leaden Sludge of Algiers Refinery by Electrooxidation

Authors: K. Ighilahriz, M. Taleb Ahmed, R. Maachi

Abstract:

Oil industries are responsible for most cases of contamination of our ecosystem by oil and heavy metals. They are toxic and considered carcinogenic and dangerous even when they exist in trace amounts. At Algiers refinery, production, transportation, and refining of crude oil generate considerable waste in storage tanks; these residues result from the gravitational settling. The composition of these residues is essentially a mixture of hydrocarbon and lead. We propose in this work the application of electrooxidation treatment for the leachate of the leaden sludge. The effect of pH, current density and the electrolysis time were studied, the effectiveness of the processes is evaluated by measuring the chemical oxygen demand (COD). The dissolution is the best way to mobilize pollutants from leaden mud, so we conducted leaching before starting the electrochemical treatment. The process was carried out in batch mode using graphite anode and a stainless steel cathode. The results clearly demonstrate the compatibility of the technique used with the type of pollution studied. In fact, it allowed COD removal about 80%.

Keywords: electrooxidation, leaching, leaden sludge, oil industry

Procedia PDF Downloads 216
24576 Genetic Data of Deceased People: Solving the Gordian Knot

Authors: Inigo de Miguel Beriain

Abstract:

Genetic data of deceased persons are of great interest for both biomedical research and clinical use. This is due to several reasons. On the one hand, many of our diseases have a genetic component; on the other hand, we share genes with a good part of our biological family. Therefore, it would be possible to improve our response considerably to these pathologies if we could use these data. Unfortunately, at the present moment, the status of data on the deceased is far from being satisfactorily resolved by the EU data protection regulation. Indeed, the General Data Protection Regulation has explicitly excluded these data from the category of personal data. This decision has given rise to a fragmented legal framework on this issue. Consequently, each EU member state offers very different solutions. For instance, Denmark considers the data as personal data of the deceased person for a set period of time while some others, such as Spain, do not consider this data as such, but have introduced some specifically focused regulations on this type of data and their access by relatives. This is an extremely dysfunctional scenario from multiple angles, not least of which is scientific cooperation at the EU level. This contribution attempts to outline a solution to this dilemma through an alternative proposal. Its main hypothesis is that, in reality, health data are, in a sense, a rara avis within data in general because they do not refer to one person but to several. Hence, it is possible to think that all of them can be considered data subjects (although not all of them can exercise the corresponding rights in the same way). When the person from whom the data were obtained dies, the data remain as personal data of his or her biological relatives. Hence, the general regime provided for in the GDPR may apply to them. As these are personal data, we could go back to thinking in terms of a general prohibition of data processing, with the exceptions provided for in Article 9.2 and on the legal bases included in Article 6. This may be complicated in practice, given that, since we are dealing with data that refer to several data subjects, it may be complex to refer to some of these bases, such as consent. Furthermore, there are theoretical arguments that may oppose this hypothesis. In this contribution, it is shown, however, that none of these objections is of sufficient substance to delegitimize the argument exposed. Therefore, the conclusion of this contribution is that we can indeed build a general framework on the processing of personal data of deceased persons in the context of the GDPR. This would constitute a considerable improvement over the current regulatory framework, although it is true that some clarifications will be necessary for its practical application.

Keywords: collective data conceptual issues, data from deceased people, genetic data protection issues, GDPR and deceased people

Procedia PDF Downloads 141
24575 Steps towards the Development of National Health Data Standards in Developing Countries

Authors: Abdullah I. Alkraiji, Thomas W. Jackson, Ian Murray

Abstract:

The proliferation of health data standards today is somewhat overlapping and conflicting, resulting in market confusion and leading to increasing proprietary interests. The government role and support in standardization for health data are thought to be crucial in order to establish credible standards for the next decade, to maximize interoperability across the health sector, and to decrease the risks associated with the implementation of non-standard systems. The normative literature missed out the exploration of the different steps required to be undertaken by the government towards the development of national health data standards. Based on the lessons learned from a qualitative study investigating the different issues to the adoption of health data standards in the major tertiary hospitals in Saudi Arabia and the opinions and feedback from different experts in the areas of data exchange and standards and medical informatics in Saudi Arabia and UK, a list of steps required towards the development of national health data standards was constructed. Main steps are the existence of: a national formal reference for health data standards, an agreed national strategic direction for medical data exchange, a national medical information management plan and a national accreditation body, and more important is the change management at the national and organizational level. The outcome of this study can be used by academics and practitioners to develop the planning of health data standards, and in particular those in developing countries.

Keywords: interoperabilty, medical data exchange, health data standards, case study, Saudi Arabia

Procedia PDF Downloads 327
24574 A Proposal for U-City (Smart City) Service Method Using Real-Time Digital Map

Authors: SangWon Han, MuWook Pyeon, Sujung Moon, DaeKyo Seo

Abstract:

Recently, technologies based on three-dimensional (3D) space information are being developed and quality of life is improving as a result. Research on real-time digital map (RDM) is being conducted now to provide 3D space information. RDM is a service that creates and supplies 3D space information in real time based on location/shape detection. Research subjects on RDM include the construction of 3D space information with matching image data, complementing the weaknesses of image acquisition using multi-source data, and data collection methods using big data. Using RDM will be effective for space analysis using 3D space information in a U-City and for other space information utilization technologies.

Keywords: RDM, multi-source data, big data, U-City

Procedia PDF Downloads 419
24573 Agile Methodology for Modeling and Design of Data Warehouses -AM4DW-

Authors: Nieto Bernal Wilson, Carmona Suarez Edgar

Abstract:

The organizations have structured and unstructured information in different formats, sources, and systems. Part of these come from ERP under OLTP processing that support the information system, however these organizations in OLAP processing level, presented some deficiencies, part of this problematic lies in that does not exist interesting into extract knowledge from their data sources, as also the absence of operational capabilities to tackle with these kind of projects.  Data Warehouse and its applications are considered as non-proprietary tools, which are of great interest to business intelligence, since they are repositories basis for creating models or patterns (behavior of customers, suppliers, products, social networks and genomics) and facilitate corporate decision making and research. The following paper present a structured methodology, simple, inspired from the agile development models as Scrum, XP and AUP. Also the models object relational, spatial data models, and the base line of data modeling under UML and Big data, from this way sought to deliver an agile methodology for the developing of data warehouses, simple and of easy application. The methodology naturally take into account the application of process for the respectively information analysis, visualization and data mining, particularly for patterns generation and derived models from the objects facts structured.

Keywords: data warehouse, model data, big data, object fact, object relational fact, process developed data warehouse

Procedia PDF Downloads 392
24572 Identifying Model to Predict Deterioration of Water Mains Using Robust Analysis

Authors: Go Bong Choi, Shin Je Lee, Sung Jin Yoo, Gibaek Lee, Jong Min Lee

Abstract:

In South Korea, it is difficult to obtain data for statistical pipe assessment. In this paper, to address these issues, we find that various statistical model presented before is how data mixed with noise and are whether apply in South Korea. Three major type of model is studied and if data is presented in the paper, we add noise to data, which affects how model response changes. Moreover, we generate data from model in paper and analyse effect of noise. From this we can find robustness and applicability in Korea of each model.

Keywords: proportional hazard model, survival model, water main deterioration, ecological sciences

Procedia PDF Downloads 730
24571 Automated Testing to Detect Instance Data Loss in Android Applications

Authors: Anusha Konduru, Zhiyong Shan, Preethi Santhanam, Vinod Namboodiri, Rajiv Bagai

Abstract:

Mobile applications are increasing in a significant amount, each to address the requirements of many users. However, the quick developments and enhancements are resulting in many underlying defects. Android apps create and handle a large variety of 'instance' data that has to persist across runs, such as the current navigation route, workout results, antivirus settings, or game state. Due to the nature of Android, an app can be paused, sent into the background, or killed at any time. If the instance data is not saved and restored between runs, in addition to data loss, partially-saved or corrupted data can crash the app upon resume or restart. However, it is difficult for the programmer to manually test this issue for all the activities. This results in the issue of data loss that the data entered by the user are not saved when there is any interruption. This issue can degrade user experience because the user needs to reenter the information each time there is an interruption. Automated testing to detect such data loss is important to improve the user experience. This research proposes a tool, DroidDL, a data loss detector for Android, which detects the instance data loss from a given android application. We have tested 395 applications and found 12 applications with the issue of data loss. This approach is proved highly accurate and reliable to find the apps with this defect, which can be used by android developers to avoid such errors.

Keywords: Android, automated testing, activity, data loss

Procedia PDF Downloads 224
24570 Near Bottom Concentrations of Krill in Two Arctic Fjords, Spitsbergen

Authors: Kajetan Deja, Katarzyna Draganska-Deja, Mateusz Ormanczyk, Michał Procajlo

Abstract:

Two glaciated fjords on Spitsbergen (Hornsund 77°N) and Kongsfjorden (79°N) were studied for the occurrence of macroplankton (mostly euphausids, hyperiids, chaetognaths) with the use of drop down the camera. The underwater imagery demonstrates that closer to the glacier front, where turbid and freshwater occurs, most of the macroplankters leave the upper water column and descends to the bottom (about 100m depth). Concentrations of macroplankton in the immediate vicinity of the sediment reach over 500 specimens per m² - what corresponds to the biomass of 10g C/m³. Such concentrations of macroplankton are of prime interest for fish, seals and other carnivores. Conditions in the near-bottom waters are in many respects better than in the upper water column- better oxygenated, cold, fully saline and transparent waters with rich food deposited on the seabed from the surface (sinking microplankton). We suggest that near bottom occurrence of macroplankton is related to the increase of glacier melt and freshwater discharge intensity.

Keywords: arctic, ecosystem, fjords, Krill

Procedia PDF Downloads 255
24569 Alternative Ways to Measure Impacts of Dam Closure to the Structure of Fish Communities of a Neotropical River

Authors: Ana Carolina Lima, Carlos Sérgio Agostinho, Amadeu M. V. M. Soares, Kieran A. Monaghan

Abstract:

Neotropical freshwaters host some of the most biodiverse ecosystems in the world and are among the most threatened by habitat alterations. The high number of species and lack of basic ecological knowledge provides a major obstacle to understanding the effects of environmental change. We assessed the impact of dam closure on the fish communities of a neotropical river by applying simple descriptions of community organizations: Species Abundance Distribution (SAD) and Abundance Biomass Comparison (ABC) curves. Fish data were collected during three distinct time periods (one year before, one year after and five years after closure), at eight sites located downstream of the dam, in the reservoir and reservoir transition zone and upstream of the regulated flow. Dam closure was associated with changes in the structural and functional organization of fish communities at all sites. Species richness tended to increase immediately after dam closure while evenness decreased. Changes in taxonomic structure were accompanied by a change in the distribution of biomass with the proportionate contribution by smaller individuals significantly increased relative to larger individuals. Five years on, richness had fallen to below pre-closure levels at all sites, while the comparative stability of the transformed habitats was reflected by biomass-abundance distribution patterns that approximated pre-disturbance ratios. Despite initial generality, respective sites demonstrated distinct ecological responses that were related to the environmental characteristics of their transformed habitats. This simplistic analysis provides a sensitive and informative assessment of ecological conditions that highlights the impact to ecosystem process and ecological networks and has particular value in regions where detailed ecological knowledge precludes the application of traditional bioassessment methods.

Keywords: ABC curves, SADs, biodiversity, damming, tropical fish

Procedia PDF Downloads 374
24568 Big Data: Appearance and Disappearance

Authors: James Moir

Abstract:

The mainstay of Big Data is prediction in that it allows practitioners, researchers, and policy analysts to predict trends based upon the analysis of large and varied sources of data. These can range from changing social and political opinions, patterns in crimes, and consumer behaviour. Big Data has therefore shifted the criterion of success in science from causal explanations to predictive modelling and simulation. The 19th-century science sought to capture phenomena and seek to show the appearance of it through causal mechanisms while 20th-century science attempted to save the appearance and relinquish causal explanations. Now 21st-century science in the form of Big Data is concerned with the prediction of appearances and nothing more. However, this pulls social science back in the direction of a more rule- or law-governed reality model of science and away from a consideration of the internal nature of rules in relation to various practices. In effect Big Data offers us no more than a world of surface appearance and in doing so it makes disappear any context-specific conceptual sensitivity.

Keywords: big data, appearance, disappearance, surface, epistemology

Procedia PDF Downloads 402
24567 From Data Processing to Experimental Design and Back Again: A Parameter Identification Problem Based on FRAP Images

Authors: Stepan Papacek, Jiri Jablonsky, Radek Kana, Ctirad Matonoha, Stefan Kindermann

Abstract:

FRAP (Fluorescence Recovery After Photobleaching) is a widely used measurement technique to determine the mobility of fluorescent molecules within living cells. While the experimental setup and protocol for FRAP experiments are usually fixed, data processing part is still under development. In this paper, we formulate and solve the problem of data selection which enhances the processing of FRAP images. We introduce the concept of the irrelevant data set, i.e., the data which are almost not reducing the confidence interval of the estimated parameters and thus could be neglected. Based on sensitivity analysis, we both solve the problem of the optimal data space selection and we find specific conditions for optimizing an important experimental design factor, e.g., the radius of bleach spot. Finally, a theorem announcing less precision of the integrated data approach compared to the full data case is proven; i.e., we claim that the data set represented by the FRAP recovery curve lead to a larger confidence interval compared to the spatio-temporal (full) data.

Keywords: FRAP, inverse problem, parameter identification, sensitivity analysis, optimal experimental design

Procedia PDF Downloads 264
24566 Representation Data without Lost Compression Properties in Time Series: A Review

Authors: Nabilah Filzah Mohd Radzuan, Zalinda Othman, Azuraliza Abu Bakar, Abdul Razak Hamdan

Abstract:

Uncertain data is believed to be an important issue in building up a prediction model. The main objective in the time series uncertainty analysis is to formulate uncertain data in order to gain knowledge and fit low dimensional model prior to a prediction task. This paper discusses the performance of a number of techniques in dealing with uncertain data specifically those which solve uncertain data condition by minimizing the loss of compression properties.

Keywords: compression properties, uncertainty, uncertain time series, mining technique, weather prediction

Procedia PDF Downloads 417
24565 Data Mining As A Tool For Knowledge Management: A Review

Authors: Maram Saleh

Abstract:

Knowledge has become an essential resource in today’s economy and become the most important asset of maintaining competition advantage in organizations. The importance of knowledge has made organizations to manage their knowledge assets and resources through all multiple knowledge management stages such as: Knowledge Creation, knowledge storage, knowledge sharing and knowledge use. Researches on data mining are continues growing over recent years on both business and educational fields. Data mining is one of the most important steps of the knowledge discovery in databases process aiming to extract implicit, unknown but useful knowledge and it is considered as significant subfield in knowledge management. Data miming have the great potential to help organizations to focus on extracting the most important information on their data warehouses. Data mining tools and techniques can predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. This review paper explores the applications of data mining techniques in supporting knowledge management process as an effective knowledge discovery technique. In this paper, we identify the relationship between data mining and knowledge management, and then focus on introducing some application of date mining techniques in knowledge management for some real life domains.

Keywords: Data Mining, Knowledge management, Knowledge discovery, Knowledge creation.

Procedia PDF Downloads 196
24564 Energy Recovery from Swell with a Height Inferior to 1.5 m

Authors: A. Errasti, F. Doffagne, O. Foucrier, S. Kao, A. Meigne, H. Pellae, T. Rouland

Abstract:

Renewable energy recovery is an important domain of research in past few years in view of protection of our ecosystem. Several industrial companies are setting up widespread recovery systems to exploit wave energy. Most of them have a large size, are implanted near the shores and exploit current flows. However, as oceans represent 70% of Earth surface, a huge space is still unexploited to produce energy. Present analysis focuses on surface small scale wave energy recovery. The principle is exactly the opposite of wheel damper for a car on a road. Instead of maintaining the car body as non-oscillatory as possible by adapted control, a system is designed so that its oscillation amplitude under wave action will be maximized with respect to a boat carrying it in view of differential potential energy recuperation. From parametric analysis of system equations, interesting domains have been selected and expected energy output has been evaluated.

Keywords: small scale wave, potential energy, optimized energy recovery, auto-adaptive system

Procedia PDF Downloads 240
24563 Anomaly Detection Based Fuzzy K-Mode Clustering for Categorical Data

Authors: Murat Yazici

Abstract:

Anomalies are irregularities found in data that do not adhere to a well-defined standard of normal behavior. The identification of outliers or anomalies in data has been a subject of study within the statistics field since the 1800s. Over time, a variety of anomaly detection techniques have been developed in several research communities. The cluster analysis can be used to detect anomalies. It is the process of associating data with clusters that are as similar as possible while dissimilar clusters are associated with each other. Many of the traditional cluster algorithms have limitations in dealing with data sets containing categorical properties. To detect anomalies in categorical data, fuzzy clustering approach can be used with its advantages. The fuzzy k-Mode (FKM) clustering algorithm, which is one of the fuzzy clustering approaches, by extension to the k-means algorithm, is reported for clustering datasets with categorical values. It is a form of clustering: each point can be associated with more than one cluster. In this paper, anomaly detection is performed on two simulated data by using the FKM cluster algorithm. As a significance of the study, the FKM cluster algorithm allows to determine anomalies with their abnormality degree in contrast to numerous anomaly detection algorithms. According to the results, the FKM cluster algorithm illustrated good performance in the anomaly detection of data, including both one anomaly and more than one anomaly.

Keywords: fuzzy k-mode clustering, anomaly detection, noise, categorical data

Procedia PDF Downloads 40
24562 Big Data Analytics and Data Security in the Cloud via Fully Homomorphic Encyption Scheme

Authors: Victor Onomza Waziri, John K. Alhassan, Idris Ismaila, Noel Dogonyara

Abstract:

This paper describes the problem of building secure computational services for encrypted information in the Cloud. Computing without decrypting the encrypted data; therefore, it meets the yearning of computational encryption algorithmic aspiration model that could enhance the security of big data for privacy or confidentiality, availability and integrity of the data and user’s security. The cryptographic model applied for the computational process of the encrypted data is the Fully Homomorphic Encryption Scheme. We contribute a theoretical presentations in a high-level computational processes that are based on number theory that is derivable from abstract algebra which can easily be integrated and leveraged in the Cloud computing interface with detail theoretic mathematical concepts to the fully homomorphic encryption models. This contribution enhances the full implementation of big data analytics based on cryptographic security algorithm.

Keywords: big data analytics, security, privacy, bootstrapping, Fully Homomorphic Encryption Scheme

Procedia PDF Downloads 466
24561 An Approximation of Daily Rainfall by Using a Pixel Value Data Approach

Authors: Sarisa Pinkham, Kanyarat Bussaban

Abstract:

The research aims to approximate the amount of daily rainfall by using a pixel value data approach. The daily rainfall maps from the Thailand Meteorological Department in period of time from January to December 2013 were the data used in this study. The results showed that this approach can approximate the amount of daily rainfall with RMSE=3.343.

Keywords: daily rainfall, image processing, approximation, pixel value data

Procedia PDF Downloads 380
24560 Evaluation Of A Start Up Business Strategy In Movie Industry: Case Study Of Visinema

Authors: Stacia E. H. Sitohang, S.Mn., Socrates Rudy Sirait

Abstract:

The first movie theater in Indonesia was established in December 1900. The movie industry started with international movie penetration. After a while, local movie producers started to rise and created local Indonesian movies. The industry is growing through ups and downs in Indonesia. In 2008, Visinema was founded in Jakarta, Indonesia, by AnggaDwimasSasongko, one of the most respected movie director in Indonesia. After getting achievements and recognition, Visinema chose to grow the company horizontally as opposed to only grow vertically and gain another similar achievement. Visinemachose to build the ecosystem that enables them to obtain many more opportunities and generatebusiness sustainability. The company proceed as an agile company. They created several business subsidiaries to support the company’s Intellectual Property (IP) development. This research was done through interview with the key persons in the company and questionnaire to get market insights regarding Visinema. The is able to transform their IP that initially started from movies to different kinds of business model. Interestingly, Angga chose to use the start up approach to create Visinema. In 2019, the company successfully gained Series A funding from Intudo Ventures and got other various investment schemes to support the business. In early 2020, Covid-19 pandemic negatively impacted many industries in Indonesia, especially the entertainment and leisure businesses. Fortunately, Visinema did not face any significant problem regarding survival during the pandemic, there were nolay-offs nor work hour reductions. Instead, they were thinking of much bigger opportunities and problems. While other companies suffer during the pandemic, Visinema created the first focused Transactional Video On Demand (TVOD) in Indonesia named Bioskop Online. The platform was created to keep the company innovating and adapting with the new online market as the result of the Covid-19 pandemic. Other than a digital platform, Visinemainvested heavily in animation to target kids and family business. They believed that penetrating the technology and animation market is going to be the biggest opportunity in Visinema’s road map. Besides huge opportunities, Visinema is also facing problems. The first is company brand positioning. Angga, as the founder, felt the need to detach his name from the brand image of Visinema to create system sustainability and scalability. Second, the company has to create a strategy to refocus in a particular business area to maintain and improve the competitive advantages. The third problem, IP piracy is a huge structural problem in Indonesia, the company considers IP thieves as their biggest competitors as opposed to other production company. As the recommendation, we suggest a set of branding and management strategy to detach the founder’s name from Visinema’s brand and improve the competitive advantages. We also suggest Visinema invest in system building to prevent IP piracy in the entertainment industry, which later can be another business subsidiary of Visinema.

Keywords: business ecosystem, agile, sustainability, scalability, start Up, intellectual property, digital platform

Procedia PDF Downloads 125
24559 The Effect of Measurement Distribution on System Identification and Detection of Behavior of Nonlinearities of Data

Authors: Mohammad Javad Mollakazemi, Farhad Asadi, Aref Ghafouri

Abstract:

In this paper, we considered and applied parametric modeling for some experimental data of dynamical system. In this study, we investigated the different distribution of output measurement from some dynamical systems. Also, with variance processing in experimental data we obtained the region of nonlinearity in experimental data and then identification of output section is applied in different situation and data distribution. Finally, the effect of the spanning the measurement such as variance to identification and limitation of this approach is explained.

Keywords: Gaussian process, nonlinearity distribution, particle filter, system identification

Procedia PDF Downloads 498