Search results for: data anomalies
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24287

Search results for: data anomalies

24107 Prioritization of Mutation Test Generation with Centrality Measure

Authors: Supachai Supmak, Yachai Limpiyakorn

Abstract:

Mutation testing can be applied for the quality assessment of test cases. Prioritization of mutation test generation has been a critical element of the industry practice that would contribute to the evaluation of test cases. The industry generally delivers the product under the condition of time to the market and thus, inevitably sacrifices software testing tasks, even though many test cases are required for software verification. This paper presents an approach of applying a social network centrality measure, PageRank, to prioritize mutation test generation. The source code with the highest values of PageRank will be focused first when developing their test cases as these modules are vulnerable to defects or anomalies which may cause the consequent defects in many other associated modules. Moreover, the approach would help identify the reducible test cases in the test suite, still maintaining the same criteria as the original number of test cases.

Keywords: software testing, mutation test, network centrality measure, test case prioritization

Procedia PDF Downloads 75
24106 A Review on Big Data Movement with Different Approaches

Authors: Nay Myo Sandar

Abstract:

With the growth of technologies and applications, a large amount of data has been producing at increasing rate from various resources such as social media networks, sensor devices, and other information serving devices. This large collection of massive, complex and exponential growth of dataset is called big data. The traditional database systems cannot store and process such data due to large and complexity. Consequently, cloud computing is a potential solution for data storage and processing since it can provide a pool of resources for servers and storage. However, moving large amount of data to and from is a challenging issue since it can encounter a high latency due to large data size. With respect to big data movement problem, this paper reviews the literature of previous works, discusses about research issues, finds out approaches for dealing with big data movement problem.

Keywords: Big Data, Cloud Computing, Big Data Movement, Network Techniques

Procedia PDF Downloads 55
24105 Reworking of the Anomalies in the Discounted Utility Model as a Combination of Cognitive Bias and Decrease in Impatience: Decision Making in Relation to Bounded Rationality and Emotional Factors in Intertemporal Choices

Authors: Roberta Martino, Viviana Ventre

Abstract:

Every day we face choices whose consequences are deferred in time. These types of choices are the intertemporal choices and play an important role in the social, economic, and financial world. The Discounted Utility Model is the mathematical model of reference to calculate the utility of intertemporal prospects. The discount rate is the main element of the model as it describes how the individual perceives the indeterminacy of subsequent periods. Empirical evidence has shown a discrepancy between the behavior expected from the predictions of the model and the effective choices made from the decision makers. In particular, the term temporal inconsistency indicates those choices that do not remain optimal with the passage of time. This phenomenon has been described with hyperbolic models of the discount rate which, unlike the linear or exponential nature assumed by the discounted utility model, is not constant over time. This paper explores the problem of inconsistency by tracing the decision-making process through the concept of impatience. The degree of impatience and the degree of decrease of impatience are two parameters that allow to quantify the weight of emotional factors and cognitive limitations during the evaluation and selection of alternatives. In fact, although the theory assumes perfectly rational decision makers, behavioral finance and cognitive psychology have made it possible to understand that distortions in the decision-making process and emotional influence have an inevitable impact on the decision-making process. The degree to which impatience is diminished is the focus of the first part of the study. By comparing consistent and inconsistent preferences over time, it was possible to verify that some anomalies in the discounted utility model are a result of the combination of cognitive bias and emotional factors. In particular: the delay effect and the interval effect are compared through the concept of misperception of time; starting from psychological considerations, a criterion is proposed to identify the causes of the magnitude effect that considers the differences in outcomes rather than their ratio; the sign effect is analyzed by integrating in the evaluation of prospects with negative outcomes the psychological aspects of loss aversion provided by Prospect Theory. An experiment implemented confirms three findings: the greatest variation in the degree of decrease in impatience corresponds to shorter intervals close to the present; the greatest variation in the degree of impatience occurs for outcomes of lower magnitude; the variation in the degree of impatience is greatest for negative outcomes. The experimental phase was implemented with the construction of the hyperbolic factor through the administration of questionnaires constructed for each anomaly. This work formalizes the underlying causes of the discrepancy between the discounted utility model and the empirical evidence of preference reversal.

Keywords: decreasing impatience, discount utility model, hyperbolic discount, hyperbolic factor, impatience

Procedia PDF Downloads 79
24104 Optimized Approach for Secure Data Sharing in Distributed Database

Authors: Ahmed Mateen, Zhu Qingsheng, Ahmad Bilal

Abstract:

In the current age of technology, information is the most precious asset of a company. Today, companies have a large amount of data. As the data become larger, access to data for some particular information is becoming slower day by day. Faster data processing to shape it in the form of information is the biggest issue. The major problems in distributed databases are the efficiency of data distribution and response time of data distribution. The security of data distribution is also a big issue. For these problems, we proposed a strategy that can maximize the efficiency of data distribution and also increase its response time. This technique gives better results for secure data distribution from multiple heterogeneous sources. The newly proposed technique facilitates the companies for secure data sharing efficiently and quickly.

Keywords: ER-schema, electronic record, P2P framework, API, query formulation

Procedia PDF Downloads 302
24103 Spectral Mapping of Hydrothermal Alteration Minerals for Geothermal Exploration Using Advanced Spaceborne Thermal Emission and Reflection Radiometer Short Wave Infrared Data

Authors: Aliyu J. Abubakar, Mazlan Hashim, Amin B. Pour

Abstract:

Exploiting geothermal resources for either power, home heating, Spa, greenhouses, industrial or tourism requires an initial identification of suitable areas. This can be done cost-effectively using remote sensing satellite imagery which has synoptic capabilities of covering large areas in real time and by identifying possible areas of hydrothermal alteration and minerals related to Geothermal systems. Earth features and minerals are known to have unique diagnostic spectral reflectance characteristics that can be used to discriminate them. The focus of this paper is to investigate the applicability of mapping hydrothermal alteration in relation to geothermal systems (thermal springs) at Yankari Park Northeastern Nigeria, using Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) satellite data for resource exploration. The ASTER Short Wave Infrared (SWIR) bands are used to highlight and discriminate alteration areas by employing sophisticated digital image processing techniques including image transformations and spectral mapping methods. Field verifications are conducted at the Yankari Park using hand held Global Positioning System (GPS) monterra to identify locations of hydrothermal alteration and rock samples obtained at the vicinity and surrounding areas of the ‘Mawulgo’ and ‘Wikki’ thermal springs. X-Ray Diffraction (XRD) results of rock samples obtained from the field validated hydrothermal alteration by the presence of indicator minerals including; Dickite, Kaolinite, Hematite and Quart. The study indicated the applicability of mapping geothermal anomalies for resource exploration in unmapped sparsely vegetated savanna environment characterized by subtle surface manifestations such as thermal springs. The results could have implication for geothermal resource exploration especially at the prefeasibility stages by narrowing targets for comprehensive surveys and in unexplored savanna regions where expensive airborne surveys are unaffordable.

Keywords: geothermal exploration, image enhancement, minerals, spectral mapping

Procedia PDF Downloads 335
24102 Data Mining Algorithms Analysis: Case Study of Price Predictions of Lands

Authors: Julio Albuja, David Zaldumbide

Abstract:

Data analysis is an important step before taking a decision about money. The aim of this work is to analyze the factors that influence the final price of the houses through data mining algorithms. To our best knowledge, previous work was researched just to compare results. Furthermore, before using the data of the data set, the Z-Transformation were used to standardize the data in the same range. Hence, the data was classified into two groups to visualize them in a readability format. A decision tree was built, and graphical data is displayed where clearly is easy to see the results and the factors' influence in these graphics. The definitions of these methods are described, as well as the descriptions of the results. Finally, conclusions and recommendations are presented related to the released results that our research showed making it easier to apply these algorithms using a customized data set.

Keywords: algorithms, data, decision tree, transformation

Procedia PDF Downloads 345
24101 Soil Moisture Regulation in Irrigated Agriculture

Authors: I. Kruashvili, I. Inashvili, K. Bziava, M. Lomishvili

Abstract:

Seepage capillary anomalies in the active layer of soil, related to the soil water movement, often cause variation of soil hydrophysical properties and become one of the main objectives of the hydroecology. It is necessary to mention that all existing equations for computing the seepage flow particularly from soil channels, through dams, bulkheads, and foundations of hydraulic engineering structures are preferable based on the linear seepage law. Regarding the existing beliefs, anomalous seepage is based on postulates according to which the fluid in free volume is characterized by resistance against shear deformation and is presented in the form of initial gradient. According to the above-mentioned information, we have determined: Equation to calculate seepage coefficient when the velocity of transition flow is equal to seepage flow velocity; by means of power function, equations for the calculation of average and maximum velocities of seepage flow have been derived; taking into consideration the fluid continuity condition, average velocity for calculation of average velocity in capillary tube has been received.

Keywords: seepage, soil, velocity, water

Procedia PDF Downloads 433
24100 Harnessing Artificial Intelligence for Early Detection and Management of Infectious Disease Outbreaks

Authors: Amarachukwu B. Isiaka, Vivian N. Anakwenze, Chinyere C. Ezemba, Chiamaka R. Ilodinso, Chikodili G. Anaukwu, Chukwuebuka M. Ezeokoli, Ugonna H. Uzoka

Abstract:

Infectious diseases continue to pose significant threats to global public health, necessitating advanced and timely detection methods for effective outbreak management. This study explores the integration of artificial intelligence (AI) in the early detection and management of infectious disease outbreaks. Leveraging vast datasets from diverse sources, including electronic health records, social media, and environmental monitoring, AI-driven algorithms are employed to analyze patterns and anomalies indicative of potential outbreaks. Machine learning models, trained on historical data and continuously updated with real-time information, contribute to the identification of emerging threats. The implementation of AI extends beyond detection, encompassing predictive analytics for disease spread and severity assessment. Furthermore, the paper discusses the role of AI in predictive modeling, enabling public health officials to anticipate the spread of infectious diseases and allocate resources proactively. Machine learning algorithms can analyze historical data, climatic conditions, and human mobility patterns to predict potential hotspots and optimize intervention strategies. The study evaluates the current landscape of AI applications in infectious disease surveillance and proposes a comprehensive framework for their integration into existing public health infrastructures. The implementation of an AI-driven early detection system requires collaboration between public health agencies, healthcare providers, and technology experts. Ethical considerations, privacy protection, and data security are paramount in developing a framework that balances the benefits of AI with the protection of individual rights. The synergistic collaboration between AI technologies and traditional epidemiological methods is emphasized, highlighting the potential to enhance a nation's ability to detect, respond to, and manage infectious disease outbreaks in a proactive and data-driven manner. The findings of this research underscore the transformative impact of harnessing AI for early detection and management, offering a promising avenue for strengthening the resilience of public health systems in the face of evolving infectious disease challenges. This paper advocates for the integration of artificial intelligence into the existing public health infrastructure for early detection and management of infectious disease outbreaks. The proposed AI-driven system has the potential to revolutionize the way we approach infectious disease surveillance, providing a more proactive and effective response to safeguard public health.

Keywords: artificial intelligence, early detection, disease surveillance, infectious diseases, outbreak management

Procedia PDF Downloads 35
24099 Application of Blockchain Technology in Geological Field

Authors: Mengdi Zhang, Zhenji Gao, Ning Kang, Rongmei Liu

Abstract:

Management and application of geological big data is an important part of China's national big data strategy. With the implementation of a national big data strategy, geological big data management becomes more and more critical. At present, there are still a lot of technology barriers as well as cognition chaos in many aspects of geological big data management and application, such as data sharing, intellectual property protection, and application technology. Therefore, it’s a key task to make better use of new technologies for deeper delving and wider application of geological big data. In this paper, we briefly introduce the basic principle of blockchain technology at the beginning and then make an analysis of the application dilemma of geological data. Based on the current analysis, we bring forward some feasible patterns and scenarios for the blockchain application in geological big data and put forward serval suggestions for future work in geological big data management.

Keywords: blockchain, intellectual property protection, geological data, big data management

Procedia PDF Downloads 54
24098 Experiment of Geophysical Exploration in Egypt

Authors: Ramadan Fayez Zowaid Hussein

Abstract:

Exploration geophysics is an applied branch of geophysics, and it is very important to use such a method in Egypt and not just Egypt but in Africa and the Middle East. This research aims to work deeply on the importance of this method, and this paper focuses more on the benefits of the exploration of geophysics and how to apply it to scientific methods. It helps to discover earthquakes and assist in seismology. It also helps to map the surface structure of a region and also magnetic techniques, including aeromagnetic surveys to map magnetic anomalies. This is known that having a great experience in this field as it was very interesting reading a lot and searching about this matter and this technology, and all was found made this fantastic: as the method is existing and we do not use it. It costs a lot, but one believes that this method is very important; for example, in discovering earthquakes, check the surface of the ground easily; it makes us see the surface of the ground clearly so we can find the elements of the earth easily. In conclusion, geophysical exploration use is very important, and it must be highlighted and considered to be discussed in the Middle East, not just in the Middle East but also in Africa.

Keywords: geophysics, magnetic, gravitational, hydrocarbon exploration

Procedia PDF Downloads 42
24097 Frequent Item Set Mining for Big Data Using MapReduce Framework

Authors: Tamanna Jethava, Rahul Joshi

Abstract:

Frequent Item sets play an essential role in many data Mining tasks that try to find interesting patterns from the database. Typically it refers to a set of items that frequently appear together in transaction dataset. There are several mining algorithm being used for frequent item set mining, yet most do not scale to the type of data we presented with today, so called “BIG DATA”. Big Data is a collection of large data sets. Our approach is to work on the frequent item set mining over the large dataset with scalable and speedy way. Big Data basically works with Map Reduce along with HDFS is used to find out frequent item sets from Big Data on large cluster. This paper focuses on using pre-processing & mining algorithm as hybrid approach for big data over Hadoop platform.

Keywords: frequent item set mining, big data, Hadoop, MapReduce

Procedia PDF Downloads 390
24096 Impact of Insect-Feeding and Fire-Heating Wounding on Wood Properties of Lodgepole Pine

Authors: Estelle Arbellay, Lori D. Daniels, Shawn D. Mansfield, Alice S. Chang

Abstract:

Mountain pine beetle (MPB) outbreaks are currently devastating lodgepole pine forests in western North America, which are also widely disturbed by frequent wildfires. Both MPB and fire can leave scars on lodgepole pine trees, thereby diminishing their commercial value and possibly compromising their utilization in solid wood products. In order to fully exploit the affected resource, it is crucial to understand how wounding from these two disturbance agents impact wood properties. Moreover, previous research on lodgepole pine has focused solely on sound wood and stained wood resulting from the MPB-transmitted blue fungi. By means of a quantitative multi-proxy approach, we tested the hypotheses that (i) wounding (of either MPB or fire origin) caused significant changes in wood properties of lodgepole pine and that (ii) MPB-induced wound effects could differ from those induced by fire in type and magnitude. Pith-to-bark strips were extracted from 30 MPB scars and 30 fire scars. Strips were cut immediately adjacent to the wound margin and encompassed 12 rings from normal wood formed prior to wounding and 12 rings from wound wood formed after wounding. Wood properties evaluated within this 24-year window included ring width, relative wood density, cellulose crystallinity, fibre dimensions, and carbon and nitrogen concentrations. Methods used to measure these proxies at a (sub-)annual resolution included X-ray densitometry, X-ray diffraction, fibre quality analysis, and elemental analysis. Results showed a substantial growth release in wound wood compared to normal wood, as both earlywood and latewood width increased over a decade following wounding. Wound wood was also shown to have a significantly different latewood density than normal wood 4 years after wounding. Latewood density decreased in MPB scars while the opposite was true in fire scars. By contrast, earlywood density was presented only minor variations following wounding. Cellulose crystallinity decreased in wound wood compared to normal wood, being especially diminished in MPB scars the first year after wounding. Fibre dimensions also decreased following wounding. However, carbon and nitrogen concentrations did not substantially differ between wound wood and normal wood. Nevertheless, insect-feeding and fire-heating wounding were shown to significantly alter most wood properties of lodgepole pine, as demonstrated by the existence of several morphological anomalies in wound wood. MPB and fire generally elicited similar anomalies, with the major exception of latewood density. In addition to providing quantitative criteria for differentiating between biotic (MPB) and abiotic (fire) disturbances, this study provides the wood industry with fundamental information on the physiological response of lodgepole pine to wounding in order to evaluate the utilization of scarred trees in solid wood products.

Keywords: elemental analysis, fibre quality analysis, lodgepole pine, wood properties, wounding, X-ray densitometry, X-ray diffraction

Procedia PDF Downloads 295
24095 The Role Of Data Gathering In NGOs

Authors: Hussaini Garba Mohammed

Abstract:

Background/Significance: The lack of data gathering is affecting NGOs world-wide in general to have good data information about educational and health related issues among communities in any country and around the world. For example, HIV/AIDS smoking (Tuberculosis diseases) and COVID-19 virus carriers is becoming a serious public health problem, especially among old men and women. But there is no full details data survey assessment from communities, villages, and rural area in some countries to show the percentage of victims and patients, especial with this world COVID-19 virus among the people. These data are essential to inform programming targets, strategies, and priorities in getting good information about data gathering in any society.

Keywords: reliable information, data assessment, data mining, data communication

Procedia PDF Downloads 156
24094 Dimensionality Reduction in Modal Analysis for Structural Health Monitoring

Authors: Elia Favarelli, Enrico Testi, Andrea Giorgetti

Abstract:

Autonomous structural health monitoring (SHM) of many structures and bridges became a topic of paramount importance for maintenance purposes and safety reasons. This paper proposes a set of machine learning (ML) tools to perform automatic feature selection and detection of anomalies in a bridge from vibrational data and compare different feature extraction schemes to increase the accuracy and reduce the amount of data collected. As a case study, the Z-24 bridge is considered because of the extensive database of accelerometric data in both standard and damaged conditions. The proposed framework starts from the first four fundamental frequencies extracted through operational modal analysis (OMA) and clustering, followed by density-based time-domain filtering (tracking). The fundamental frequencies extracted are then fed to a dimensionality reduction block implemented through two different approaches: feature selection (intelligent multiplexer) that tries to estimate the most reliable frequencies based on the evaluation of some statistical features (i.e., mean value, variance, kurtosis), and feature extraction (auto-associative neural network (ANN)) that combine the fundamental frequencies to extract new damage sensitive features in a low dimensional feature space. Finally, one class classifier (OCC) algorithms perform anomaly detection, trained with standard condition points, and tested with normal and anomaly ones. In particular, a new anomaly detector strategy is proposed, namely one class classifier neural network two (OCCNN2), which exploit the classification capability of standard classifiers in an anomaly detection problem, finding the standard class (the boundary of the features space in normal operating conditions) through a two-step approach: coarse and fine boundary estimation. The coarse estimation uses classics OCC techniques, while the fine estimation is performed through a feedforward neural network (NN) trained that exploits the boundaries estimated in the coarse step. The detection algorithms vare then compared with known methods based on principal component analysis (PCA), kernel principal component analysis (KPCA), and auto-associative neural network (ANN). In many cases, the proposed solution increases the performance with respect to the standard OCC algorithms in terms of F1 score and accuracy. In particular, by evaluating the correct features, the anomaly can be detected with accuracy and an F1 score greater than 96% with the proposed method.

Keywords: anomaly detection, frequencies selection, modal analysis, neural network, sensor network, structural health monitoring, vibration measurement

Procedia PDF Downloads 94
24093 Evaluation of Agricultural Drought Impact in the Crop Productivity of East Gojjam Zone

Authors: Walelgn Dilnesa Cherie, Fasikaw Atanaw Zimale, Bekalu W. Asres

Abstract:

The most catastrophic condition for agricultural production is a drought event, which is also one of the most hydro-metrological-related hazards. According to the combined susceptibility of plants to meteorological and hydrological conditions, agricultural drought is defined as the magnitude, severity, and duration of a drought that affects crop production. The accurate and timely assessment of agricultural drought can lead to the development of risk management strategies, appropriate proactive mechanisms for the protection of farmers, and the improvement of food security. The evaluation of agricultural drought in the East Gojjam zone was the primary subject of this study. To identify the agricultural drought, soil moisture anomalies, soil moisture deficit indices, and Normalized Difference Vegetation Indices (NDVI) are used. The measured welting point, field capacity, and soil moisture were utilized to validate the soil water deficit indices computed from the satellite data. The soil moisture and soil water deficit indices in 2013 in all woredas were minimum; this makes vegetation stress also in all woredas. The soil moisture content decreased in 2013/2014/2019, and 2021 in Dejen, 2014, and 2019 in Awobel Woreda. The max/ min values of NDVI in 2013 are minimum; it dominantly shows vegetation stress and an observed agricultural drought that happened in all woredas. The validation process of satellite and in-situ soil moisture and soil water deficit indices shows a good agreement with a value of R²=0.87 and 0.56, respectively. The study area becomes drought detected region, so government officials, policymakers, and environmentalists pay attention to the protection of drought effects.

Keywords: NDVI, agricultural drought, SWDI, soil moisture

Procedia PDF Downloads 43
24092 The Application of Data Mining Technology in Building Energy Consumption Data Analysis

Authors: Liang Zhao, Jili Zhang, Chongquan Zhong

Abstract:

Energy consumption data, in particular those involving public buildings, are impacted by many factors: the building structure, climate/environmental parameters, construction, system operating condition, and user behavior patterns. Traditional methods for data analysis are insufficient. This paper delves into the data mining technology to determine its application in the analysis of building energy consumption data including energy consumption prediction, fault diagnosis, and optimal operation. Recent literature are reviewed and summarized, the problems faced by data mining technology in the area of energy consumption data analysis are enumerated, and research points for future studies are given.

Keywords: data mining, data analysis, prediction, optimization, building operational performance

Procedia PDF Downloads 818
24091 To Handle Data-Driven Software Development Projects Effectively

Authors: Shahnewaz Khan

Abstract:

Machine learning (ML) techniques are often used in projects for creating data-driven applications. These tasks typically demand additional research and analysis. The proper technique and strategy must be chosen to ensure the success of data-driven projects. Otherwise, even exerting a lot of effort, the necessary development might not always be possible. In this post, an effort to examine the workflow of data-driven software development projects and its implementation process in order to describe how to manage a project successfully. Which will assist in minimizing the added workload.

Keywords: data, data-driven projects, data science, NLP, software project

Procedia PDF Downloads 56
24090 Geological, Geochronological, Geochemical, and Geophysical Characteristics of the Dalli Porphyry Cu-Au Deposit in Central Iran; Implications for Exploration

Authors: Hooshag Asadi Haroni, Maryam Veiskarami, Yongjun Lu

Abstract:

The Dalli gold-rich porphyry deposit (17 Mt @ 0.5% Cu and 0.65 g/t Au) is located in the Urumieh-Dokhtar Magmatic Arc (UDMA), a small segment of the Tethyan metallogenic belt, hosting several porphyry Cu (Mo-Au) systems in Iran. This research characterizes the Dalli deposit to define exploration criteria in advanced exploration such as the drilling of possible blind porphyry centers. Geological map, trench/drill hole geochemical and ground magnetic data, and age dating and isotope trace element analyses, carried out at the John De Laeter Research Center of Curtin University, were used to characterize the Delli deposit. Mineralization at Dalli is hosted by NE-trending quartz-diorite porphyry stocks (~ 200m in diameter) intruded by a wall-rock andesite porphyry. Disseminated and stockwork Cu-Au mineralization is related to potassic alteration, comprising magnetite, late K-feldspar and biotite, and quartz-sericite-specularite overprint, surrounded by extensive barren argillic and propylitic alterations. In the peripheries of the porphyry centers, there are N-trending vuggy quartz veins, hosting epithermal Au-Ag-As-Sb mineralization. Geochemical analyses of drill core samples showed that the core of the porphyry stocks is low-grade, whereas the high-grade disseminated and stockwork mineralization (~ 1% Cu and ~ 1.2 g/t Au) occurred at the contact of the porphyry stocks and andesite porphyry. Geochemical studies of the drill hole and trench samples showed a strong correlation between Cu and Au and both show a second-order correlation with Fe and As. Magnetic survey revealed two significant magnetic anomalies, associated with intensive potassic alteration, in the reduced-to-the-pole magnetic map of the area. A relatively weaker magnetic anomaly, showing no surface porphyry expressions, is located on a lithocap, consisting of advanced argillic alteration, vuggy quartz veins, and surface expressions of epithermal geochemical signatures. The association of the lithocap and the weak magnetic anomaly could be indicative of a hidden mineralized porphyry center. Litho-geochemical analyses of the least altered Dalli intrusions and volcanic rocks indicated high Sr/Y (49-61) and Eu/Eu* (0.89-0.92), features typical of Cu porphyries. The U-Pb dating of zircons of the mineralized quartz diorite and andesite porphyry, carried out by laser ablation inductively coupled plasma mass spectrometry, yielded magmatic crystallization ages of 15.4-16.0 Ma (Middle Miocene). The zircon trace element concentrations of Dalli are characterized by high Eu/Eu* (0.3-0.8), (Ce/Nd)/Y (0.01-0.3), and 10000*(Eu/Eu*)/Y (2-15) ratios, similar to fertile porphyry suites such as the giant Sar-Cheshmeh and Qulong porphyry Cu deposits along the Tethyan belt. This suggests that the Middle Miocene Dalli intrusions are fertile and require extensive deep drillings to define their potential. Chondrite-normalized rare earth element (REE) patterns show no significant Eu anomalies, and are characterized by light-REE enrichments (La/Sm)n = 2.57–6.40). In normalized multi-element diagrams, analyzed rocks are characterized by enrichments in large ion lithophile elements (LILE) and depletions in high field strength elements (HFSE), and display typical features of subduction-related calc-alkaline magmas. The characteristics of the Dalli deposit provided several recognition criteria for detailed exploration of Cu-Au porphyry deposits and highlighted the importance of the UDMA as a potentially significant, economically important, but relatively underexplored porphyry province.

Keywords: porphyry, gold, geochronology, magnetic, exploration

Procedia PDF Downloads 23
24089 The Relationship Between Artificial Intelligence, Data Science, and Privacy

Authors: M. Naidoo

Abstract:

Artificial intelligence often requires large amounts of good quality data. Within important fields, such as healthcare, the training of AI systems predominately relies on health and personal data; however, the usage of this data is complicated by various layers of law and ethics that seek to protect individuals’ privacy rights. This research seeks to establish the challenges AI and data sciences pose to (i) informational rights, (ii) privacy rights, and (iii) data protection. To solve some of the issues presented, various methods are suggested, such as embedding values in technological development, proper balancing of rights and interests, and others.

Keywords: artificial intelligence, data science, law, policy

Procedia PDF Downloads 83
24088 Robust Inference with a Skew T Distribution

Authors: M. Qamarul Islam, Ergun Dogan, Mehmet Yazici

Abstract:

There is a growing body of evidence that non-normal data is more prevalent in nature than the normal one. Examples can be quoted from, but not restricted to, the areas of Economics, Finance and Actuarial Science. The non-normality considered here is expressed in terms of fat-tailedness and asymmetry of the relevant distribution. In this study a skew t distribution that can be used to model a data that exhibit inherent non-normal behavior is considered. This distribution has tails fatter than a normal distribution and it also exhibits skewness. Although maximum likelihood estimates can be obtained by solving iteratively the likelihood equations that are non-linear in form, this can be problematic in terms of convergence and in many other respects as well. Therefore, it is preferred to use the method of modified maximum likelihood in which the likelihood estimates are derived by expressing the intractable non-linear likelihood equations in terms of standardized ordered variates and replacing the intractable terms by their linear approximations obtained from the first two terms of a Taylor series expansion about the quantiles of the distribution. These estimates, called modified maximum likelihood estimates, are obtained in closed form. Hence, they are easy to compute and to manipulate analytically. In fact the modified maximum likelihood estimates are equivalent to maximum likelihood estimates, asymptotically. Even in small samples the modified maximum likelihood estimates are found to be approximately the same as maximum likelihood estimates that are obtained iteratively. It is shown in this study that the modified maximum likelihood estimates are not only unbiased but substantially more efficient than the commonly used moment estimates or the least square estimates that are known to be biased and inefficient in such cases. Furthermore, in conventional regression analysis, it is assumed that the error terms are distributed normally and, hence, the well-known least square method is considered to be a suitable and preferred method for making the relevant statistical inferences. However, a number of empirical researches have shown that non-normal errors are more prevalent. Even transforming and/or filtering techniques may not produce normally distributed residuals. Here, a study is done for multiple linear regression models with random error having non-normal pattern. Through an extensive simulation it is shown that the modified maximum likelihood estimates of regression parameters are plausibly robust to the distributional assumptions and to various data anomalies as compared to the widely used least square estimates. Relevant tests of hypothesis are developed and are explored for desirable properties in terms of their size and power. The tests based upon modified maximum likelihood estimates are found to be substantially more powerful than the tests based upon least square estimates. Several examples are provided from the areas of Economics and Finance where such distributions are interpretable in terms of efficient market hypothesis with respect to asset pricing, portfolio selection, risk measurement and capital allocation, etc.

Keywords: least square estimates, linear regression, maximum likelihood estimates, modified maximum likelihood method, non-normality, robustness

Procedia PDF Downloads 381
24087 Simulation Data Summarization Based on Spatial Histograms

Authors: Jing Zhao, Yoshiharu Ishikawa, Chuan Xiao, Kento Sugiura

Abstract:

In order to analyze large-scale scientific data, research on data exploration and visualization has gained popularity. In this paper, we focus on the exploration and visualization of scientific simulation data, and define a spatial V-Optimal histogram for data summarization. We propose histogram construction algorithms based on a general binary hierarchical partitioning as well as a more specific one, the l-grid partitioning. For effective data summarization and efficient data visualization in scientific data analysis, we propose an optimal algorithm as well as a heuristic algorithm for histogram construction. To verify the effectiveness and efficiency of the proposed methods, we conduct experiments on the massive evacuation simulation data.

Keywords: simulation data, data summarization, spatial histograms, exploration, visualization

Procedia PDF Downloads 155
24086 Algorithms used in Spatial Data Mining GIS

Authors: Vahid Bairami Rad

Abstract:

Extracting knowledge from spatial data like GIS data is important to reduce the data and extract information. Therefore, the development of new techniques and tools that support the human in transforming data into useful knowledge has been the focus of the relatively new and interdisciplinary research area ‘knowledge discovery in databases’. Thus, we introduce a set of database primitives or basic operations for spatial data mining which are sufficient to express most of the spatial data mining algorithms from the literature. This approach has several advantages. Similar to the relational standard language SQL, the use of standard primitives will speed-up the development of new data mining algorithms and will also make them more portable. We introduced a database-oriented framework for spatial data mining which is based on the concepts of neighborhood graphs and paths. A small set of basic operations on these graphs and paths were defined as database primitives for spatial data mining. Furthermore, techniques to efficiently support the database primitives by a commercial DBMS were presented.

Keywords: spatial data base, knowledge discovery database, data mining, spatial relationship, predictive data mining

Procedia PDF Downloads 427
24085 Data Stream Association Rule Mining with Cloud Computing

Authors: B. Suraj Aravind, M. H. M. Krishna Prasad

Abstract:

There exist emerging applications of data streams that require association rule mining, such as network traffic monitoring, web click streams analysis, sensor data, data from satellites etc. Data streams typically arrive continuously in high speed with huge amount and changing data distribution. This raises new issues that need to be considered when developing association rule mining techniques for stream data. This paper proposes to introduce an improved data stream association rule mining algorithm by eliminating the limitation of resources. For this, the concept of cloud computing is used. Inclusion of this may lead to additional unknown problems which needs further research.

Keywords: data stream, association rule mining, cloud computing, frequent itemsets

Procedia PDF Downloads 471
24084 A Comprehensive Survey and Improvement to Existing Privacy Preserving Data Mining Techniques

Authors: Tosin Ige

Abstract:

Ethics must be a condition of the world, like logic. (Ludwig Wittgenstein, 1889-1951). As important as data mining is, it possess a significant threat to ethics, privacy, and legality, since data mining makes it difficult for an individual or consumer (in the case of a company) to control the accessibility and usage of his data. This research focuses on Current issues and the latest research and development on Privacy preserving data mining methods as at year 2022. It also discusses some advances in those techniques while at the same time highlighting and providing a new technique as a solution to an existing technique of privacy preserving data mining methods. This paper also bridges the wide gap between Data mining and the Web Application Programing Interface (web API), where research is urgently needed for an added layer of security in data mining while at the same time introducing a seamless and more efficient way of data mining.

Keywords: data, privacy, data mining, association rule, privacy preserving, mining technique

Procedia PDF Downloads 129
24083 Big Data: Concepts, Technologies and Applications in the Public Sector

Authors: A. Alexandru, C. A. Alexandru, D. Coardos, E. Tudora

Abstract:

Big Data (BD) is associated with a new generation of technologies and architectures which can harness the value of extremely large volumes of very varied data through real time processing and analysis. It involves changes in (1) data types, (2) accumulation speed, and (3) data volume. This paper presents the main concepts related to the BD paradigm, and introduces architectures and technologies for BD and BD sets. The integration of BD with the Hadoop Framework is also underlined. BD has attracted a lot of attention in the public sector due to the newly emerging technologies that allow the availability of network access. The volume of different types of data has exponentially increased. Some applications of BD in the public sector in Romania are briefly presented.

Keywords: big data, big data analytics, Hadoop, cloud

Procedia PDF Downloads 279
24082 Access Control System for Big Data Application

Authors: Winfred Okoe Addy, Jean Jacques Dominique Beraud

Abstract:

Access control systems (ACs) are some of the most important components in safety areas. Inaccuracies of regulatory frameworks make personal policies and remedies more appropriate than standard models or protocols. This problem is exacerbated by the increasing complexity of software, such as integrated Big Data (BD) software for controlling large volumes of encrypted data and resources embedded in a dedicated BD production system. This paper proposes a general access control strategy system for the diffusion of Big Data domains since it is crucial to secure the data provided to data consumers (DC). We presented a general access control circulation strategy for the Big Data domain by describing the benefit of using designated access control for BD units and performance and taking into consideration the need for BD and AC system. We then presented a generic of Big Data access control system to improve the dissemination of Big Data.

Keywords: access control, security, Big Data, domain

Procedia PDF Downloads 107
24081 A Data Envelopment Analysis Model in a Multi-Objective Optimization with Fuzzy Environment

Authors: Michael Gidey Gebru

Abstract:

Most of Data Envelopment Analysis models operate in a static environment with input and output parameters that are chosen by deterministic data. However, due to ambiguity brought on shifting market conditions, input and output data are not always precisely gathered in real-world scenarios. Fuzzy numbers can be used to address this kind of ambiguity in input and output data. Therefore, this work aims to expand crisp Data Envelopment Analysis into Data Envelopment Analysis with fuzzy environment. In this study, the input and output data are regarded as fuzzy triangular numbers. Then, the Data Envelopment Analysis model with fuzzy environment is solved using a multi-objective method to gauge the Decision Making Units' efficiency. Finally, the developed Data Envelopment Analysis model is illustrated with an application on real data 50 educational institutions.

Keywords: efficiency, Data Envelopment Analysis, fuzzy, higher education, input, output

Procedia PDF Downloads 17
24080 Major Mechanisms of Atmospheric Moisture Transport and Their Role in Precipitation Extreme Events in the Amazonia

Authors: Luis Gimeno, Rosmeri da Rocha, Raquel Nieto, Tercio Ambrizzi, Alex Ramos, Anita Drumond

Abstract:

The transport of moisture from oceanic sources to the continents represents the atmospheric branch of the water cycle, forming the connection between evaporation from the ocean and precipitation over the continents. In this regard two large scale dynamical/meteorological structures appear to play a key role, namely Low Level Jet (LLJ) systems and Atmospheric Rivers (ARs). The former are particularly important in tropical and subtropical regions; the latter is mostly confined to extratropical regions. A key question relates to the anomalies in the transport of moisture observed during natural hazards related to extremes of precipitation (i.e., drought or wet spells). In this study we will be focused on these two major atmospheric moisture transport mechanisms (LLJs and ARs) and its role in precipitation extreme events (droughts and wet spells) in the Amazonia paying particular attention to i) intensification (decreasing) of moisture transport by them and its role in wet spells (droughts), and ii) changes in their positions and occurrence with associated flooding and wet spells.

Keywords: droughts, wet spells, amazonia, LLJs, atmospheric rivers

Procedia PDF Downloads 273
24079 A Retrospective Study of Pain Management Strategies for Pediatric Hypospadias Surgery in a Tertiary Care Hospital in Western Rajasthan

Authors: Darshana Rathod, Kirtikumar Rathod, Kamlesh Kumari, Abhilasha Motghare

Abstract:

Background and Aims: Hypospadias is one of the common congenital anomalies in males. Various modalities are used for pain management, including caudal, penile, pudendal, ring blocks, and systemic analgesics. There has yet to be a consensus regarding the most effective and safe analgesic method for controlling pain in these children. We planned this study to determine our institute's pain management practices for hypospadias surgeries. Material and Methods: This retrospective cohort study reviewed 150 children with hypospadias undergoing surgery from January 2020 to December 2023. Data regarding the mode of pain management, postoperative opioid requirement, PACU discharge, and complications was collected from the records. Results: For postoperative pain, 33 (22%) children received caudal block, 60 (40%) penile block, and 57 (38%) were managed by intravenous analgesics. A significant difference was found in the three groups, with the IV analgesic group requiring significantly higher opioid boluses in PACU [43 (75.4%) required two boluses (p < 0.05)]. The difference in PACU discharge time among the three groups was statistically significant (p< 0.05), with IV analgesics groups having the highest (55 mins [47, 60]), the Caudal group at 35mins (30, 40), and the dorsal penile block group at 35mins (25, 40). There was no significant difference in complications like edema, meatal stenosis, urethra-cutaneous fistula, or wound dehiscence among all three groups. Conclusion: Intravenous analgesics and regional blocks like caudal and penile blocks are the common pain management modalities in our institute. The regional blocks are effective in managing pain in the postoperative period and are not significantly associated with complications.

Keywords: caudal block, hypospadias, pain management, penile block

Procedia PDF Downloads 15
24078 Evaluation of Fetal brain using Magnetic Resonance Imaging

Authors: Mahdi Farajzadeh Ajirlou

Abstract:

Ordinary fetal brain development can be considered by in vivo attractive reverberation imaging (MRI) from the 18th gestational week (GW) to term and depends fundamentally on T2-weighted and diffusion-weighted (DW) arrangements. The foremost commonly suspected brain pathologies alluded to fetal MRI for assist assessment are ventriculomegaly, lost corpus callosum, and anomalies of the posterior fossa. Brain division could be a crucial to begin with step in neuroimage examination. Within the case of fetal MRI it is especially challenging and critical due to the subjective introduction of the hatchling, organs that encompass the fetal head, and irregular fetal movement. A few promising strategies have been proposed but are constrained in their execution in challenging cases and in realtime division. Fetal MRI is routinely performed on a 1.5-Tesla scanner without maternal or fetal sedation. The mother lies recumbent amid the course of the examination, the length of which is ordinarily 45 to 60 minutes. The accessibility and continuous approval of standardizing fetal brain development directions will give critical devices for early discovery of impeded fetal brain development upon which to oversee high-risk pregnancies.

Keywords: brain, fetal, MRI, imaging

Procedia PDF Downloads 46