Search results for: data analyses
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26910

Search results for: data analyses

22680 Data Analysis for Taxonomy Prediction and Annotation of 16S rRNA Gene Sequences from Metagenome Data

Authors: Suchithra V., Shreedhanya, Kavya Menon, Vidya Niranjan

Abstract:

Skin metagenomics has a wide range of applications with direct relevance to the health of the organism. It gives us insight to the diverse community of microorganisms (the microbiome) harbored on the skin. In the recent years, it has become increasingly apparent that the interaction between skin microbiome and the human body plays a prominent role in immune system development, cancer development, disease pathology, and many other biological implications. Next Generation Sequencing has led to faster and better understanding of environmental organisms and their mutual interactions. This project is studying the human skin microbiome of different individuals having varied skin conditions. Bacterial 16S rRNA data of skin microbiome is downloaded from SRA toolkit provided by NCBI to perform metagenomics analysis. Twelve samples are selected with two controls, and 3 different categories, i.e., sex (male/female), skin type (moist/intermittently moist/sebaceous) and occlusion (occluded/intermittently occluded/exposed). Quality of the data is increased using Cutadapt, and its analysis is done using FastQC. USearch, a tool used to analyze an NGS data, provides a suitable platform to obtain taxonomy classification and abundance of bacteria from the metagenome data. The statistical tool used for analyzing the USearch result is METAGENassist. The results revealed that the top three abundant organisms found were: Prevotella, Corynebacterium, and Anaerococcus. Prevotella is known to be an infectious bacterium found on wound, tooth cavity, etc. Corynebacterium and Anaerococcus are opportunist bacteria responsible for skin odor. This result infers that Prevotella thrives easily in sebaceous skin conditions. Therefore it is better to undergo intermittently occluded treatment such as applying ointments, creams, etc. to treat wound for sebaceous skin type. Exposing the wound should be avoided as it leads to an increase in Prevotella abundance. Moist skin type individuals can opt for occluded or intermittently occluded treatment as they have shown to decrease the abundance of bacteria during treatment.

Keywords: bacterial 16S rRNA , next generation sequencing, skin metagenomics, skin microbiome, taxonomy

Procedia PDF Downloads 167
22679 The Nexus between Socio-Economic Inequalities and the Talibanization in Pakistan’s Federally Administrated Tribal Areas

Authors: Sajjad Ahmed

Abstract:

Since September 2001, the Federally Administered Tribal Areas (FATA) have become a hotbed of Talibanization. The eruption of Talibanization has caused a catastrophic human and socio-economic cost on Pakistan ever since. The vast majority of extant studies have tended to focus on assessing the current disparaging and destructive condition of FATA as a product of the notorious 'Global War on Terrorism' and its consequences in the form of the Afghan war and the rising socio-political unrest in the region. This, however, is not the case. This study argues that the Talibanization has not happened overnight, the magma of current militant volcanic outburst has been stockpiled since the inception of Pakistan in 1947. The study claims that the Talibanization is the expression of the conflict between the privileged and the underprivileged. The prevailing situation in FATA warrants an in-depth analysis of the problem. By using a qualitative and quantitative research principle, this paper attempts to critically examine 'How is Talibanization in Pakistan connected with the political, social, and economic conditions in FATA?' The critical analyses of this study would assist to policymakers in order to formulate all-encompassing anti-radicalization policies to effectively root out Talibanization in FATA. This research intends to explore the undiscovered root causes of the problem and to suggest remedial measures.

Keywords: exclusion, FATA (Federally Administrated Tribal Areas), inequalities, marginalization, Pakistan, socio-economic, talibanization

Procedia PDF Downloads 134
22678 Development of a Predictive Model to Prevent Financial Crisis

Authors: Tengqin Han

Abstract:

Delinquency has been a crucial factor in economics throughout the years. Commonly seen in credit card and mortgage, it played one of the crucial roles in causing the most recent financial crisis in 2008. In each case, a delinquency is a sign of the loaner being unable to pay off the debt, and thus may cause a lost of property in the end. Individually, one case of delinquency seems unimportant compared to the entire credit system. China, as an emerging economic entity, the national strength and economic strength has grown rapidly, and the gross domestic product (GDP) growth rate has remained as high as 8% in the past decades. However, potential risks exist behind the appearance of prosperity. Among the risks, the credit system is the most significant one. Due to long term and a large amount of balance of the mortgage, it is critical to monitor the risk during the performance period. In this project, about 300,000 mortgage account data are analyzed in order to develop a predictive model to predict the probability of delinquency. Through univariate analysis, the data is cleaned up, and through bivariate analysis, the variables with strong predictive power are detected. The project is divided into two parts. In the first part, the analysis data of 2005 are split into 2 parts, 60% for model development, and 40% for in-time model validation. The KS of model development is 31, and the KS for in-time validation is 31, indicating the model is stable. In addition, the model is further validation by out-of-time validation, which uses 40% of 2006 data, and KS is 33. This indicates the model is still stable and robust. In the second part, the model is improved by the addition of macroeconomic economic indexes, including GDP, consumer price index, unemployment rate, inflation rate, etc. The data of 2005 to 2010 is used for model development and validation. Compared with the base model (without microeconomic variables), KS is increased from 41 to 44, indicating that the macroeconomic variables can be used to improve the separation power of the model, and make the prediction more accurate.

Keywords: delinquency, mortgage, model development, model validation

Procedia PDF Downloads 219
22677 Self-Supervised Learning for Hate-Speech Identification

Authors: Shrabani Ghosh

Abstract:

Automatic offensive language detection in social media has become a stirring task in today's NLP. Manual Offensive language detection is tedious and laborious work where automatic methods based on machine learning are only alternatives. Previous works have done sentiment analysis over social media in different ways such as supervised, semi-supervised, and unsupervised manner. Domain adaptation in a semi-supervised way has also been explored in NLP, where the source domain and the target domain are different. In domain adaptation, the source domain usually has a large amount of labeled data, while only a limited amount of labeled data is available in the target domain. Pretrained transformers like BERT, RoBERTa models are fine-tuned to perform text classification in an unsupervised manner to perform further pre-train masked language modeling (MLM) tasks. In previous work, hate speech detection has been explored in Gab.ai, which is a free speech platform described as a platform of extremist in varying degrees in online social media. In domain adaptation process, Twitter data is used as the source domain, and Gab data is used as the target domain. The performance of domain adaptation also depends on the cross-domain similarity. Different distance measure methods such as L2 distance, cosine distance, Maximum Mean Discrepancy (MMD), Fisher Linear Discriminant (FLD), and CORAL have been used to estimate domain similarity. Certainly, in-domain distances are small, and between-domain distances are expected to be large. The previous work finding shows that pretrain masked language model (MLM) fine-tuned with a mixture of posts of source and target domain gives higher accuracy. However, in-domain performance of the hate classifier on Twitter data accuracy is 71.78%, and out-of-domain performance of the hate classifier on Gab data goes down to 56.53%. Recently self-supervised learning got a lot of attention as it is more applicable when labeled data are scarce. Few works have already been explored to apply self-supervised learning on NLP tasks such as sentiment classification. Self-supervised language representation model ALBERTA focuses on modeling inter-sentence coherence and helps downstream tasks with multi-sentence inputs. Self-supervised attention learning approach shows better performance as it exploits extracted context word in the training process. In this work, a self-supervised attention mechanism has been proposed to detect hate speech on Gab.ai. This framework initially classifies the Gab dataset in an attention-based self-supervised manner. On the next step, a semi-supervised classifier trained on the combination of labeled data from the first step and unlabeled data. The performance of the proposed framework will be compared with the results described earlier and also with optimized outcomes obtained from different optimization techniques.

Keywords: attention learning, language model, offensive language detection, self-supervised learning

Procedia PDF Downloads 100
22676 IPO Valuation and Profitability Expectations: Evidence from the Italian Exchange

Authors: Matteo Bonaventura, Giancarlo Giudici

Abstract:

This paper analyses the valuation process of companies listed on the Italian Exchange in the period 2000-2009 at their Initial Public Offering (IPO). One the most common valuation techniques declared in the IPO prospectus to determine the offer price is the Discounted Cash Flow (DCF) method. We develop a ‘reverse engineering’ model to discover the short term profitability implied in the offer prices. We show that there is a significant optimistic bias in the estimation of future profitability compared to ex-post actual realization and the mean forecast error is substantially large. Yet we show that such error characterizes also the estimations carried out by analysts evaluating non-IPO companies. The forecast error is larger the faster has been the recent growth of the company, the higher is the leverage of the IPO firm, the more companies issued equity on the market. IPO companies generally exhibit better operating performance before the listing, with respect to comparable listed companies, while after the flotation they do not perform significantly different in term of return on invested capital. Pre-IPO book building activity plays a significant role in partially reducing the forecast error and revising expectations, while the market price of the first day of trading does not contain information for further reducing forecast errors.

Keywords: initial public offerings, DCF, book building, post-IPO profitability drop

Procedia PDF Downloads 348
22675 Time and Cost Prediction Models for Language Classification Over a Large Corpus on Spark

Authors: Jairson Barbosa Rodrigues, Paulo Romero Martins Maciel, Germano Crispim Vasconcelos

Abstract:

This paper presents an investigation of the performance impacts regarding the variation of five factors (input data size, node number, cores, memory, and disks) when applying a distributed implementation of Naïve Bayes for text classification of a large Corpus on the Spark big data processing framework. Problem: The algorithm's performance depends on multiple factors, and knowing before-hand the effects of each factor becomes especially critical as hardware is priced by time slice in cloud environments. Objectives: To explain the functional relationship between factors and performance and to develop linear predictor models for time and cost. Methods: the solid statistical principles of Design of Experiments (DoE), particularly the randomized two-level fractional factorial design with replications. This research involved 48 real clusters with different hardware arrangements. The metrics were analyzed using linear models for screening, ranking, and measurement of each factor's impact. Results: Our findings include prediction models and show some non-intuitive results about the small influence of cores and the neutrality of memory and disks on total execution time, and the non-significant impact of data input scale on costs, although notably impacts the execution time.

Keywords: big data, design of experiments, distributed machine learning, natural language processing, spark

Procedia PDF Downloads 110
22674 The Developing of Teaching Materials Online for Students in Thailand

Authors: Pitimanus Bunlue

Abstract:

The objectives of this study were to identify the unique characteristics of Salaya Old market, Phutthamonthon, Nakhon Pathom and develop the effective video media to promote the homeland awareness among local people and the characteristic features of this community were collectively summarized based on historical data, community observation, and people’s interview. The acquired data were used to develop a media describing prominent features of the community. The quality of the media was later assessed by interviewing local people in the old market in terms of content accuracy, video, and narration qualities, and sense of homeland awareness after watching the video. The result shows a 6-minute video media containing historical data and outstanding features of this community was developed. Based on the interview, the content accuracy was good. The picture quality and the narration were very good. Most people developed a sense of homeland awareness after watching the video also as well.

Keywords: audio-visual, creating homeland awareness, Phutthamonthon Nakhon Pathom, research and development

Procedia PDF Downloads 288
22673 Design and Validation of a Darrieus Type Hydrokinetic Turbine for South African Irrigation Canals Experimentally and Computationally

Authors: Maritz Lourens Van Rensburg, Chantel Niebuhr

Abstract:

Utilizing all available renewable energy sources is an ever-growing necessity, this includes a newfound interest into hydrokinetic energy systems, which open the door to installations where conventional hydropower shows no potential. Optimization and obtaining high efficiencies are key in these installations. In this study a vertical axis Darrieus hydrokinetic turbine is designed and constructed to address certain drawbacks experience by axial flow horizontal axis turbines in an irrigation channel. Many horizontal axis turbines have been well developed and optimized to have high efficiencies but depending on the conditions experienced in an open channel, the performance of these turbines may be adversely affected. The study analyses how the designed vertical axis turbine addresses the problems experienced by a horizontal axis turbine while still achieving a satisfactory efficiency. To be able to optimize the vertical axis turbine, a computational fluid dynamics model was validated to the experimental results obtained from the power generated from a test turbine installation operating at various rotational speeds. It was found that an accurate validated model can be obtained through validation of generated power output.

Keywords: hydrokinetic, Darrieus, computational fluid dynamics, vertical axis turbine

Procedia PDF Downloads 111
22672 A Decision Support System for the Detection of Illicit Substance Production Sites

Authors: Krystian Chachula, Robert Nowak

Abstract:

Manufacturing home-made explosives and synthetic drugs is an increasing problem in Europe. To combat that, a data fusion system is proposed for the detection and localization of production sites in urban environments. The data consists of measurements of properties of wastewater performed by various sensors installed in a sewage network. A four-stage fusion strategy allows detecting sources of waste products from known chemical reactions. First, suspicious measurements are used to compute the amount and position of discharged compounds. Then, this information is propagated through the sewage network to account for missing sensors. The next step is clustering and the formation of tracks. Eventually, tracks are used to reconstruct discharge events. Sensor measurements are simulated by a subsystem based on real-world data. In this paper, different discharge scenarios are considered to show how the parameters of used algorithms affect the effectiveness of the proposed system. This research is a part of the SYSTEM project (SYnergy of integrated Sensors and Technologies for urban sEcured environMent).

Keywords: continuous monitoring, information fusion and sensors, internet of things, multisensor fusion

Procedia PDF Downloads 109
22671 Implementation of CNV-CH Algorithm Using Map-Reduce Approach

Authors: Aishik Deb, Rituparna Sinha

Abstract:

We have developed an algorithm to detect the abnormal segment/"structural variation in the genome across a number of samples. We have worked on simulated as well as real data from the BAM Files and have designed a segmentation algorithm where abnormal segments are detected. This algorithm aims to improve the accuracy and performance of the existing CNV-CH algorithm. The next-generation sequencing (NGS) approach is very fast and can generate large sequences in a reasonable time. So the huge volume of sequence information gives rise to the need for Big Data and parallel approaches of segmentation. Therefore, we have designed a map-reduce approach for the existing CNV-CH algorithm where a large amount of sequence data can be segmented and structural variations in the human genome can be detected. We have compared the efficiency of the traditional and map-reduce algorithms with respect to precision, sensitivity, and F-Score. The advantages of using our algorithm are that it is fast and has better accuracy. This algorithm can be applied to detect structural variations within a genome, which in turn can be used to detect various genetic disorders such as cancer, etc. The defects may be caused by new mutations or changes to the DNA and generally result in abnormally high or low base coverage and quantification values.

Keywords: cancer detection, convex hull segmentation, map reduce, next generation sequencing

Procedia PDF Downloads 127
22670 Inferring Human Mobility in India Using Machine Learning

Authors: Asra Yousuf, Ajaykumar Tannirkulum

Abstract:

Inferring rural-urban migration trends can help design effective policies that promote better urban planning and rural development. In this paper, we describe how machine learning algorithms can be applied to predict internal migration decisions of people. We consider data collected from household surveys in Tamil Nadu to train our model. To measure the performance of the model, we use data on past migration from National Sample Survey Organisation of India. The factors for training the model include socioeconomic characteristic of each individual like age, gender, place of residence, outstanding loans, strength of the household, etc. and his past migration history. We perform a comparative analysis of the performance of a number of machine learning algorithm to determine their prediction accuracy. Our results show that machine learning algorithms provide a stronger prediction accuracy as compared to statistical models. Our goal through this research is to propose the use of data science techniques in understanding human decisions and behaviour in developing countries.

Keywords: development, migration, internal migration, machine learning, prediction

Procedia PDF Downloads 267
22669 Effect of Measured and Calculated Static Torque on Instantaneous Torque Profile of Switched Reluctance Motor

Authors: Ali Asghar Memon

Abstract:

The simulation modeling of switched reluctance (SR) machine often relies and uses the three data tables identified as static torque characteristics that include flux linkage characteristics, co energy characteristics and static torque characteristics separately. It has been noticed from the literature that the data of static torque used in the simulation model is often calculated so far the literature is concerned. This paper presents the simulation model that include the data of measured and calculated static torque separately to see its effect on instantaneous torque profile of the machine. This is probably for the first time so far the literature review is concerned that static torque from co energy information, and measured static torque directly from experiments are separately used in the model. This research is helpful for accurate modeling of switched reluctance drive.

Keywords: static characteristics, current chopping, flux linkage characteristics, switched reluctance motor

Procedia PDF Downloads 286
22668 Effects of Kenaf and Rice Husk on Water Absorption and Flexural Properties of Kenaf/CaCO3/HDPE and Rice Husk/CaCO3/HDPE Hybrid Composites

Authors: Noor Zuhaira Abd Aziz

Abstract:

Rice husk and kenaf filled with calcium carbonate (CaCO3) and high density polyethylene (HDPE) composite were prepared separately using twin-screw extruder at 50rpm. Different filler loading up to 30 parts of rice husk particulate and kenaf fiber were mixed with the fixed 30% amount of CaCO3 mineral filler to produce rice husk/CaCO3/HDPE and kenaf/CaCO3/HDPE hybrid composites. In this study, the effects of natural fiber for both rice husk and kenaf in CaCO3/HDPE composite on physical and mechanical properties were investigated. The property analyses showed that water absorption increased with the presence of kenaf and rice husk fillers. Natural fibers in composite significantly influence water absorption properties due to natural characters of fibers which contain cellulose, hemicellulose and lignin structures. The result showed that 10% of additional natural fibers into hybrid composite had caused decreased flexural strength, however additional of high natural fiber (>10%) filler loading has proved to increase its flexural strength.

Keywords: Hybrid composites, Water absorption, Mechanical properties

Procedia PDF Downloads 459
22667 Hardware Implementation on Field Programmable Gate Array of Two-Stage Algorithm for Rough Set Reduct Generation

Authors: Tomasz Grzes, Maciej Kopczynski, Jaroslaw Stepaniuk

Abstract:

The rough sets theory developed by Prof. Z. Pawlak is one of the tools that can be used in the intelligent systems for data analysis and processing. Banking, medicine, image recognition and security are among the possible fields of utilization. In all these fields, the amount of the collected data is increasing quickly, but with the increase of the data, the computation speed becomes the critical factor. Data reduction is one of the solutions to this problem. Removing the redundancy in the rough sets can be achieved with the reduct. A lot of algorithms of generating the reduct were developed, but most of them are only software implementations, therefore have many limitations. Microprocessor uses the fixed word length, consumes a lot of time for either fetching as well as processing of the instruction and data; consequently, the software based implementations are relatively slow. Hardware systems don’t have these limitations and can process the data faster than a software. Reduct is the subset of the decision attributes that provides the discernibility of the objects. For the given decision table there can be more than one reduct. Core is the set of all indispensable condition attributes. None of its elements can be removed without affecting the classification power of all condition attributes. Moreover, every reduct consists of all the attributes from the core. In this paper, the hardware implementation of the two-stage greedy algorithm to find the one reduct is presented. The decision table is used as an input. Output of the algorithm is the superreduct which is the reduct with some additional removable attributes. First stage of the algorithm is calculating the core using the discernibility matrix. Second stage is generating the superreduct by enriching the core with the most common attributes, i.e., attributes that are more frequent in the decision table. Described above algorithm has two disadvantages: i) generating the superreduct instead of reduct, ii) additional first stage may be unnecessary if the core is empty. But for the systems focused on the fast computation of the reduct the first disadvantage is not the key problem. The core calculation can be achieved with a combinational logic block, and thus add respectively little time to the whole process. Algorithm presented in this paper was implemented in Field Programmable Gate Array (FPGA) as a digital device consisting of blocks that process the data in a single step. Calculating the core is done by the comparators connected to the block called 'singleton detector', which detects if the input word contains only single 'one'. Calculating the number of occurrences of the attribute is performed in the combinational block made up of the cascade of the adders. The superreduct generation process is iterative and thus needs the sequential circuit for controlling the calculations. For the research purpose, the algorithm was also implemented in C language and run on a PC. The times of execution of the reduct calculation in a hardware and software were considered. Results show increase in the speed of data processing.

Keywords: data reduction, digital systems design, field programmable gate array (FPGA), reduct, rough set

Procedia PDF Downloads 214
22666 Digitally Mapping Aboriginal Journey Ways

Authors: Paul Longley Arthur

Abstract:

This paper reports on an Australian Research Council-funded project utilising the Australian digital research infrastructure the ‘Time-Layered Cultural Map of Australia’ (TLCMap) (https://www.tlcmap.org/) [1]. This resource has been developed to help researchers create digital maps from cultural, textual, and historical data, layered with datasets registered on the platform. TLCMap is a set of online tools that allows humanities researchers to compile humanities data using spatio-temporal coordinates – to upload, gather, analyse and visualise data. It is the only purpose-designed, Australian-developed research tool for humanities and social science researchers to identify geographical clusters and parallel journeys by sight. This presentation discusses a series of Aboriginal mapping and visualisation experiments using TLCMap to show how Indigenous knowledge can reconfigure contemporary understandings of space including the urbanised landscape [2, 3]. The research data being generated – investigating the historical movements of Aboriginal people, the distribution of networks, and their relation to land – lends itself to mapping and geo-spatial visualisation and analysis. TLCMap allows researchers to create layers on a 3D map which pinpoint locations with accompanying information, and this has enabled our research team to plot out traditional historical journeys undertaken by Aboriginal people as well as to compile a gazetteer of Aboriginal place names, many of which have largely been undocumented until now [4]. The documented journeys intersect with and overlay many of today’s urban formations including main roads, municipal boundaries, and state borders. The paper questions how such data can be incorporated into a more culturally and ethically responsive understanding of contemporary urban spaces and as well as natural environments [5].

Keywords: spatio-temporal mapping, visualisation, Indigenous knowledge, mobility and migration, research infrastructure

Procedia PDF Downloads 11
22665 Japanese and Europe Legal Frameworks on Data Protection and Cybersecurity: Asymmetries from a Comparative Perspective

Authors: S. Fantin

Abstract:

This study is the result of the legal research on cybersecurity and data protection within the EUNITY (Cybersecurity and Privacy Dialogue between Europe and Japan) project, aimed at fostering the dialogue between the European Union and Japan. Based on the research undertaken therein, the author offers an outline of the main asymmetries in the laws governing such fields in the two regions. The research is a comparative analysis of the two legal frameworks, taking into account specific provisions, ratio legis and policy initiatives. Recent doctrine was taken into account, too, as well as empirical interviews with EU and Japanese stakeholders and project partners. With respect to the protection of personal data, the European Union has recently reformed its legal framework with a package which includes a regulation (General Data Protection Regulation), and a directive (Directive 680 on personal data processing in the law enforcement domain). In turn, the Japanese law under scrutiny for this study has been the Act on Protection of Personal Information. Based on a comparative analysis, some asymmetries arise. The main ones refer to the definition of personal information and the scope of the two frameworks. Furthermore, the rights of the data subjects are differently articulated in the two regions, while the nature of sanctions take two opposite approaches. Regarding the cybersecurity framework, the situation looks similarly misaligned. Japan’s main text of reference is the Basic Cybersecurity Act, while the European Union has a more fragmented legal structure (to name a few, Network and Information Security Directive, Critical Infrastructure Directive and Directive on the Attacks at Information Systems). On an relevant note, unlike a more industry-oriented European approach, the concept of cyber hygiene seems to be neatly embedded in the Japanese legal framework, with a number of provisions that alleviate operators’ liability by turning such a burden into a set of recommendations to be primarily observed by citizens. With respect to the reasons to fill such normative gaps, these are mostly grounded on three basis. Firstly, the cross-border nature of cybercrime brings to consider both magnitude of the issue and its regulatory stance globally. Secondly, empirical findings from the EUNITY project showed how recent data breaches and cyber-attacks had shared implications between Europe and Japan. Thirdly, the geopolitical context is currently going through the direction of bringing the two regions to significant agreements from a trade standpoint, but also from a data protection perspective (with an imminent signature by both parts of a so-called ‘Adequacy Decision’). The research conducted in this study reveals two asymmetric legal frameworks on cyber security and data protection. With a view to the future challenges presented by the strengthening of the collaboration between the two regions and the trans-national fashion of cybercrime, it is urged that solutions are found to fill in such gaps, in order to allow European Union and Japan to wisely increment their partnership.

Keywords: cybersecurity, data protection, European Union, Japan

Procedia PDF Downloads 118
22664 The Causality between Corruption and Economic Growth in MENA Countries: A Dynamic Panel-Data Analysis

Authors: Nour Mohamad Fayad

Abstract:

Complex and extensively researched, the impact of corruption on economic growth seems to be intricate. Many experts believe that corruption reduces economic development. However, counterarguments have suggested that corruption either promotes growth and development or has no significant impact on economic performance. Clearly, there is no consensus in the economics literature regarding the possible relationship between corruption and economic development. Corruption's complex and clandestine nature, which makes it difficult to define and measure, is one of the obstacles that must be overcome when investigating its effect on an economy. In an attempt to contribute to the ongoing debate, this study examines the impact of corruption on economic growth in the Middle East and North Africa (MENA) region between 2000 and 2021 using a Customized Corruption Index-CCI and panel data on MENA countries. These countries were selected because they are understudied in the economic literature, and despite the World Bank's recent emphasis on corruption in the developing world, the MENA countries have received little attention. The researcher used Cobb-Douglas functional form to test corruption in MENA using a customized index known as Customized Corruption Index-CCI to track corruption over almost 20 years, then used the dynamic panel data. The findings indicate that there is a positive correlation between corruption and economic growth, but this is not consistent across all MENA nations. First, the relatively recent lack of data from MENA nations. This issue is related to the inaccessibility of data for many MENA countries, particularly regarding the returns on resources, private malfeasance, and other variables in Gulf countries. In addition, the researcher encountered several restrictions, such as electricity and internet outages, due to the fact that he is from Lebanon, a country whose citizens have endured difficult living conditions since the Lebanese crisis began in 2019. Demonstrating a customized index known as Customized Corruption Index-CCI that suits the characteristics of MENA countries to peculiarly measure corruption in this region, the outcome of the Customized Corruption Index-CCI is then compared to the Corruption Perception Index-CPI and Control of Corruption from World Governance Indicator-CC from WGI.

Keywords: corruption, economic growth, corruption measurements, empirical review, impact of corruption

Procedia PDF Downloads 69
22663 Enhancing the Rollability of Cu-Ge-Ni Alloy through Heat Treatment Methods

Authors: Morteza Hadi

Abstract:

This research investigates the potential enhancement of the rollability of Cu-Ge-Ni alloy through the mitigation of microstructural and compositional inhomogeneities via two distinct heat treatment methods: homogenization and solution treatment. To achieve this objective, the alloy with the desired composition was fabricated using a vacuum arc remelting furnace (VAR), followed by sample preparation for microstructural, compositional, and heat treatment analyses at varying temperatures and durations. Characterization was conducted employing optical and scanning electron microscopy (SEM), X-ray diffraction (XRD), and Vickers hardness testing. The results obtained indicate that a minimum duration of 10 hours is necessary for adequate homogenization of the alloy at 750°C. This heat treatment effectively removes coarse dendrites from the casting microstructure and significantly reduces elemental separations. However, despite these improvements, the presence of a second phase with markedly different hardness from the matrix results in poor rolling ability for the alloy. The optimal time for solution treatment at various temperatures was determined, with the most effective cycle identified as 750°C for 2 hours, followed by rapid quenching in water. This process induces the formation of a single-phase microstructure and complete elimination of the second  phase, as confirmed by X-ray diffraction analysis. Results demonstrate a reduction in hardness by 30 Vickers, and the elimination of microstructural unevenness enables successful thickness reduction by up to 50% through rolling without encountering cracking.

Keywords: Cu-Ge-Ni alloy, homogenization. solution treatment, rollability

Procedia PDF Downloads 50
22662 Evaluation of Illegal Hunting of Red Deer and Conservation Policy of Department of Environment in Iran

Authors: Tahere Fazilat

Abstract:

Caspian red deer or maral (Cervus elaphus maral) is the largest type of deer in iran. Maral in the past has lived in the north forests of Iran from the Caspian sea coast, Alborz mountains chain and oak forest of Zagros margin from the Azarbaijan up to fars province. However, the generation of them was completely destroyed in the north west and west of Iran. According to reports about 50 years and out of reach of humans. In the present studies, data were collected from 2004 to 2014 in the Mazandaran state Hyrcanian forest by means of guard of environment and justiciary office of department of environment of Mazandaran in this process the all arrested illegal hunting of red deer and the population census, estimation and the correlation of these data was assayed. We provide a first evaluation of how suitable these methods are by comparing the results with population estimates obtained using cohort analysis, and by analyzing the within-season variation in number of seen deer. The data gave us the future of red deer in northern forest of Iran and the results of policy of department of environment in Iran in red deer conservation.

Keywords: illegal hunting, red deer, census, concervation

Procedia PDF Downloads 546
22661 Thermodynamic Performance Tests for 3D Printed Steel Slag Powder Concrete Walls

Authors: Li Guoyou, Zhang Tao, Ji Wenzhan, Huo Liang, Lin Xiqiang, Zhang Nan

Abstract:

The three dimensional (3D) printing technology has undergone rapid development in the last few years and it is possible to print engineering structures. 3D printing buildings use wastes from constructions, industries and mine tailings as “ink”, and mix it with property improved materials, such as cement, fiber etc. This paper presents a study of the Thermodynamic performance of 3D printed walls using cement and steel slag powder. Analyses the thermal simulation regarding 3D printed walls and solid brick wall by the way of the hot-box methods and the infrared technology, and the results were contrasted with theoretical calculation. The results show that the excellent thermodynamic performance of 3D printed concrete wall made it suitable as the partial materials for self-thermal insulation walls in residential buildings. The thermodynamic performance of 3D printed concrete walls depended on the density of materials, distribution of holes, and the filling materials. Decreasing the density of materials, increasing the number of holes or replacing the filling materials with foamed concrete could improve its thermodynamic performance significantly. The average of heat transfer coefficient and thermal inertia index of 3D printed steel slag powder concrete wall all better than the traditional solid brick wall with a thickness of 240mm.

Keywords: concrete, 3D printed walls, thermodynamic performance, steel slag powder

Procedia PDF Downloads 179
22660 Development of High Quality Refractory Bricks from Fireclays for Industrial Applications

Authors: David E. Esezobor, Friday I. Apeh, Harrison O. Onovo, Ademola A. Agbeleye

Abstract:

Available indigenous refractory bricks in Nigeria can only be used in the lining of furnaces for melting of cast iron operating at less than 1,400°C or in preheating furnaces due to their low refractoriness less than 1,500°C. The bricks crack and shatter on heating at 1350 to 1450°C. In this paper, a simple and adaptable technology of manufacturing high-quality refractory bricks from selected Nigerian clays for furnace linings was developed. Fireclays from Onibode, Owode-Ketu in Ogun State and Kwoi in Kaduna State were crushed, ground, and sieved into various grain sizes using standard techniques. The pulverized clays were blended with alumina in various mix ratios and indurated in the furnace at 900 – 16000C. Their chemical, microstructure and mineralogical properties were characterized using atomic absorption spectrophotometry, scanning electron microscopy and x-ray diffraction spectrometry respectively. The mineralogical and spectrochemical analyses suggested that the clays are of siliceous alumino-silicate and acidic in nature. The appropriate blending of fireclays with alumina provided the tremendous improvement in the refractoriness of the bricks and other acceptable service properties comparable with imported refractory bricks. The change in microstructure from pseudo-hexagonal grains to equiaxed grains of well – ordered sequence of structural layers could be responsible for the improved properties.

Keywords: alumina, furnace, industry, manufacturing, refractoriness

Procedia PDF Downloads 247
22659 Research on Straightening Process Model Based on Iteration and Self-Learning

Authors: Hong Lu, Xiong Xiao

Abstract:

Shaft parts are widely used in machinery industry, however, bending deformation often occurred when this kind of parts is being heat treated. This parts needs to be straightened to meet the requirement of straightness. As for the pressure straightening process, a good straightening stroke algorithm is related to the precision and efficiency of straightening process. In this paper, the relationship between straightening load and deflection during the straightening process is analyzed, and the mathematical model of the straightening process has been established. By the mathematical model, the iterative method is used to solve the straightening stroke. Compared to the traditional straightening stroke algorithm, straightening stroke calculated by this method is much more precise; because it can adapt to the change of material performance parameters. Considering that the straightening method is widely used in the mass production of the shaft parts, knowledge base is used to store the data of the straightening process, and a straightening stroke algorithm based on empirical data is set up. In this paper, the straightening process control model which combine the straightening stroke method based on iteration and straightening stroke algorithm based on empirical data has been set up. Finally, an experiment has been designed to verify the straightening process control model.

Keywords: straightness, straightening stroke, deflection, shaft parts

Procedia PDF Downloads 321
22658 Numerical Analysis of Rainfall-Induced Roadside Slope Failures and Their Stabilizing Solution

Authors: Muhammad Suradi, Sugiarto, Abdullah Latip

Abstract:

Many roadside slope failures occur during the rainy season, particularly in the period of extreme rainfall along Connecting National Road of Salubatu-Mambi, West Sulawesi, Indonesia. These occurrences cause traffic obstacles and endanger people along and around the road. Research collaboration between P2JN (National Road Construction Board) West Sulawesi Province, who authorize to supervise the road condition, and Ujung Pandang State Polytechnic (Applied University) was established to cope with the landslide problem. This research aims to determine factors triggering roadside slope failures and their optimum stabilizing solution. To achieve this objective, site observation and soil investigation were carried out to obtain parameters for analyses of rainfall-induced slope instability and reinforcement design using the SV Flux and SV Slope software. The result of this analysis will be taken into account for the next analysis to get an optimum design of the slope reinforcement. The result indicates some factors such as steep slopes, sandy soils, and unvegetated slope surface mainly contribute to the slope failures during intense rainfall. With respect to the contributing factors as well as construction material and technology, cantilever/butressing retaining wall becomes the optimum solution for the roadside slope reinforcement.

Keywords: roadside slope, failure, rainfall, slope reinforcement, optimum solution

Procedia PDF Downloads 95
22657 Effects of Elastic, Plyometric and Strength Training on Selected Anaerobic Factors in Sanandaj Elite Volleyball Players

Authors: Majed Zobairy, Fardin Kalvandi, Kamal Azizbaigi

Abstract:

This research was carried out for evaluation of elastic, plyometric and resistance training on selected anaerobic factors in men volleyball players. For these reason 30 elite volleyball players of Sanandaj city randomly divided into 3 groups as follow: elastic training, plyometric training and resistance training. Pre-exercise tests which include vertical jumping, 50 yard speed running and scat test were done and data were recorded. Specific exercise protocol regimen was done for each group and then post-exercise tests again were done. Data analysis showed that there were significant increases in exercise test in each group. One way ANOVA analysis showed that increases in speed records in elastic group were significantly higher than the other groups (p<0/05),based on research data it seems that elastic training can be a useful method and new approach in improving functional test and training regimen.

Keywords: elastic training, plyometric training, strength training, anaerobic power

Procedia PDF Downloads 519
22656 Evaluation and Comparison of Seismic Performance of Structural Trusses under Cyclic Loading with Finite Element Method

Authors: Masoud Mahdavi

Abstract:

The structure is made using different members and combining them with each other. These members are basically based on technical and engineering principles and are combined in different ways and have their own unique effects on the building. Trusses are one of the most common and important members of the structure, accounting for a large percentage of the power transmission structure in the building. Different types of trusses are based on structural needs and evaluating and making complete comparisons between them is one of the most important engineering analyses. In the present study, four types of trusses have been studied; 1) Hawe truss, 2) Pratt truss, 3) k truss, and 4) warren truss, under cyclic loading for 80 seconds. The trusses are modeled in 3d using st37 steel. The results showed that Hawe trusses had higher values ​​than all other trusses (k, Pratt and Warren) in all the studied indicators. Indicators examined in the study include; 1) von Mises stresses, 2) displacement, 3) support force, 4) velocity, 5) acceleration, 6) capacity (hysteresis curve) and 7) energy diagram. Pratt truss in indicators; Mises stress, displacement, energy have the least amount compared to other trusses. K truss in indicators; support force, speed and acceleration are the lowest compared to other trusses.

Keywords: hawe truss, pratt truss, K truss, warren truss, cyclic loading, finite element method

Procedia PDF Downloads 139
22655 Integrating Data Envelopment Analysis and Variance Inflation Factor to Measure the Efficiency of Decision Making Units

Authors: Mostafa Kazemi, Zahra N. Farkhani

Abstract:

This paper proposes an integrated Data Envelopment Analysis (DEA) and Variance Inflation Factor (VIF) model for measuring the technical efficiency of decision making units. The model is validated using a set of 69% sales representatives’ dairy products. The analysis is done in two stages, in the first stage, VIF technique is used to distinguish independent effective factors of resellers, and in the second stage we used DEA for measuring efficiency for both constant and variable return to scales status. Further DEA is used to examine the utilization of environmental factors on efficiency. Results of this paper indicated an average managerial efficiency of 83% in the whole sales representatives’ dairy products. In addition, technical and scale efficiency were counted 96% and 80% respectively. 38% of sales representative have the technical efficiency of 100% and 72% of the sales representative in terms of managerial efficiency are quite efficient.High levels of relative efficiency indicate a good condition for sales representative efficiency.

Keywords: data envelopment analysis (DEA), relative efficiency, sales representatives’ dairy products, variance inflation factor (VIF)

Procedia PDF Downloads 556
22654 Circular Polarized and Surface Compatible Microstrip Array Antenna Design for Image and Telemetric Data Transfer in UAV and Armed UAV Systems

Authors: Kübra Taşkıran, Bahattin Türetken

Abstract:

In this paper, a microstrip array antenna with circular polarization at 2.4 GHz frequency has been designed using the in order to provide image and telemetric data transmission in Unmanned Aerial Vehicle and Armed Unmanned Aerial Vehicle Systems. In addition to the antenna design, the power divider design was made and the antennas were fed in phase. As a result of the analysis, it was observed that the antenna operates at a frequency of 2.4016 GHz with 12.2 dBi directing gain. In addition, this designed array antenna was transformed into a form compatible with the rocket surface used in A-UAV Systems, and analyzes were made. As a result of these analyzes, it has been observed that the antenna operates on the surface of the missile at a frequency of 2.372 GHz with a directivity gain of 10.2 dBi.

Keywords: cicrostrip array antenna, circular polarization, 2.4 GHz, image and telemetric data, transmission, surface compatible, UAV and armed UAV

Procedia PDF Downloads 97
22653 Prediction of Live Birth in a Matched Cohort of Elective Single Embryo Transfers

Authors: Mohsen Bahrami, Banafsheh Nikmehr, Yueqiang Song, Anuradha Koduru, Ayse K. Vuruskan, Hongkun Lu, Tamer M. Yalcinkaya

Abstract:

In recent years, we have witnessed an explosion of studies aimed at using a combination of artificial intelligence (AI) and time-lapse imaging data on embryos to improve IVF outcomes. However, despite promising results, no study has used a matched cohort of transferred embryos which only differ in pregnancy outcome, i.e., embryos from a single clinic which are similar in parameters, such as: morphokinetic condition, patient age, and overall clinic and lab performance. Here, we used time-lapse data on embryos with known pregnancy outcomes to see if the rich spatiotemporal information embedded in this data would allow the prediction of the pregnancy outcome regardless of such critical parameters. Methodology—We did a retrospective analysis of time-lapse data from our IVF clinic utilizing Embryoscope 100% of the time for embryo culture to blastocyst stage with known clinical outcomes, including live birth vs nonpregnant (embryos with spontaneous abortion outcomes were excluded). We used time-lapse data from 200 elective single transfer embryos randomly selected from January 2019 to June 2021. Our sample included 100 embryos in each group with no significant difference in patient age (P=0.9550) and morphokinetic scores (P=0.4032). Data from all patients were combined to make a 4th order tensor, and feature extraction were subsequently carried out by a tensor decomposition methodology. The features were then used in a machine learning classifier to classify the two groups. Major Findings—The performance of the model was evaluated using 100 random subsampling cross validation (train (80%) - test (20%)). The prediction accuracy, averaged across 100 permutations, exceeded 80%. We also did a random grouping analysis, in which labels (live birth, nonpregnant) were randomly assigned to embryos, which yielded 50% accuracy. Conclusion—The high accuracy in the main analysis and the low accuracy in random grouping analysis suggest a consistent spatiotemporal pattern which is associated with pregnancy outcomes, regardless of patient age and embryo morphokinetic condition, and beyond already known parameters, such as: early cleavage or early blastulation. Despite small samples size, this ongoing analysis is the first to show the potential of AI methods in capturing the complex morphokinetic changes embedded in embryo time-lapse data, which contribute to successful pregnancy outcomes, regardless of already known parameters. The results on a larger sample size with complementary analysis on prediction of other key outcomes, such as: euploidy and aneuploidy of embryos will be presented at the meeting.

Keywords: IVF, embryo, machine learning, time-lapse imaging data

Procedia PDF Downloads 88
22652 Association of Zinc with New Generation Cardiovascular Risk Markers in Childhood Obesity

Authors: Mustafa M. Donma, Orkide Donma

Abstract:

Zinc is a vital element required for growth and development. This fact makes zinc important, particularly for children. It maintains normal cellular structure and functions. This essential element appears to have protective effects against coronary artery disease and cardiomyopathy. Higher serum zinc levels are associated with lower risk of cardiovascular diseases (CVDs). There is a significant association between low serum zinc levels and heart failure. Zinc may be a potential biomarker of cardiovascular health. High sensitive cardiac troponin T (hs-cTnT) and cardiac myosin binding protein C (cMyBP-C) are new generation markers used for prediagnosis, diagnosis, and prognosis of CVDs. The aim of this study is to determine zinc as well as new generation cardiac markers profiles in children with normal body mass index (N-BMI), obese (OB), morbid obese (MO) children, and children with metabolic syndrome (MetS) findings. The association among them will also be investigated. Four study groups were constituted. The study protocol was approved by the institutional Ethics Committee of Tekirdag Namik Kemal University. Parents of the participants filled informed consent forms to participate in the study. Group 1 is composed of 44 children with N-BMI. Group 2 and Group 3 comprised 43 OB and 45 MO children, respectively. Forty-five MO children with MetS findings were included in Group 4. World Health Organization age- and sex-adjusted BMI percentile tables were used to constitute groups. These values were 15-85, 95-99, and above 99 for N-BMI, OB, and MO, respectively. Criteria for MetS findings were determined. Routine biochemical analyses, including zinc, were performed. High sensitive-cTnT and cMyBP-C concentrations were measured by kits based on enzyme-linked immunosorbent assay principle. Appropriate statistical tests within the scope of SPSS were used for the evaluation of the study data. p<0.05 was accepted as statistically significant. Four groups were matched for age and gender. Decreased zinc concentrations were measured in Groups 2, 3, and 4 compared to Group 1. Groups did not differ from one another in terms of hs-cTnT. There were statistically significant differences between cMyBP-C levels of MetS group and N-BMI as well as OB groups. There was an increasing trend going from N-BMI group to MetS group. There were statistically significant negative correlations between zinc and hs-cTnT as well as cMyBP-C concentrations in MetS group. In conclusion, inverse correlations detected between zinc and new generation cardiac markers (hs-TnT and cMyBP-C) have pointed out that decreased levels of this physiologically essential trace element accompany increased levels of hs-cTnT as well as cMyBP-C in children with MetS. This finding emphasizes that both zinc and these new generation cardiac markers may be evaluated as biomarkers of cardiovascular health during severe childhood obesity precipitated with MetS findings and also suggested as the messengers of the future risk in the adulthood periods of children with MetS.

Keywords: cardiac myosin binding protein-C, cardiovascular diseases, children, high sensitive cardiac troponin T, obesity

Procedia PDF Downloads 105
22651 Application of Gamma Frailty Model in Survival of Liver Cirrhosis Patients

Authors: Elnaz Saeedi, Jamileh Abolaghasemi, Mohsen Nasiri Tousi, Saeedeh Khosravi

Abstract:

Goals and Objectives: A typical analysis of survival data involves the modeling of time-to-event data, such as the time till death. A frailty model is a random effect model for time-to-event data, where the random effect has a multiplicative influence on the baseline hazard function. This article aims to investigate the use of gamma frailty model with concomitant variable in order to individualize the prognostic factors that influence the liver cirrhosis patients’ survival times. Methods: During the one-year study period (May 2008-May 2009), data have been used from the recorded information of patients with liver cirrhosis who were scheduled for liver transplantation and were followed up for at least seven years in Imam Khomeini Hospital in Iran. In order to determine the effective factors for cirrhotic patients’ survival in the presence of latent variables, the gamma frailty distribution has been applied. In this article, it was considering the parametric model, such as Exponential and Weibull distributions for survival time. Data analysis is performed using R software, and the error level of 0.05 was considered for all tests. Results: 305 patients with liver cirrhosis including 180 (59%) men and 125 (41%) women were studied. The age average of patients was 39.8 years. At the end of the study, 82 (26%) patients died, among them 48 (58%) were men and 34 (42%) women. The main cause of liver cirrhosis was found hepatitis 'B' with 23%, followed by cryptogenic with 22.6% were identified as the second factor. Generally, 7-year’s survival was 28.44 months, for dead patients and for censoring was 19.33 and 31.79 months, respectively. Using multi-parametric survival models of progressive and regressive, Exponential and Weibull models with regard to the gamma frailty distribution were fitted to the cirrhosis data. In both models, factors including, age, bilirubin serum, albumin serum, and encephalopathy had a significant effect on survival time of cirrhotic patients. Conclusion: To investigate the effective factors for the time of patients’ death with liver cirrhosis in the presence of latent variables, gamma frailty model with parametric distributions seems desirable.

Keywords: frailty model, latent variables, liver cirrhosis, parametric distribution

Procedia PDF Downloads 257