Search results for: data mining technique
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 30162

Search results for: data mining technique

27372 Incidence of Disasters and Coping Mechanism among Farming Households in South West Nigeria

Authors: Fawehinmi Olabisi Alaba, O. R. Adeniyi

Abstract:

Farming households faces lots of disaster which contribute to endemic poverty. Anticipated increases in extreme weather events will exacerbate this. Primary data was administered to farming household using multi-stage random sampling technique. The result of the analysis shows that majority of the respondents (69.9%) are male, have mean household size, years of formal education and age of 5±1.14, 6±3.41, and 51.06±10.43 respectively. The major (48.9%) type of disaster experienced is flooding. Major coping mechanism adopted is sourcing for support from family and friends. Age, education, experience, access to extension agent, and mitigation control method contribute significantly to vulnerability to disaster. The major adaptation method (62.3%) is construction of drainage. The study revealed that the coping mechanisms employed may become less effective as increasingly fragile livelihood systems struggle to withstand disaster shocks. Thus there is need for training of the farmers on measures to adapt to mitigate the shock from disasters.

Keywords: adaptation, disasters, flooding, vulnerability

Procedia PDF Downloads 261
27371 Development of Generalized Correlation for Liquid Thermal Conductivity of N-Alkane and Olefin

Authors: A. Ishag Mohamed, A. A. Rabah

Abstract:

The objective of this research is to develop a generalized correlation for the prediction of thermal conductivity of n-Alkanes and Alkenes. There is a minority of research and lack of correlation for thermal conductivity of liquids in the open literature. The available experimental data are collected covering the groups of n-Alkanes and Alkenes.The data were assumed to correlate to temperature using Filippov correlation. Nonparametric regression of Grace Algorithm was used to develop the generalized correlation model. A spread sheet program based on Microsoft Excel was used to plot and calculate the value of the coefficients. The results obtained were compared with the data that found in Perry's Chemical Engineering Hand Book. The experimental data correlated to the temperature ranged "between" 273.15 to 673.15 K, with R2 = 0.99.The developed correlation reproduced experimental data that which were not included in regression with absolute average percent deviation (AAPD) of less than 7 %. Thus the spread sheet was quite accurate which produces reliable data.

Keywords: N-Alkanes, N-Alkenes, nonparametric, regression

Procedia PDF Downloads 655
27370 Development of PVA/polypyrrole Scaffolds by Supercritical CO₂ for Its Application in Biomedicine

Authors: Antonio Montes, Antonio Cozar, Clara Pereyra, Diego Valor, Enrique Martinez de la Ossa

Abstract:

Tissues and organs can be damaged because of traumatism, congenital illnesses, or cancer and the traditional therapeutic alternatives, such as surgery, cannot usually completely repair the damaged tissues. Tissue engineering allows regeneration of the patient's tissues, reducing the problems caused by the traditional methods. Scaffolds, polymeric structures with interconnected porosity, can be promoted the proliferation and adhesion of the patient’s cells in the damaged area. Furthermore, by means of impregnation of the scaffold with beneficial active substances, tissue regeneration can be induced through a drug delivery process. The objective of the work is the fabrication of a PVA scaffold coated with Gallic Acid and polypyrrole through a one-step foaming and impregnation process using the SSI technique (Supercritical Solvent Impregnation). In this technique, supercritical CO₂ penetrates into the polymer chains producing the plasticization of the polymer. In the depressurization step a CO₂ cellular nucleation and growing to take place to an interconnected porous structure of the polymer. The foaming process using supercritical CO₂ as solvent and expansion agent presents advantages compared to the traditional scaffolds’ fabrication methods, such as the polymer’s high solubility in the solvent or the possibility of carrying out the process at a low temperature, avoiding the inactivation of the active substance. In this sense, the supercritical CO₂ avoids the use of organic solvents and reduces the solvent residues in the final product. Moreover, this process does not require long processing time that could cause the stratification of substance inside the scaffold reducing the therapeutic efficiency of the formulation. An experimental design has been carried out to optimize the SSI technique operating conditions, as well as a study of the morphological characteristics of the scaffold for its use in tissue engineerings, such as porosity, conductivity or the release profiles of the active substance. It has been proved that the obtained scaffolds are partially porous, conductors of electricity and are able to release Gallic Acid in the long term.

Keywords: scaffold, foaming, supercritical, PVA, polypyrrole, gallic acid

Procedia PDF Downloads 183
27369 Laser Powder Bed Fusion Awareness for Engineering Students in France and Qatar

Authors: Hiba Naccache, Rima Hleiss

Abstract:

Additive manufacturing AM or 3D printing is one of the pillars of Industry 4.0. Compared to traditional manufacturing, AM provides a prototype before production in order to optimize the design and avoid the stock market and uses strictly necessary material which can be recyclable, for the benefit of leaning towards local production, saving money, time and resources. Different types of AM exist and it has a broad range of applications across several industries like aerospace, automotive, medicine, education and else. The Laser Powder Bed Fusion (LPBF) is a metal AM technique that uses a laser to liquefy metal powder, layer by layer, to build a three-dimensional (3D) object. In industry 4.0 and aligned with the numbers 9 (Industry, Innovation and Infrastructure) and 12 (Responsible Production and Consumption) of the Sustainable Development Goals of the UNESCO 2030 Agenda, the AM’s manufacturers committed to minimizing the environmental impact by being sustainable in every production. The LPBF has several environmental advantages, like reduced waste production, lower energy consumption, and greater flexibility in creating components with lightweight and complex geometries. However, LPBF also have environmental drawbacks, like energy consumption, gas consumption and emissions. It is critical to recognize the environmental impacts of LPBF in order to mitigate them. To increase awareness and promote sustainable practices regarding LPBF, the researchers use the Elaboration Likelihood Model (ELM) theory where people from multiple universities in France and Qatar process information in two ways: peripherally and centrally. The peripheral campaigns use superficial cues to get attention, and the central campaigns provide clear and concise information. The authors created a seminar including a video showing LPBF production and a website with educational resources. The data is collected using questionnaire to test attitude about the public awareness before and after the seminar. The results reflected a great shift on the awareness toward LPBF and its impact on the environment. With no presence of similar research, to our best knowledge, this study will add to the literature on the sustainability of the LPBF production technique.

Keywords: additive manufacturing, laser powder bed fusion, elaboration likelihood model theory, sustainable development goals, education-awareness, France, Qatar, specific energy consumption, environmental impact, lightweight components

Procedia PDF Downloads 90
27368 Single Cell and Spatial Transcriptomics: A Beginners Viewpoint from the Conceptual Pipeline

Authors: Leo Nnamdi Ozurumba-Dwight

Abstract:

Messenger ribooxynucleic acid (mRNA) molecules are compositional, protein-based. These proteins, encoding mRNA molecules (which collectively connote the transcriptome), when analyzed by RNA sequencing (RNAseq), unveils the nature of gene expression in the RNA. The obtained gene expression provides clues of cellular traits and their dynamics in presentations. These can be studied in relation to function and responses. RNAseq is a practical concept in Genomics as it enables detection and quantitative analysis of mRNA molecules. Single cell and spatial transcriptomics both present varying avenues for expositions in genomic characteristics of single cells and pooled cells in disease conditions such as cancer, auto-immune diseases, hematopoietic based diseases, among others, from investigated biological tissue samples. Single cell transcriptomics helps conduct a direct assessment of each building unit of tissues (the cell) during diagnosis and molecular gene expressional studies. A typical technique to achieve this is through the use of a single-cell RNA sequencer (scRNAseq), which helps in conducting high throughput genomic expressional studies. However, this technique generates expressional gene data for several cells which lack presentations on the cells’ positional coordinates within the tissue. As science is developmental, the use of complimentary pre-established tissue reference maps using molecular and bioinformatics techniques has innovatively sprung-forth and is now used to resolve this set back to produce both levels of data in one shot of scRNAseq analysis. This is an emerging conceptual approach in methodology for integrative and progressively dependable transcriptomics analysis. This can support in-situ fashioned analysis for better understanding of tissue functional organization, unveil new biomarkers for early-stage detection of diseases, biomarkers for therapeutic targets in drug development, and exposit nature of cell-to-cell interactions. Also, these are vital genomic signatures and characterizations of clinical applications. Over the past decades, RNAseq has generated a wide array of information that is igniting bespoke breakthroughs and innovations in Biomedicine. On the other side, spatial transcriptomics is tissue level based and utilized to study biological specimens having heterogeneous features. It exposits the gross identity of investigated mammalian tissues, which can then be used to study cell differentiation, track cell line trajectory patterns and behavior, and regulatory homeostasis in disease states. Also, it requires referenced positional analysis to make up of genomic signatures that will be sassed from the single cells in the tissue sample. Given these two presented approaches to RNA transcriptomics study in varying quantities of cell lines, with avenues for appropriate resolutions, both approaches have made the study of gene expression from mRNA molecules interesting, progressive, developmental, and helping to tackle health challenges head-on.

Keywords: transcriptomics, RNA sequencing, single cell, spatial, gene expression.

Procedia PDF Downloads 125
27367 Solar Power Monitoring and Control System using Internet of Things

Authors: Oladapo Tolulope Ibitoye

Abstract:

It has become imperative to harmonize energy poverty alleviation and carbon footprint reduction. This is geared towards embracing independent power generation at local levels to reduce the popular ambiguity in the transmission of generated power. Also, it will contribute towards the total adoption of electric vehicles and direct current (DC) appliances that are currently flooding the global market. Solar power system is gaining momentum as it is now an affordable and less complex alternative to fossil fuel-based power generation. Although, there are many issues associated with solar power system, which resulted in deprivation of optimum working capacity. One of the key problems is inadequate monitoring of the energy pool from solar irradiance, which can then serve as a foundation for informed energy usage decisions and appropriate solar system control for effective energy pooling. The proposed technique utilized Internet of Things (IoT) in developing a system to automate solar irradiance pooling by controlling solar photovoltaic panels autonomously for optimal usage. The technique is potent with better solar irradiance exposure which results into 30% voltage pooling capacity than a system with static solar panels. The evaluation of the system show that the developed system possesses higher voltage pooling capacity than a system of static positioning of solar panel.

Keywords: solar system, internet of things, renewable energy, power monitoring

Procedia PDF Downloads 85
27366 Survey on Arabic Sentiment Analysis in Twitter

Authors: Sarah O. Alhumoud, Mawaheb I. Altuwaijri, Tarfa M. Albuhairi, Wejdan M. Alohaideb

Abstract:

Large-scale data stream analysis has become one of the important business and research priorities lately. Social networks like Twitter and other micro-blogging platforms hold an enormous amount of data that is large in volume, velocity and variety. Extracting valuable information and trends out of these data would aid in a better understanding and decision-making. Multiple analysis techniques are deployed for English content. Moreover, one of the languages that produce a large amount of data over social networks and is least analyzed is the Arabic language. The proposed paper is a survey on the research efforts to analyze the Arabic content in Twitter focusing on the tools and methods used to extract the sentiments for the Arabic content on Twitter.

Keywords: big data, social networks, sentiment analysis, twitter

Procedia PDF Downloads 580
27365 Estimating Current Suicide Rates Using Google Trends

Authors: Ladislav Kristoufek, Helen Susannah Moat, Tobias Preis

Abstract:

Data on the number of people who have committed suicide tends to be reported with a substantial time lag of around two years. We examine whether online activity measured by Google searches can help us improve estimates of the number of suicide occurrences in England before official figures are released. Specifically, we analyse how data on the number of Google searches for the terms “depression” and “suicide” relate to the number of suicides between 2004 and 2013. We find that estimates drawing on Google data are significantly better than estimates using previous suicide data alone. We show that a greater number of searches for the term “depression” is related to fewer suicides, whereas a greater number of searches for the term “suicide” is related to more suicides. Data on suicide related search behaviour can be used to improve current estimates of the number of suicide occurrences.

Keywords: nowcasting, search data, Google Trends, official statistics

Procedia PDF Downloads 361
27364 Experimental and Finite Element Analysis for Mechanics of Soil-Tool Interaction

Authors: A. Armin, R. Fotouhi, W. Szyszkowski

Abstract:

In this paper a 3-D finite element (FE) investigation of soil-blade interaction is described. The effects of blade’s shape and rake angle are examined both numerically and experimentally. The soil is considered as an elastic-plastic granular material with non-associated Drucker-Prager material model. Contact elements with different properties are used to mimic soil-blade sliding and soil-soil cutting phenomena. A separation criterion is presented and a procedure to evaluate the forces acting on the blade is given and discussed in detail. Experimental results were derived from tests using soil bin facility and instruments at the University of Saskatchewan. During motion of the blade, load cells collect data and send them to a computer. The measured forces using load cells had noisy signals which are needed to be filtered. The FE results are compared with experimental results for verification. This technique can be used in blade shape optimization and design of more complicated blade’s shape.

Keywords: finite element analysis, experimental results, blade force, soil-blade contact modeling

Procedia PDF Downloads 322
27363 The Causes and Effects of Delinquent Behaviour among Students in Juvenile Home: A Case Study of Osun State

Authors: Baleeqs, O. Adegoke, Adeola, O. Aburime

Abstract:

Juvenile delinquency is fast becoming one of the largest problems facing many societies due to many different factors ranging from parental factors to bullying at schools all which had led to different theoretical notions by different scholars. Delinquency is an illegal or immoral behaviour, especially by the young person who behaves in a way that is illegal or that society does not approve of. The purpose of the study was to investigate causes and effects of delinquent behaviours among adolescent in juvenile home in Osun State. A descriptive survey research type was employed. The random sampling technique was used to select 100 adolescents in Juvenile home in Osun State. Questionnaires were developed and given to them. The data collected from this study were analyzed using frequency counts and percentage for the demographic data in section A, while the two research hypotheses postulated for this study were tested using t-test statistics at the significance level of 0.05. Findings revealed that the greatest school effects of delinquent behaviours among adolescent in juvenile home in Osun by respondents were their aggressive behaviours. Findings revealed that there was a significant difference in the causes and effects of delinquent behaviours among adolescent in juvenile home in Osun State. It was also revealed that there was no significant difference in the causes and effects of delinquent behaviours among secondary school students in Osun based on gender. These recommendations were made in order to address the findings of this study: More number of teachers should be appointed in the observation home so that it will be possible to provide teaching to the different age group of delinquents. Developing the infrastructure facilities of short stay homes and observation home is a top priority. Proper counseling session’s interval is highly essential for these juveniles.

Keywords: behaviour, delinquency, juvenile, random sampling, statistical techniques, survey

Procedia PDF Downloads 193
27362 On the Network Packet Loss Tolerance of SVM Based Activity Recognition

Authors: Gamze Uslu, Sebnem Baydere, Alper K. Demir

Abstract:

In this study, data loss tolerance of Support Vector Machines (SVM) based activity recognition model and multi activity classification performance when data are received over a lossy wireless sensor network is examined. Initially, the classification algorithm we use is evaluated in terms of resilience to random data loss with 3D acceleration sensor data for sitting, lying, walking and standing actions. The results show that the proposed classification method can recognize these activities successfully despite high data loss. Secondly, the effect of differentiated quality of service performance on activity recognition success is measured with activity data acquired from a multi hop wireless sensor network, which introduces high data loss. The effect of number of nodes on the reliability and multi activity classification success is demonstrated in simulation environment. To the best of our knowledge, the effect of data loss in a wireless sensor network on activity detection success rate of an SVM based classification algorithm has not been studied before.

Keywords: activity recognition, support vector machines, acceleration sensor, wireless sensor networks, packet loss

Procedia PDF Downloads 477
27361 Analysis of Risks in Financing Agriculture a Case of Agricultural Cooperatives in Benue State, Nigeria

Authors: Odey Moses Ogah, Felix Terhemba Ikyereve

Abstract:

The study was carried out to analyzed risks in financing agriculture by agricultural cooperatives in Benue State, Nigeria. The study made use of research questionnaires for data collection. A multistage sampling technique was used to select a sample of 210 respondents from 21 agricultural cooperatives. Both descriptive and inferential statistics were employed in data analysis. Loan defaulting (66.7%) and reduction in savings by members (51.4%) were the major causes of risks faced by agricultural cooperatives in financing agriculture in the study area. Other causes include adverse changes in commodity prices (48.6%), disaster (45.7%), among others. It was found that risks adversely influence the profitability and competition of agricultural cooperatives (82.9%). Multiple regression analysis results showed that the coefficient of multiple determinations was 0.67, implying that the explanatory variables included in the model accounted for 67% of the variation in the level of profitability of agricultural cooperatives. The number of loans, average amount of loan and the interest rate were significant and important determinants of profitability of the cooperatives. The majority of the respondents (88.6%) made use of loan guarantors as a strategy of managing loan default/no repayment. It was found that the majority (70%) of the respondents were faced with the challenge of lack of insurance cover. The study recommends that agricultural cooperative officials should be encouraged to undergo formal training and education to easily acquire administrative skills in the management of agricultural loans; Farmer's loan size should be increased and released on time to enable them to use it effectively. Policies that enhance insuring farm activities should be put in place to discourage farmers from risk aversion.

Keywords: agriculture, analysis, cooperative, finance, risks

Procedia PDF Downloads 115
27360 GIS Data Governance: GIS Data Submission Process for Build-in Project, Replacement Project at Oman electricity Transmission Company

Authors: Rahma Saleh Hussein Al Balushi

Abstract:

Oman Electricity Transmission Company's (OETC) vision is to be a renowned world-class transmission grid by 2025, and one of the indications of achieving the vision is obtaining Asset Management ISO55001 certification, which required setting out a documented Standard Operating Procedures (SOP). Hence, documented SOP for the Geographical information system data process has been established. Also, to effectively manage and improve OETC power transmission, asset data and information need to be governed as such by Asset Information & GIS department. This paper will describe in detail the current GIS data submission process and the journey for developing it. The methodology used to develop the process is based on three main pillars, which are system and end-user requirements, Risk evaluation, data availability, and accuracy. The output of this paper shows the dramatic change in the used process, which results subsequently in more efficient, accurate, and updated data. Furthermore, due to this process, GIS has been and is ready to be integrated with other systems as well as the source of data for all OETC users. Some decisions related to issuing No objection certificates (NOC) for excavation permits and scheduling asset maintenance plans in Computerized Maintenance Management System (CMMS) have been made consequently upon GIS data availability. On the Other hand, defining agreed and documented procedures for data collection, data systems update, data release/reporting and data alterations has also contributed to reducing the missing attributes and enhance data quality index of GIS transmission data. A considerable difference in Geodatabase (GDB) completeness percentage was observed between the years 2017 and year 2022. Overall, concluding that by governance, asset information & GIS department can control the GIS data process; collect, properly record, and manage asset data and information within the OETC network. This control extends to other applications and systems integrated with/related to GIS systems.

Keywords: asset management ISO55001, standard procedures process, governance, CMMS

Procedia PDF Downloads 126
27359 Production of Energetic Nanomaterials by Spray Flash Evaporation

Authors: Martin Klaumünzer, Jakob Hübner, Denis Spitzer

Abstract:

Within this paper, latest results on processing of energetic nanomaterials by means of the Spray Flash Evaporation technique are presented. This technology constitutes a highly effective and continuous way to prepare fascinating materials on the nano- and micro-scale. Within the process, a solution is set under high pressure and sprayed into an evacuated atomization chamber. Subsequent ultrafast evaporation of the solvent leads to an aerosol stream, which is separated by cyclones or filters. No drying gas is required, so the present technique should not be confused with spray dying. Resulting nanothermites, insensitive explosives or propellants and compositions are foreseen to replace toxic (according to REACH) and very sensitive matter in military and civil applications. Diverse examples are given in detail: nano-RDX (n-Cyclotrimethylentrinitramin) and nano-aluminum based systems, mixtures (n-RDX/n-TNT - trinitrotoluene) or even cocrystalline matter like n-CL-20/HMX (Hexanitrohexaazaisowurtzitane/ Cyclotetra-methylentetranitramin). These nanomaterials show reduced sensitivity by trend without losing effectiveness and performance. An analytical study for material characterization was performed by using Atomic Force Microscopy, X-Ray Diffraction, and combined techniques as well as spectroscopic methods. As a matter of course, sensitivity tests regarding electrostatic discharge, impact, and friction are provided.

Keywords: continuous synthesis, energetic material, nanoscale, nanoexplosive, nanothermite

Procedia PDF Downloads 266
27358 Efects of Data Corelation in a Sparse-View Compresive Sensing Based Image Reconstruction

Authors: Sajid Abas, Jon Pyo Hong, Jung-Ryun Le, Seungryong Cho

Abstract:

Computed tomography and laminography are heavily investigated in a compressive sensing based image reconstruction framework to reduce the dose to the patients as well as to the radiosensitive devices such as multilayer microelectronic circuit boards. Nowadays researchers are actively working on optimizing the compressive sensing based iterative image reconstruction algorithm to obtain better quality images. However, the effects of the sampled data’s properties on reconstructed the image’s quality, particularly in an insufficient sampled data conditions have not been explored in computed laminography. In this paper, we investigated the effects of two data properties i.e. sampling density and data incoherence on the reconstructed image obtained by conventional computed laminography and a recently proposed method called spherical sinusoidal scanning scheme. We have found that in a compressive sensing based image reconstruction framework, the image quality mainly depends upon the data incoherence when the data is uniformly sampled.

Keywords: computed tomography, computed laminography, compressive sending, low-dose

Procedia PDF Downloads 465
27357 Fuzzy Wavelet Model to Forecast the Exchange Rate of IDR/USD

Authors: Tri Wijayanti Septiarini, Agus Maman Abadi, Muhammad Rifki Taufik

Abstract:

The exchange rate of IDR/USD can be the indicator to analysis Indonesian economy. The exchange rate as a important factor because it has big effect in Indonesian economy overall. So, it needs the analysis data of exchange rate. There is decomposition data of exchange rate of IDR/USD to be frequency and time. It can help the government to monitor the Indonesian economy. This method is very effective to identify the case, have high accurate result and have simple structure. In this paper, data of exchange rate that used is weekly data from December 17, 2010 until November 11, 2014.

Keywords: the exchange rate, fuzzy mamdani, discrete wavelet transforms, fuzzy wavelet

Procedia PDF Downloads 575
27356 Humanising Digital Healthcare to Build Capacity by Harnessing the Power of Patient Data

Authors: Durhane Wong-Rieger, Kawaldip Sehmi, Nicola Bedlington, Nicole Boice, Tamás Bereczky

Abstract:

Patient-generated health data should be seen as the expression of the experience of patients, including the outcomes reflecting the impact a treatment or service had on their physical health and wellness. We discuss how the healthcare system can reach a place where digital is a determinant of health - where data is generated by patients and is respected and which acknowledges their contribution to science. We explore the biggest barriers facing this. The International Experience Exchange with Patient Organisation’s Position Paper is based on a global patient survey conducted in Q3 2021 that received 304 responses. Results were discussed and validated by the 15 patient experts and supplemented with literature research. Results are a subset of this. Our research showed patient communities want to influence how their data is generated, shared, and used. Our study concludes that a reasonable framework is needed to protect the integrity of patient data and minimise abuse, and build trust. Results also demonstrated a need for patient communities to have more influence and control over how health data is generated, shared, and used. The results clearly highlight that the community feels there is a lack of clear policies on sharing data.

Keywords: digital health, equitable access, humanise healthcare, patient data

Procedia PDF Downloads 86
27355 Chelator-assisted Phytoextraction of Nickel from Nickeliferous Lateritic Soil by Phyllanthus sp. nov.

Authors: Grecco M. Ante, Princess Rochelle O. Gan

Abstract:

Plants that can absorb greater than 10,000 µg Ni/g dry mass in their stems and leaves are termed as ‘hypernickelophores’. Chelators are chemicals that make the metals in the soil more soluble, making them a potential enhancer for phytoextraction. This study aims to observe the effect of different concentrations of the chelating agent ethylene diamine tetraacetate (EDTA) on the metal uptake (or rate of phytoextraction) of Nickel by Phyllanthus sp. nov. The plant is found to be a hyperickelophore in normal conditions. The addition of EDTA increased the metal uptake of the plant. The increasing amount of the chelating agent causes a decrease in the phytoextraction of the plant but moves the onset of its peak of maximum nickel content in its tissue to an earlier time. The chelator-assisted phytoextraction of nickel by Phyllanthus sp. nov. is proven to be an efficient auxiliary mining operation for nickel laterite mines.

Keywords: phytomining, Phyllanthus sp. nov., EDTA, nickel, laterite

Procedia PDF Downloads 469
27354 Monitoring of Belt-Drive Defects Using the Vibration Signals and Simulation Models

Authors: A. Nabhan, Mohamed R. El-Sharkawy, A. Rashed

Abstract:

The main aim of this paper is to dedicate the belt drive system faults like cogs missing, misalignment and belt worm using vibration analysis technique. Experimentally, the belt drive test-rig is equipped to measure vibrations signals under different operating conditions. Finite element 3D model of belt drive system is created and vibration response analyzed using commercial finite element software ABAQUS/CAE.  Root mean square (RMS) and Crest Factor will serve as indicators of average amplitude of envelope analysis signals. The vibration signals pattern obtained from the simulation model and experimental data have the same characteristics. It can be concluded that each case of the RMS is more effective in detecting the defect for acceleration response. While Crest Factor parameter has a response with the displacement and velocity of vibration signals. Also it can be noticed that the model has difficulty in completing the solution when the misalignment angle is higher than 1 degree.

Keywords: simulation model, misalignment, cogs missing, vibration analysis

Procedia PDF Downloads 285
27353 Factor to Elicit Spatial Presence: Calmness

Authors: Nadia Diyana Mohd Muhaiyuddin, Dayang Rohaya Awang Rambli

Abstract:

The aim of our work is to identify whether user’s calmness can be a factor to elicit user’s spatial presence experience. Hence, a systematic mental model technique called repertory grid was selected to collect data because users can freely give their opinions in this approach. Three image-based virtual reality (IBVR) environments were created to satisfy the requirement of the repertory grid. Different virtual environments were necessary to allow users to compare and give feedback. Result was analyzed by using descriptive analysis through the SPSS software. The result revealed that ‘users feel calm’ is accepted as one of the factors that can elicit spatial presence. Users also highlighted five IBVR characteristics that could elicit spatial presence, namely, calm sound, calm content, calm color, calm story line, and the calm feeling of the user.

Keywords: spatial presence, presence, virtual reality, image-based virtual reality, human-computer interaction

Procedia PDF Downloads 290
27352 Why is the Recurrence Rate of Residual or Recurrent Disease Following Endoscopic Mucosal Resection (EMR) of the Oesophageal Dysplasia’s and T1 Tumours Higher in the Greater Midlands Cancer Network?

Authors: Harshadkumar Rajgor, Jeff Butterworth

Abstract:

Background: Barretts oesophagus increases the risk of developing oesophageal adenocarcinoma. Over the last 40 years, there has been a 6 fold increase in the incidence of oesophageal adenocarcinoma in the western world and the incidence rates are increasing at a greater rate than cancers of the colon, breast and lung. Endoscopic mucosal resection (EMR) is a relatively new technique being used by 2 centres in the greater midlands cancer network. EMR can be used for curative or staging purposes, for high-grade dysplasia’s and T1 tumours of the oesophagus. EMR is also suitable for those who are deemed high risk for oesophagectomy. EMR has a recurrence rate of 21% according to the Wiesbaden data. Method: A retrospective study of prospectively collected data was carried out involving 24 patients who had EMR for curative or staging purposes. Complications of residual or recurrent disease following EMR that required further treatment were investigated. Results: In 54% of cases residual or recurrent disease was suspected. 96% of patients were given clear and concise information regarding their diagnosis of high-grade dysplasia or T1 tumours. All 24 patients consulted the same specialist healthcare team. Conclusion: EMR is a safe and effective treatment for patients who have high-grade dysplasia and T1NO tumours. In 54% of cases residual or recurrent disease was suspected. Initially, only single resections were undertaken. Multiple resections are now being carried out to reduce the risk of recurrence. Complications from EMR remain low in this series and consisted of a single episode of post procedural bleeding.

Keywords: endoscopic mucosal resection, oesophageal dysplasia, T1 tumours, cancer network

Procedia PDF Downloads 318
27351 Use of Machine Learning in Data Quality Assessment

Authors: Bruno Pinto Vieira, Marco Antonio Calijorne Soares, Armando Sérgio de Aguiar Filho

Abstract:

Nowadays, a massive amount of information has been produced by different data sources, including mobile devices and transactional systems. In this scenario, concerns arise on how to maintain or establish data quality, which is now treated as a product to be defined, measured, analyzed, and improved to meet consumers' needs, which is the one who uses these data in decision making and companies strategies. Information that reaches low levels of quality can lead to issues that can consume time and money, such as missed business opportunities, inadequate decisions, and bad risk management actions. The step of selecting, identifying, evaluating, and selecting data sources with significant quality according to the need has become a costly task for users since the sources do not provide information about their quality. Traditional data quality control methods are based on user experience or business rules limiting performance and slowing down the process with less than desirable accuracy. Using advanced machine learning algorithms, it is possible to take advantage of computational resources to overcome challenges and add value to companies and users. In this study, machine learning is applied to data quality analysis on different datasets, seeking to compare the performance of the techniques according to the dimensions of quality assessment. As a result, we could create a ranking of approaches used, besides a system that is able to carry out automatically, data quality assessment.

Keywords: machine learning, data quality, quality dimension, quality assessment

Procedia PDF Downloads 151
27350 Early Diagnosis of Myocardial Ischemia Based on Support Vector Machine and Gaussian Mixture Model by Using Features of ECG Recordings

Authors: Merve Begum Terzi, Orhan Arikan, Adnan Abaci, Mustafa Candemir

Abstract:

Acute myocardial infarction is a major cause of death in the world. Therefore, its fast and reliable diagnosis is a major clinical need. ECG is the most important diagnostic methodology which is used to make decisions about the management of the cardiovascular diseases. In patients with acute myocardial ischemia, temporary chest pains together with changes in ST segment and T wave of ECG occur shortly before the start of myocardial infarction. In this study, a technique which detects changes in ST/T sections of ECG is developed for the early diagnosis of acute myocardial ischemia. For this purpose, a database of real ECG recordings that contains a set of records from 75 patients presenting symptoms of chest pain who underwent elective percutaneous coronary intervention (PCI) is constituted. 12-lead ECG’s of the patients were recorded before and during the PCI procedure. Two ECG epochs, which are the pre-inflation ECG which is acquired before any catheter insertion and the occlusion ECG which is acquired during balloon inflation, are analyzed for each patient. By using pre-inflation and occlusion recordings, ECG features that are critical in the detection of acute myocardial ischemia are identified and the most discriminative features for the detection of acute myocardial ischemia are extracted. A classification technique based on support vector machine (SVM) approach operating with linear and radial basis function (RBF) kernels to detect ischemic events by using ST-T derived joint features from non-ischemic and ischemic states of the patients is developed. The dataset is randomly divided into training and testing sets and the training set is used to optimize SVM hyperparameters by using grid-search method and 10fold cross-validation. SVMs are designed specifically for each patient by tuning the kernel parameters in order to obtain the optimal classification performance results. As a result of implementing the developed classification technique to real ECG recordings, it is shown that the proposed technique provides highly reliable detections of the anomalies in ECG signals. Furthermore, to develop a detection technique that can be used in the absence of ECG recording obtained during healthy stage, the detection of acute myocardial ischemia based on ECG recordings of the patients obtained during ischemia is also investigated. For this purpose, a Gaussian mixture model (GMM) is used to represent the joint pdf of the most discriminating ECG features of myocardial ischemia. Then, a Neyman-Pearson type of approach is developed to provide detection of outliers that would correspond to acute myocardial ischemia. Neyman – Pearson decision strategy is used by computing the average log likelihood values of ECG segments and comparing them with a range of different threshold values. For different discrimination threshold values and number of ECG segments, probability of detection and probability of false alarm values are computed, and the corresponding ROC curves are obtained. The results indicate that increasing number of ECG segments provide higher performance for GMM based classification. Moreover, the comparison between the performances of SVM and GMM based classification showed that SVM provides higher classification performance results over ECG recordings of considerable number of patients.

Keywords: ECG classification, Gaussian mixture model, Neyman–Pearson approach, support vector machine

Procedia PDF Downloads 163
27349 Microstructure of Ti – AlN Composite Produced by Selective Laser Melting

Authors: Jaroslaw Mizera, Pawel Wisniewski, Ryszard Sitek

Abstract:

Selective Laser Melting (SLM) is an advanced additive manufacturing technique used for producing parts made of wide range of materials such as: austenitic steel, titanium, nickel etc. In the our experiment we produced a Ti-AlN composite from a mixture of titanium and aluminum nitride respectively 70% at. and 30% at. using SLM technique. In order to define the size of powder particles, laser diffraction tests were performed on HORIBA LA-950 device. The microstructure and chemical composition of the composite was examined by Scanning Electron Microscopy (SEM). The chemical composition in micro areas of the obtained samples was determined by of EDS. The phase composition was analyzed by X-ray phase analysis (XRD). Microhardness Vickers tests were performed using Zwick/Roell microhardness machine under the load of 0.2kG (HV0.2). Hardness measurements were made along the building (xy) and along the plane of the lateral side of the cuboid (xz). The powder used for manufacturing of the samples had a mean particle size of 41μm. It was homogenous with a spherical shape. The specimens were built chiefly from Ti, TiN and AlN. The dendritic microstructure was porous and fine-grained. Some of the aluminum nitride remained unmelted but no porosity was observed in the interface. The formed material was characterized by high hardness exceeding 700 HV0.2 over the entire cross-section.

Keywords: Selective Laser Melting, Composite, SEM, microhardness

Procedia PDF Downloads 141
27348 Exploring Data Leakage in EEG Based Brain-Computer Interfaces: Overfitting Challenges

Authors: Khalida Douibi, Rodrigo Balp, Solène Le Bars

Abstract:

In the medical field, applications related to human experiments are frequently linked to reduced samples size, which makes the training of machine learning models quite sensitive and therefore not very robust nor generalizable. This is notably the case in Brain-Computer Interface (BCI) studies, where the sample size rarely exceeds 20 subjects or a few number of trials. To address this problem, several resampling approaches are often used during the data preparation phase, which is an overly critical step in a data science analysis process. One of the naive approaches that is usually applied by data scientists consists in the transformation of the entire database before the resampling phase. However, this can cause model’ s performance to be incorrectly estimated when making predictions on unseen data. In this paper, we explored the effect of data leakage observed during our BCI experiments for device control through the real-time classification of SSVEPs (Steady State Visually Evoked Potentials). We also studied potential ways to ensure optimal validation of the classifiers during the calibration phase to avoid overfitting. The results show that the scaling step is crucial for some algorithms, and it should be applied after the resampling phase to avoid data leackage and improve results.

Keywords: data leackage, data science, machine learning, SSVEP, BCI, overfitting

Procedia PDF Downloads 154
27347 Hemodialysis Technique in a Diabetic Population

Authors: Daniel Thompson, Sophie Cerutti, Muhammad Peerbux, Hansraj Bookun

Abstract:

Introduction: Diabetic nephropathy is the leading cause end stage renal failure in Australia, responsible for 36% of cases. Patients who require dialysis may be suitable for haemodialysis through an arteriovenous fistula (AVF), and preoperatively careful planning is required to select suitable vessels for a long-lasting fistula that provides suitable dialysis access. Due to high levels of vascular disease in diabetic patients, we sought to investigate whether there is a difference in the types of autologous AVFs created for diabetic patients in renal failure compared to their non-diabetic counterparts. Method: Data was collected from the Australasian Vascular Audit, for all vascular surgery completed at St. Vincent’s Hospital Melbourne between 2011-2020. Patients were selected by operative type, creation of AVF, and compared in two groups, diabetic patients and patients without diabetes. Chi-squared test was utilised to determine significance. Results: Data analysis is ongoing and will be complete with updated abstract in time for the conference. Discussion: Diabetic nephropathy is the cause for roughly a third of end stage renal failure in Australia. Diabetic patients present with a unique set of challenges when it comes to dialysis access due to increased risk of peripheral vascular disease and arterial calcification. Care must be taken in the creation of fistulas to minimise complications and increase the chance of long-lasting access. Our study investigates the difference in autologous AVFs between diabetics and non-diabetics, and results may be used to influence location of fistula creation. Further research may be used to investigate patency rates of fistulas in diabetics vs non-diabetics which would further influence treatment decisions.

Keywords: dialysis, diabetes, renal access, fistula

Procedia PDF Downloads 140
27346 An Entropy Stable Three Dimensional Ideal MHD Solver with Guaranteed Positive Pressure

Authors: Andrew R. Winters, Gregor J. Gassner

Abstract:

A high-order numerical magentohydrodynamics (MHD) solver built upon a non-linear entropy stable numerical flux function that supports eight traveling wave solutions will be described. The method is designed to treat the divergence-free constraint on the magnetic field in a similar fashion to a hyperbolic divergence cleaning technique. The solver is especially well-suited for flows involving strong discontinuities due to its strong stability without the need to enforce artificial low density or energy limits. Furthermore, a new formulation of the numerical algorithm to guarantee positivity of the pressure during the simulation is described and presented. By construction, the solver conserves mass, momentum, and energy and is entropy stable. High spatial order is obtained through the use of a third order limiting technique. High temporal order is achieved by utilizing the family of strong stability preserving (SSP) Runge-Kutta methods. Main attributes of the solver are presented as well as details on an implementation of the new solver into the multi-physics, multi-scale simulation code FLASH. The accuracy, robustness, and computational efficiency is demonstrated with a variety of numerical tests. Comparisons are also made between the new solver and existing methods already present in FLASH framework.

Keywords: entropy stability, finite volume scheme, magnetohydrodynamics, pressure positivity

Procedia PDF Downloads 344
27345 Software Tool Design for Heavy Oil Upgrading by Hydrogen Donor Addition in a Hydrodynamic Cavitation Process

Authors: Munoz A. Tatiana, Solano R. Brandon, Montes C. Juan, Cierco G. Javier

Abstract:

The hydrodynamic cavitation is a process in which the energy that the fluids have in the phase changes is used. From this energy, local temperatures greater than 5000 °C are obtained where thermal cracking of the fluid molecules takes place. The process applied to heavy oil affects variables such as viscosity, density, and composition, which constitutes an important improvement in the quality of crude oil. In this study, the need to design a software through mathematical integration models of mixing, cavitation, kinetics, and reactor, allows modeling changes in density, viscosity, and composition of a heavy oil crude, when the fluid passes through a hydrodynamic cavitation reactor. In order to evaluate the viability of this technique in the industry, a heavy oil of 18° API gravity, was simulated using naphtha as a hydrogen donor at concentrations of 1, 2 and 5% vol, where the simulation results showed an API gravity increase to 0.77, 1.21 and 1.93° respectively and a reduction viscosity by 9.9, 12.9 and 15.8%. The obtained results allow to have a favorable panorama on this technological development, an appropriate visualization on the generation of innovative knowledge of this technique and the technical-economic opportunity that benefits the development of the hydrocarbon sector related to heavy crude oil that includes the largest world oil production.

Keywords: hydrodynamic cavitation, thermal cracking, hydrogen donor, heavy oil upgrading, simulator

Procedia PDF Downloads 152
27344 Teaching Kindness as Moral Virtue in Preschool Children: The Effectiveness of Picture-Storybook Reading and Hand-Puppet Storytelling

Authors: Rose Mini Agoes Salim, Shahnaz Safitri

Abstract:

The aim of this study is to test the effectiveness of teaching kindness in preschool children by using several techniques. Kindness is a physical act or emotional support aimed to build or maintain relationships with others. Kindness is known to be essential in the development of moral reasoning to distinguish between the good and bad things. In this study, kindness is operationalized as several acts including helping friends, comforting sad friends, inviting friends to play, protecting others, sharing, saying hello, saying thank you, encouraging others, and apologizing. It is mentioned that kindness is crucial to be developed in preschool children because this is the time the children begin to interact with their social environment through play. Furthermore, preschool children's cognitive development makes them begin to represent the world with words, which then allows them to interact with others. On the other hand, preschool children egocentric thinking makes them still need to learn to consider another person's perspective. In relation to social interaction, preschool children need to be stimulated and assisted by adult to be able to pay attention to other and act with kindness toward them. On teaching kindness to children, the quality of interaction between children and their significant others is the key factor. It is known that preschool children learn about kindness by imitating adults on their two way interaction. Specifically, this study examines two types of teaching techniques that can be done by parents as a way to teach kindness, namely the picture-storybook reading and hand-puppet storytelling. These techniques were examined because both activities are easy to do and both also provide a model of behavior for the child based on the character in the story. To specifically examine those techniques effectiveness in teaching kindness, two studies were conducted. Study I involves 31 children aged 5-6 years old with picture-storybook reading technique, where the intervention is done by reading 8 picture books for 8 days. In study II, hand-puppet storytelling technique is examined to 32 children aged 3-5 years old. The treatments effectiveness are measured using an instrument in the form of nine colored cards that describe the behavior of kindness. Data analysis using Wilcoxon Signed-rank test shows a significant difference on the average score of kindness (p < 0.05) before and after the intervention has been held. For daily observation, a ‘kindness tree’ and observation sheets are used which are filled out by the teacher. Two weeks after interventions, an improvement on all kindness behaviors measured is intact. The same result is also gained from both ‘kindness tree’ and observational sheets.

Keywords: kindness, moral teaching, storytelling, hand puppet

Procedia PDF Downloads 253
27343 Genetic Algorithms for Feature Generation in the Context of Audio Classification

Authors: José A. Menezes, Giordano Cabral, Bruno T. Gomes

Abstract:

Choosing good features is an essential part of machine learning. Recent techniques aim to automate this process. For instance, feature learning intends to learn the transformation of raw data into a useful representation to machine learning tasks. In automatic audio classification tasks, this is interesting since the audio, usually complex information, needs to be transformed into a computationally convenient input to process. Another technique tries to generate features by searching a feature space. Genetic algorithms, for instance, have being used to generate audio features by combining or modifying them. We find this approach particularly interesting and, despite the undeniable advances of feature learning approaches, we wanted to take a step forward in the use of genetic algorithms to find audio features, combining them with more conventional methods, like PCA, and inserting search control mechanisms, such as constraints over a confusion matrix. This work presents the results obtained on particular audio classification problems.

Keywords: feature generation, feature learning, genetic algorithm, music information retrieval

Procedia PDF Downloads 437