Search results for: food composition data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 28931

Search results for: food composition data

24371 Changing Patterns of Colorectal Cancer in Hail Region

Authors: Laila Salah Seada, Ashraf Ibrahim, Fawaz Al Rashid, Ihab Abdo, Hassan Kasim, Waleed Al Mansi, Saud Al Shabli

Abstract:

Background and Objectives: Colorectal carcinoma is increasing among both men and women worldwide. It has a multifactorial etiology including genetic factors, environmental factors and inflammatory conditions of the digestive tract. A clinicopathologic assessment of colorectal carcinoma in Hail region is done, considering any changing patterns in two 5-year periods from 2005-2009 (A) and from 2012 to 2017 (B). All data had been retrieved from histopathology files of King Khalid Hospital, Hail. Results: During period (A), 75 cases were diagnosed as colorectal carcinoma. Male patients comprised 56/75 (74.7%) of the study, with a mean age of 58.4 (36-97), while females were 19/75 (25.3%) with a mean age of 50.3(30-85) and the difference was significant (p = 0.05). M:F ratio was 2.9:1. Most common histological type was adenocarcioma in 68/75 (90.7%) patients mostly well differentiated in 44/68 (64.7%). Mucinous neoplasms comprised only 7/75 (9.3%) of cases and tended to have a higher stage (p = 0.04). During period (B), 115 cases were diagnosed with an increase of 53.3% in number of cases than period (A). Male to female ratio also decreased to 1.35:1, females being 44.83% more affected. Adenocarcinoma remained the prevalent type (93.9%), while mucinous type was still rare (5.2%). No distal metastases found at time of presentation. Localization of tumors was rectosigmoid in group (A) in 41.4%, which increased to 56.6% in group (B), with an increase of 15.2%. Iliocecal location also decreased from 8% to 3.5%, being 56.25% less. Other proximal areas of the colon were decreased by 25.75%, from 53.9% in group (A) to 40% in group (B). Conclusion: Colorectal carcinoma in Hail region has increased by 53.3% in the past 5 years, with more females being diagnosed. Localization has also shifted distally by 15.2%. These findings are different from Western world patterns which experienced a decrease in incidence and proximal shift of the colon cancer localization. This might be due to better diagnostic tools, population awareness of the disease, as well as changing of life style and/or food habits in the region.

Keywords: colorectal cancer, Hail Region, changing pattern, distal shift

Procedia PDF Downloads 186
24370 Multiple Query Optimization in Wireless Sensor Networks Using Data Correlation

Authors: Elaheh Vaezpour

Abstract:

Data sensing in wireless sensor networks is done by query deceleration the network by the users. In many applications of the wireless sensor networks, many users send queries to the network simultaneously. If the queries are processed separately, the network’s energy consumption will increase significantly. Therefore, it is very important to aggregate the queries before sending them to the network. In this paper, we propose a multiple query optimization framework based on sensors physical and temporal correlation. In the proposed method, queries are merged and sent to network by considering correlation among the sensors in order to reduce the communication cost between the sensors and the base station.

Keywords: wireless sensor networks, multiple query optimization, data correlation, reducing energy consumption

Procedia PDF Downloads 320
24369 Efficient Tuning Parameter Selection by Cross-Validated Score in High Dimensional Models

Authors: Yoonsuh Jung

Abstract:

As DNA microarray data contain relatively small sample size compared to the number of genes, high dimensional models are often employed. In high dimensional models, the selection of tuning parameter (or, penalty parameter) is often one of the crucial parts of the modeling. Cross-validation is one of the most common methods for the tuning parameter selection, which selects a parameter value with the smallest cross-validated score. However, selecting a single value as an "optimal" value for the parameter can be very unstable due to the sampling variation since the sample sizes of microarray data are often small. Our approach is to choose multiple candidates of tuning parameter first, then average the candidates with different weights depending on their performance. The additional step of estimating the weights and averaging the candidates rarely increase the computational cost, while it can considerably improve the traditional cross-validation. We show that the selected value from the suggested methods often lead to stable parameter selection as well as improved detection of significant genetic variables compared to the tradition cross-validation via real data and simulated data sets.

Keywords: cross validation, parameter averaging, parameter selection, regularization parameter search

Procedia PDF Downloads 400
24368 Digital Image Steganography with Multilayer Security

Authors: Amar Partap Singh Pharwaha, Balkrishan Jindal

Abstract:

In this paper, a new method is developed for hiding image in a digital image with multilayer security. In the proposed method, the secret image is encrypted in the first instance using a flexible matrix based symmetric key to add first layer of security. Then another layer of security is added to the secret data by encrypting the ciphered data using Pythagorean Theorem method. The ciphered data bits (4 bits) produced after double encryption are then embedded within digital image in the spatial domain using Least Significant Bits (LSBs) substitution. To improve the image quality of the stego-image, an improved form of pixel adjustment process is proposed. To evaluate the effectiveness of the proposed method, image quality metrics including Peak Signal-to-Noise Ratio (PSNR), Mean Square Error (MSE), entropy, correlation, mean value and Universal Image Quality Index (UIQI) are measured. It has been found experimentally that the proposed method provides higher security as well as robustness. In fact, the results of this study are quite promising.

Keywords: Pythagorean theorem, pixel adjustment, ciphered data, image hiding, least significant bit, flexible matrix

Procedia PDF Downloads 324
24367 MapReduce Logistic Regression Algorithms with RHadoop

Authors: Byung Ho Jung, Dong Hoon Lim

Abstract:

Logistic regression is a statistical method for analyzing a dataset in which there are one or more independent variables that determine an outcome. Logistic regression is used extensively in numerous disciplines, including the medical and social science fields. In this paper, we address the problem of estimating parameters in the logistic regression based on MapReduce framework with RHadoop that integrates R and Hadoop environment applicable to large scale data. There exist three learning algorithms for logistic regression, namely Gradient descent method, Cost minimization method and Newton-Rhapson's method. The Newton-Rhapson's method does not require a learning rate, while gradient descent and cost minimization methods need to manually pick a learning rate. The experimental results demonstrated that our learning algorithms using RHadoop can scale well and efficiently process large data sets on commodity hardware. We also compared the performance of our Newton-Rhapson's method with gradient descent and cost minimization methods. The results showed that our newton's method appeared to be the most robust to all data tested.

Keywords: big data, logistic regression, MapReduce, RHadoop

Procedia PDF Downloads 259
24366 Iterative Panel RC Extraction for Capacitive Touchscreen

Authors: Chae Hoon Park, Jong Kang Park, Jong Tae Kim

Abstract:

Electrical characteristics of capacitive touchscreen need to be accurately analyzed to result in better performance for multi-channel capacitance sensing. In this paper, we extracted the panel resistances and capacitances of the touchscreen by comparing measurement data and model data. By employing a lumped RC model for driver-to-receiver paths in touchscreen, we estimated resistance and capacitance values according to the physical lengths of channel paths which are proportional to the RC model. As a result, we obtained the model having 95.54% accuracy of the measurement data.

Keywords: electrical characteristics of capacitive touchscreen, iterative extraction, lumped RC model, physical lengths of channel paths

Procedia PDF Downloads 324
24365 Combining Shallow and Deep Unsupervised Machine Learning Techniques to Detect Bad Actors in Complex Datasets

Authors: Jun Ming Moey, Zhiyaun Chen, David Nicholson

Abstract:

Bad actors are often hard to detect in data that imprints their behaviour patterns because they are comparatively rare events embedded in non-bad actor data. An unsupervised machine learning framework is applied here to detect bad actors in financial crime datasets that record millions of transactions undertaken by hundreds of actors (<0.01% bad). Specifically, the framework combines ‘shallow’ (PCA, Isolation Forest) and ‘deep’ (Autoencoder) methods to detect outlier patterns. Detection performance analysis for both the individual methods and their combination is reported.

Keywords: detection, machine learning, deep learning, unsupervised, outlier analysis, data science, fraud, financial crime

Procedia PDF Downloads 76
24364 A Dissolution Mechanism of the Silicon Carbide in HF/K₂Cr₂O₇ Solutions

Authors: Karima Bourenane, Aissa Keffous

Abstract:

In this paper, we present an experimental method on the etching reaction of p-type 6H-SiC, etching that was carried out in HF/K₂Cr₂O₇ solutions. The morphology of the etched surface was examined with varying K₂Cr₂O₇ concentrations, etching time and temperature solution. The surfaces of the etched samples were analyzed using Scanning electron microscopy (SEM), Fourier transform infrared spectroscopy (FT-IR) and Photoluminescence. The surface morphology of samples etched in HF/K₂Cr₂O₇ is shown to depend on the solution composition and bath temperature. The investigation of the HF/K₂Cr₂O₇ solutions on 6H-SiC surface shows that as K₂Cr₂O₇ concentration increases, the etch rate increases to reach a maximum value at about 0.75 M and then decreases. Similar behavior has been observed when the temperature of the solution is increased. The maximum etch rate is found for 80 °C. Taking into account the result, a polishing etching solution of 6H-SiC has been developed. In addition, the result is very interesting when, to date, no chemical polishing solution has been developed on silicon carbide (SiC). Finally, we have proposed a dissolution mechanism of the silicon carbide in HF/K₂Cr₂O₇ solutions.

Keywords: silicon carbide, dissolution, Chemical etching, mechanism

Procedia PDF Downloads 36
24363 Machine Learning Development Audit Framework: Assessment and Inspection of Risk and Quality of Data, Model and Development Process

Authors: Jan Stodt, Christoph Reich

Abstract:

The usage of machine learning models for prediction is growing rapidly and proof that the intended requirements are met is essential. Audits are a proven method to determine whether requirements or guidelines are met. However, machine learning models have intrinsic characteristics, such as the quality of training data, that make it difficult to demonstrate the required behavior and make audits more challenging. This paper describes an ML audit framework that evaluates and reviews the risks of machine learning applications, the quality of the training data, and the machine learning model. We evaluate and demonstrate the functionality of the proposed framework by auditing an steel plate fault prediction model.

Keywords: audit, machine learning, assessment, metrics

Procedia PDF Downloads 251
24362 Efficient Sampling of Probabilistic Program for Biological Systems

Authors: Keerthi S. Shetty, Annappa Basava

Abstract:

In recent years, modelling of biological systems represented by biochemical reactions has become increasingly important in Systems Biology. Biological systems represented by biochemical reactions are highly stochastic in nature. Probabilistic model is often used to describe such systems. One of the main challenges in Systems biology is to combine absolute experimental data into probabilistic model. This challenge arises because (1) some molecules may be present in relatively small quantities, (2) there is a switching between individual elements present in the system, and (3) the process is inherently stochastic on the level at which observations are made. In this paper, we describe a novel idea of combining absolute experimental data into probabilistic model using tool R2. Through a case study of the Transcription Process in Prokaryotes we explain how biological systems can be written as probabilistic program to combine experimental data into the model. The model developed is then analysed in terms of intrinsic noise and exact sampling of switching times between individual elements in the system. We have mainly concentrated on inferring number of genes in ON and OFF states from experimental data.

Keywords: systems biology, probabilistic model, inference, biology, model

Procedia PDF Downloads 330
24361 A Real-time Classification of Lying Bodies for Care Application of Elderly Patients

Authors: E. Vazquez-Santacruz, M. Gamboa-Zuniga

Abstract:

In this paper, we show a methodology for bodies classification in lying state using HOG descriptors and pressures sensors positioned in a matrix form (14 x 32 sensors) on the surface where bodies lie down. it will be done in real time. Our system is embedded in a care robot that can assist the elderly patient and medical staff around to get a better quality of life in and out of hospitals. Due to current technology a limited number of sensors is used, wich results in low-resolution data array, that will be used as image of 14 x 32 pixels. Our work considers the problem of human posture classification with few information (sensors), applying digital process to expand the original data of the sensors and so get more significant data for the classification, however, this is done with low-cost algorithms to ensure the real-time execution.

Keywords: real-time classification, sensors, robots, health care, elderly patients, artificial intelligence

Procedia PDF Downloads 851
24360 Disidentification of Historical City Centers: A Comparative Study of the Old and New Settlements of Mardin, Turkey

Authors: Fatma Kürüm Varolgüneş, Fatih Canan

Abstract:

Mardin is one of the unique cities in Turkey with its rich cultural and historical heritage. Mardin’s traditional dwellings have been affected both by natural data such as climate and topography and by cultural data like lifestyle and belief. However, in the new settlements, housing is formed with modern approaches and unsuitable forms clashing with Mardin’s culture and environment. While the city is expanding, traditional textures are ignored. Thus, traditional settlements are losing their identity and are vanishing because of the rapid change and transformation. The main aim of this paper is to determine the physical and social data needed to define the characteristic features of Mardin’s old and new settlements. In this context, based on social and cultural data, old and new settlement formations of Mardin have been investigated from various aspects. During this research, the following methods have been utilized: observations, interviews, public surveys, literature review, as well as site examination via maps, photographs and questionnaire methodology. In conclusion, this paper focuses on how changes in the physical forms of cities affect the typology and the identity of cities, as in the case of Mardin.

Keywords: urban and local identity, historical city center, traditional settlements, Mardin

Procedia PDF Downloads 313
24359 Pediatric Hearing Aid Use: A Study Based on Data Logging Information

Authors: Mina Salamatmanesh, Elizabeth Fitzpatrick, Tim Ramsay, Josee Lagacé, Lindsey Sikora, JoAnne Whittingham

Abstract:

Introduction: Hearing loss (HL) is one of the most common disorders that presents at birth and in early childhood. Universal newborn hearing screening (UNHS) has been adopted based on the assumption that with early identification of HL, children will have access to optimal amplification and intervention at younger ages, therefore, taking advantage of the brain’s maximal plasticity. One particular challenge for parents in the early years is achieving consistent hearing aid (HA) use which is critical to the child’s development and constitutes the first step in the rehabilitation process. This study examined the consistency of hearing aid use in young children based on data logging information documented during audiology sessions in the first three years after hearing aid fitting. Methodology: The first 100 children who were diagnosed with bilateral HL before 72 months of age since 2003 to 2015 in a pediatric audiology clinic and who had at least two hearing aid follow-up sessions with available data logging information were included in the study. Data from each audiology session (age of child at the session, average hours of use per day (for each ear) in the first three years after HA fitting) were collected. Clinical characteristics (degree of hearing loss, age of HA fitting) were also documented to further understanding of factors that impact HA use. Results: Preliminary analysis of the results of the first 20 children shows that all of them (100%) have at least one data logging session recorded in the clinical audiology system (Noah). Of the 20 children, 17(85%) have three data logging events recorded in the first three years after HA fitting. Based on the statistical analysis of the first 20 cases, the median hours of use in the first follow-up session after the hearing aid fitting in the right ear is 3.9 hours with an interquartile range (IQR) of 10.2h. For the left ear the median is 4.4 and the IQR is 9.7h. In the first session 47% of the children use their hearing aids ≤5 hours, 12% use them between 5 to 10 hours and 22% use them ≥10 hours a day. However, these children showed increased use by the third follow-up session with a median (IQR) of 9.1 hours for the right ear and 2.5, and of 8.2 hours for left ear (IQR) IQR is 5.6 By the third follow-up session, 14% of children used hearing aids ≤5 hours, while 38% of children used them ≥10 hours. Based on the primary results, factors like age and level of HL significantly impact the hours of use. Conclusion: The use of data logging information to assess the actual hours of HA provides an opportunity to examine the: a) challenges of families of young children with HAs, b) factors that impact use in very young children. Data logging when used collaboratively with parents, can be a powerful tool to identify problems and to encourage and assist families in maximizing their child’s hearing potential.

Keywords: hearing loss, hearing aid, data logging, hours of use

Procedia PDF Downloads 213
24358 Towards End-To-End Disease Prediction from Raw Metagenomic Data

Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker

Abstract:

Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.

Keywords: deep learning, disease prediction, end-to-end machine learning, metagenomics, multiple instance learning, precision medicine

Procedia PDF Downloads 106
24357 Public and Private Involvement in Agricultural Extension Services: Factors of Farmers’ Preference in Southwestern Nigeria

Authors: S. O. Ayansina, O. A. Adekunle

Abstract:

There is an increasing demand for a functional extension delivery services in Nigeria with a view to meet up with the food and fiber needs of the ever growing population of human and animal respectively. The study was therefore designed to examine the farmers’ preference for public and private extension services in Southwestern Nigeria, specifically to determine the farmers’ level of participation in the two types of organizations involved and also to evaluate the Performance level of personnel in the two organizations in order to ascertain the beneficiaries’ satisfaction. A multi-stage random sampling technique was used to samples 30 respondents from each of the three selected organizations in each of the three states sampled in Southwestern Nigeria. Hence, 270 respondents were sampled for the study. Data collected were analyzed with Kruskal Wallis one-way Analysis of variance to test the difference between the participation of beneficiaries in the public and private extension services and the level of benefit accrued from the two organizations involved in the study. Results generally revealed that private organizations were performing better and were more preferred by the beneficiaries. Results of the tested hypotheses as shown by Kruskal Wallis test of difference (x2=0.709) indicates no significant difference between farmers’ participation in the extension services of public and private organizations but however shows significant difference (X2=12.074) in the benefits achieved by respondents in the two organizations, such benefits include: increased quantity of Crop produced, farm income, skill acquisition, and improved Education in private extension organizations. Based on this result, it could be inferred that beneficiaries generally preferred private extension organizations because of their effectiveness and vibrancy in programme administration. Public extension is therefore recommended for general overhauling and possibly “merging” of public and private sectors in order to cater for teeming population of farmers demanding for efficient and functional extension services to better their lots both in production and processing.

Keywords: public and private involvement, extension services, farmers’ preferences, Kruskal Wallis Test

Procedia PDF Downloads 313
24356 The Role of Waqf Forestry for Sustainable Economic Development: A Panel Logit Analysis

Authors: Patria Yunita

Abstract:

Kuznets’ environmental curve analysis suggests sacrificing economic development to reduce environmental problems. However, we hope to achieve sustainable economic development. In this case, Islamic social finance, especially that of waqf in Indonesia, can be used as a solution to bridge the problem of environmental damage to the sustainability of economic development. The Panel Logit Regression method was used to analyze the probability of increasing economic growth and the role of waqf in the environmental impact of CO₂ emissions. This study uses panel data from 33 Indonesian provinces. The data used were the National Waqf Index, Forest Area, Waqf Land Area, Growth Rate of Regional Gross Domestic Product (YoY), and CO₂ Emissions for 2018-2022. Data were obtained from the Indonesian Waqf Board, Climate World Data, the Ministry of the Environment, and the Bank of Indonesia. The results prove that CO₂ emissions have a negative effect on regional economic growth and that waqf governance in the waqf index has a positive effect on regional economic growth in 33 provinces.

Keywords: waqf, CO₂ emissions, panel logit analysis, sustainable economic development

Procedia PDF Downloads 11
24355 Intelligent Human Pose Recognition Based on EMG Signal Analysis and Machine 3D Model

Authors: Si Chen, Quanhong Jiang

Abstract:

In the increasingly mature posture recognition technology, human movement information is widely used in sports rehabilitation, human-computer interaction, medical health, human posture assessment, and other fields today; this project uses the most original ideas; it is proposed to use the collection equipment for the collection of myoelectric data, reflect the muscle posture change on a degree of freedom through data processing, carry out data-muscle three-dimensional model joint adjustment, and realize basic pose recognition. Based on this, bionic aids or medical rehabilitation equipment can be further developed with the help of robotic arms and cutting-edge technology, which has a bright future and unlimited development space.

Keywords: pose recognition, 3D animation, electromyography, machine learning, bionics

Procedia PDF Downloads 59
24354 Optimizing Energy Efficiency: Leveraging Big Data Analytics and AWS Services for Buildings and Industries

Authors: Gaurav Kumar Sinha

Abstract:

In an era marked by increasing concerns about energy sustainability, this research endeavors to address the pressing challenge of energy consumption in buildings and industries. This study delves into the transformative potential of AWS services in optimizing energy efficiency. The research is founded on the recognition that effective management of energy consumption is imperative for both environmental conservation and economic viability. Buildings and industries account for a substantial portion of global energy use, making it crucial to develop advanced techniques for analysis and reduction. This study sets out to explore the integration of AWS services with big data analytics to provide innovative solutions for energy consumption analysis. Leveraging AWS's cloud computing capabilities, scalable infrastructure, and data analytics tools, the research aims to develop efficient methods for collecting, processing, and analyzing energy data from diverse sources. The core focus is on creating predictive models and real-time monitoring systems that enable proactive energy management. By harnessing AWS's machine learning and data analytics capabilities, the research seeks to identify patterns, anomalies, and optimization opportunities within energy consumption data. Furthermore, this study aims to propose actionable recommendations for reducing energy consumption in buildings and industries. By combining AWS services with metrics-driven insights, the research strives to facilitate the implementation of energy-efficient practices, ultimately leading to reduced carbon emissions and cost savings. The integration of AWS services not only enhances the analytical capabilities but also offers scalable solutions that can be customized for different building and industrial contexts. The research also recognizes the potential for AWS-powered solutions to promote sustainable practices and support environmental stewardship.

Keywords: energy consumption analysis, big data analytics, AWS services, energy efficiency

Procedia PDF Downloads 49
24353 Bandwidth Efficient Cluster Based Collision Avoidance Multicasting Protocol in VANETs

Authors: Navneet Kaur, Amarpreet Singh

Abstract:

In Vehicular Adhoc Networks, Data Dissemination is a challenging task. There are number of techniques, types and protocols available for disseminating the data but in order to preserve limited bandwidth and to disseminate maximum data over networks makes it more challenging. There are broadcasting, multicasting and geocasting based protocols. Multicasting based protocols are found to be best for conserving the bandwidth. One such protocol named BEAM exists that improves the performance of Vehicular Adhoc Networks by reducing the number of in-network message transactions and thereby efficiently utilizing the bandwidth during an emergency situation. But this protocol may result in multicar chain collision as there was no V2V communication. So, this paper proposes a new protocol named Enhanced Bandwidth Efficient Cluster Based Multicasting Protocol (EBECM) that will overcome the limitations of existing BEAM protocol. And Simulation results will show the improved performance of EBECM in terms of Routing overhead, throughput and PDR when compared with BEAM protocol.

Keywords: BEAM, data dissemination, emergency situation, vehicular adhoc network

Procedia PDF Downloads 331
24352 Machine Learning-Based Workflow for the Analysis of Project Portfolio

Authors: Jean Marie Tshimula, Atsushi Togashi

Abstract:

We develop a data-science approach for providing an interactive visualization and predictive models to find insights into the projects' historical data in order for stakeholders understand some unseen opportunities in the African market that might escape them behind the online project portfolio of the African Development Bank. This machine learning-based web application identifies the market trend of the fastest growing economies across the continent as well skyrocketing sectors which have a significant impact on the future of business in Africa. Owing to this, the approach is tailored to predict where the investment needs are the most required. Moreover, we create a corpus that includes the descriptions of over more than 1,200 projects that approximately cover 14 sectors designed for some of 53 African countries. Then, we sift out this large amount of semi-structured data for extracting tiny details susceptible to contain some directions to follow. In the light of the foregoing, we have applied the combination of Latent Dirichlet Allocation and Random Forests at the level of the analysis module of our methodology to highlight the most relevant topics that investors may focus on for investing in Africa.

Keywords: machine learning, topic modeling, natural language processing, big data

Procedia PDF Downloads 159
24351 On the Existence of Homotopic Mapping Between Knowledge Graphs and Graph Embeddings

Authors: Jude K. Safo

Abstract:

Knowledge Graphs KG) and their relation to Graph Embeddings (GE) represent a unique data structure in the landscape of machine learning (relative to image, text and acoustic data). Unlike the latter, GEs are the only data structure sufficient for representing hierarchically dense, semantic information needed for use-cases like supply chain data and protein folding where the search space exceeds the limits traditional search methods (e.g. page-rank, Dijkstra, etc.). While GEs are effective for compressing low rank tensor data, at scale, they begin to introduce a new problem of ’data retreival’ which we observe in Large Language Models. Notable attempts by transE, TransR and other prominent industry standards have shown a peak performance just north of 57% on WN18 and FB15K benchmarks, insufficient practical industry applications. They’re also limited, in scope, to next node/link predictions. Traditional linear methods like Tucker, CP, PARAFAC and CANDECOMP quickly hit memory limits on tensors exceeding 6.4 million nodes. This paper outlines a topological framework for linear mapping between concepts in KG space and GE space that preserve cardinality. Most importantly we introduce a traceable framework for composing dense linguistic strcutures. We demonstrate performance on WN18 benchmark this model hits. This model does not rely on Large Langauge Models (LLM) though the applications are certainy relevant here as well.

Keywords: representation theory, large language models, graph embeddings, applied algebraic topology, applied knot theory, combinatorics

Procedia PDF Downloads 57
24350 The Names of the Traditional Motif of Batik Solo

Authors: Annisa D. Febryandini

Abstract:

Batik is a unique cultural heritage that strongly linked with its community. As a product of current culture in Solo, Batik Solo not only has a specific design and color to represent the cultural identity, cultural values, and spirituality of the community, but also has some specific names given by its community which are not arbitrary. This qualitative research paper uses the primary data by interview method as well as the secondary data to support it. Based on the data, this paper concludes that the names consist of a word or words taken from a current name of things in Javanese language. They indicate the cultural meaning such as a specific event, a hope, and the social status of the people who use the motif. Different from the other research, this paper takes a look at the names of traditional motif of Batik Solo which analyzed linguistically to reveal the cultural meaning.

Keywords: traditional motif, Batik, solo, anthropological linguistics

Procedia PDF Downloads 259
24349 SEM Image Classification Using CNN Architectures

Authors: Güzi̇n Ti̇rkeş, Özge Teki̇n, Kerem Kurtuluş, Y. Yekta Yurtseven, Murat Baran

Abstract:

A scanning electron microscope (SEM) is a type of electron microscope mainly used in nanoscience and nanotechnology areas. Automatic image recognition and classification are among the general areas of application concerning SEM. In line with these usages, the present paper proposes a deep learning algorithm that classifies SEM images into nine categories by means of an online application to simplify the process. The NFFA-EUROPE - 100% SEM data set, containing approximately 21,000 images, was used to train and test the algorithm at 80% and 20%, respectively. Validation was carried out using a separate data set obtained from the Middle East Technical University (METU) in Turkey. To increase the accuracy in the results, the Inception ResNet-V2 model was used in view of the Fine-Tuning approach. By using a confusion matrix, it was observed that the coated-surface category has a negative effect on the accuracy of the results since it contains other categories in the data set, thereby confusing the model when detecting category-specific patterns. For this reason, the coated-surface category was removed from the train data set, hence increasing accuracy by up to 96.5%.

Keywords: convolutional neural networks, deep learning, image classification, scanning electron microscope

Procedia PDF Downloads 105
24348 Nearest Neighbor Investigate Using R+ Tree

Authors: Rutuja Desai

Abstract:

Search engine is fundamentally a framework used to search the data which is pertinent to the client via WWW. Looking close-by spot identified with the keywords is an imperative concept in developing web advances. For such kind of searching, extent pursuit or closest neighbor is utilized. In range search the forecast is made whether the objects meet to query object. Nearest neighbor is the forecast of the focuses close to the query set by the client. Here, the nearest neighbor methodology is utilized where Data recovery R+ tree is utilized rather than IR2 tree. The disadvantages of IR2 tree is: The false hit number can surpass the limit and the mark in Information Retrieval R-tree must have Voice over IP bit for each one of a kind word in W set is recouped by Data recovery R+ tree. The inquiry is fundamentally subordinate upon the key words and the geometric directions.

Keywords: information retrieval, nearest neighbor search, keyword search, R+ tree

Procedia PDF Downloads 273
24347 Classical and Bayesian Inference of the Generalized Log-Logistic Distribution with Applications to Survival Data

Authors: Abdisalam Hassan Muse, Samuel Mwalili, Oscar Ngesa

Abstract:

A generalized log-logistic distribution with variable shapes of the hazard rate was introduced and studied, extending the log-logistic distribution by adding an extra parameter to the classical distribution, leading to greater flexibility in analysing and modeling various data types. The proposed distribution has a large number of well-known lifetime special sub-models such as; Weibull, log-logistic, exponential, and Burr XII distributions. Its basic mathematical and statistical properties were derived. The method of maximum likelihood was adopted for estimating the unknown parameters of the proposed distribution, and a Monte Carlo simulation study is carried out to assess the behavior of the estimators. The importance of this distribution is that its tendency to model both monotone (increasing and decreasing) and non-monotone (unimodal and bathtub shape) or reversed “bathtub” shape hazard rate functions which are quite common in survival and reliability data analysis. Furthermore, the flexibility and usefulness of the proposed distribution are illustrated in a real-life data set and compared to its sub-models; Weibull, log-logistic, and BurrXII distributions and other parametric survival distributions with 3-parmaeters; like the exponentiated Weibull distribution, the 3-parameter lognormal distribution, the 3- parameter gamma distribution, the 3-parameter Weibull distribution, and the 3-parameter log-logistic (also known as shifted log-logistic) distribution. The proposed distribution provided a better fit than all of the competitive distributions based on the goodness-of-fit tests, the log-likelihood, and information criterion values. Finally, Bayesian analysis and performance of Gibbs sampling for the data set are also carried out.

Keywords: hazard rate function, log-logistic distribution, maximum likelihood estimation, generalized log-logistic distribution, survival data, Monte Carlo simulation

Procedia PDF Downloads 184
24346 Phytodiversity and Phytogeographic Characterization Stands of Pistacia lentiscus L. in the Coastal Region of Honaine, Tlemcen, Western Algeria

Authors: I. Benmehdi, O. Hasnaoui, N. Hachemi, M. Bouazza

Abstract:

The Understanding of the mechanisms structuring of plant diversity in the region of Tlemcen (western Algeria) is a related problem. The current floristic composition of different groups in Pistacia lentiscus L. resulting from the combination of human and climate action. This study is devoted to biodiversity inventory and phytogeographic characterization of Pistacia lentiscus groups in the Honaine coastal (western Algeria). The floristic inventory (150 levels) made in three stations of the study area allowed to count a 109 species belonging to 44 families of vascular plants. The biogeographical analysis of the Pistacia lentiscus groups reveals the most representative elements. The Mediterranean elements are numerically the most dominant with 39.45% represented by: Pistacia lentiscus, Cistus monspeliensis, Plantago lagopus, Linum strictum, Echium vulgare; followed by the western Mediterranean elements with 10.09% and are represented by: Chamaerops humilis, Lavandula dentata, Ampelodesma mauritanicum and Iris xyphium. However, this phytotaxonomic wealth is exposed to anthropogenic impact causing its disruption see its decline.

Keywords: Pistacia lentiscus L., phytodiversity, phytogeography, honaine, western Algeria

Procedia PDF Downloads 382
24345 Antibacterial and Anti-Biofilm Activity of Vaccinium meridionale S. Pomace Extract Against Staphylococcus aureus, Escherichia coli and Salmonella Enterica

Authors: Carlos Y. Soto, Camila A. Lota, G. Astrid Garzón

Abstract:

Bacterial biofilms cause an ongoing problem for food safety. They are formed when microorganisms aggregate to form a community that attaches to solid surfaces. Biofilms increase the resistance of pathogens to cleaning, disinfection and antibacterial products. This resistance gives rise to problems for human health, industry, and agriculture. At present, plant extracts rich in polyphenolics are being investigated as natural alternatives to degrade bacterial biofilms. The pomace of the tropical Berry Vaccinium meridionale S. contains high amounts of phenolic compounds. Therefore, in the current study, the antimicrobial and antibiofilm effects of extracts from the pomace of Vaccinium meridionale S. were tested on three foodborne pathogens: Enterohaemorrhagic Escherichia coli O157:H7 (ATCC®700728TM), Staphylococcus aureus subsp. aureus (ATCC® 6538TM), and Salmonella enterica serovar Enteritidis (ATCC® 13076TM). Microwave-assisted extraction was used to extract polyphenols with aqueous methanol (80% v/v) at a solid to solvent ratio of 1:10 (w/v) for 20 min. The magnetic stirring was set at 400 rpm, and the microwave power was adjusted to 400 W. The antimicrobial effect of the extract was assessed by determining the half maximal inhibitory concentration (IC50) against the three food poisoning pathogens at concentrations ranging from 50 to 2,850 μg gallic acid equivalents (GAE)/mL of the extract. Biofilm inhibition was assessed using a crystal violet assay applying the same range of concentration. Three replications of the experiments were carried out, and all analyses were run in triplicate. IC50 values were determined using the GraphPad Prism8® program. Significant differences (P<0.05) among means were identified using one-factor analysis of variance (ANOVA) and the post-hoc least significant difference (LSD) test using the Statgraphics plus program, version 2.1.There was significant difference among the mean IC50 values for the tested bacteria. The IC50 for S. aureus was 48 ± 9 μg GAE/mL, followed by 123 ± 49 μg GAE/mL for Salmonella and 376 ± 32 μg GAE/mL for E. coli. The percent inhibition of the extract on biofilm formation was significantly higher for S. aureus (85.8  0.3), followed by E. coli (74.5  1.0) and Salmonella (53.6  9.7). These findings suggest that polyphenolic extracts obtained from the pomace of V. meridionale S. might be used as natural antimicrobial and anti-biofilm natural agents, effective against S. aureus, E. coli and Salmonella enterica.

Keywords: antibiofilm, antimicrobial, E. coli, S. aureus, salmonella, IC50, pomace, V. meridionale

Procedia PDF Downloads 48
24344 Fuzzy Inference-Assisted Saliency-Aware Convolution Neural Networks for Multi-View Summarization

Authors: Tanveer Hussain, Khan Muhammad, Amin Ullah, Mi Young Lee, Sung Wook Baik

Abstract:

The Big Data generated from distributed vision sensors installed on large scale in smart cities create hurdles in its efficient and beneficial exploration for browsing, retrieval, and indexing. This paper presents a three-folded framework for effective video summarization of such data and provide a compact and representative format of Big Video Data. In the first fold, the paper acquires input video data from the installed cameras and collect clues such as type and count of objects and clarity of the view from a chunk of pre-defined number of frames of each view. The decision of representative view selection for a particular interval is based on fuzzy inference system, acquiring a precise and human resembling decision, reinforced by the known clues as a part of the second fold. In the third fold, the paper forwards the selected view frames to the summary generation mechanism that is supported by a saliency-aware convolution neural network (CNN) model. The new trend of fuzzy rules for view selection followed by CNN architecture for saliency computation makes the multi-view video summarization (MVS) framework a suitable candidate for real-world practice in smart cities.

Keywords: big video data analysis, fuzzy logic, multi-view video summarization, saliency detection

Procedia PDF Downloads 172
24343 Electrochemical Coagulation of Synthetic Textile Dye Wastewater

Authors: H. B. Rekha, Usha N. Murthy, Prashanth, Ashoka

Abstract:

Dyes are manufactured to have high chemical resistance because they are normally species, very difficult to degrade (reactive dyes). It damages flora and fauna. Furthermore, coloured components are highly hazardous. So removal of dyes becomes a challenge for both textile industry and water treatment facility. Dyeing wastewater is usually treated by conventional methods such as biological oxidation and adsorption but nowadays them becoming in-adequate because of large variability of composition of waste water. In the present investigation, mild steel electrodes of varying surface area were used for treatment of synthetic textile dye. It appears that electro-chemical coagulation could be very effective in removing coloured from wastewater; it could also be used to remove other parameters like chlorides, COD, and solids to some extent. In the present study, coloured removal up to 99% was obtained for surface area of mild steel electrode of 80 cm2 and 96% of surface area of mild steel electrode of 50 cm2. The findings from this study could be used to improve the design of electro-chemical treatment systems and modify existing systems to improve efficiency.

Keywords: electrochemical coagulation, mild steel, colour, environmental engineering

Procedia PDF Downloads 289
24342 Relation between Pavement Roughness and Distress Parameters for Highways

Authors: Suryapeta Harini

Abstract:

Road surface roughness is one of the essential aspects of the road's functional condition, indicating riding comfort in both the transverse and longitudinal directions. The government of India has made maintaining good surface evenness a prerequisite for all highway projects. Pavement distress data was collected with a Network Survey Vehicle (NSV) on a National Highway. It determines the smoothness and frictional qualities of the pavement surface, which are related to driving safety and ease. Based on the data obtained in the field, a regression equation was created with the IRI value and the visual distresses. The suggested system can use wireless acceleration sensors and GPS to gather vehicle status and location data, as well as calculate the international roughness index (IRI). Potholes, raveling, rut depth, cracked area, and repair work are all affected by pavement roughness, according to the current study. The study was carried out in one location. Data collected through using Bump integrator was used for the validation. The bump integrator (BI) obtained using deflection from the network survey vehicle was correlated with the distress parameter to establish an equation.

Keywords: roughness index, network survey vehicle, regression, correlation

Procedia PDF Downloads 161