Search results for: virtual machine migration
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4764

Search results for: virtual machine migration

1704 Investigating the Role of Artificial Intelligence in Developing Creativity in Architecture Education in Egypt: A Case Study of Design Studios

Authors: Ahmed Radwan, Ahmed Abdel Ghaney

Abstract:

This paper delves into the transformative potential of artificial intelligence (AI) in fostering creativity within the domain of architecture education, especially with a specific emphasis on its implications within the Design Studios; the convergence of AI and architectural pedagogy has introduced avenues for redefining the boundaries of creative expression and problem-solving. By harnessing AI-driven tools, students and educators can collaboratively explore a spectrum of design possibilities, stimulate innovative ideation, and engage in multidimensional design processes. This paper investigates the ways in which AI contributes to architectural creativity by facilitating generative design, pattern recognition, virtual reality experiences, and sustainable design optimization. Furthermore, the study examines the balance between AI-enhanced creativity and the preservation of core principles of architectural design/education, ensuring that technology is harnessed to augment rather than replace foundational design skills. Through an exploration of Egypt's architectural heritage and contemporary challenges, this research underscores how AI can synergize with cultural context and historical insights to inspire cutting-edge architectural solutions. By analyzing AI's impact on nurturing creativity among Egyptian architecture students, this paper seeks to contribute to the ongoing discourse on the integration of technology within global architectural education paradigms. It is hoped that this research will guide the thoughtful incorporation of AI in fostering creativity while preserving the authenticity and richness of architectural design education in Egypt and beyond.

Keywords: architecture, artificial intelligence, architecture education, Egypt

Procedia PDF Downloads 57
1703 Transfer Knowledge From Multiple Source Problems to a Target Problem in Genetic Algorithm

Authors: Terence Soule, Tami Al Ghamdi

Abstract:

To study how to transfer knowledge from multiple source problems to the target problem, we modeled the Transfer Learning (TL) process using Genetic Algorithms as the model solver. TL is the process that aims to transfer learned data from one problem to another problem. The TL process aims to help Machine Learning (ML) algorithms find a solution to the problems. The Genetic Algorithms (GA) give researchers access to information that we have about how the old problem is solved. In this paper, we have five different source problems, and we transfer the knowledge to the target problem. We studied different scenarios of the target problem. The results showed combined knowledge from multiple source problems improves the GA performance. Also, the process of combining knowledge from several problems results in promoting diversity of the transferred population.

Keywords: transfer learning, genetic algorithm, evolutionary computation, source and target

Procedia PDF Downloads 119
1702 Regression Approach for Optimal Purchase of Hosts Cluster in Fixed Fund for Hadoop Big Data Platform

Authors: Haitao Yang, Jianming Lv, Fei Xu, Xintong Wang, Yilin Huang, Lanting Xia, Xuewu Zhu

Abstract:

Given a fixed fund, purchasing fewer hosts of higher capability or inversely more of lower capability is a must-be-made trade-off in practices for building a Hadoop big data platform. An exploratory study is presented for a Housing Big Data Platform project (HBDP), where typical big data computing is with SQL queries of aggregate, join, and space-time condition selections executed upon massive data from more than 10 million housing units. In HBDP, an empirical formula was introduced to predict the performance of host clusters potential for the intended typical big data computing, and it was shaped via a regression approach. With this empirical formula, it is easy to suggest an optimal cluster configuration. The investigation was based on a typical Hadoop computing ecosystem HDFS+Hive+Spark. A proper metric was raised to measure the performance of Hadoop clusters in HBDP, which was tested and compared with its predicted counterpart, on executing three kinds of typical SQL query tasks. Tests were conducted with respect to factors of CPU benchmark, memory size, virtual host division, and the number of element physical host in cluster. The research has been applied to practical cluster procurement for housing big data computing.

Keywords: Hadoop platform planning, optimal cluster scheme at fixed-fund, performance predicting formula, typical SQL query tasks

Procedia PDF Downloads 213
1701 A Petri Net Model to Obtain the Throughput of Unreliable Production Lines in the Buffer Allocation Problem

Authors: Joselito Medina-Marin, Alexandr Karelin, Ana Tarasenko, Juan Carlos Seck-Tuoh-Mora, Norberto Hernandez-Romero, Eva Selene Hernandez-Gress

Abstract:

A production line designer faces with several challenges in manufacturing system design. One of them is the assignment of buffer slots in between every machine of the production line in order to maximize the throughput of the whole line, which is known as the Buffer Allocation Problem (BAP). The BAP is a combinatorial problem that depends on the number of machines and the total number of slots to be distributed on the production line. In this paper, we are proposing a Petri Net (PN) Model to obtain the throughput in unreliable production lines, based on PN mathematical tools and the decomposition method. The results obtained by this methodology are similar to those presented in previous works, and the number of machines is not a hard restriction.

Keywords: buffer allocation problem, Petri Nets, throughput, production lines

Procedia PDF Downloads 285
1700 Harnessing Artificial Intelligence for Early Detection and Management of Infectious Disease Outbreaks

Authors: Amarachukwu B. Isiaka, Vivian N. Anakwenze, Chinyere C. Ezemba, Chiamaka R. Ilodinso, Chikodili G. Anaukwu, Chukwuebuka M. Ezeokoli, Ugonna H. Uzoka

Abstract:

Infectious diseases continue to pose significant threats to global public health, necessitating advanced and timely detection methods for effective outbreak management. This study explores the integration of artificial intelligence (AI) in the early detection and management of infectious disease outbreaks. Leveraging vast datasets from diverse sources, including electronic health records, social media, and environmental monitoring, AI-driven algorithms are employed to analyze patterns and anomalies indicative of potential outbreaks. Machine learning models, trained on historical data and continuously updated with real-time information, contribute to the identification of emerging threats. The implementation of AI extends beyond detection, encompassing predictive analytics for disease spread and severity assessment. Furthermore, the paper discusses the role of AI in predictive modeling, enabling public health officials to anticipate the spread of infectious diseases and allocate resources proactively. Machine learning algorithms can analyze historical data, climatic conditions, and human mobility patterns to predict potential hotspots and optimize intervention strategies. The study evaluates the current landscape of AI applications in infectious disease surveillance and proposes a comprehensive framework for their integration into existing public health infrastructures. The implementation of an AI-driven early detection system requires collaboration between public health agencies, healthcare providers, and technology experts. Ethical considerations, privacy protection, and data security are paramount in developing a framework that balances the benefits of AI with the protection of individual rights. The synergistic collaboration between AI technologies and traditional epidemiological methods is emphasized, highlighting the potential to enhance a nation's ability to detect, respond to, and manage infectious disease outbreaks in a proactive and data-driven manner. The findings of this research underscore the transformative impact of harnessing AI for early detection and management, offering a promising avenue for strengthening the resilience of public health systems in the face of evolving infectious disease challenges. This paper advocates for the integration of artificial intelligence into the existing public health infrastructure for early detection and management of infectious disease outbreaks. The proposed AI-driven system has the potential to revolutionize the way we approach infectious disease surveillance, providing a more proactive and effective response to safeguard public health.

Keywords: artificial intelligence, early detection, disease surveillance, infectious diseases, outbreak management

Procedia PDF Downloads 40
1699 Investigation of Topic Modeling-Based Semi-Supervised Interpretable Document Classifier

Authors: Dasom Kim, William Xiu Shun Wong, Yoonjin Hyun, Donghoon Lee, Minji Paek, Sungho Byun, Namgyu Kim

Abstract:

There have been many researches on document classification for classifying voluminous documents automatically. Through document classification, we can assign a specific category to each unlabeled document on the basis of various machine learning algorithms. However, providing labeled documents manually requires considerable time and effort. To overcome the limitations, the semi-supervised learning which uses unlabeled document as well as labeled documents has been invented. However, traditional document classifiers, regardless of supervised or semi-supervised ones, cannot sufficiently explain the reason or the process of the classification. Thus, in this paper, we proposed a methodology to visualize major topics and class components of each document. We believe that our methodology for visualizing topics and classes of each document can enhance the reliability and explanatory power of document classifiers.

Keywords: data mining, document classifier, text mining, topic modeling

Procedia PDF Downloads 381
1698 Eudesmane-Type Sesquiterpenes from Laggera alata Inhibiting Angiogenesis

Authors: Liang Ning, Chung Hau Yin

Abstract:

Angiogenesis is the process of new blood vessel development. It has been recognized as a therapeutic target for blocking cancer growth four decades ago. Vascular sprouting is initiated by pro-angiogenic factors. Vascular endothelial cell growth factor (VEGF) plays a central role in angiogenic initiation, many patients with cancer or ocular neovascularization have been benefited from anti-VEGF therapy. Emerging approaches impacting in the later stages of vessel remodeling and maturation are expected to improve clinical efficacy. TIE receptor as well as the corresponding angiopoietin ligands, were identified as another endothelial cell specific receptor tyrosine kinase signaling system. Much efforts were made to reduce the activity of angiopoietin-TIE receptor axis. Two eudesmane-type sesquiterpenes from laggera alata, namely, 15-dihydrocostic acid and ilicic acid were found with strong anti-angiogenic properties in zebrafish model. Meanwhile, the mRNA expression levels of VEGFR2 and TIE2 pathway related genes were down-regulated in the sesquiterpenes treated zebrafish embryos. Besides, in human umbilical vein endothelial cells (HUVECs), the sesquiterpenes have the ability to inhibit VEGF-induced HUVECs proliferation and migration at non-toxic concentration. Moreover, angiopoietin-2 induced TIE2 phosphorylation was inhibited by the sesquiterpenes, the inhibitory effect was detected in angiopoietin-1 induced HUVECs proliferation as well. Thus, we hypothesized the anti-angiogenic activity of the compounds may via the inhibition of VEGF and TIE2 related pathways. How the compounds come into play as the pathways inhibitors need to be evaluated in the future.

Keywords: Laggera alata, eudesmane-type sesquiterpene, anti-angiogenesis, VEGF, angiopoietin, TIE2

Procedia PDF Downloads 187
1697 Core Loss Influence on MTPA Current Vector Variation of Synchronous Reluctance Machine

Authors: Huai-Cong Liu, Tae Chul Jeong, Ju Lee

Abstract:

The aim of this study was to develop an electric circuit method (ECM) to ascertain the core loss influence on a Synchronous Reluctance Motor (SynRM) in the condition of the maximum torque per ampere (MTPA). SynRM for fan usually operates on the constant torque region, at synchronous speed the MTPA control is adopted due to current vector. However, finite element analysis (FEA) program is not sufficient exactly to reflect how the core loss influenced on the current vector. This paper proposed a method to calculate the current vector with consideration of core loss. The precision of current vector by ECM is useful for MTPA control. The result shows that ECM analysis is closer to the actual motor’s characteristics by testing with a 7.5kW SynRM drive System.

Keywords: core loss, SynRM, current vector, magnetic saturation, maximum torque per ampere (MTPA)

Procedia PDF Downloads 494
1696 Preoperative 3D Planning and Reconstruction of Mandibular Defects for Patients with Oral Cavity Tumors

Authors: Janis Zarins, Kristaps Blums, Oskars Radzins, Renars Deksnis, Atis Svare, Santa Salaka

Abstract:

Wide tumor resection remains the first choice method for tumors of the oral cavity. Nevertheless, remained tissue defect impacts patients functional and aesthetical outcome, which could be improved using microvascular tissue transfers. Mandibular reconstruction is challenging due to the complexity of composite tissue defects and occlusal relationships for normal eating, chewing, and pain free jaw motions. Individual 3-D virtual planning would provide better symmetry and functional outcome. The main goal of preoperative planning is to develop a customized surgical approach with patient specific cutting guides of the mandible, osteotomy guides of the fibula, pre-bended osteosynthesis plates to perform more precise reconstruction, to decrease the surgery time and reach the best outcome. Our study is based on the analysis of 32 patients operated on between 2019 to 2021. All patients underwent mandible reconstruction with vascularized fibula flaps. Patients characteristics, surgery profile, survival, functional outcome, and quality of life was evaluated. Preoperative planning provided a significant decrease of surgery time and the best arrangement of bone closely similar as before the surgery. In cases of bone asymmetry, deformity and malposition, a new mandible was created using 3D planning to restore the appearance of lower jaw anatomy and functionality.

Keywords: mandibular, 3D planning, cutting guides, fibula flap, reconstruction

Procedia PDF Downloads 105
1695 The Use Support Vector Machine and Back Propagation Neural Network for Prediction of Daily Tidal Levels Along The Jeddah Coast, Saudi Arabia

Authors: E. A. Mlybari, M. S. Elbisy, A. H. Alshahri, O. M. Albarakati

Abstract:

Sea level rise threatens to increase the impact of future storms and hurricanes on coastal communities. Accurate sea level change prediction and supplement is an important task in determining constructions and human activities in coastal and oceanic areas. In this study, support vector machines (SVM) is proposed to predict daily tidal levels along the Jeddah Coast, Saudi Arabia. The optimal parameter values of kernel function are determined using a genetic algorithm. The SVM results are compared with the field data and with back propagation (BP). Among the models, the SVM is superior to BPNN and has better generalization performance.

Keywords: tides, prediction, support vector machines, genetic algorithm, back-propagation neural network, risk, hazards

Procedia PDF Downloads 442
1694 mm-Wave Wearable Edge Computing Module Hosted by Printed Ridge Gap Waveguide Structures: A Physical Layer Study

Authors: Matthew Kostawich, Mohammed Elmorsy, Mohamed Sayed Sifat, Shoukry Shams, Mahmoud Elsaadany

Abstract:

6G communication systems represent the nominal future extension of current wireless technology, where its impact is extended to touch upon all human activities, including medical, security, and entertainment applications. As a result, human needs are allocated among the highest priority aspects of the system design and requirements. 6G communications is expected to replace all the current video conferencing with interactive virtual reality meetings involving high data-rate transmission merged with massive distributed computing resources. In addition, the current expansion of IoT applications must be mitigated with significant network changes to provide a reasonable Quality of Service (QoS). This directly implies a high demand for Human-Computer Interaction (HCI) through mobile computing modules in future wireless communication systems. This article proposes the utilization of a Printed Ridge Gap Waveguide (PRGW) to host the wearable nodes. To the best of our knowledge, we propose for the first time a physical layer analysis within the context of a complete architecture. A thorough study is provided on the impact of the distortion of the guiding structure on the overall system performance. The proposed structure shows small latency and small losses, highlighting its compatibility with future applications.

Keywords: ridge gap waveguide, edge computing module, 6G, multimedia IoT applications

Procedia PDF Downloads 50
1693 Music Genre Classification Based on Non-Negative Matrix Factorization Features

Authors: Soyon Kim, Edward Kim

Abstract:

In order to retrieve information from the massive stream of songs in the music industry, music search by title, lyrics, artist, mood, and genre has become more important. Despite the subjectivity and controversy over the definition of music genres across different nations and cultures, automatic genre classification systems that facilitate the process of music categorization have been developed. Manual genre selection by music producers is being provided as statistical data for designing automatic genre classification systems. In this paper, an automatic music genre classification system utilizing non-negative matrix factorization (NMF) is proposed. Short-term characteristics of the music signal can be captured based on the timbre features such as mel-frequency cepstral coefficient (MFCC), decorrelated filter bank (DFB), octave-based spectral contrast (OSC), and octave band sum (OBS). Long-term time-varying characteristics of the music signal can be summarized with (1) the statistical features such as mean, variance, minimum, and maximum of the timbre features and (2) the modulation spectrum features such as spectral flatness measure, spectral crest measure, spectral peak, spectral valley, and spectral contrast of the timbre features. Not only these conventional basic long-term feature vectors, but also NMF based feature vectors are proposed to be used together for genre classification. In the training stage, NMF basis vectors were extracted for each genre class. The NMF features were calculated in the log spectral magnitude domain (NMF-LSM) as well as in the basic feature vector domain (NMF-BFV). For NMF-LSM, an entire full band spectrum was used. However, for NMF-BFV, only low band spectrum was used since high frequency modulation spectrum of the basic feature vectors did not contain important information for genre classification. In the test stage, using the set of pre-trained NMF basis vectors, the genre classification system extracted the NMF weighting values of each genre as the NMF feature vectors. A support vector machine (SVM) was used as a classifier. The GTZAN multi-genre music database was used for training and testing. It is composed of 10 genres and 100 songs for each genre. To increase the reliability of the experiments, 10-fold cross validation was used. For a given input song, an extracted NMF-LSM feature vector was composed of 10 weighting values that corresponded to the classification probabilities for 10 genres. An NMF-BFV feature vector also had a dimensionality of 10. Combined with the basic long-term features such as statistical features and modulation spectrum features, the NMF features provided the increased accuracy with a slight increase in feature dimensionality. The conventional basic features by themselves yielded 84.0% accuracy, but the basic features with NMF-LSM and NMF-BFV provided 85.1% and 84.2% accuracy, respectively. The basic features required dimensionality of 460, but NMF-LSM and NMF-BFV required dimensionalities of 10 and 10, respectively. Combining the basic features, NMF-LSM and NMF-BFV together with the SVM with a radial basis function (RBF) kernel produced the significantly higher classification accuracy of 88.3% with a feature dimensionality of 480.

Keywords: mel-frequency cepstral coefficient (MFCC), music genre classification, non-negative matrix factorization (NMF), support vector machine (SVM)

Procedia PDF Downloads 274
1692 Identification of Breast Anomalies Based on Deep Convolutional Neural Networks and K-Nearest Neighbors

Authors: Ayyaz Hussain, Tariq Sadad

Abstract:

Breast cancer (BC) is one of the widespread ailments among females globally. The early prognosis of BC can decrease the mortality rate. Exact findings of benign tumors can avoid unnecessary biopsies and further treatments of patients under investigation. However, due to variations in images, it is a tough job to isolate cancerous cases from normal and benign ones. The machine learning technique is widely employed in the classification of BC pattern and prognosis. In this research, a deep convolution neural network (DCNN) called AlexNet architecture is employed to get more discriminative features from breast tissues. To achieve higher accuracy, K-nearest neighbor (KNN) classifiers are employed as a substitute for the softmax layer in deep learning. The proposed model is tested on a widely used breast image database called MIAS dataset for experimental purposes and achieved 99% accuracy.

Keywords: breast cancer, DCNN, KNN, mammography

Procedia PDF Downloads 114
1691 Cognition Technique for Developing a World Music

Authors: Haider Javed Uppal, Javed Yunas Uppal

Abstract:

In today's globalized world, it is necessary to develop a form of music that is able to evoke equal emotional responses among people from diverse cultural backgrounds. Indigenous cultures throughout history have developed their own music cognition, specifically in terms of the connections between music and mood. With the advancements in artificial intelligence technologies, it has become possible to analyze and categorize music features such as timbre, harmony, melody, and rhythm and relate them to the resulting mood effects experienced by listeners. This paper presents a model that utilizes a screenshot translator to convert music from different origins into waveforms, which are then analyzed using machine learning and information retrieval techniques. By connecting these waveforms with Thayer's matrix of moods, a mood classifier has been developed using fuzzy logic algorithms to determine the emotional impact of different types of music on listeners from various cultures.

Keywords: cognition, world music, artificial intelligence, Thayer’s matrix

Procedia PDF Downloads 57
1690 Development of an Optimization Method for Myoelectric Signal Processing by Active Matrix Sensing in Robot Rehabilitation

Authors: Noriyoshi Yamauchi, Etsuo Horikawa, Takunori Tsuji

Abstract:

Training by exoskeleton robot is drawing attention as a rehabilitation method for body paralysis seen in many cases, and there are many forms that assist with the myoelectric signal generated by exercise commands from the brain. Rehabilitation requires more frequent training, but it is one of the reasons that the technology is required for the identification of the myoelectric potential derivation site and attachment of the device is preventing the spread of paralysis. In this research, we focus on improving the efficiency of gait training by exoskeleton type robots, improvement of myoelectric acquisition and analysis method using active matrix sensing method, and improvement of walking rehabilitation and walking by optimization of robot control.

Keywords: active matrix sensing, brain machine interface (BMI), the central pattern generator (CPG), myoelectric signal processing, robot rehabilitation

Procedia PDF Downloads 368
1689 A Zero-Flaring Flowback Solution to Revive Liquid Loaded Gas Wells

Authors: Elsayed Amer, Tarek Essam, Abdullah Hella, Mohammed Al-Ajmi

Abstract:

Hydrocarbon production decline in mature gas fields is inevitable, and mitigating these circumstances is essential to ensure a longer production period. Production decline is not only influenced by reservoir pressure and wellbore integrity; however, associated liquids in the reservoir rock have a considerable impact on the production process. The associated liquid may result in liquid loading, near wellbore damage, condensate banking, fine sand migration, and wellhead pressure depletion. Consequently, the producing well will suffocate, and the liquid column will seize the well from flowing. A common solution in such circumstances is reducing the surface pressure by opening the well to the atmospheric pressure and flaring the produced liquids. This practice may not be applicable to many cases since the atmospheric pressure is not low enough to create a sufficient driving force to flow the well. In addition, flaring the produced hydrocarbon is solving the issue on account of the environment, which is against the world's efforts to mitigate the impact of climate change. This paper presents a novel approach and a case study that utilizes a multi-phase mobile wellhead gas compression unit (MMWGC) to reduce surface pressure to the sub-atmospheric level and transfer the produced hydrocarbons to the sales line. As a result, the liquid column will unload in a zero-flaring manner, and the life of the producing well will extend considerably. The MMWGC unit was able to successfully kick off a dead well to produce up to 10 MMSCFD after reducing the surface pressure for 3 hours. Applying such novelty on a broader scale will not only extend the life of the producing wells yet will also provide a zero-flaring, economically and environmentally preferred solution.

Keywords: petroleum engineering, zero-flaring, liquid loading, well revival

Procedia PDF Downloads 79
1688 Model Based Development of a Processing Map for Friction Stir Welding of AA7075

Authors: Elizabeth Hoyos, Hernán Alvarez, Diana Lopez, Yesid Montoya

Abstract:

The main goal of this research relates to the modeling of FSW from a different or unusual perspective coming from mechanical engineering, particularly looking for a way to establish process windows by assessing soundness of the joints as a priority and with the added advantage of lower computational time. This paper presents the use of a previously developed model applied to specific aspects of soundness evaluation of AA7075 FSW welds. EMSO software (Environment for Modeling, Simulation, and Optimization) was used for simulation and an adapted CNC machine was used for actual welding. This model based approach showed good agreement with the experimental data, from which it is possible to set a window of operation for commercial aluminum alloy AA7075, all with low computational costs and employing simple quality indicators that can be used by non-specialized users in process modeling.

Keywords: aluminum AA7075, friction stir welding, phenomenological based semiphysical model, processing map

Procedia PDF Downloads 239
1687 General Mathematical Framework for Analysis of Cattle Farm System

Authors: Krzysztof Pomorski

Abstract:

In the given work we present universal mathematical framework for modeling of cattle farm system that can set and validate various hypothesis that can be tested against experimental data. The presented work is preliminary but it is expected to be valid tool for future deeper analysis that can result in new class of prediction methods allowing early detection of cow dieseaes as well as cow performance. Therefore the presented work shall have its meaning in agriculture models and in machine learning as well. It also opens the possibilities for incorporation of certain class of biological models necessary in modeling of cow behavior and farm performance that might include the impact of environment on the farm system. Particular attention is paid to the model of coupled oscillators that it the basic building hypothesis that can construct the model showing certain periodic or quasiperiodic behavior.

Keywords: coupled ordinary differential equations, cattle farm system, numerical methods, stochastic differential equations

Procedia PDF Downloads 127
1686 Optimizing of Machining Parameters of Plastic Material Using Taguchi Method

Authors: Jumazulhisham Abdul Shukor, Mohd. Sazali Said, Roshanizah Harun, Shuib Husin, Ahmad Razlee Ab Kadir

Abstract:

This paper applies Taguchi Optimization Method in determining the best machining parameters for pocket milling process on Polypropylene (PP) using CNC milling machine where the surface roughness is considered and the Carbide inserts cutting tool are used. Three machining parameters; speed, feed rate and depth of cut are investigated along three levels; low, medium and high of each parameter (Taguchi Orthogonal Arrays). The setting of machining parameters were determined by using Taguchi Method and the Signal-to-Noise (S/N) ratio are assessed to define the optimal levels and to predict the effect of surface roughness with assigned parameters based on L9. The final experimental outcomes are presented to prove the optimization parameters recommended by manufacturer are accurate.

Keywords: inserts, milling process, signal-to-noise (S/N) ratio, surface roughness, Taguchi Optimization Method

Procedia PDF Downloads 606
1685 The Methanotrophic Activity in a Landfill Bio-Cover through a Subzero Winter

Authors: Parvin Berenjkar, Qiuyan Yuan, Richard Sparling, Stan Lozecznik

Abstract:

Landfills highly contribute to anthropological global warming through CH₄ emissions. Landfills are usually capped by a conventional soil cover to control the migration of gases. Methane is consumed by CH₄-oxidizing microorganisms known as methanotrophs that naturally exist in the landfill soil cover. The growth of methanotrophs can be optimized in a bio-cover that typically consists of a gas distribution layer (GDL) to homogenize landfill gas fluxes and an overlying oxidation layer composed of suitable materials that support methanotrophic populations. Materials such as mature yard waste composts can provide an inexpensive and favourable porous support for the growth and activity of methanotrophs. In areas with seasonal cold climates, it is valuable to know if methanotrophs in a bio-cover can survive in winter until the next spring, and how deep they are active in the bio-cover to mitigate CH₄. In this study, a pilot bio-cover was constructed in a closed landfill cell in Winnipeg that has a very cold climate in Canada. The bio-cover has a surface area of 2.5 m x 3.5 m and 1.5 m of depth, filled with 50 cm of gravel as a GDL and 70 cm of biosolids compost amended with yard and leaf waste compost. The observed in situ potential of methanotrophs for CH₄ oxidation was investigated at a specific period of time from December 2016 to April 2017 as well as November 2017 to April 2018, when the transition to surface frost and thawing happens in the bio-cover. Compost samples taken from different depths of the bio-cover were incubated in the laboratory under standardized conditions; an optimal air: methane atmosphere, at 22ºC, but at in situ moisture content. Results showed that the methanotrophs were alive oxidizing methane without a lag, indicating that there was the potential for methanotrophic activity at some depths of the bio-cover.

Keywords: bio-cover, global warming, landfill, methanotrophic activity

Procedia PDF Downloads 105
1684 Design of a Small and Medium Enterprise Growth Prediction Model Based on Web Mining

Authors: Yiea Funk Te, Daniel Mueller, Irena Pletikosa Cvijikj

Abstract:

Small and medium enterprises (SMEs) play an important role in the economy of many countries. When the overall world economy is considered, SMEs represent 95% of all businesses in the world, accounting for 66% of the total employment. Existing studies show that the current business environment is characterized as highly turbulent and strongly influenced by modern information and communication technologies, thus forcing SMEs to experience more severe challenges in maintaining their existence and expanding their business. To support SMEs at improving their competitiveness, researchers recently turned their focus on applying data mining techniques to build risk and growth prediction models. However, data used to assess risk and growth indicators is primarily obtained via questionnaires, which is very laborious and time-consuming, or is provided by financial institutes, thus highly sensitive to privacy issues. Recently, web mining (WM) has emerged as a new approach towards obtaining valuable insights in the business world. WM enables automatic and large scale collection and analysis of potentially valuable data from various online platforms, including companies’ websites. While WM methods have been frequently studied to anticipate growth of sales volume for e-commerce platforms, their application for assessment of SME risk and growth indicators is still scarce. Considering that a vast proportion of SMEs own a website, WM bears a great potential in revealing valuable information hidden in SME websites, which can further be used to understand SME risk and growth indicators, as well as to enhance current SME risk and growth prediction models. This study aims at developing an automated system to collect business-relevant data from the Web and predict future growth trends of SMEs by means of WM and data mining techniques. The envisioned system should serve as an 'early recognition system' for future growth opportunities. In an initial step, we examine how structured and semi-structured Web data in governmental or SME websites can be used to explain the success of SMEs. WM methods are applied to extract Web data in a form of additional input features for the growth prediction model. The data on SMEs provided by a large Swiss insurance company is used as ground truth data (i.e. growth-labeled data) to train the growth prediction model. Different machine learning classification algorithms such as the Support Vector Machine, Random Forest and Artificial Neural Network are applied and compared, with the goal to optimize the prediction performance. The results are compared to those from previous studies, in order to assess the contribution of growth indicators retrieved from the Web for increasing the predictive power of the model.

Keywords: data mining, SME growth, success factors, web mining

Procedia PDF Downloads 244
1683 Learning Grammars for Detection of Disaster-Related Micro Events

Authors: Josef Steinberger, Vanni Zavarella, Hristo Tanev

Abstract:

Natural disasters cause tens of thousands of victims and massive material damages. We refer to all those events caused by natural disasters, such as damage on people, infrastructure, vehicles, services and resource supply, as micro events. This paper addresses the problem of micro - event detection in online media sources. We present a natural language grammar learning algorithm and apply it to online news. The algorithm in question is based on distributional clustering and detection of word collocations. We also explore the extraction of micro-events from social media and describe a Twitter mining robot, who uses combinations of keywords to detect tweets which talk about effects of disasters.

Keywords: online news, natural language processing, machine learning, event extraction, crisis computing, disaster effects, Twitter

Procedia PDF Downloads 461
1682 Analysis of Matching Pursuit Features of EEG Signal for Mental Tasks Classification

Authors: Zin Mar Lwin

Abstract:

Brain Computer Interface (BCI) Systems have developed for people who suffer from severe motor disabilities and challenging to communicate with their environment. BCI allows them for communication by a non-muscular way. For communication between human and computer, BCI uses a type of signal called Electroencephalogram (EEG) signal which is recorded from the human„s brain by means of an electrode. The electroencephalogram (EEG) signal is an important information source for knowing brain processes for the non-invasive BCI. Translating human‟s thought, it needs to classify acquired EEG signal accurately. This paper proposed a typical EEG signal classification system which experiments the Dataset from “Purdue University.” Independent Component Analysis (ICA) method via EEGLab Tools for removing artifacts which are caused by eye blinks. For features extraction, the Time and Frequency features of non-stationary EEG signals are extracted by Matching Pursuit (MP) algorithm. The classification of one of five mental tasks is performed by Multi_Class Support Vector Machine (SVM). For SVMs, the comparisons have been carried out for both 1-against-1 and 1-against-all methods.

Keywords: BCI, EEG, ICA, SVM

Procedia PDF Downloads 254
1681 Business Intelligent to a Decision Support Tool for Green Entrepreneurship: Meso and Macro Regions

Authors: Anishur Rahman, Maria Areias, Diogo Simões, Ana Figeuiredo, Filipa Figueiredo, João Nunes

Abstract:

The circular economy (CE) has gained increased awareness among academics, businesses, and decision-makers as it stimulates resource circularity in the production and consumption systems. A large epistemological study has explored the principles of CE, but scant attention eagerly focused on analysing how CE is evaluated, consented to, and enforced using economic metabolism data and business intelligent framework. Economic metabolism involves the ongoing exchange of materials and energy within and across socio-economic systems and requires the assessment of vast amounts of data to provide quantitative analysis related to effective resource management. Limited concern, the present work has focused on the regional flows pilot region from Portugal. By addressing this gap, this study aims to promote eco-innovation and sustainability in the regions of Intermunicipal Communities Região de Coimbra, Viseu Dão Lafões and Beiras e Serra da Estrela, using this data to find precise synergies in terms of material flows and give companies a competitive advantage in form of valuable waste destinations, access to new resources and new markets, cost reduction and risk sharing benefits. In our work, emphasis on applying artificial intelligence (AI) and, more specifically, on implementing state-of-the-art deep learning algorithms is placed, contributing to construction a business intelligent approach. With the emergence of new approaches generally highlighted under the sub-heading of AI and machine learning (ML), the methods for statistical analysis of complex and uncertain production systems are facing significant changes. Therefore, various definitions of AI and its differences from traditional statistics are presented, and furthermore, ML is introduced to identify its place in data science and the differences in topics such as big data analytics and in production problems that using AI and ML are identified. A lifecycle-based approach is then taken to analyse the use of different methods in each phase to identify the most useful technologies and unifying attributes of AI in manufacturing. Most of macroeconomic metabolisms models are mainly direct to contexts of large metropolis, neglecting rural territories, so within this project, a dynamic decision support model coupled with artificial intelligence tools and information platforms will be developed, focused on the reality of these transition zones between the rural and urban. Thus, a real decision support tool is under development, which will surpass the scientific developments carried out to date and will allow to overcome imitations related to the availability and reliability of data.

Keywords: circular economy, artificial intelligence, economic metabolisms, machine learning

Procedia PDF Downloads 50
1680 Audio-Visual Co-Data Processing Pipeline

Authors: Rita Chattopadhyay, Vivek Anand Thoutam

Abstract:

Speech is the most acceptable means of communication where we can quickly exchange our feelings and thoughts. Quite often, people can communicate orally but cannot interact or work with computers or devices. It’s easy and quick to give speech commands than typing commands to computers. In the same way, it’s easy listening to audio played from a device than extract output from computers or devices. Especially with Robotics being an emerging market with applications in warehouses, the hospitality industry, consumer electronics, assistive technology, etc., speech-based human-machine interaction is emerging as a lucrative feature for robot manufacturers. Considering this factor, the objective of this paper is to design the “Audio-Visual Co-Data Processing Pipeline.” This pipeline is an integrated version of Automatic speech recognition, a Natural language model for text understanding, object detection, and text-to-speech modules. There are many Deep Learning models for each type of the modules mentioned above, but OpenVINO Model Zoo models are used because the OpenVINO toolkit covers both computer vision and non-computer vision workloads across Intel hardware and maximizes performance, and accelerates application development. A speech command is given as input that has information about target objects to be detected and start and end times to extract the required interval from the video. Speech is converted to text using the Automatic speech recognition QuartzNet model. The summary is extracted from text using a natural language model Generative Pre-Trained Transformer-3 (GPT-3). Based on the summary, essential frames from the video are extracted, and the You Only Look Once (YOLO) object detection model detects You Only Look Once (YOLO) objects on these extracted frames. Frame numbers that have target objects (specified objects in the speech command) are saved as text. Finally, this text (frame numbers) is converted to speech using text to speech model and will be played from the device. This project is developed for 80 You Only Look Once (YOLO) labels, and the user can extract frames based on only one or two target labels. This pipeline can be extended for more than two target labels easily by making appropriate changes in the object detection module. This project is developed for four different speech command formats by including sample examples in the prompt used by Generative Pre-Trained Transformer-3 (GPT-3) model. Based on user preference, one can come up with a new speech command format by including some examples of the respective format in the prompt used by the Generative Pre-Trained Transformer-3 (GPT-3) model. This pipeline can be used in many projects like human-machine interface, human-robot interaction, and surveillance through speech commands. All object detection projects can be upgraded using this pipeline so that one can give speech commands and output is played from the device.

Keywords: OpenVINO, automatic speech recognition, natural language processing, object detection, text to speech

Procedia PDF Downloads 57
1679 Networked Implementation of Milling Stability Optimization with Bayesian Learning

Authors: Christoph Ramsauer, Jaydeep Karandikar, Tony Schmitz, Friedrich Bleicher

Abstract:

Machining stability is an important limitation to discrete part machining. In this work, a networked implementation of milling stability optimization with Bayesian learning is presented. The milling process was monitored with a wireless sensory tool holder instrumented with an accelerometer at the Vienna University of Technology, Vienna, Austria. The recorded data from a milling test cut is used to classify the cut as stable or unstable based on the frequency analysis. The test cut result is fed to a Bayesian stability learning algorithm at the University of Tennessee, Knoxville, Tennessee, USA. The algorithm calculates the probability of stability as a function of axial depth of cut and spindle speed and recommends the parameters for the next test cut. The iterative process between two transatlantic locations repeats until convergence to a stable optimal process parameter set is achieved.

Keywords: machining stability, machine learning, sensor, optimization

Procedia PDF Downloads 186
1678 Social Media Marketing and Blog Usage in Business Schools: An Exploratory Study

Authors: Grzegorz Mazurek, Michal Kucia

Abstract:

The following study of a preliminary character, presents a first step of multifaceted study on the usage of social media in HEIs. It examines a significance, potential, and managerial implications of social media marketing and blogs usage in HEIs – namely in the sphere of business schools. Social media – particularly: blogging and virtual platforms such as Facebook, Twitter or Instagram have been covered at length in publications of both theoretical and practical nature as of late. Still, the amount of information related to the framework of application of social media in HEIs is rather limited. A pre-designed observation matrix has been used to collect primary data found at websites of different HEIs and to include blog observations. Additionally, a pilot study based on on-line questionnaires with marketing officers of HEI schools has been conducted. The main aim of the study was to identify and elaborate on matters like the scope of social media usage (and blogs in particular) in practice, recognition of the functions fulfilled by social media and blogs, or the anticipated potential of social media for HEIs. The study reveals that the majority of business schools highly ranked in Financial Times rankings use social media and interactive functionalities of their web sites, however, mostly for promotional reasons, and they are targeted at new students. The usage of blogs, though, is not so common and in most cases, blogs are independent platforms, not managed but supported by organizations. Managers and specialists point to lack of resources, insufficient users’ engagement and lack of strategic approach to social media as the main reasons of not advancing in the usage of blogs and social media platforms.

Keywords: blogs, social media marketing, higher education institutions, business schools, value co-creation

Procedia PDF Downloads 245
1677 3D Classification Optimization of Low-Density Airborne Light Detection and Ranging Point Cloud by Parameters Selection

Authors: Baha Eddine Aissou, Aichouche Belhadj Aissa

Abstract:

Light detection and ranging (LiDAR) is an active remote sensing technology used for several applications. Airborne LiDAR is becoming an important technology for the acquisition of a highly accurate dense point cloud. A classification of airborne laser scanning (ALS) point cloud is a very important task that still remains a real challenge for many scientists. Support vector machine (SVM) is one of the most used statistical learning algorithms based on kernels. SVM is a non-parametric method, and it is recommended to be used in cases where the data distribution cannot be well modeled by a standard parametric probability density function. Using a kernel, it performs a robust non-linear classification of samples. Often, the data are rarely linearly separable. SVMs are able to map the data into a higher-dimensional space to become linearly separable, which allows performing all the computations in the original space. This is one of the main reasons that SVMs are well suited for high-dimensional classification problems. Only a few training samples, called support vectors, are required. SVM has also shown its potential to cope with uncertainty in data caused by noise and fluctuation, and it is computationally efficient as compared to several other methods. Such properties are particularly suited for remote sensing classification problems and explain their recent adoption. In this poster, the SVM classification of ALS LiDAR data is proposed. Firstly, connected component analysis is applied for clustering the point cloud. Secondly, the resulting clusters are incorporated in the SVM classifier. Radial basic function (RFB) kernel is used due to the few numbers of parameters (C and γ) that needs to be chosen, which decreases the computation time. In order to optimize the classification rates, the parameters selection is explored. It consists to find the parameters (C and γ) leading to the best overall accuracy using grid search and 5-fold cross-validation. The exploited LiDAR point cloud is provided by the German Society for Photogrammetry, Remote Sensing, and Geoinformation. The ALS data used is characterized by a low density (4-6 points/m²) and is covering an urban area located in residential parts of the city Vaihingen in southern Germany. The class ground and three other classes belonging to roof superstructures are considered, i.e., a total of 4 classes. The training and test sets are selected randomly several times. The obtained results demonstrated that a parameters selection can orient the selection in a restricted interval of (C and γ) that can be further explored but does not systematically lead to the optimal rates. The SVM classifier with hyper-parameters is compared with the most used classifiers in literature for LiDAR data, random forest, AdaBoost, and decision tree. The comparison showed the superiority of the SVM classifier using parameters selection for LiDAR data compared to other classifiers.

Keywords: classification, airborne LiDAR, parameters selection, support vector machine

Procedia PDF Downloads 130
1676 The Influence of Brands in E-Sports Spectators

Authors: Rene Kasper, Hyago Ribeiro, Marcelo Curth

Abstract:

Electronic sports, or just e-sports, boast an exponential growth in the interest of the public and large investors. The e-sports teams are equal to classic sports teams, like football, since in their structure they have, besides the athletes, administrators, coaches and even doctors. The concept of team games arises with a very strong social interaction, as it is perceived that users interact with real peers rather than competing with intelligent software. In this sense, electronic games are established as a sociocultural phenomenon and as multidimensional media. Thus, the research aims to identify the profile of users and the importance of brands in the Brazilian electronic sports scene, as well as the relationship of consumers (called fans) with the products and services that occupy the media spaces of the transmissions of sports championships. The research used descriptive quantitative methodology, applied in different e-sports communities, with 160 respondents. The data collection instrument was a survey containing seven questions, which addressed the profile of the participants and their perception on the proposed theme in research. Regarding the profile, the age ranged from 17 to 31 years, of which 93.3% were male and 6.7% female. It was found that 93.3% of the participants had contact with the Brazilian electronic sports scene for at least 2 years, of which 26.7% played between 6 and 12 hours a week and 46.7% played more than 12 hours a week. In addition, it was noticed that income was not a deciding factor to enjoy electronic sports games, because the percentage distribution of participants ranged from 1 to 3 minimum wages (33.3%) and greater than 6 salaries (46.7 %). Regarding the brands, 85.6% emphasized that brands should support the scenario through sponsorship and publicity and 28.6% are attracted to consume brands that advertise in e-sports championships.

Keywords: brands, consumer behavior, e-sports, virtual games

Procedia PDF Downloads 251
1675 Transforming ESL Teaching and Learning with ICT

Authors: Helena Sit

Abstract:

Developing skills in using ICT in the language classroom has been discussed at all educational levels. Digital tools and learning management systems enable teachers to transform their instructional activities while giving learners the opportunity to engage with virtual communities. In the field of English as a second language (ESL) teaching and learning, the use of technology-enhanced learning and diverse pedagogical practices continues to grow. Whilst technology and multimodal learning is a way of the future for education, second language teachers now face the predicament as to whether implementing these newer ways of learning is, in fact, beneficial or disadvantageous to learners. Research has shown that integrating multimodality and technology can improve students’ engagement and participation in their English language learning. However, students can experience anxiety or misunderstanding when engaging with E-learning or digital-mediated learning. This paper aims to explore how ESL teaching and learning are transformed via the use of educational technology and what impact it has had on student teachers. Case study is employed in this research. The study reviews the growing presence of technology and multimodality in university language classrooms, discusses their impact on teachers’ pedagogical practices, and proposes scaffolding strategies to help design effective English language courses in the Australian education context. The study sheds light on how pedagogical integration today may offer a way forward for language teachers of tomorrow and provides implications to implement an evidence-informed approach that blends knowledge from research, practice and people experiencing the practice in the digital era.

Keywords: educational technology, ICT in higher education, curriculum design and innovation, teacher education, multiliteracies pedagogy

Procedia PDF Downloads 53