Search results for: collaborative learning approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18803

Search results for: collaborative learning approach

8123 Determinants of Artificial Intelligence Capabilities in Healthcare: The Case of Ethiopia

Authors: Dereje Ferede, Solomon Negash

Abstract:

Artificial Intelligence (AI) is a key enabler and driver to transform and revolutionize the healthcare industries. However, utilizing AI and achieving these benefits is challenging for different sectors in wide-ranging, more difficult for developing economy healthcare. Due to this, real-world clinical execution and implementation of AI have not yet aged. We believe that examining the determinants is key to addressing these challenges. Furthermore, the literature does not yet particularize determinants of AI capabilities and ways of empowering the healthcare ecosystem to develop AI capabilities in a developing economy. Thus, this study aims to position AI as a digital transformation weapon for the healthcare ecosystem by examining AI capability determinants and providing insights on better empowering the healthcare industry to develop AI capabilities. To do so, we base on the technology-organization-environment (TOE) model and will apply a mixed research approach. We will conclude with recommendations based on findings for future practitioners and researchers.

Keywords: artificial intelligence, capability, digital transformation, developing economies, healthcare

Procedia PDF Downloads 211
8122 R-Killer: An Email-Based Ransomware Protection Tool

Authors: B. Lokuketagoda, M. Weerakoon, U. Madushan, A. N. Senaratne, K. Y. Abeywardena

Abstract:

Ransomware has become a common threat in past few years and the recent threat reports show an increase of growth in Ransomware infections. Researchers have identified different variants of Ransomware families since 2015. Lack of knowledge of the user about the threat is a major concern. Ransomware detection methodologies are still growing through the industry. Email is the easiest method to send Ransomware to its victims. Uninformed users tend to click on links and attachments without much consideration assuming the emails are genuine. As a solution to this in this paper R-Killer Ransomware detection tool is introduced. Tool can be integrated with existing email services. The core detection Engine (CDE) discussed in the paper focuses on separating suspicious samples from emails and handling them until a decision is made regarding the suspicious mail. It has the capability of preventing execution of identified ransomware processes. On the other hand, Sandboxing and URL analyzing system has the capability of communication with public threat intelligence services to gather known threat intelligence. The R-Killer has its own mechanism developed in its Proactive Monitoring System (PMS) which can monitor the processes created by downloaded email attachments and identify potential Ransomware activities. R-killer is capable of gathering threat intelligence without exposing the user’s data to public threat intelligence services, hence protecting the confidentiality of user data.

Keywords: ransomware, deep learning, recurrent neural networks, email, core detection engine

Procedia PDF Downloads 193
8121 Risk Assessment of Building Information Modelling Adoption in Construction Projects

Authors: Amirhossein Karamoozian, Desheng Wu, Behzad Abbasnejad

Abstract:

Building information modelling (BIM) is a new technology to enhance the efficiency of project management in the construction industry. In addition to the potential benefits of this useful technology, there are various risks and obstacles to applying it in construction projects. In this study, a decision making approach is presented for risk assessment in BIM adoption in construction projects. Various risk factors of exerting BIM during different phases of the project lifecycle are identified with the help of Delphi method, experts’ opinions and related literature. Afterward, Shannon’s entropy and Fuzzy TOPSIS (Technique for Order Preference by Similarity to Ideal Situation) are applied to derive priorities of the identified risk factors. Results indicated that lack of knowledge between professional engineers about workflows in BIM and conflict of opinions between different stakeholders are the risk factors with the highest priority.

Keywords: risk, BIM, fuzzy TOPSIS, construction projects

Procedia PDF Downloads 213
8120 Electroencephalography (EEG) Analysis of Alcoholic and Control Subjects Using Multiscale Permutation Entropy

Authors: Lal Hussain, Wajid Aziz, Sajjad Ahmed Nadeem, Saeed Arif Shah, Abdul Majid

Abstract:

Brain electrical activity as reflected in Electroencephalography (EEG) have been analyzed and diagnosed using various techniques. Among them, complexity measure, nonlinearity, disorder, and unpredictability play vital role due to the nonlinear interconnection between functional and anatomical subsystem emerged in brain in healthy state and during various diseases. There are many social and economical issues of alcoholic abuse as memory weakness, decision making, impairments, and concentrations etc. Alcoholism not only defect the brains but also associated with emotional, behavior, and cognitive impairments damaging the white and gray brain matters. A recently developed signal analysis method i.e. Multiscale Permutation Entropy (MPE) is proposed to estimate the complexity of long-range temporal correlation time series EEG of Alcoholic and Control subjects acquired from University of California Machine Learning repository and results are compared with MSE. Using MPE, coarsed grained series is first generated and the PE is computed for each coarsed grained time series against the electrodes O1, O2, C3, C4, F2, F3, F4, F7, F8, Fp1, Fp2, P3, P4, T7, and T8. The results computed against each electrode using MPE gives higher significant values as compared to MSE as well as mean rank differences accordingly. Likewise, ROC and Area under the ROC also gives higher separation against each electrode using MPE in comparison to MSE.

Keywords: electroencephalogram (EEG), multiscale permutation entropy (MPE), multiscale sample entropy (MSE), permutation entropy (PE), mann whitney test (MMT), receiver operator curve (ROC), complexity measure

Procedia PDF Downloads 478
8119 The Qualification and Quality of Space Sciences and Space Engineering Education in Turkey

Authors: Hatice Canan Gungor, Ahmet Akdemir

Abstract:

The fields of engineering and technological sciences are increasing in quality and quantity day by day all over the world. Countries have to follow, implement and adapt these developments in order to economical empowerments. In our era, it's possible to follow the rapidly developing technology and to produce new technologies by inquisitive, curious, numerical thinking individuals who can show several approaches to problem solving. In this case, countries should develop te result oriented and need-focused curriculums in university education. As in the whole world, there are more space studies in our country as well. Universities should undertake the task of supply the need for staff of this technological race. In this context, questions about the purpose, content and learning outcomes of the space sciences and space engineering departments in our country will be researched answers to reveal the characteristic of this section. In this study, it was determined in which universities the space engineering and the departments of basic sciences educate with formal education and the contents of this education, and the universities were compared with each other as of 2017. In our country three universities provide Aeronautical and Aerospace Engineering, two universities provide Space Sciences and Technologies, two universities provide Aerospace Engineering, two universities provide Aeronautics and Astronautics Engineering education. In all universities, specialized courses are taught after basic engineering education. But the question that needs to be answered is, do the lessons benefit in practice? The answer of this question will reveal the quality of the education. This paper suggests that surveys be conducted to search for the answer to this question. It's thought to be the base for the next works.

Keywords: education, space engineering, space science, quality of systems

Procedia PDF Downloads 268
8118 Student Project on Using a Spreadsheet for Solving Differential Equations by Euler's Method

Authors: Andriy Didenko, Zanin Kavazovic

Abstract:

Engineering students often have certain difficulties in mastering major theoretical concepts in mathematical courses such as differential equations. Student projects were proposed to motivate students’ learning and can be used as a tool to promote students’ interest in the material. Authors propose a student project that includes the use of Microsoft Excel. This instructional tool is often overlooked by both educators and students. An integral component of the experimental part of such a project is the exploration of an interactive spreadsheet. The aim is to assist engineering students in better understanding of Euler’s method. This method is employed to numerically solve first order differential equations. At first, students are invited to select classic equations from a list presented in a form of a drop-down menu. For each of these equations, students can select and modify certain key parameters and observe the influence of initial condition on the solution. This will give students an insight into the behavior of the method in different configurations as solutions to equations are given in numerical and graphical forms. Further, students could also create their own equations by providing functions of their own choice and a variety of initial conditions. Moreover, they can visualize and explore the impact of the length of the time step on the convergence of a sequence of numerical solutions to the exact solution of the equation. As a final stage of the project, students are encouraged to develop their own spreadsheets for other numerical methods and other types of equations. Such projects promote students’ interest in mathematical applications and further improve their mathematical and programming skills.

Keywords: student project, Euler's method, spreadsheet, engineering education

Procedia PDF Downloads 116
8117 Ideological Manipulations and Cultural-Norm Constraints

Authors: Masoud Hassanzade Novin, Bahloul Salmani

Abstract:

Translation cannot be considered as a simple linguistic act. Through the rise of descriptive approach in the late 1970s and 1980s, translation process managed to meet the requirements of social aspects as well as linguistic approaches. To have the translation considered as the cross-cultural communication through which various cultures communicate in ideological and cultural constraints, the contrastive analysis was conducted in this paper to reveal the distortions imposed in the translated texts. The corpus of the study involved the novel 1984 written by George Orwell and its Persian translated texts which were analyzed through the qualitative type of the research based on critical discourse analysis (CDA) and Toury's norms as well as Lefever's concepts of ideology. Results of the study revealed the point that ideology and the cultural constraints were considered as an important stimulus which can control the process of the translation.

Keywords: critical discourse analysis, ideology, norms, translated texts

Procedia PDF Downloads 324
8116 A Novel Image Steganography Method Based on Mandelbrot Fractal

Authors: Adnan H. M. Al-Helali, Hamza A. Ali

Abstract:

The growth of censorship and pervasive monitoring on the Internet, Steganography arises as a new means of achieving secret communication. Steganography is the art and science of embedding information within electronic media used by common applications and systems. Generally, hiding information of multimedia within images will change some of their properties that may introduce few degradation or unusual characteristics. This paper presents a new image steganography approach for hiding information of multimedia (images, text, and audio) using generated Mandelbrot Fractal image as a cover. The proposed technique has been extensively tested with different images. The results show that the method is a very secure means of hiding and retrieving steganographic information. Experimental results demonstrate that an effective improvement in the values of the Peak Signal to Noise Ratio (PSNR), Mean Square Error (MSE), Normalized Cross Correlation (NCC), and Image Fidelity (IF) over the pervious techniques.

Keywords: fractal image, information hiding, Mandelbrot set fractal, steganography

Procedia PDF Downloads 606
8115 Sloshing-Induced Overflow Assessment of the Seismically-Isolated Nuclear Tanks

Authors: Kihyon Kwon, Hyun T. Park, Gil Y. Chung, Sang-Hoon Lee

Abstract:

This paper focuses on assessing sloshing-induced overflow of the seismically-isolated nuclear tanks based on Fluid-Structure Interaction (FSI) analysis. Typically, fluid motion in the seismically-isolated nuclear tank systems may be rather amplified and even overflowed under earthquake. Sloshing-induced overflow in those structures has to be reliably assessed and predicted since it can often cause critical damages to humans and environments. FSI analysis is herein performed to compute the total cumulative overflowed water volume more accurately, by coupling ANSYS with CFX for structural and fluid analyses, respectively. The approach is illustrated on a nuclear liquid storage tank, Spent Fuel Pool (SFP), forgiven conditions under consideration: different liquid levels, Peak Ground Accelerations (PGAs), and post earthquakes.

Keywords: FSI analysis, seismically-isolated nuclear tank system, sloshing-induced overflow

Procedia PDF Downloads 460
8114 Project Management Agile Model Based on Project Management Body of Knowledge Guideline

Authors: Mehrzad Abdi Khalife, Iraj Mahdavi

Abstract:

This paper presents the agile model for project management process. For project management process, the Project Management Body of Knowledge (PMBOK) guideline has been selected as platform. Combination of computational science and artificial intelligent methodology has been added to the guideline to transfer the standard to agile project management process. The model is the combination of practical standard, computational science and artificial intelligent. In this model, we present communication model and protocols to keep process agile. Here, we illustrate the collaboration man and machine in project management area with artificial intelligent approach.

Keywords: artificial intelligent, conceptual model, man-machine collaboration, project management, standard

Procedia PDF Downloads 324
8113 Reliability Analysis of Geometric Performance of Onboard Satellite Sensors: A Study on Location Accuracy

Authors: Ch. Sridevi, A. Chalapathi Rao, P. Srinivasulu

Abstract:

The location accuracy of data products is a critical parameter in assessing the geometric performance of satellite sensors. This study focuses on reliability analysis of onboard sensors to evaluate their performance in terms of location accuracy performance over time. The analysis utilizes field failure data and employs the weibull distribution to determine the reliability and in turn to understand the improvements or degradations over a period of time. The analysis begins by scrutinizing the location accuracy error which is the root mean square (RMS) error of differences between ground control point coordinates observed on the product and the map and identifying the failure data with reference to time. A significant challenge in this study is to thoroughly analyze the possibility of an infant mortality phase in the data. To address this, the Weibull distribution is utilized to determine if the data exhibits an infant stage or if it has transitioned into the operational phase. The shape parameter beta plays a crucial role in identifying this stage. Additionally, determining the exact start of the operational phase and the end of the infant stage poses another challenge as it is crucial to eliminate residual infant mortality or wear-out from the model, as it can significantly increase the total failure rate. To address this, an approach utilizing the well-established statistical Laplace test is applied to infer the behavior of sensors and to accurately ascertain the duration of different phases in the lifetime and the time required for stabilization. This approach also helps in understanding if the bathtub curve model, which accounts for the different phases in the lifetime of a product, is appropriate for the data and whether the thresholds for the infant period and wear-out phase are accurately estimated by validating the data in individual phases with Weibull distribution curve fitting analysis. Once the operational phase is determined, reliability is assessed using Weibull analysis. This analysis not only provides insights into the reliability of individual sensors with regards to location accuracy over the required period of time, but also establishes a model that can be applied to automate similar analyses for various sensors and parameters using field failure data. Furthermore, the identification of the best-performing sensor through this analysis serves as a benchmark for future missions and designs, ensuring continuous improvement in sensor performance and reliability. Overall, this study provides a methodology to accurately determine the duration of different phases in the life data of individual sensors. It enables an assessment of the time required for stabilization and provides insights into the reliability during the operational phase and the commencement of the wear-out phase. By employing this methodology, designers can make informed decisions regarding sensor performance with regards to location accuracy, contributing to enhanced accuracy in satellite-based applications.

Keywords: bathtub curve, geometric performance, Laplace test, location accuracy, reliability analysis, Weibull analysis

Procedia PDF Downloads 58
8112 Integration of Knowledge and Metadata for Complex Data Warehouses and Big Data

Authors: Jean Christian Ralaivao, Fabrice Razafindraibe, Hasina Rakotonirainy

Abstract:

This document constitutes a resumption of work carried out in the field of complex data warehouses (DW) relating to the management and formalization of knowledge and metadata. It offers a methodological approach for integrating two concepts, knowledge and metadata, within the framework of a complex DW architecture. The objective of the work considers the use of the technique of knowledge representation by description logics and the extension of Common Warehouse Metamodel (CWM) specifications. This will lead to a fallout in terms of the performance of a complex DW. Three essential aspects of this work are expected, including the representation of knowledge in description logics and the declination of this knowledge into consistent UML diagrams while respecting or extending the CWM specifications and using XML as pivot. The field of application is large but will be adapted to systems with heteroge-neous, complex and unstructured content and moreover requiring a great (re)use of knowledge such as medical data warehouses.

Keywords: data warehouse, description logics, integration, knowledge, metadata

Procedia PDF Downloads 125
8111 Automated Test Data Generation For some types of Algorithm

Authors: Hitesh Tahbildar

Abstract:

The cost of test data generation for a program is computationally very high. In general case, no algorithm to generate test data for all types of algorithms has been found. The cost of generating test data for different types of algorithm is different. Till date, people are emphasizing the need to generate test data for different types of programming constructs rather than different types of algorithms. The test data generation methods have been implemented to find heuristics for different types of algorithms. Some algorithms that includes divide and conquer, backtracking, greedy approach, dynamic programming to find the minimum cost of test data generation have been tested. Our experimental results say that some of these types of algorithm can be used as a necessary condition for selecting heuristics and programming constructs are sufficient condition for selecting our heuristics. Finally we recommend the different heuristics for test data generation to be selected for different types of algorithms.

Keywords: ongest path, saturation point, lmax, kL, kS

Procedia PDF Downloads 388
8110 Material Detection by Phase Shift Cavity Ring-Down Spectroscopy

Authors: Rana Muhammad Armaghan Ayaz, Yigit Uysallı, Nima Bavili, Berna Morova, Alper Kiraz

Abstract:

Traditional optical methods for example resonance wavelength shift and cavity ring-down spectroscopy used for material detection and sensing have disadvantages, for example, less resistance to laser noise, temperature fluctuations and extraction of the required information can be a difficult task like ring downtime in case of cavity ring-down spectroscopy. Phase shift cavity ring down spectroscopy is not only easy to use but is also capable of overcoming the said problems. This technique compares the phase difference between the signal coming out of the cavity with the reference signal. Detection of any material is made by the phase difference between them. By using this technique, air, water, and isopropyl alcohol can be recognized easily. This Methodology has far-reaching applications and can be used in air pollution detection, human breath analysis and many more.

Keywords: materials, noise, phase shift, resonance wavelength, sensitivity, time domain approach

Procedia PDF Downloads 138
8109 Reactivity Study on South African Calcium Based Material Using a pH-Stat and Citric Acid: A Statistical Approach

Authors: Hilary Rutto, Mbali Chiliza, Tumisang Seodigeng

Abstract:

The study on reactivity of calcined calcium-based material is very important in dry flue gas desulphurisation (FGD) process, so as to produce absorbent with high sulphur dioxide capture capacity during the hydration process. The effect of calcining temperature and time on the reactivity of calcined limestone material were investigated. In this study, the reactivity was measured using a pH stat apparatus and also confirming the result by performing citric acid reactivity test. The reactivity was calculated using the shrinking core model. Based on the experiments, a mathematical model is developed to correlate the effect of time and temperature to the reactivity of absorbent. The calcination process variables were temperature (700 -1000°C) and time (1-6 hrs). It was found that reactivity increases with an increase in time and temperature.

Keywords: reactivity, citric acid, calcination, time

Procedia PDF Downloads 205
8108 Analyzing Claude Debussy’s Piano Preludes by Focusing on His Recordings

Authors: Parham Bakhtiari

Abstract:

Between 1910 and 1912, Claude Debussy recorded twelve of his solo piano pieces. Although Debussy frequently provided advice to his students on performing while they followed the written notes when performing, his personal recordings are characterized by creative liberties and unique freedom interpretations. Debussy's use of numerous interpretive gestures in these recordings is fascinating and corresponds with the techniques utilized by French Baroque keyboard performers. This paper will situate Debussy's presentation in the Baroque musical approach. Initially, we will discuss the recording by analyzing Welte-Mignon's used technology to guarantee the reliability of these recordings. Then, we will find commonalities in the intricate performances of harpsichord musicians who played in the 1600s and 1700s and recordings of Debussy. Finally, by drawing comparisons, we will review the patterns by contrasting Debussy's execution with recordings of the same pieces from the latter half of the 20th century as striving for improved presentations while limiting artistic freedom.

Keywords: music, Debussy, piano, performance, prelude

Procedia PDF Downloads 24
8107 Combining the Deep Neural Network with the K-Means for Traffic Accident Prediction

Authors: Celso L. Fernando, Toshio Yoshii, Takahiro Tsubota

Abstract:

Understanding the causes of a road accident and predicting their occurrence is key to preventing deaths and serious injuries from road accident events. Traditional statistical methods such as the Poisson and the Logistics regressions have been used to find the association of the traffic environmental factors with the accident occurred; recently, an artificial neural network, ANN, a computational technique that learns from historical data to make a more accurate prediction, has emerged. Although the ability to make accurate predictions, the ANN has difficulty dealing with highly unbalanced attribute patterns distribution in the training dataset; in such circumstances, the ANN treats the minority group as noise. However, in the real world data, the minority group is often the group of interest; e.g., in the road traffic accident data, the events of the accident are the group of interest. This study proposes a combination of the k-means with the ANN to improve the predictive ability of the neural network model by alleviating the effect of the unbalanced distribution of the attribute patterns in the training dataset. The results show that the proposed method improves the ability of the neural network to make a prediction on a highly unbalanced distributed attribute patterns dataset; however, on an even distributed attribute patterns dataset, the proposed method performs almost like a standard neural network.

Keywords: accident risks estimation, artificial neural network, deep learning, k-mean, road safety

Procedia PDF Downloads 140
8106 Cooperative Coevolution for Neuro-Evolution of Feed Forward Networks for Time Series Prediction Using Hidden Neuron Connections

Authors: Ravneil Nand

Abstract:

Cooperative coevolution uses problem decomposition methods to solve a larger problem. The problem decomposition deals with breaking down the larger problem into a number of smaller sub-problems depending on their method. Different problem decomposition methods have their own strengths and limitations depending on the neural network used and application problem. In this paper we are introducing a new problem decomposition method known as Hidden-Neuron Level Decomposition (HNL). The HNL method is competing with established problem decomposition method in time series prediction. The results show that the proposed approach has improved the results in some benchmark data sets when compared to the standalone method and has competitive results when compared to methods from literature.

Keywords: cooperative coevaluation, feed forward network, problem decomposition, neuron, synapse

Procedia PDF Downloads 316
8105 Care and Support for Infants and Toddlers with Special Needs

Authors: Florence A. Undiyaundeye, Aniashie Akpanke

Abstract:

Early identification of developmental disorders in infants and toddlers is critical for the well being of children. It is also an integral function of the primary care medical provider and the early care given in the home or crèche. This paper is focused at providing information on special need infants and toddlers and strategies to support them in developmental concern to cope with the challenges in and out of the classroom and to interact with their peers without stigmatization and inferiority complex. The target children are from birth through three years of age. There is a strong recommendation for developmental surveillance to be incorporated at every well child preventive care program in training and practical stage of formal school settings. The paper posits that any concerns raised during surveillance should be promptly addressed with standardized developmental screening by appropriate health service providers. In addition screening tests should be administered regularly at age 9+, 19+ and 30 months of these infants. The paper also establishes that the early identification of these developmental challenges of the infants and toddlers should lead to further developmental and medical evaluation, diagnosis and treatment, including early developmental school intervention, control and teaching and learning integration and inclusion for proper career build up. Children diagnosed with developmental disorders should be identified as children with special needs so that management is initiated and its underlying etiology may also drive a range of treatment of the child, to parents. Conselling and school integration as applicable to the child’s specific need and care for sustenance in societal functioning.

Keywords: care, special need, support, infants and toddlers, management and developmental disorders

Procedia PDF Downloads 375
8104 Efficient Position Based Operation Code Authentication

Authors: Hashim Ali, Sheheryar Khan

Abstract:

Security for applications is always been a keen issue of concern. In general, security is to allow access of grant to legal user or to deny non-authorized access to the system. Shoulder surfing is an observation technique to hack an account or to enter into a system. When a malicious observer is capturing or recording the fingers of a user while he is entering sensitive inputs (PIN, Passwords etc.) and may be able to observe user’s password credential. It is very rigorous for a novice user to prevent himself from shoulder surfing or unaided observer in a public place while accessing his account. In order to secure the user account, there are five factors of authentication; they are: “(i) something you have, (ii) something you are, (iii) something you know, (iv) somebody you know, (v) something you process”. A technique has been developed of fifth-factor authentication “something you process” to provide novel approach to the user. In this paper, we have applied position based operational code authentication in such a way to more easy and user friendly to the user.

Keywords: shoulder surfing, malicious observer, sensitive inputs, authentication

Procedia PDF Downloads 256
8103 Model Order Reduction Using Hybrid Genetic Algorithm and Simulated Annealing

Authors: Khaled Salah

Abstract:

Model order reduction has been one of the most challenging topics in the past years. In this paper, a hybrid solution of genetic algorithm (GA) and simulated annealing algorithm (SA) are used to approximate high-order transfer functions (TFs) to lower-order TFs. In this approach, hybrid algorithm is applied to model order reduction putting in consideration improving accuracy and preserving the properties of the original model which are two important issues for improving the performance of simulation and computation and maintaining the behavior of the original complex models being reduced. Compared to conventional mathematical methods that have been used to obtain a reduced order model of high order complex models, our proposed method provides better results in terms of reducing run-time. Thus, the proposed technique could be used in electronic design automation (EDA) tools.

Keywords: genetic algorithm, simulated annealing, model reduction, transfer function

Procedia PDF Downloads 131
8102 Optimization Design of Superposition Wave Form Automotive Exhaust Bellows Structure

Authors: Zhang Jianrun, He Tangling

Abstract:

Superposition wave form automotive exhaust bellows is a new type of bellows, which has the characteristics of large compensation, good vibration isolation performance and long life. It has been paid more and more attention and applications in automotive exhaust pipe system. Aiming at the lack of current design methods of superposition wave form automotive exhaust bellows, this paper proposes a response surface parameter optimization method where the fatigue life and vibration transmissibility of the bellows are set as objectives. The parametric modeling of bellow structure is also adopted to achieve the high efficiency in the design. The approach proposed in this paper provides a new way for the design of superposition wave form automotive exhaust bellows. It embodies good engineering application value.

Keywords: superposition wave form, exhaust bellows, optimization, vibration, fatigue life

Procedia PDF Downloads 78
8101 Improoving Readability for Tweet Contextualization Using Bipartite Graphs

Authors: Amira Dhokar, Lobna Hlaoua, Lotfi Ben Romdhane

Abstract:

Tweet contextualization (TC) is a new issue that aims to answer questions of the form 'What is this tweet about?' The idea of this task was imagined as an extension of a previous area called multi-document summarization (MDS), which consists in generating a summary from many sources. In both TC and MDS, the summary should ideally contain the most relevant information of the topic that is being discussed in the source texts (for MDS) and related to the query (for TC). Furthermore of being informative, a summary should be coherent, i.e. well written to be readable and grammatically compact. Hence, coherence is an essential characteristic in order to produce comprehensible texts. In this paper, we propose a new approach to improve readability and coherence for tweet contextualization based on bipartite graphs. The main idea of our proposed method is to reorder sentences in a given paragraph by combining most expressive words detection and HITS (Hyperlink-Induced Topic Search) algorithm to make up a coherent context.

Keywords: bipartite graphs, readability, summarization, tweet contextualization

Procedia PDF Downloads 179
8100 Defining a Holistic Approach for Model-Based System Engineering: Paradigm and Modeling Requirements

Authors: Hycham Aboutaleb, Bruno Monsuez

Abstract:

Current systems complexity has reached a degree that requires addressing conception and design issues while taking into account all the necessary aspects. Therefore, one of the main challenges is the way complex systems are specified and designed. The exponential growing effort, cost and time investment of complex systems in modeling phase emphasize the need for a paradigm, a framework and a environment to handle the system model complexity. For that, it is necessary to understand the expectations of the human user of the model and his limits. This paper presents a generic framework for designing complex systems, highlights the requirements a system model needs to fulfill to meet human user expectations, and defines the refined functional as well as non functional requirements modeling tools needs to meet to be useful in model-based system engineering.

Keywords: system modeling, modeling language, modeling requirements, framework

Procedia PDF Downloads 517
8099 Estimation and Forecasting with a Quantile AR Model for Financial Returns

Authors: Yuzhi Cai

Abstract:

This talk presents a Bayesian approach to quantile autoregressive (QAR) time series model estimation and forecasting. We establish that the joint posterior distribution of the model parameters and future values is well defined. The associated MCMC algorithm for parameter estimation and forecasting converges to the posterior distribution quickly. We also present a combining forecasts technique to produce more accurate out-of-sample forecasts by using a weighted sequence of fitted QAR models. A moving window method to check the quality of the estimated conditional quantiles is developed. We verify our methodology using simulation studies and then apply it to currency exchange rate data. An application of the method to the USD to GBP daily currency exchange rates will also be discussed. The results obtained show that an unequally weighted combining method performs better than other forecasting methodology.

Keywords: combining forecasts, MCMC, quantile modelling, quantile forecasting, predictive density functions

Procedia PDF Downloads 336
8098 The International Prohibition of Religiously-Motivated 'Incitement' to Violence

Authors: J. D. Temperman

Abstract:

Introduction: In particular, in relation to religion, the meaning and scope of freedom of expression have been tested in recent times. This paper investigates the legal justifications for restrictions that have been suggested in this area and asks whether they are sustainable from an international human rights perspective. The universal human rights instruments, particularly the UN International Covenant on Civil and Political Rights (ICCPR), are increasingly geared towards eradicating ‘incitement’ to contingent harms like violence or discrimination, whilst forms of extreme speech that fall short of such incitement are to be protected rather than countered by states. Human Rights Committee’s draft-General Comment on freedom of expression, adopted in 2011, provides another strong indication that this is the envisaged way forward: repealing anti-blasphemy and anti-religious defamation laws, whilst simultaneously increasing efforts to combat ‘incitement’. Within regional human rights frameworks, notably the European Convention system, judgments have in fact supported legal restrictions on both hate speech, holocaust denial, and blasphemy or religious defamation. Major contributions to scholarship: This paper proposes an actus reus for the offense of ‘advocacy of religious hatred that constitutes incitement to discrimination or violence’, as enshrined in Article 20(2) of the UN ICCPR. In underscoring the high threshold of ‘incitement’, the author distinguishes this offense from such notions as ‘blasphemy’ or ‘defamation of religions’. In addition to treating the said provision as a sui generis prohibition, the question is addresses whether a ‘right to be protected against incitement’ may be distilled from the ICCPR. Furthermore, the author will discuss the question of how to judge incitement; notably, is mens rea required to convict someone of incitement, and if so, what degree of mens rea? This analysis also includes the question how to balance content and context factors when addressing alleged instances of incitement, notably what factors make provide for a likelihood that imminent acts of violence or discrimination will ensue from an inciteful speech act? Methodology: This paper takes a double comparative approach: (i) it endeavours to compare and contrast monitoring bodies’ approach to incitement (notably, the UN Human Rights Committee, but also the UN Committee on the Elimination of Racial Discrimination which monitors states’ compliance with Article 4 of ICERD on incitement); and (ii) it endeavours to chart and compare and analyse from an international human rights perspective recent forms of state practice in the field of dealing with incitement (i.e. a comparative legal analysis and vertical human rights analysis of newly emerging incitement legislation in the light of the said international standards). Conclusion: This paper conceptualizes a legal notion – ‘incitement’ – encapsulated in international human rights law that may have a profound bearing on contemporary challenges of radicalization and religious strife.

Keywords: incitement, international human rights law, religious hatred, violence

Procedia PDF Downloads 294
8097 Structural Identification for Layered Composite Structures through a Wave and Finite Element Methodology

Authors: Rilwan Kayode Apalowo, Dimitrios Chronopoulos

Abstract:

An approach for identifying the geometric and material characteristics of layered composite structures through an inverse wave and finite element methodology is proposed. These characteristics are obtained through multi-frequency single shot measurements. However, it is established that the frequency regime of the measurements does not matter, meaning that both ultrasonic and structural dynamics frequency spectra can be employed. Taking advantage of a full FE (finite elements) description of the periodic composite, the scheme is able to account for arbitrarily complex structures. In order to demonstrate the robustness of the presented scheme, it is applied to a sandwich composite panel and results are compared with that of experimental characterization techniques. Excellent agreement is obtained with the experimental measurements.

Keywords: structural identification, non-destructive evaluation, finite elements, wave propagation, layered structures, ultrasound

Procedia PDF Downloads 125
8096 Optimizing Approach for Sifting Process to Solve a Common Type of Empirical Mode Decomposition Mode Mixing

Authors: Saad Al-Baddai, Karema Al-Subari, Elmar Lang, Bernd Ludwig

Abstract:

Empirical mode decomposition (EMD), a new data-driven of time-series decomposition, has the advantage of supposing that a time series is non-linear or non-stationary, as is implicitly achieved in Fourier decomposition. However, the EMD suffers of mode mixing problem in some cases. The aim of this paper is to present a solution for a common type of signals causing of EMD mode mixing problem, in case a signal suffers of an intermittency. By an artificial example, the solution shows superior performance in terms of cope EMD mode mixing problem comparing with the conventional EMD and Ensemble Empirical Mode decomposition (EEMD). Furthermore, the over-sifting problem is also completely avoided; and computation load is reduced roughly six times compared with EEMD, an ensemble number of 50.

Keywords: empirical mode decomposition (EMD), mode mixing, sifting process, over-sifting

Procedia PDF Downloads 377
8095 A Method for Modeling Flexible Manipulators: Transfer Matrix Method with Finite Segments

Authors: Haijie Li, Xuping Zhang

Abstract:

This paper presents a computationally efficient method for the modeling of robot manipulators with flexible links and joints. This approach combines the Discrete Time Transfer Matrix Method with the Finite Segment Method, in which the flexible links are discretized by a number of rigid segments connected by torsion springs; and the flexibility of joints are modeled by torsion springs. The proposed method avoids the global dynamics and has the advantage of modeling non-uniform manipulators. Experiments and simulations of a single-link flexible manipulator are conducted for verifying the proposed methodologies. The simulations of a three-link robot arm with links and joints flexibility are also performed.

Keywords: flexible manipulator, transfer matrix method, linearization, finite segment method

Procedia PDF Downloads 419
8094 Municipal Solid Waste Management in an Unplanned Hill Station in India

Authors: Moanaro Ao, Nzanthung Ngullie

Abstract:

Municipal solid waste management (MSWM) has unique challenges in hilly urban settlements. Efforts have been taken by municipalities, private players, non-governmental organizations, etc. for managing solid waste by preventing its generation, reusing, and recovering them into useful products to the extent possible, thereby minimizing its impact on the environment and human health. However, there are many constraints that lead to inadequate management of solid waste. Kohima is an unplanned hill station city in the North Eastern Region of India. The city is facing numerous issues due to the mismanagement of the MSW generated. Kohima Municipal Council (KMC) is the Urban Local Body (ULB) responsible for providing municipal services. The present MSWM system in Kohima comprises of collection, transportation, and disposal of waste without any treatment. Several efforts and experimental projects on waste management have been implemented without any success. Waste management in Kohima city is challenging due to its remote location, difficult topography, dispersed settlements within the city, sensitive ecosystem, etc. Furthermore, the narrow road network in Kohima with limited scope for expansion, inadequate infrastructure facilities, and financial constraints of the ULB add up to the problems faced in managing solid waste. This hill station also has a unique system of traditional local self-governance. Thus, shifting from a traditional system to a modern system in implementing systematic and scientific waste management is also a challenge in itself. This study aims to analyse the existing situation of waste generation, evaluate the effectiveness of the existing management system of MSW, and evolve a strategic approach to achieve a sustainable and resilient MSWM system. The results from the study show that a holistic approach, including social aspects, technical aspects, environmental aspects, and financial aspects, is needed to reform the MSWM system. Stringent adherence to source segregation is required by encouraging public participation through awareness programs. Active involvement of community-based organizations (CBOs) has brought a positive change in sensitizing the public. A waste management model was designed to be adopted at a micro-level such as composting household biodegradable waste and incinerator plants at the community level for non-biodegradable waste. Suitable locations for small waste stations were identified using geographical information system (GIS) tools for waste recovery and recycling. Inculcating the sense of responsibility in every waste generator towards waste management by implementing incentive-based strategies at the Ward level was explored. Initiatives based on the ‘polluters pay principle’ were also explored to make the solid waste management model “self-sustaining”.

Keywords: municipal solid waste management, public participation, source segregation, sustainable

Procedia PDF Downloads 57