Search results for: Full Bayes models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8663

Search results for: Full Bayes models

5843 An Analysis of How Students Perceive Their Self-Efficacy in Online Speaking Classes

Authors: Heny Hartono, Cecilia Titiek Murniati

Abstract:

The pandemic has given teachers and students no other choice but having full online learning. In such an emergency situation as the time of the covid-19 pandemic, the application of LMS (Learner Management System) in higher education is the most reasonable solution for students and teachers. In fact, the online learning requires all elements of a higher education systems, including the human resources, infrastructure, and supporting systems such as the application, server, and stable internet connection. The readiness of the higher education institution in preparing the online system may secure those who are involved in the online learning process. It may also result in students’ self-efficacy in online learning. This research aimed to investigate how students perceive their self-efficacy in online English learning, especially in speaking classes which is considered as a productive language skill. This research collects qualitative data with narrative inquiry involving 25 students of speaking classes as the respondents. The results of this study show that students perceive their self-efficacy in speaking online classes as not very high.

Keywords: self-efficacy, online learning, speaking class, college students, e-learning

Procedia PDF Downloads 87
5842 The Grammatical Dictionary Compiler: A System for Kartvelian Languages

Authors: Liana Lortkipanidze, Nino Amirezashvili, Nino Javashvili

Abstract:

The purpose of the grammatical dictionary is to provide information on the morphological and syntactic characteristics of the basic word in the dictionary entry. The electronic grammatical dictionaries are used as a tool of automated morphological analysis for texts processing. The Georgian Grammatical Dictionary should contain grammatical information for each word: part of speech, type of declension/conjugation, grammatical forms of the word (paradigm), alternative variants of basic word/lemma. In this paper, we present the system for compiling the Georgian Grammatical Dictionary automatically. We propose dictionary-based methods for extending grammatical lexicons. The input lexicon contains only a few number of words with identical grammatical features. The extension is based on similarity measures between features of words; more precisely, we add words to the extended lexicons, which are similar to those, which are already in the grammatical dictionary. Our dictionaries are corpora-based, and for the compiling, we introduce the method for lemmatization of unknown words, i.e., words of which neither full form nor lemma is in the grammatical dictionary.

Keywords: acquisition of lexicon, Georgian grammatical dictionary, lemmatization rules, morphological processor

Procedia PDF Downloads 129
5841 Discrete-Event Modeling and Simulation Methodologies: Past, Present and Future

Authors: Gabriel Wainer

Abstract:

Modeling and Simulation methods have been used to better analyze the behavior of complex physical systems, and it is now common to use simulation as a part of the scientific and technological discovery process. M&S advanced thanks to the improvements in computer technology, which, in many cases, resulted in the development of simulation software using ad-hoc techniques. Formal M&S appeared in order to try to improve the development task of very complex simulation systems. Some of these techniques proved to be successful in providing a sound base for the development of discrete-event simulation models, improving the ease of model definition and enhancing the application development tasks; reducing costs and favoring reuse. The DEVS formalism is one of these techniques, which proved to be successful in providing means for modeling while reducing development complexity and costs. DEVS model development is based on a sound theoretical framework. The independence of M&S tasks made possible to run DEVS models on different environments (personal computers, parallel computers, real-time equipment, and distributed simulators) and middleware. We will present a historical perspective of discrete-event M&S methodologies, showing different modeling techniques. We will introduce DEVS origins and general ideas, and compare it with some of these techniques. We will then show the current status of DEVS M&S, and we will discuss a technological perspective to solve current M&S problems (including real-time simulation, interoperability, and model-centered development techniques). We will show some examples of the current use of DEVS, including applications in different fields. We will finally show current open topics in the area, which include advanced methods for centralized, parallel or distributed simulation, the need for real-time modeling techniques, and our view in these fields.

Keywords: modeling and simulation, discrete-event simulation, hybrid systems modeling, parallel and distributed simulation

Procedia PDF Downloads 309
5840 Establishment and Validation of Correlation Equations to Estimate Volumetric Oxygen Mass Transfer Coefficient (KLa) from Process Parameters in Stirred-Tank Bioreactors Using Response Surface Methodology

Authors: Jantakan Jullawateelert, Korakod Haonoo, Sutipong Sananseang, Sarun Torpaiboon, Thanunthon Bowornsakulwong, Lalintip Hocharoen

Abstract:

Process scale-up is essential for the biological process to increase production capacity from bench-scale bioreactors to either pilot or commercial production. Scale-up based on constant volumetric oxygen mass transfer coefficient (KLa) is mostly used as a scale-up factor since oxygen supply is one of the key limiting factors for cell growth. However, to estimate KLa of culture vessels operated with different conditions are time-consuming since it is considerably influenced by a lot of factors. To overcome the issue, this study aimed to establish correlation equations of KLa and operating parameters in 0.5 L and 5 L bioreactor employed with pitched-blade impeller and gas sparger. Temperature, gas flow rate, agitation speed, and impeller position were selected as process parameters and equations were created using response surface methodology (RSM) based on central composite design (CCD). In addition, the effects of these parameters on KLa were also investigated. Based on RSM, second-order polynomial models for 0.5 L and 5 L bioreactor were obtained with an acceptable determination coefficient (R²) as 0.9736 and 0.9190, respectively. These models were validated, and experimental values showed differences less than 10% from the predicted values. Moreover, RSM revealed that gas flow rate is the most significant parameter while temperature and agitation speed were also found to greatly affect the KLa in both bioreactors. Nevertheless, impeller position was shown to influence KLa in only 5L system. To sum up, these modeled correlations can be used to accurately predict KLa within the specified range of process parameters of two different sizes of bioreactors for further scale-up application.

Keywords: response surface methodology, scale-up, stirred-tank bioreactor, volumetric oxygen mass transfer coefficient

Procedia PDF Downloads 186
5839 Collaborative Management Approach for Logistics Flow Management of Cuban Medicine Supply Chain

Authors: Ana Julia Acevedo Urquiaga, Jose A. Acevedo Suarez, Ana Julia Urquiaga Rodriguez, Neyfe Sablon Cossio

Abstract:

Despite the progress made in logistics and supply chains fields, it is unavoidable the development of business models that use efficiently information to facilitate the integrated logistics flows management between partners. Collaborative management is an important tool for materializing the cooperation between companies, as a way to achieve the supply chain efficiency and effectiveness. The first face of this research was a comprehensive analysis of the collaborative planning on the Cuban companies. It is evident that they have difficulties in supply chains planning where production, supplies and replenishment planning are independent tasks, as well as logistics and distribution operations. Large inventories generate serious financial and organizational problems for entities, demanding increasing levels of working capital that cannot be financed. Problems were found in the efficient application of Information and Communication Technology on business management. The general objective of this work is to develop a methodology that allows the deployment of a planning and control system in a coordinated way on the medicine’s logistics system in Cuba. To achieve these objectives, several mechanisms of supply chain coordination, mathematical programming models, and other management techniques were analyzed to meet the requirements of collaborative logistics management in Cuba. One of the findings is the practical and theoretical inadequacies of the studied models to solve the current situation of the Cuban logistics systems management. To contribute to the tactical-operative management of logistics, the Collaborative Logistics Flow Management Model (CLFMM) is proposed as a tool for the balance of cycles, capacities, and inventories, always to meet the final customers’ demands in correspondence with the service level expected by these. The CLFMM has as center the supply chain planning and control system as a unique information system, which acts on the processes network. The development of the model is based on the empirical methods of analysis-synthesis and the study cases. Other finding is the demonstration of the use of a single information system to support the supply chain logistics management, allows determining the deadlines and quantities required in each process. This ensures that medications are always available to patients and there are no faults that put the population's health at risk. The simulation of planning and control with the CLFMM in medicines such as dipyrone and chlordiazepoxide, during 5 months of 2017, permitted to take measures to adjust the logistic flow, eliminate delayed processes and avoid shortages of the medicines studied. As a result, the logistics cycle efficiency can be increased to 91%, the inventory rotation would increase, and this results in a release of financial resources.

Keywords: collaborative management, medicine logistic system, supply chain planning, tactical-operative planning

Procedia PDF Downloads 161
5838 Transforming Maternity and Neonatal Services in a Middle Eastern Country

Authors: M. A. Brown, K. Hugill, D. Meredith

Abstract:

Since the establishment of midwifery, as a professional identity in its own right, in the early years of the 20th century, midwifery-led models of childbirth have prevailed in many parts of the world. However, in many locations midwives’ scope of practice remains underdeveloped or absent. In Qatar, all births take place in hospital and are under the professional jurisdiction of obstetricians, predominately supported by internationally trained nurse-midwives and obstetric nurses. The strategic vision for health services in Qatar endorsed a desire to provide women with the ‘Best Care Always’ and the introduction of midwifery was seen as a way to achieve this. In 2015 the process of recruiting postgraduate educated Clinical Midwife Specialists from international sources began. The midwives were brought together to initiate an in hospital and community service transformation plan. This plan set out a series of wide-ranging actions to transform maternity and neonatal services to make care safer and give women more health choices. Change in any organization is a complex and dynamic process. This is made even more complex when multifaceted professional and cross cultural factors are involved. This presentation reports upon the motivations and challenges that exist and the progress around introducing a multicultural midwifery model of childbirth care in the state of Qatar. The paper examines and reflects upon the drivers and unique features of childbirth in the country. Despite accomplishments, progress still needs to be made in order to fully implement sustainable changes to further improve care and ensure women and neonates get the ‘Best Care Always’. The progress within the transformation plan highlights how midwifery may coexist with competing models of maternity care to create an innovative, eclectic and culturally sensitive paradigm that can best serve women and neonatal health needs.

Keywords: culture, managing change, midwifery, neonatal, service transformation plan

Procedia PDF Downloads 139
5837 Modeling Socioeconomic and Political Dynamics of Terrorism in Pakistan

Authors: Syed Toqueer, Omer Younus

Abstract:

Terrorism, today, has emerged as a global menace with Pakistan being the most adversely affected state. Therefore, the motive behind this study is to empirically establish the linkage of terrorism with socio-economic (uneven income distribution, poverty and unemployment) and political nexuses so that a policy recommendation can be put forth to better approach this issue in Pakistan. For this purpose, the study employs two competing models, namely, the distributed lag model and OLS, so that findings of the model may be consolidated comprehensively, over the reference period of 1984-2012. The findings of both models are indicative of the fact that uneven income distribution of Pakistan is rather a contributing factor towards terrorism when measured through GDP per capita. This supports the hypothesis that immiserizing modernization theory is applicable for the state of Pakistan where the underprivileged are marginalized. Results also suggest that other socio-economic variables (poverty, unemployment and consumer confidence) can condense the brutality of terrorism once these conditions are catered to and improved. The rational of opportunity cost is at the base of this argument. Poor conditions of employment and poverty reduces the opportunity cost for individuals to be recruited by terrorist organizations as economic returns are considerably low and thus increasing the supply of volunteers and subsequently increasing the intensity of terrorism. The argument of political freedom as a means of lowering terrorism stands true. The more the people are politically repressed the more alternative and illegal means they will find to make their voice heard. Also, the argument that politically transitioning economy faces more terrorism is found applicable for Pakistan. Finally, the study contributes to an ongoing debate on which of the two set of factors are more significant with relation to terrorism by suggesting that socio-economic factors are found to be the primary causes of terrorism for Pakistan.

Keywords: terrorism, socioeconomic conditions, political freedom, distributed lag model, ordinary least square

Procedia PDF Downloads 311
5836 Crude Oil and Stocks Markets: Prices and Uncertainty Transmission Analysis

Authors: Kamel Malik Bensafta, Gervasio Semedo

Abstract:

The purpose of this paper is to investigate the relationship between oil prices and socks markets. The empirical analysis in this paper is conducted within the context of Multivariate GARCH models, using a transform version of the so-called BEKK parameterization. We show that mean and uncertainty of US market are transmitted to oil market and European market. We also identify an important transmission from WTI prices to Brent Prices.

Keywords: oil volatility, stock markets, MGARCH, transmission, structural break

Procedia PDF Downloads 509
5835 RASPE: Risk Advisory Smart System for Pipeline Projects in Egypt

Authors: Nael Y. Zabel, Maged E. Georgy, Moheeb E. Ibrahim

Abstract:

A knowledge-based expert system with the acronym RASPE is developed as an application tool to help decision makers in construction companies make informed decisions about managing risks in pipeline construction projects. Choosing to use expert systems from all available artificial intelligence techniques is due to the fact that an expert system is more suited to representing a domain’s knowledge and the reasoning behind domain-specific decisions. The knowledge-based expert system can capture the knowledge in the form of conditional rules which represent various project scenarios and potential risk mitigation/response actions. The built knowledge in RASPE is utilized through the underlying inference engine that allows the firing of rules relevant to a project scenario into consideration. This paper provides an overview of the knowledge acquisition process and goes about describing the knowledge structure which is divided up into four major modules. The paper shows one module in full detail for illustration purposes and concludes with insightful remarks.

Keywords: expert system, knowledge management, pipeline projects, risk mismanagement

Procedia PDF Downloads 298
5834 Optimization, Yield and Chemical Composition of Essential Oil from Cymbopogon citratus: Comparative Study with Microwave Assisted Extraction and Hydrodistillation

Authors: Irsha Dhotre

Abstract:

Cymbopogon citratus is generally known as Indian Lemongrass and is widely applicable in the cosmetic, pharmaceutical, dairy puddings, and food industries. To enhance the quality of extraction, microwave-oven-aided hydro distillation processes were implemented. The basic parameter which influences the rate of extraction is considered, such as the temperature of extraction, the time required for extraction, and microwave-oven power applied. Locally available CKP 25 Cymbopogon citratus was used for the extraction of essential oil. Optimization of Extractions Parameters and full factorial Box–Behnken design (BBD) evaluated by using Design expert 13 software. The regression model revealed that the optimum parameters required for extractions are a temperature of 35℃, a time of extraction of 130 minutes, and microwave-oven power of 700 W. The extraction efficiency of yield is 4.76%. Gas Chromatography-Mass Spectroscopy (GC-MS) analysis confirmed the significant components present in the extraction of lemongrass oil.

Keywords: Box–Behnken design, Cymbopogon citratus, hydro distillation, microwave-oven, response surface methodology

Procedia PDF Downloads 71
5833 Expanded Access through Open and Distance Learning in Nigeria

Authors: Okoro Ngozi Priscilla

Abstract:

Education is the bedrock of development in every nation of the world, and it is very useful in ensuring quality of life for every individual and a better world for the people. Education, therefore, is the basic instrument of economic growth and technological advancement in any society. It is in recognition of this fact that the Nigerian government commits immense resources to ensuring that its citizens acquire education and also policies are being made to ensure the accessibility of education, qualitative higher education is highly recognized as a vital driving force for the socio-economic growth and technological development of nations yet the problem of access to University education in the country persists and therefore brought about the introduction of Open and Distance Learning (ODL) which has as its main objective, the attainment of mass literacy and providing opportunities for those who could not gain admission through designated entrance examination agencies as well as those who could not afford to leave their job to attend a full-time educational programme. Open and distance learning seeks to improve skilled manpower and also improve the skills for those already at work.

Keywords: accessibility, open and distant learning programme, fulltime educational programme, distance learning

Procedia PDF Downloads 445
5832 Developing a Comprehensive Model for the Prevention of Tension Neck Syndrome: A Focus on Musculoskeletal Disorder Prevention Strategies

Authors: Behnaz Sohani, Ifeoluwa Joshua Adigun, Amir Rahmani, Khaled Goher

Abstract:

This paper provides initial results on the efficacy of the designed ergonomic-oriented neck support to mitigate and alleviate tension neck syndrome musculoskeletal disorder. This is done using both simulations and measurements. Tension Neck Syndrome Musculoskeletal Disorder (TNS MSD) causes discomfort in the muscles around the neck and shoulder. TNS MSD is one of the leading causes of early retirement. This research focuses on the design of an adaptive neck supporter by integrating a soft actuator massager to help deliver a soothing massage. The massager and adaptive neck supporter prototype were validated by finite element analysis prior to fabrication to envisage the feasibility of the design concept. Then a prototype for the massager was fabricated and tested for concept validation. Future work will be focused on fabricating the full-scale prototype and upgrading and optimizing the design concept for the adaptive neck supporter.

Keywords: adaptive neck supporter, tension neck syndrome, musculoskeletal disorder, soft actuator massager, soft robotics

Procedia PDF Downloads 88
5831 High Purity Germanium Detector Characterization by Means of Monte Carlo Simulation through Application of Geant4 Toolkit

Authors: Milos Travar, Jovana Nikolov, Andrej Vranicar, Natasa Todorovic

Abstract:

Over the years, High Purity Germanium (HPGe) detectors proved to be an excellent practical tool and, as such, have established their today's wide use in low background γ-spectrometry. One of the advantages of gamma-ray spectrometry is its easy sample preparation as chemical processing and separation of the studied subject are not required. Thus, with a single measurement, one can simultaneously perform both qualitative and quantitative analysis. One of the most prominent features of HPGe detectors, besides their excellent efficiency, is their superior resolution. This feature virtually allows a researcher to perform a thorough analysis by discriminating photons of similar energies in the studied spectra where otherwise they would superimpose within a single-energy peak and, as such, could potentially scathe analysis and produce wrongly assessed results. Naturally, this feature is of great importance when the identification of radionuclides, as well as their activity concentrations, is being practiced where high precision comes as a necessity. In measurements of this nature, in order to be able to reproduce good and trustworthy results, one has to have initially performed an adequate full-energy peak (FEP) efficiency calibration of the used equipment. However, experimental determination of the response, i.e., efficiency curves for a given detector-sample configuration and its geometry, is not always easy and requires a certain set of reference calibration sources in order to account for and cover broader energy ranges of interest. With the goal of overcoming these difficulties, a lot of researches turned towards the application of different software toolkits that implement the Monte Carlo method (e.g., MCNP, FLUKA, PENELOPE, Geant4, etc.), as it has proven time and time again to be a very powerful tool. In the process of creating a reliable model, one has to have well-established and described specifications of the detector. Unfortunately, the documentation that manufacturers provide alongside the equipment is rarely sufficient enough for this purpose. Furthermore, certain parameters tend to evolve and change over time, especially with older equipment. Deterioration of these parameters consequently decreases the active volume of the crystal and can thus affect the efficiencies by a large margin if they are not properly taken into account. In this study, the optimisation method of two HPGe detectors through the implementation of the Geant4 toolkit developed by CERN is described, with the goal of further improving simulation accuracy in calculations of FEP efficiencies by investigating the influence of certain detector variables (e.g., crystal-to-window distance, dead layer thicknesses, inner crystal’s void dimensions, etc.). Detectors on which the optimisation procedures were carried out were a standard traditional co-axial extended range detector (XtRa HPGe, CANBERRA) and a broad energy range planar detector (BEGe, CANBERRA). Optimised models were verified through comparison with experimentally obtained data from measurements of a set of point-like radioactive sources. Acquired results of both detectors displayed good agreement with experimental data that falls under an average statistical uncertainty of ∼ 4.6% for XtRa and ∼ 1.8% for BEGe detector within the energy range of 59.4−1836.1 [keV] and 59.4−1212.9 [keV], respectively.

Keywords: HPGe detector, γ spectrometry, efficiency, Geant4 simulation, Monte Carlo method

Procedia PDF Downloads 103
5830 Linguistic Misinterpretation and the Dialogue of Civilizations

Authors: Oleg Redkin, Olga Bernikova

Abstract:

Globalization and migrations have made cross-cultural contacts more frequent and intensive. Sometimes, these contacts may lead to misunderstanding between partners of communication and misinterpretations of the verbal messages that some researchers tend to consider as the 'clash of civilizations'. In most cases, reasons for that may be found in cultural and linguistic differences and hence misinterpretations of intentions and behavior. The current research examines factors of verbal and non-verbal communication that should be taken into consideration in verbal and non-verbal contacts. Language is one of the most important manifestations of the cultural code, and it is often considered as one of the special features of a civilization. The Arabic language, in particular, is commonly associated with Islam and the language and the Arab-Muslim civilization. It is one of the most important markers of self-identification for more than 200 million of native speakers. Arabic is the language of the Quran and hence the symbol of religious affiliation for more than one billion Muslims around the globe. Adequate interpretation of Arabic texts requires profound knowledge of its grammar, semantics of its vocabulary. Communicating sides who belong to different cultural groups are guided by different models of behavior and hierarchy of values, besides that the vocabulary each of them uses in the dialogue may convey different semantic realities and vary in connotations. In this context direct, literal translation in most cases cannot adequately convey the original meaning of the original message. Besides that peculiarities and diversities of the extralinguistic information, such as the body language, communicative etiquette, cultural background and religious affiliations may make the dialogue even more difficult. It is very likely that the so called 'clash of civilizations' in most cases is due to misinterpretation of counterpart's means of discourse such as language, cultural codes, and models of behavior rather than lies in basic contradictions between partners of communication. In the process of communication, one has to rely on universal values rather than focus on cultural or religious peculiarities, to take into account current linguistic and extralinguistic context.

Keywords: Arabic, civilization, discourse, language, linguistic

Procedia PDF Downloads 209
5829 Building Information Modeling Implementation for Managing an Extra Large Governmental Building Renovation Project

Authors: Pornpote Nusen, Manop Kaewmoracharoen

Abstract:

In recent years, there was an observable shift in fully developed countries from constructing new buildings to modifying existing buildings. The issue was that although an effective instrument like BIM (Building Information Modeling) was well developed for constructing new buildings, it was not widely used to renovate old buildings. BIM was accepted as an effective means to overcome common managerial problems such as project delay, cost overrun, and poor quality of the project life cycle. It was recently introduced in Thailand and rarely used in a renovation project. Today, in Thailand, BIM is mostly used for creating aesthetic 3D models and quantity takeoff purposes, though it can be an effective tool to use as a project management tool in planning and scheduling. Now the governmental sector in Thailand begins to recognize the uses of using BIM to manage a construction project, but the knowledge about the BIM implementation to governmental construction projects is underdeveloped. Further studies need to be conducted to maximize its advantages for the governmental sector. An educational extra large governmental building of 17,000 square-meters was used in this research. It is currently under construction for a two-year renovation project. BIM models of the building for the exterior and interior areas were created for the whole five floors. Then 4D BIM with combination of 3D BIM plus time was created for planning and scheduling. Three focus groups had been done with executive committee, contractors, and officers of the building to discuss the possibility of usage and usefulness of BIM approach over the traditional process. Several aspects were discussed in the positive sides, especially several foreseen problems, such as the inadequate accessibility of ways, the altered ceiling levels, the impractical construction plan created through a traditional approach, and the lack of constructability information. However, for some parties, the cost of BIM implementation was a concern, though, this study believes, its uses outweigh the cost.

Keywords: building information modeling, extra large building, governmental building renovation, project management, renovation, 4D BIM

Procedia PDF Downloads 136
5828 In Silico Exploration of Quinazoline Derivatives as EGFR Inhibitors for Lung Cancer: A Multi-Modal Approach Integrating QSAR-3D, ADMET, Molecular Docking, and Molecular Dynamics Analyses

Authors: Mohamed Moussaoui

Abstract:

A series of thirty-one potential inhibitors targeting the epidermal growth factor receptor kinase (EGFR), derived from quinazoline, underwent 3D-QSAR analysis using CoMFA and CoMSIA methodologies. The training and test sets of quinazoline derivatives were utilized to construct and validate the QSAR models, respectively, with dataset alignment performed using the lowest energy conformer of the most active compound. The best-performing CoMFA and CoMSIA models demonstrated impressive determination coefficients, with R² values of 0.981 and 0.978, respectively, and Leave One Out cross-validation determination coefficients, Q², of 0.645 and 0.729, respectively. Furthermore, external validation using a test set of five compounds yielded predicted determination coefficients, R² test, of 0.929 and 0.909 for CoMFA and CoMSIA, respectively. Building upon these promising results, eighteen new compounds were designed and assessed for drug likeness and ADMET properties through in silico methods. Additionally, molecular docking studies were conducted to elucidate the binding interactions between the selected compounds and the enzyme. Detailed molecular dynamics simulations were performed to analyze the stability, conformational changes, and binding interactions of the quinazoline derivatives with the EGFR kinase. These simulations provided deeper insights into the dynamic behavior of the compounds within the active site. This comprehensive analysis enhances the understanding of quinazoline derivatives as potential anti-cancer agents and provides valuable insights for lead optimization in the early stages of drug discovery, particularly for developing highly potent anticancer therapeutics

Keywords: 3D-QSAR, CoMFA, CoMSIA, ADMET, molecular docking, quinazoline, molecular dynamic, egfr inhibitors, lung cancer, anticancer

Procedia PDF Downloads 27
5827 Hierarchical Tree Long Short-Term Memory for Sentence Representations

Authors: Xiuying Wang, Changliang Li, Bo Xu

Abstract:

A fixed-length feature vector is required for many machine learning algorithms in NLP field. Word embeddings have been very successful at learning lexical information. However, they cannot capture the compositional meaning of sentences, which prevents them from a deeper understanding of language. In this paper, we introduce a novel hierarchical tree long short-term memory (HTLSTM) model that learns vector representations for sentences of arbitrary syntactic type and length. We propose to split one sentence into three hierarchies: short phrase, long phrase and full sentence level. The HTLSTM model gives our algorithm the potential to fully consider the hierarchical information and long-term dependencies of language. We design the experiments on both English and Chinese corpus to evaluate our model on sentiment analysis task. And the results show that our model outperforms several existing state of the art approaches significantly.

Keywords: deep learning, hierarchical tree long short-term memory, sentence representation, sentiment analysis

Procedia PDF Downloads 341
5826 Underground Coal Gasification Technology in Türkiye: A Techno-Economic Assessment

Authors: Fatma Ünal, Hasancan Okutan

Abstract:

Increasing worldwide population and technological requirements lead to an increase in energy demand every year. The demand has been mainly supplied from fossil fuels such as coal and petroleum due to insufficient natural gas resources. In recent years, the amount of coal reserves has reached almost 21 billion tons in Türkiye. These are mostly lignite (%92,7), that contains high levels of moisture and sulfur components. Underground coal gasification technology is one of the most suitable methods in comparison with direct combustion techniques for the evaluation of such coal types. In this study, the applicability of the underground coal gasification process is investigated in the Eskişehir-Alpu lignite reserve as a pilot region, both technologically and economically. It is assumed that the electricity is produced from the obtained synthesis gas in an integrated gasification combined cycle (IGCC). Firstly, an equilibrium model has been developed by using the thermodynamic properties of the gasification reactions. The effect of the type of oxidizing gas, the sulfur content of coal, the rate of water vapor/air, and the pressure of the system have been investigated to find optimum process conditions. Secondly, the parallel and linear controlled recreation and injection point (CRIP) models were implemented as drilling methods, and costs were calculated under the different oxidizing agents (air and high-purity O2). In Parallel CRIP (P-CRIP), drilling cost is found to be lower than the linear CRIP (L-CRIP) since two coal beds simultaneously are gasified. It is seen that CO2 Capture and Storage (CCS) technology was the most effective unit on the total cost in both models. The cost of the synthesis gas produced varies between 0,02 $/Mcal and 0,09 $/Mcal. This is the promising result when considering the selling price of Türkiye natural gas for Q1-2023 (0.103 $ /Mcal).

Keywords: energy, lignite reserve, techno-economic analysis, underground coal gasification.

Procedia PDF Downloads 51
5825 Gas Flow, Time, Distance Dynamic Modelling

Authors: A. Abdul-Ameer

Abstract:

The equations governing the distance, pressure- volume flow relationships for the pipeline transportation of gaseous mixtures, are considered. A derivation based on differential calculus, for an element of this system model, is addressed. Solutions, yielding the input- output response following pressure changes, are reviewed. The technical problems associated with these analytical results are identified. Procedures resolving these difficulties providing thereby an attractive, simple, analysis route are outlined. Computed responses, validating thereby calculated predictions, are presented.

Keywords: pressure, distance, flow, dissipation, models

Procedia PDF Downloads 464
5824 Development of a Fuzzy Logic Based Model for Monitoring Child Pornography

Authors: Mariam Ismail, Kazeem Rufai, Jeremiah Balogun

Abstract:

A study was conducted to apply fuzzy logic to the development of a monitoring model for child pornography based on associated risk factors, which can be used by forensic experts or integrated into forensic systems for the early detection of child pornographic activities. A number of methods were adopted in the study, which includes an extensive review of related works was done in order to identify the factors that are associated with child pornography following which they were validated by an expert sex psychologist and guidance counselor, and relevant data was collected. Fuzzy membership functions were used to fuzzify the associated variables identified alongside the risk of the occurrence of child pornography based on the inference rules that were provided by the experts consulted, and the fuzzy logic expert system was simulated using the Fuzzy Logic Toolbox available in the MATLAB Software Release 2016. The results of the study showed that there were 4 categories of risk factors required for assessing the risk of a suspect committing child pornography offenses. The results of the study showed that 2 and 3 triangular membership functions were used to formulate the risk factors based on the 2 and 3 number of labels assigned, respectively. The results of the study showed that 5 fuzzy logic models were formulated such that the first 4 was used to assess the impact of each category on child pornography while the last one takes the 4 outputs from the 4 fuzzy logic models as inputs required for assessing the risk of child pornography. The following conclusion was made; there were factors that were related to personal traits, social traits, history of child pornography crimes, and self-regulatory deficiency traits by the suspects required for the assessment of the risk of child pornography crimes committed by a suspect. Using the values of the identified risk factors selected for this study, the risk of child pornography can be easily assessed from their values in order to determine the likelihood of a suspect perpetuating the crime.

Keywords: fuzzy, membership functions, pornography, risk factors

Procedia PDF Downloads 113
5823 Efficacy of Deep Learning for Below-Canopy Reconstruction of Satellite and Aerial Sensing Point Clouds through Fractal Tree Symmetry

Authors: Dhanuj M. Gandikota

Abstract:

Sensor-derived three-dimensional (3D) point clouds of trees are invaluable in remote sensing analysis for the accurate measurement of key structural metrics, bio-inventory values, spatial planning/visualization, and ecological modeling. Machine learning (ML) holds the potential in addressing the restrictive tradeoffs in cost, spatial coverage, resolution, and information gain that exist in current point cloud sensing methods. Terrestrial laser scanning (TLS) remains the highest fidelity source of both canopy and below-canopy structural features, but usage is limited in both coverage and cost, requiring manual deployment to map out large, forested areas. While aerial laser scanning (ALS) remains a reliable avenue of LIDAR active remote sensing, ALS is also cost-restrictive in deployment methods. Space-borne photogrammetry from high-resolution satellite constellations is an avenue of passive remote sensing with promising viability in research for the accurate construction of vegetation 3-D point clouds. It provides both the lowest comparative cost and the largest spatial coverage across remote sensing methods. However, both space-borne photogrammetry and ALS demonstrate technical limitations in the capture of valuable below-canopy point cloud data. Looking to minimize these tradeoffs, we explored a class of powerful ML algorithms called Deep Learning (DL) that show promise in recent research on 3-D point cloud reconstruction and interpolation. Our research details the efficacy of applying these DL techniques to reconstruct accurate below-canopy point clouds from space-borne and aerial remote sensing through learned patterns of tree species fractal symmetry properties and the supplementation of locally sourced bio-inventory metrics. From our dataset, consisting of tree point clouds obtained from TLS, we deconstructed the point clouds of each tree into those that would be obtained through ALS and satellite photogrammetry of varying resolutions. We fed this ALS/satellite point cloud dataset, along with the simulated local bio-inventory metrics, into the DL point cloud reconstruction architectures to generate the full 3-D tree point clouds (the truth values are denoted by the full TLS tree point clouds containing the below-canopy information). Point cloud reconstruction accuracy was validated both through the measurement of error from the original TLS point clouds as well as the error of extraction of key structural metrics, such as crown base height, diameter above root crown, and leaf/wood volume. The results of this research additionally demonstrate the supplemental performance gain of using minimum locally sourced bio-inventory metric information as an input in ML systems to reach specified accuracy thresholds of tree point cloud reconstruction. This research provides insight into methods for the rapid, cost-effective, and accurate construction of below-canopy tree 3-D point clouds, as well as the supported potential of ML and DL to learn complex, unmodeled patterns of fractal tree growth symmetry.

Keywords: deep learning, machine learning, satellite, photogrammetry, aerial laser scanning, terrestrial laser scanning, point cloud, fractal symmetry

Procedia PDF Downloads 85
5822 Behavior of Common Philippine-Made Concrete Hollow Block Structures Subjected to Seismic Load Using Rigid Body Spring-Discrete Element Method

Authors: Arwin Malabanan, Carl Chester Ragudo, Jerome Tadiosa, John Dee Mangoba, Eric Augustus Tingatinga, Romeo Eliezer Longalong

Abstract:

Concrete hollow blocks (CHB) are the most commonly used masonry block for walls in residential houses, school buildings and public buildings in the Philippines. During the recent 2013 Bohol earthquake (Mw 7.2), it has been proven that CHB walls are very vulnerable to severe external action like strong ground motion. In this paper, a numerical model of CHB structures is proposed, and seismic behavior of CHB houses is presented. In modeling, the Rigid Body Spring-Discrete Element method (RBS-DEM)) is used wherein masonry blocks are discretized into rigid elements and connected by nonlinear springs at preselected contact points. The shear and normal stiffness of springs are derived from the material properties of CHB unit incorporating the grout and mortar fillings through the volumetric transformation of the dimension using material ratio. Numerical models of reinforced and unreinforced walls are first subjected to linearly-increasing in plane loading to observe the different failure mechanisms. These wall models are then assembled to form typical model masonry houses and then subjected to the El Centro and Pacoima earthquake records. Numerical simulations show that the elastic, failure and collapse behavior of the model houses agree well with shaking table tests results. The effectiveness of the method in replicating failure patterns will serve as a basis for the improvement of the design and provides a good basis of strengthening the structure.

Keywords: concrete hollow blocks, discrete element method, earthquake, rigid body spring model

Procedia PDF Downloads 346
5821 Strength Evaluation by Finite Element Analysis of Mesoscale Concrete Models Developed from CT Scan Images of Concrete Cube

Authors: Nirjhar Dhang, S. Vinay Kumar

Abstract:

Concrete is a non-homogeneous mix of coarse aggregates, sand, cement, air-voids and interfacial transition zone (ITZ) around aggregates. Adoption of these complex structures and material properties in numerical simulation would lead us to better understanding and design of concrete. In this work, the mesoscale model of concrete has been prepared from X-ray computerized tomography (CT) image. These images are converted into computer model and numerically simulated using commercially available finite element software. The mesoscale models are simulated under the influence of compressive displacement. The effect of shape and distribution of aggregates, continuous and discrete ITZ thickness, voids, and variation of mortar strength has been investigated. The CT scan of concrete cube consists of series of two dimensional slices. Total 49 slices are obtained from a cube of 150mm and the interval of slices comes approximately 3mm. In CT scan images, the same cube can be CT scanned in a non-destructive manner and later the compression test can be carried out in a universal testing machine (UTM) for finding its strength. The image processing and extraction of mortar and aggregates from CT scan slices are performed by programming in Python. The digital colour image consists of red, green and blue (RGB) pixels. The conversion of RGB image to black and white image (BW) is carried out, and identification of mesoscale constituents is made by putting value between 0-255. The pixel matrix is created for modeling of mortar, aggregates, and ITZ. Pixels are normalized to 0-9 scale considering the relative strength. Here, zero is assigned to voids, 4-6 for mortar and 7-9 for aggregates. The value between 1-3 identifies boundary between aggregates and mortar. In the next step, triangular and quadrilateral elements for plane stress and plane strain models are generated depending on option given. Properties of materials, boundary conditions, and analysis scheme are specified in this module. The responses like displacement, stresses, and damages are evaluated by ABAQUS importing the input file. This simulation evaluates compressive strengths of 49 slices of the cube. The model is meshed with more than sixty thousand elements. The effect of shape and distribution of aggregates, inclusion of voids and variation of thickness of ITZ layer with relation to load carrying capacity, stress-strain response and strain localizations of concrete have been studied. The plane strain condition carried more load than plane stress condition due to confinement. The CT scan technique can be used to get slices from concrete cores taken from the actual structure, and the digital image processing can be used for finding the shape and contents of aggregates in concrete. This may be further compared with test results of concrete cores and can be used as an important tool for strength evaluation of concrete.

Keywords: concrete, image processing, plane strain, interfacial transition zone

Procedia PDF Downloads 231
5820 Improving the Training for Civil Engineers by Introducing Virtual Reality Technique

Authors: Manar Al-Ateeq

Abstract:

The building construction industry plays a major role in the economy of the word and the state of Kuwait. This paper evaluates existing new civil site engineers, describes a new system for improvement and insures the importance of prequalifying and developing for new engineers. In order to have a strong base in engineering, educational institutes and workplaces should be responsible to continuously train engineers and update them with new methods and techniques in engineering. As to achieve that, school of engineering should constantly update computational resources to be used in the professions. A survey was prepared for graduated Engineers based on stated objectives to understand the status of graduate engineers in both the public and private sector. Interviews were made with different sectors in Kuwait, and several visits were made to different training centers within different workplaces in Kuwait to evaluate training process and try to improve it. Virtual Reality (VR) technology could be applied as a complement to three-dimensional (3D) modeling, leading to better communication whether in job training, in education or in professional practice. Techniques of 3D modeling and VR can be applied to develop the models related to the construction process. The 3D models can support rehabilitation design as it can be considered as a great tool for monitoring failure and defaults in structures; also it can support decisions based on the visual analyses of alternative solutions. Therefore, teaching computer-aided design (CAD) and VR techniques in school will help engineering students in order to prepare them to site work and also will assist them to consider these technologies as important supports in their later professional practice. This teaching technique will show how the construction works developed, allow the visual simulation of progression of each type of work and help them to know more about the necessary equipment needed for tasks and how it works on site.

Keywords: three dimensional modeling (3DM), civil engineers (CE), professional practice (PP), virtual reality (VR)

Procedia PDF Downloads 159
5819 Comparison between Two Software Packages GSTARS4 and HEC-6 about Prediction of the Sedimentation Amount in Dam Reservoirs and to Estimate Its Efficient Life Time in the South of Iran

Authors: Fatemeh Faramarzi, Hosein Mahjoob

Abstract:

Building dams on rivers for utilization of water resources causes problems in hydrodynamic equilibrium and results in leaving all or part of the sediments carried by water in dam reservoir. This phenomenon has also significant impacts on water and sediment flow regime and in the long term can cause morphological changes in the environment surrounding the river, reducing the useful life of the reservoir which threatens sustainable development through inefficient management of water resources. In the past, empirical methods were used to predict the sedimentation amount in dam reservoirs and to estimate its efficient lifetime. But recently the mathematical and computational models are widely used in sedimentation studies in dam reservoirs as a suitable tool. These models usually solve the equations using finite element method. This study compares the results from tow software packages, GSTARS4 & HEC-6, in the prediction of the sedimentation amount in Dez dam, southern Iran. The model provides a one-dimensional, steady-state simulation of sediment deposition and erosion by solving the equations of momentum, flow and sediment continuity and sediment transport. GSTARS4 (Generalized Sediment Transport Model for Alluvial River Simulation) which is based on a one-dimensional mathematical model that simulates bed changes in both longitudinal and transverse directions by using flow tubes in a quasi-two-dimensional scheme to calibrate a period of 47 years and forecast the next 47 years of sedimentation in Dez Dam, Southern Iran. This dam is among the highest dams all over the world (with its 203 m height), and irrigates more than 125000 square hectares of downstream lands and plays a major role in flood control in the region. The input data including geometry, hydraulic and sedimentary data, starts from 1955 to 2003 on a daily basis. To predict future river discharge, in this research, the time series data were assumed to be repeated after 47 years. Finally, the obtained result was very satisfactory in the delta region so that the output from GSTARS4 was almost identical to the hydrographic profile in 2003. In the Dez dam due to the long (65 km) and a large tank, the vertical currents are dominant causing the calculations by the above-mentioned method to be inaccurate. To solve this problem, we used the empirical reduction method to calculate the sedimentation in the downstream area which led to very good answers. Thus, we demonstrated that by combining these two methods a very suitable model for sedimentation in Dez dam for the study period can be obtained. The present study demonstrated successfully that the outputs of both methods are the same.

Keywords: Dez Dam, prediction, sedimentation, water resources, computational models, finite element method, GSTARS4, HEC-6

Procedia PDF Downloads 302
5818 Theoretical Comparisons and Empirical Illustration of Malmquist, Hicks–Moorsteen, and Luenberger Productivity Indices

Authors: Fatemeh Abbasi, Sahand Daneshvar

Abstract:

Productivity is one of the essential goals of companies to improve performance, which as a strategy-oriented method, determines the basis of the company's economic growth. The history of productivity goes back centuries, but most researchers defined productivity as the relationship between a product and the factors used in production in the early twentieth century. Productivity as the optimal use of available resources means that "more output using less input" can increase companies' economic growth and prosperity capacity. Also, having a quality life based on economic progress depends on productivity growth in that society. Therefore, productivity is a national priority for any developed country. There are several methods for calculating productivity growth measurements that can be divided into parametric and non-parametric methods. Parametric methods rely on the existence of a function in their hypotheses, while non-parametric methods do not require a function based on empirical evidence. One of the most popular non-parametric methods is Data Envelopment Analysis (DEA), which measures changes in productivity over time. The DEA evaluates the productivity of decision-making units (DMUs) based on mathematical models. This method uses multiple inputs and outputs to compare the productivity of similar DMUs such as banks, government agencies, companies, airports, Etc. Non-parametric methods are themselves divided into the frontier and non frontier approaches. The Malmquist productivity index (MPI) proposed by Caves, Christensen, and Diewert (1982), the Hicks–Moorsteen productivity index (HMPI) proposed by Bjurek (1996), or the Luenberger productivity indicator (LPI) proposed by Chambers (2002) are powerful tools for measuring productivity changes over time. This study will compare the Malmquist, Hicks–Moorsteen, and Luenberger indices theoretically and empirically based on DEA models and review their strengths and weaknesses.

Keywords: data envelopment analysis, Hicks–Moorsteen productivity index, Leuenberger productivity indicator, malmquist productivity index

Procedia PDF Downloads 180
5817 Investigating the Effect of Height on Essential Oils of Urtica diocia L.: Case Study of Ramsar, Mazandaran, Iran

Authors: Keivan Saeb, Azade Kakouei, Razieh Jafari Hajati, Khalil Pourshamsian, Babak Babakhani

Abstract:

Urtica Diocia L. from the Urticaceae family is a plant of herbal value and of a noticeable distribution in the north of Iran. The growth of different plants in various natural environments and ecosystems seems to be affected by factors such as the height (from sea surface).To investigate the effect of height on Urtica Diocia L. medicine compounds in its natural environment, three areas with the height of zero, 800 and 1800m were selected.The samples were randomly gathered three times and were dried; also, their compounds was extracted using the Clivenger with the water-distilling method. To determine the medicine compounds, the GC-MS as well as the GC machines were used. The analysis of variance was done in the form of the random-full-block design. The results indicated that there was a significant difference between the percent of EOs in the selected heights; however, such difference was not significant within each height. From among the eight flavors of the study, the phytol compound was more in terms of percentage. By increasing the height the percent of EOs would decrease. lower heights could be considered most appropriate for producing the studied effective materials despite of the moistened climate and soil there.

Keywords: Urtica diocia L., height, EOs, medicine

Procedia PDF Downloads 449
5816 The Type II Immune Response in Acute and Chronic Pancreatitis Mediated by STAT6 in Murine

Authors: Hager Elsheikh

Abstract:

Context: Pancreatitis is a condition characterized by inflammation in the pancreas, which can lead to serious complications if untreated. Both acute and chronic pancreatitis are associated with immune reactions and fibrosis, which further damage the pancreas. The type 2 immune response, primarily driven by alternative activated macrophages (AAMs), plays a significant role in the development of fibrosis. The IL-4/STAT6 pathway is a crucial signaling pathway for the activation of M2 macrophages. Pancreatic fibrosis is induced by dysregulated inflammatory responses and can result in the autodigestion and necrosis of pancreatic acinar cells. Research Aim: The aim of this study is to investigate the impact of STAT6, a crucial molecule in the IL-4/STAT6 pathway, on the severity and development of fibrosis during acute and chronic pancreatitis. The research also aims to understand the influence of the JAK/STAT6 signaling pathway on the balance between fibrosis and regeneration in the presence of different macrophage populations. Methodology: The research utilizes murine models of acute and chronic pancreatitis induced by cerulean injection. Animal models will be employed to study the effect of STAT6 knockout on disease severity and fibrosis. Isolation of acinar cells and cell culture techniques will be used to assess the impact of different macrophage populations on wound healing and regeneration. Various techniques such as PCR, histology, immunofluorescence, and transcriptomics will be employed to analyze the tissues and cells. Findings: The research aims to provide insights into the mechanisms underlying tissue fibrosis and wound healing during acute and chronic pancreatitis. By investigating the influence of the JAK/STAT6 signaling pathway and different macrophage populations, the study aims to understand their impact on tissue fibrosis, disease severity, and pancreatic regeneration. Theoretical Importance: This research contributes to our understanding of the role of specific signaling pathways, macrophage polarization, and the type 2 immune response in pancreatitis. It provides insights into the molecular mechanisms underlying tissue fibrosis and the potential for targeted therapies. Data Collection and Analysis Procedures: Data will be collected through the use of murine models, isolation and culture of acinar cells, and various experimental techniques such as PCR, histology, immunofluorescence, and transcriptomics. Data will be analyzed using appropriate statistical methods and techniques, and the findings will be interpreted in the context of the research objectives. Conclusion: By investigating the mechanisms of tissue fibrosis and wound healing during acute and chronic pancreatitis, this research aims to enhance our understanding of the disease progression and potential therapeutic targets. The findings have theoretical importance in expanding our knowledge of pancreatic fibrosis and the role of macrophage polarization in the context of the type 2 immune response.

Keywords: immunity in chronic diseases, pancreatitis, macrophages, immune response

Procedia PDF Downloads 23
5815 On Elastic Anisotropy of Fused Filament Fabricated Acrylonitrile Butadiene Styrene Structures

Authors: Joseph Marae Djouda, Ashraf Kasmi, François Hild

Abstract:

Fused filament fabrication is one of the most widespread additive manufacturing techniques because of its low-cost implementation. Its initial development was based on part fabrication with thermoplastic materials. The influence of the manufacturing parameters such as the filament orientation through the nozzle, the deposited layer thickness, or the speed deposition on the mechanical properties of the parts has been widely experimentally investigated. It has been recorded the remarkable variations of the anisotropy in the function of the filament path during the fabrication process. However, there is a lack in the development of constitutive models describing the mechanical properties. In this study, integrated digital image correlation (I-DIC) is used for the identification of mechanical constitutive parameters of two configurations of ABS samples: +/-45° and so-called “oriented deposition.” In this last, the filament was deposited in order to follow the principal strain of the sample. The identification scheme based on the gap reduction between simulation and the experiment directly from images recorded from a single sample (single edge notched tension specimen) is developed. The macroscopic and mesoscopic analysis are conducted from images recorded in both sample surfaces during the tensile test. The elastic and elastoplastic models in isotropic and orthotropic frameworks have been established. It appears that independently of the sample configurations (filament orientation during the fabrication), the elastoplastic isotropic model gives the correct description of the behavior of samples. It is worth noting that in this model, the number of constitutive parameters is limited to the one considered in the elastoplastic orthotropic model. This leads to the fact that the anisotropy of the architectured 3D printed ABS parts can be neglected in the establishment of the macroscopic behavior description.

Keywords: elastic anisotropy, fused filament fabrication, Acrylonitrile butadiene styrene, I-DIC identification

Procedia PDF Downloads 114
5814 Application of Argumentation for Improving the Classification Accuracy in Inductive Concept Formation

Authors: Vadim Vagin, Marina Fomina, Oleg Morosin

Abstract:

This paper contains the description of argumentation approach for the problem of inductive concept formation. It is proposed to use argumentation, based on defeasible reasoning with justification degrees, to improve the quality of classification models, obtained by generalization algorithms. The experiment’s results on both clear and noisy data are also presented.

Keywords: argumentation, justification degrees, inductive concept formation, noise, generalization

Procedia PDF Downloads 423