Search results for: integration model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18760

Search results for: integration model

16390 Supervised-Component-Based Generalised Linear Regression with Multiple Explanatory Blocks: THEME-SCGLR

Authors: Bry X., Trottier C., Mortier F., Cornu G., Verron T.

Abstract:

We address component-based regularization of a Multivariate Generalized Linear Model (MGLM). A set of random responses Y is assumed to depend, through a GLM, on a set X of explanatory variables, as well as on a set T of additional covariates. X is partitioned into R conceptually homogeneous blocks X1, ... , XR , viewed as explanatory themes. Variables in each Xr are assumed many and redundant. Thus, Generalised Linear Regression (GLR) demands regularization with respect to each Xr. By contrast, variables in T are assumed selected so as to demand no regularization. Regularization is performed searching each Xr for an appropriate number of orthogonal components that both contribute to model Y and capture relevant structural information in Xr. We propose a very general criterion to measure structural relevance (SR) of a component in a block, and show how to take SR into account within a Fisher-scoring-type algorithm in order to estimate the model. We show how to deal with mixed-type explanatory variables. The method, named THEME-SCGLR, is tested on simulated data.

Keywords: Component-Model, Fisher Scoring Algorithm, GLM, PLS Regression, SCGLR, SEER, THEME

Procedia PDF Downloads 392
16389 A Model Towards Creating Positive Accounting Classroom Conditions That Supports Successful Learning at School

Authors: Vine Petzer, Mirna Nel

Abstract:

An explanatory mixed method design was used to investigate accounting classroom conditions in the Further Education and Training (FET) Phase in South Africa. A descriptive survey research study with a heterogeneous group of learners and teachers was conducted in the first phase. In the qualitative phase, semi-structured individual interviews with learners and teachers, as well as observations in the accounting classroom, were employed to gain more in depth understanding of the learning conditions in the accounting classroom. The findings of the empirical research informed the development of a model for teachers in accounting, supporting them to use more effective teaching methods and create positive learning conditions for all learners to experience successful learning. A model towards creating positive Accounting classroom conditions that support successful learning was developed and recommended for education policy and decision-makers for use as a classroom intervention capacity building tool. The model identifies and delineates classroom practices that exert significant effect on learner attainment of quality education.

Keywords: accounting classroom conditions, positive education, successful learning, teaching accounting

Procedia PDF Downloads 145
16388 Design and Implementation of a Hardened Cryptographic Coprocessor with 128-bit RISC-V Core

Authors: Yashas Bedre Raghavendra, Pim Vullers

Abstract:

This study presents the design and implementation of an abstract cryptographic coprocessor, leveraging AMBA(Advanced Microcontroller Bus Architecture) protocols - APB (Advanced Peripheral Bus) and AHB (Advanced High-performance Bus), to enable seamless integration with the main CPU(Central processing unit) and enhance the coprocessor’s algorithm flexibility. The primary objective is to create a versatile coprocessor that can execute various cryptographic algorithms, including ECC(Elliptic-curve cryptography), RSA(Rivest–Shamir–Adleman), and AES (Advanced Encryption Standard) while providing a robust and secure solution for modern secure embedded systems. To achieve this goal, the coprocessor is equipped with a tightly coupled memory (TCM) for rapid data access during cryptographic operations. The TCM is placed within the coprocessor, ensuring quick retrieval of critical data and optimizing overall performance. Additionally, the program memory is positioned outside the coprocessor, allowing for easy updates and reconfiguration, which enhances adaptability to future algorithm implementations. Direct links are employed instead of DMA(Direct memory access) for data transfer, ensuring faster communication and reducing complexity. The AMBA-based communication architecture facilitates seamless interaction between the coprocessor and the main CPU, streamlining data flow and ensuring efficient utilization of system resources. The abstract nature of the coprocessor allows for easy integration of new cryptographic algorithms in the future. As the security landscape continues to evolve, the coprocessor can adapt and incorporate emerging algorithms, making it a future-proof solution for cryptographic processing. Furthermore, this study explores the addition of custom instructions into RISC-V ISE (Instruction Set Extension) to enhance cryptographic operations. By incorporating custom instructions specifically tailored for cryptographic algorithms, the coprocessor achieves higher efficiency and reduced cycles per instruction (CPI) compared to traditional instruction sets. The adoption of RISC-V 128-bit architecture significantly reduces the total number of instructions required for complex cryptographic tasks, leading to faster execution times and improved overall performance. Comparisons are made with 32-bit and 64-bit architectures, highlighting the advantages of the 128-bit architecture in terms of reduced instruction count and CPI. In conclusion, the abstract cryptographic coprocessor presented in this study offers significant advantages in terms of algorithm flexibility, security, and integration with the main CPU. By leveraging AMBA protocols and employing direct links for data transfer, the coprocessor achieves high-performance cryptographic operations without compromising system efficiency. With its TCM and external program memory, the coprocessor is capable of securely executing a wide range of cryptographic algorithms. This versatility and adaptability, coupled with the benefits of custom instructions and the 128-bit architecture, make it an invaluable asset for secure embedded systems, meeting the demands of modern cryptographic applications.

Keywords: abstract cryptographic coprocessor, AMBA protocols, ECC, RSA, AES, tightly coupled memory, secure embedded systems, RISC-V ISE, custom instructions, instruction count, cycles per instruction

Procedia PDF Downloads 67
16387 Resistance and Sub-Resistances of RC Beams Subjected to Multiple Failure Modes

Authors: F. Sangiorgio, J. Silfwerbrand, G. Mancini

Abstract:

Geometric and mechanical properties all influence the resistance of RC structures and may, in certain combination of property values, increase the risk of a brittle failure of the whole system. This paper presents a statistical and probabilistic investigation on the resistance of RC beams designed according to Eurocodes 2 and 8, and subjected to multiple failure modes, under both the natural variation of material properties and the uncertainty associated with cross-section and transverse reinforcement geometry. A full probabilistic model based on JCSS Probabilistic Model Code is derived. Different beams are studied through material nonlinear analysis via Monte Carlo simulations. The resistance model is consistent with Eurocode 2. Both a multivariate statistical evaluation and the data clustering analysis of outcomes are then performed. Results show that the ultimate load behaviour of RC beams subjected to flexural and shear failure modes seems to be mainly influenced by the combination of the mechanical properties of both longitudinal reinforcement and stirrups, and the tensile strength of concrete, of which the latter appears to affect the overall response of the system in a nonlinear way. The model uncertainty of the resistance model used in the analysis plays undoubtedly an important role in interpreting results.

Keywords: modelling, Monte Carlo simulations, probabilistic models, data clustering, reinforced concrete members, structural design

Procedia PDF Downloads 471
16386 Comparison of Methodologies to Compute the Probabilistic Seismic Hazard Involving Faults and Associated Uncertainties

Authors: Aude Gounelle, Gloria Senfaute, Ludivine Saint-Mard, Thomas Chartier

Abstract:

The long-term deformation rates of faults are not fully captured by Probabilistic Seismic Hazard Assessment (PSHA). PSHA that use catalogues to develop area or smoothed-seismicity sources is limited by the data available to constraint future earthquakes activity rates. The integration of faults in PSHA can at least partially address the long-term deformation. However, careful treatment of fault sources is required, particularly, in low strain rate regions, where estimated seismic hazard levels are highly sensitive to assumptions concerning fault geometry, segmentation and slip rate. When integrating faults in PSHA various constraints on earthquake rates from geologic and seismologic data have to be satisfied. For low strain rate regions where such data is scarce it would be especially challenging. Faults in PSHA requires conversion of the geologic and seismologic data into fault geometries, slip rates and then into earthquake activity rates. Several approaches exist for translating slip rates into earthquake activity rates. In the most frequently used approach, the background earthquakes are handled using a truncated approach, in which earthquakes with a magnitude lower or equal to a threshold magnitude (Mw) occur in the background zone, with a rate defined by the rate in the earthquake catalogue. Although magnitudes higher than the threshold are located on the fault with a rate defined using the average slip rate of the fault. As high-lighted by several research, seismic events with magnitudes stronger than the selected magnitude threshold may potentially occur in the background and not only at the fault, especially in regions of slow tectonic deformation. It also has been known that several sections of a fault or several faults could rupture during a single fault-to-fault rupture. It is then essential to apply a consistent modelling procedure to allow for a large set of possible fault-to-fault ruptures to occur aleatory in the hazard model while reflecting the individual slip rate of each section of the fault. In 2019, a tool named SHERIFS (Seismic Hazard and Earthquake Rates in Fault Systems) was published. The tool is using a methodology to calculate the earthquake rates in a fault system where the slip-rate budget of each fault is conversed into rupture rates for all possible single faults and faultto-fault ruptures. The objective of this paper is to compare the SHERIFS method with one other frequently used model to analyse the impact on the seismic hazard and through sensibility studies better understand the influence of key parameters and assumptions. For this application, a simplified but realistic case study was selected, which is in an area of moderate to hight seismicity (South Est of France) and where the fault is supposed to have a low strain.

Keywords: deformation rates, faults, probabilistic seismic hazard, PSHA

Procedia PDF Downloads 62
16385 Inverse Prediction of Thermal Parameters of an Annular Hyperbolic Fin Subjected to Thermal Stresses

Authors: Ashis Mallick, Rajeev Ranjan

Abstract:

The closed form solution for thermal stresses in an annular fin with hyperbolic profile is derived using Adomian decomposition method (ADM). The conductive-convective fin with variable thermal conductivity is considered in the analysis. The nonlinear heat transfer equation is efficiently solved by ADM considering insulated convective boundary conditions at the tip of fin. The constant of integration in the solution is to be estimated using minimum decomposition error method. The solution of temperature field is represented in a polynomial form for convenience to use in thermo-elasticity equation. The non-dimensional thermal stress fields are obtained using the ADM solution of temperature field coupled with the thermo-elasticity solution. The influence of the various thermal parameters in temperature field and stress fields are presented. In order to show the accuracy of the ADM solution, the present results are compared with the results available in literature. The stress fields in fin with hyperbolic profile are compared with those of uniform thickness profile. Result shows that hyperbolic fin profile is better choice for enhancing heat transfer. Moreover, less thermal stresses are developed in hyperbolic profile as compared to rectangular profile. Next, Nelder-Mead based simplex search method is employed for the inverse estimation of unknown non-dimensional thermal parameters in a given stress fields. Owing to the correlated nature of the unknowns, the best combinations of the model parameters which are satisfying the predefined stress field are to be estimated. The stress fields calculated using the inverse parameters give a very good agreement with the stress fields obtained from the forward solution. The estimated parameters are suitable to use for efficient and cost effective fin designing.

Keywords: Adomian decomposition, inverse analysis, hyperbolic fin, variable thermal conductivity

Procedia PDF Downloads 326
16384 Soft Computing Employment to Optimize Safety Stock Levels in Supply Chain Dairy Product under Supply and Demand Uncertainty

Authors: Riyadh Jamegh, Alla Eldin Kassam, Sawsan Sabih

Abstract:

In order to overcome uncertainty conditions and inability to meet customers' requests due to these conditions, organizations tend to reserve a certain safety stock level (SSL). This level must be chosen carefully in order to avoid the increase in holding cost due to excess in SSL or shortage cost due to too low SSL. This paper used soft computing fuzzy logic to identify optimal SSL; this fuzzy model uses the dynamic concept to cope with high complexity environment status. The proposed model can deal with three input variables, i.e., demand stability level, raw material availability level, and on hand inventory level by using dynamic fuzzy logic to obtain the best SSL as an output. In this model, demand stability, raw material, and on hand inventory levels are described linguistically and then treated by inference rules of the fuzzy model to extract the best level of safety stock. The aim of this research is to provide dynamic approach which is used to identify safety stock level, and it can be implanted in different industries. Numerical case study in the dairy industry with Yogurt 200 gm cup product is explained to approve the validity of the proposed model. The obtained results are compared with the current level of safety stock which is calculated by using the traditional approach. The importance of the proposed model has been demonstrated by the significant reduction in safety stock level.

Keywords: inventory optimization, soft computing, safety stock optimization, dairy industries inventory optimization

Procedia PDF Downloads 123
16383 Integrating Machine Learning and Rule-Based Decision Models for Enhanced B2B Sales Forecasting and Customer Prioritization

Authors: Wenqi Liu, Reginald Bailey

Abstract:

This study explores an advanced approach to enhancing B2B sales forecasting by integrating machine learning models with a rule-based decision framework. The methodology begins with the development of a machine learning classification model to predict conversion likelihood, aiming to improve accuracy over traditional methods like logistic regression. The classification model's effectiveness is measured using metrics such as accuracy, precision, recall, and F1 score, alongside a feature importance analysis to identify key predictors. Following this, a machine learning regression model is used to forecast sales value, with the objective of reducing mean absolute error (MAE) compared to linear regression techniques. The regression model's performance is assessed using MAE, root mean square error (RMSE), and R-squared metrics, emphasizing feature contribution to the prediction. To bridge the gap between predictive analytics and decision-making, a rule-based decision model is introduced that prioritizes customers based on predefined thresholds for conversion probability and predicted sales value. This approach significantly enhances customer prioritization and improves overall sales performance by increasing conversion rates and optimizing revenue generation. The findings suggest that this combined framework offers a practical, data-driven solution for sales teams, facilitating more strategic decision-making in B2B environments.

Keywords: sales forecasting, machine learning, rule-based decision model, customer prioritization, predictive analytics

Procedia PDF Downloads 14
16382 Learning Algorithms for Fuzzy Inference Systems Composed of Double- and Single-Input Rule Modules

Authors: Hirofumi Miyajima, Kazuya Kishida, Noritaka Shigei, Hiromi Miyajima

Abstract:

Most of self-tuning fuzzy systems, which are automatically constructed from learning data, are based on the steepest descent method (SDM). However, this approach often requires a large convergence time and gets stuck into a shallow local minimum. One of its solutions is to use fuzzy rule modules with a small number of inputs such as DIRMs (Double-Input Rule Modules) and SIRMs (Single-Input Rule Modules). In this paper, we consider a (generalized) DIRMs model composed of double and single-input rule modules. Further, in order to reduce the redundant modules for the (generalized) DIRMs model, pruning and generative learning algorithms for the model are suggested. In order to show the effectiveness of them, numerical simulations for function approximation, Box-Jenkins and obstacle avoidance problems are performed.

Keywords: Box-Jenkins's problem, double-input rule module, fuzzy inference model, obstacle avoidance, single-input rule module

Procedia PDF Downloads 352
16381 CFD Study of Subcooled Boiling Flow at Elevated Pressure Using a Mechanistic Wall Heat Partitioning Model

Authors: Machimontorn Promtong, Sherman C. P. Cheung, Guan H. Yeoh, Sara Vahaji, Jiyuan Tu

Abstract:

The wide range of industrial applications involved with boiling flows promotes the necessity of establishing fundamental knowledge in boiling flow phenomena. For this purpose, a number of experimental and numerical researches have been performed to elucidate the underlying physics of this flow. In this paper, the improved wall boiling models, implemented on ANSYS CFX 14.5, were introduced to study subcooled boiling flow at elevated pressure. At the heated wall boundary, the Fractal model, Force balance approach and Mechanistic frequency model are given for predicting the nucleation site density, bubble departure diameter, and bubble departure frequency. The presented wall heat flux partitioning closures were modified to consider the influence of bubble sliding along the wall before the lift-off, which usually happens in the flow boiling. The simulation was performed based on the Two-fluid model, where the standard k-ω SST model was selected for turbulence modelling. Existing experimental data at around 5 bars were chosen to evaluate the accuracy of the presented mechanistic approach. The void fraction and Interfacial Area Concentration (IAC) are in good agreement with the experimental data. However, the predicted bubble velocity and Sauter Mean Diameter (SMD) are over-predicted. This over-prediction may be caused by consideration of only dispersed and spherical bubbles in the simulations. In the future work, the important physical mechanisms of bubbles, such as merging and shrinking during sliding on the heated wall will be incorporated into this mechanistic model to enhance its capability for a wider range of flow prediction.

Keywords: subcooled boiling flow, computational fluid dynamics (CFD), mechanistic approach, two-fluid model

Procedia PDF Downloads 317
16380 Multi-Objective Multi-Period Allocation of Temporary Earthquake Disaster Response Facilities with Multi-Commodities

Authors: Abolghasem Yousefi-Babadi, Ali Bozorgi-Amiri, Aida Kazempour, Reza Tavakkoli-Moghaddam, Maryam Irani

Abstract:

All over the world, natural disasters (e.g., earthquakes, floods, volcanoes and hurricanes) causes a lot of deaths. Earthquakes are introduced as catastrophic events, which is accident by unusual phenomena leading to much loss around the world. Such could be replaced by disasters or any other synonyms strongly demand great long-term help and relief, which can be hard to be managed. Supplies and facilities are very important challenges after any earthquake which should be prepared for the disaster regions to satisfy the people's demands who are suffering from earthquake. This paper proposed disaster response facility allocation problem for disaster relief operations as a mathematical programming model. Not only damaged people in the earthquake victims, need the consumable commodities (e.g., food and water), but also they need non-consumable commodities (e.g., clothes) to protect themselves. Therefore, it is concluded that paying attention to disaster points and people's demands are very necessary. To deal with this objective, both commodities including consumable and need non-consumable commodities are considered in the presented model. This paper presented the multi-objective multi-period mathematical programming model regarding the minimizing the average of the weighted response times and minimizing the total operational cost and penalty costs of unmet demand and unused commodities simultaneously. Furthermore, a Chebycheff multi-objective solution procedure as a powerful solution algorithm is applied to solve the proposed model. Finally, to illustrate the model applicability, a case study of the Tehran earthquake is studied, also to show model validation a sensitivity analysis is carried out.

Keywords: facility location, multi-objective model, disaster response, commodity

Procedia PDF Downloads 257
16379 Source Identification Model Based on Label Propagation and Graph Ordinary Differential Equations

Authors: Fuyuan Ma, Yuhan Wang, Junhe Zhang, Ying Wang

Abstract:

Identifying the sources of information dissemination is a pivotal task in the study of collective behaviors in networks, enabling us to discern and intercept the critical pathways through which information propagates from its origins. This allows for the control of the information’s dissemination impact in its early stages. Numerous methods for source detection rely on pre-existing, underlying propagation models as prior knowledge. Current models that eschew prior knowledge attempt to harness label propagation algorithms to model the statistical characteristics of propagation states or employ Graph Neural Networks (GNNs) for deep reverse modeling of the diffusion process. These approaches are either deficient in modeling the propagation patterns of information or are constrained by the over-smoothing problem inherent in GNNs, which limits the stacking of sufficient model depth to excavate global propagation patterns. Consequently, we introduce the ODESI model. Initially, the model employs a label propagation algorithm to delineate the distribution density of infected states within a graph structure and extends the representation of infected states from integers to state vectors, which serve as the initial states of nodes. Subsequently, the model constructs a deep architecture based on GNNs-coupled Ordinary Differential Equations (ODEs) to model the global propagation patterns of continuous propagation processes. Addressing the challenges associated with solving ODEs on graphs, we approximate the analytical solutions to reduce computational costs. Finally, we conduct simulation experiments on two real-world social network datasets, and the results affirm the efficacy of our proposed ODESI model in source identification tasks.

Keywords: source identification, ordinary differential equations, label propagation, complex networks

Procedia PDF Downloads 18
16378 The Value of Store Choice Criteria on Perceived Patronage Intentions

Authors: Susana Marques

Abstract:

Research on how store environment cues influence consumers’ store choice decision criteria, such as store operations, product quality, monetary price, store image and sales promotion, is sparse. Especially absent research on the simultaneous impact of multiple store environment cues. The authors propose a comprehensive store choice model that includes: three types of store environment cues as exogenous constructs; various store choice criteria as possible mediating constructs, and store patronage intentions as an endogenous construct. On the basis of testing with a sample of 561 customers of hypermarkets, the model is partially supported. This study used structural equation modelling to test the proposed model.

Keywords: store choice, store patronage, structural equation modelling, retailing

Procedia PDF Downloads 270
16377 The Growth Curve of Gompertz Model in Body Weight of Slovak Mixed-Sex Goose Breeds

Authors: Cyril Hrncar, Jozef Bujko, Widya P. B. Putra

Abstract:

The growth curve of poultry is important to evaluate the farming management system. This study was aimed to estimate the growth curve of body weight in goose. The growth curve in this study was estimated with non-linear Gompertz model through CurveExpert 1.4. software. Three Slovak mixed-sex goose breeds of Landes (L), Pomeranian (P) and Steinbacher (S) were used in this study. Total of 28 geese (10 L, 8 P and 10 S) were used to estimate the growth curve. Research showed that the asymptotic weight (A) in those geese were reached of 5332.51 g (L), 6186.14 g (P) and 5048.27 g (S). Thus, the maturing rate (k) in each breed were similar (0.05 g/day). The weight of inflection was reached of 1960.48 g (L), 2274.32 g (P) and 1855.98 g (S). The time of inflection (ti) was reached of 25.6 days (L), 26.2 days (P) and 27.80 days (S). The maximum growth rate (MGR) was reached of 98.02 g/day (L), 113.72 g/day (P) and 92.80 g/day (S). Hence, the coefficient of determination (R2) in Gompertz model was 0.99 for each breed. It can be concluded that Pomeranian geese had highest of growth trait than the other breeds.

Keywords: body weight, growth curve, inflection, Slovak geese, Gompertz model

Procedia PDF Downloads 143
16376 Context-Aware Point-Of-Interests Recommender Systems Using Integrated Sentiment and Network Analysis

Authors: Ho Yeon Park, Kyoung-Jae Kim

Abstract:

Recently, user’s interests for location-based social network service increases according to the advances of social web and location-based technologies. It may be easy to recommend preferred items if we can use user’s preference, context and social network information simultaneously. In this study, we propose context-aware POI (point-of-interests) recommender systems using location-based network analysis and sentiment analysis which consider context, social network information and implicit user’s preference score. We propose a context-aware POI recommendation system consisting of three sub-modules and an integrated recommendation system of them. First, we will develop a recommendation module based on network analysis. This module combines social network analysis and cluster-indexing collaboration filtering. Next, this study develops a recommendation module using social singular value decomposition (SVD) and implicit SVD. In this research, we will develop a recommendation module that can recommend preference scores based on the frequency of POI visits of user in POI recommendation process by using social and implicit SVD which can reflect implicit feedback in collaborative filtering. We also develop a recommendation module using them that can estimate preference scores based on the recommendation. Finally, this study will propose a recommendation module using opinion mining and emotional analysis using data such as reviews of POIs extracted from location-based social networks. Finally, we will develop an integration algorithm that combines the results of the three recommendation modules proposed in this research. Experimental results show the usefulness of the proposed model in relation to the recommended performance.

Keywords: sentiment analysis, network analysis, recommender systems, point-of-interests, business analytics

Procedia PDF Downloads 248
16375 Integrating Service Learning into a Business Analytics Course: A Comparative Investigation

Authors: Gokhan Egilmez, Erika Hatfield, Julie Turner

Abstract:

In this study, we investigated the impacts of service-learning integration on an undergraduate level business analytics course from multiple perspectives, including academic proficiency, community awareness, engagement, social responsibility, and reflection. We assessed the impact of the service-learning experience by using a survey developed primarily based on the literature review and secondarily on an ad hoc group of researchers. Then, we implemented the survey in two sections, where one of the sections was a control group. We compared the results of the empirical survey visually and statistically.

Keywords: business analytics, service learning, experiential education, statistical analysis, survey research

Procedia PDF Downloads 111
16374 Food Security Model and the Role of Community Empowerment: The Case of a Marginalized Village in Mexico, Tatoxcac, Puebla

Authors: Marco Antonio Lara De la Calleja, María Catalina Ovando Chico, Eduardo Lopez Ruiz

Abstract:

Community empowerment has been proved to be a key element in the solution of the food security problem. As a result of a conceptual analysis, it was found that agricultural production, economic development and governance, are the traditional basis of food security models. Although the literature points to social inclusion as an important factor for food security, no model has considered it as the basis of it. The aim of this research is to identify different dimensions that make an integral model for food security, with emphasis on community empowerment. A diagnosis was made in the study community (Tatoxcac, Zacapoaxtla, Puebla), to know the aspects that impact the level of food insecurity. With a statistical sample integrated by 200 families, the Latin American and Caribbean Food Security Scale (ELCSA) was applied, finding that: in households composed by adults and children, have moderated food insecurity, (ELCSA scale has three levels, low, moderated and high); that result is produced mainly by the economic income capacity and the diversity of the diet on its food. With that being said, a model was developed to promote food security through five dimensions: 1. Regional context of the community; 2. Structure and system of local food; 3. Health and nutrition; 4. Information and technology access; and 5. Self-awareness and empowerment. The specific actions on each axis of the model, allowed a systemic approach needed to attend food security in the community, through the empowerment of society. It is concluded that the self-awareness of local communities is an area of extreme importance, which must be taken into account for participatory schemes to improve food security. In the long term, the model requires the integrated participation of different actors, such as government, companies and universities, to solve something such vital as food security.

Keywords: community empowerment, food security, model, systemic approach

Procedia PDF Downloads 371
16373 Exploring the Factors Affecting the Intention of Using Mobile Phone E-Book by TAM and IDT

Authors: Yen-Ku Kuo, Chie-Bein Chen, Jyh-Yi Shih, Kuang-Yi Lin, Chien-Han Peng

Abstract:

This study is primarily concerned with exploring what factors affect the consumer’s intention of using mobile phone e-book. In developing research structure, we adopted technology acceptance model (TAM) and Innovation Diffusion Theory (IDT) as a foundation. The analysis method of structural equation model (SEM) was used to carry out this study. Subjects were 261 users who are using or used the mobile phone e-book. The findings can be summed up as follows: (1) The subjective norm and job relevance has non-significant and positive influence to the perceived usefulness. This represents now the user are still in a small number and most of them used it in non-work related purpose. (2) The output quality, result demonstrability and perceived ease of use were confirmed to have positive and significant influence to the perceived usefulness. (3) The moderator “innovative diffusion” affects the relationship between the attitude and behavior intention. These findings could be a reference for the practice and future study to make further exploration.

Keywords: mobile phone e-book, technology acceptance model (TAM), innovation diffusion theory (IDT), structural equation model (SEM)

Procedia PDF Downloads 509
16372 Stability Analysis and Experimental Evaluation on Maxwell Model of Impedance Control

Authors: Le Fu, Rui Wu, Gang Feng Liu, Jie Zhao

Abstract:

Normally, impedance control methods are based on a model that connects a spring and damper in parallel. The series connection, namely the Maxwell model, has emerged as a counterpart and draw the attention of robotics researchers. In the theoretical analysis, it turns out that the two pattern are both equivalents to some extent, but notable differences of response characteristics exist, especially in the effect of damping viscosity. However, this novel impedance control design is lack of validation on realistic robot platforms. In this study, stability analysis and experimental evaluation are achieved using a 3-fingered Barrett® robotic hand BH8-282 endowed with tactile sensing, mounted on a torque-controlled lightweight and collaborative robot KUKA® LBR iiwa 14 R820. Object handover and incoming objects catching tasks are executed for validation and analysis. Experimental results show that the series connection pattern has much better performance in natural impact or shock absorption, which indicate promising applications in robots’ safe and physical interaction with humans and objects in various environments.

Keywords: impedance control, Maxwell model, force control, dexterous manipulation

Procedia PDF Downloads 496
16371 The Study of Chitosan beads Adsorption Properties for the Removal of Heavy Metals

Authors: Peter O. Osifo, Hein W. J. P. Neomagus

Abstract:

In this study, a predicted pH model was used to determine adsorption equilibrium properties of copper, lead, zinc and cadmium. Chitosan was prepared from the exoskeleton of Cape rock-lobsters, collected from the surroundings of Cape Town, South Africa. The beads were cross-linked with gluteraldehyde to restore its chemical stability in acid media. The chitosan beads were characterized; the beads water contents and pKa varied in the range of 90-96% and 4.3-6.0 respectively and the degree of crosslinking for the beads was 18%. A pH-model, which described the reversibility of the metal adsorbed onto the beads, was used to predict the equilibrium properties of copper, lead, zinc and cadmium adsorption onto the cross-linked beads. The model accounts for the effect of pH and the important model parameters; the equilibrium adsorption constant (Kads) and to a lesser extent the adsorbent adsorption capacity (qmax). The adsorption equilibrium constant for copper, lead, zinc and cadmium were found to be 2.58×10-3, 2.22×0-3, 9.55×0-3, and 4.79×0-3, respectively. The adsorbent maximum capacity was determined to be 4.2 mmol/g.

Keywords: chitosan beads, adsorption, heavy metals, waste water

Procedia PDF Downloads 378
16370 Investigation Bubble Growth and Nucleation Rates during the Pool Boiling Heat Transfer of Distilled Water Using Population Balance Model

Authors: V. Nikkhah Rashidabad, M. Manteghian, M. Masoumi, S. Mousavian

Abstract:

In this research, the changes in bubbles diameter and number that may occur due to the change in heat flux of pure water during pool boiling process. For this purpose, test equipment was designed and developed to collect test data. The bubbles were graded using Caliper Screen software. To calculate the growth and nucleation rates of bubbles under different fluxes, population balance model was employed. The results show that the increase in heat flux from q=20 kw/m2 to q=102 kw/m2 raised the growth and nucleation rates of bubbles.

Keywords: heat flux, bubble growth, bubble nucleation, population balance model

Procedia PDF Downloads 474
16369 Development of Open Source Geospatial Certification Model Based on Geospatial Technology Competency Model

Authors: Tanzeel Ur Rehman Khan, Franz Josef Behr, Phillip Davis

Abstract:

Open source geospatial certifications are needed in geospatial technology education and industry sector. In parallel with proprietary software, free and open source software solutions become important in geospatial technology research and play an important role for the growth of the geospatial industry. ESRI, GISCI (GIS Certification Institute), ASPRS (American Society of Photogrammetry and remote sensing), and Meta spatial are offering certifications on proprietary and open source software. These are portfolio and competency based certifications depending on GIS Body of Knowledge (Bok). The analysis of these certification approaches might lead to the discovery of some gaps in them and will open a new way to develop certifications related to the geospatial open source (OS). This new certification will investigate the different geospatial competencies according to open source tools that help to identify geospatial professionals and strengthen the geospatial academic content. The goal of this research is to introduce a geospatial certification model based on geospatial technology competency model (GTCM).The developed certification will not only incorporate the importance of geospatial education and production of the geospatial competency-based workforce in universities and companies (private or public) as well as describe open source solutions with tools and technology. Job analysis, market analysis, survey analysis of this certification opens a new horizon for business as well.

Keywords: geospatial certification, open source, geospatial technology competency model, geoscience

Procedia PDF Downloads 565
16368 Probability-Based Damage Detection of Structures Using Kriging Surrogates and Enhanced Ideal Gas Molecular Movement Algorithm

Authors: M. R. Ghasemi, R. Ghiasi, H. Varaee

Abstract:

Surrogate model has received increasing attention for use in detecting damage of structures based on vibration modal parameters. However, uncertainties existing in the measured vibration data may lead to false or unreliable output result from such model. In this study, an efficient approach based on Monte Carlo simulation is proposed to take into account the effect of uncertainties in developing a surrogate model. The probability of damage existence (PDE) is calculated based on the probability density function of the existence of undamaged and damaged states. The kriging technique allows one to genuinely quantify the surrogate error, therefore it is chosen as metamodeling technique. Enhanced version of ideal gas molecular movement (EIGMM) algorithm is used as main algorithm for model updating. The developed approach is applied to detect simulated damage in numerical models of 72-bar space truss and 120-bar dome truss. The simulation results show the proposed method can perform well in probability-based damage detection of structures with less computational effort compared to direct finite element model.

Keywords: probability-based damage detection (PBDD), Kriging, surrogate modeling, uncertainty quantification, artificial intelligence, enhanced ideal gas molecular movement (EIGMM)

Procedia PDF Downloads 239
16367 Dynamic Risk Model for Offshore Decommissioning Using Bayesian Belief Network

Authors: Ahmed O. Babaleye, Rafet E. Kurt

Abstract:

The global oil and gas industry is beginning to witness an increase in the number of installations moving towards decommissioning. Decommissioning of offshore installations is a complex, costly and hazardous activity, making safety one of the major concerns. Among existing removal options, complete and partial removal options pose the highest risks. Therefore, a dynamic risk model of the accidents from the two options is important to assess the risks on an overall basis. In this study, a risk-based safety model is developed to conduct quantitative risk analysis (QRA) for jacket structure systems failure. Firstly, bow-tie (BT) technique is utilised to model the causal relationship between the system failure and potential accident scenarios. Subsequently, to relax the shortcomings of BT, Bayesian Belief Networks (BBNs) were established to dynamically assess associated uncertainties and conditional dependencies. The BBN is developed through a similitude mapping of the developed bow-tie. The BBN is used to update the failure probabilities of the contributing elements through diagnostic analysis, thus, providing a case-specific and realistic safety analysis method when compared to a bow-tie. This paper presents the application of dynamic safety analysis to guide the allocation of risk control measures and consequently, drive down the avoidable cost of remediation.

Keywords: Bayesian belief network, offshore decommissioning, dynamic safety model, quantitative risk analysis

Procedia PDF Downloads 279
16366 Aerodynamic Devices Development for Model Aircraft Control and Wind-Driven Bicycle

Authors: Yuta Moriyama, Tsuyoshi Yamazaki, Etsuo Morishita

Abstract:

Several aerodynamic devices currently attract engineers and research students. The plasma actuator is one of them, and it is very effective to control the flow. The actuator recovers a separated flow to an attached one. The actuator is also inversely applied to a spoiler. The model aircraft might be controlled by this actuator. We develop a model aircraft with the plasma actuator. Another interesting device is the Wells turbine which rotates in one direction. The present authors propose a bicycle with the Wells turbine in the wheels. Power reduction is measured when the turbine is driven by an electric motor at the exit of a wind tunnel. Several Watts power reduction might be possible. This means that the torque of the bike can be augmented by the turbine in the cross wind. These devices are tested in the wind tunnel with a three-component balance and the aerodynamic forces and moment are obtained. In this paper, we introduce these devices and their aerodynamic characteristics. The control force and moment of the plasma actuator are clarified and the power reduction of the bicycle is quantified.

Keywords: aerodynamics, model aircraft, plasma actuator, Wells turbine

Procedia PDF Downloads 243
16365 The SBO/LOCA Analysis of TRACE/SNAP for Kuosheng Nuclear Power Plant

Authors: J. R. Wang, H. T. Lin, Y. Chiang, H. C. Chen, C. Shih

Abstract:

Kuosheng Nuclear Power Plant (NPP) is located on the northern coast of Taiwan. Its nuclear steam supply system is a type of BWR/6 designed and built by General Electric on a twin unit concept. First, the methodology of Kuosheng NPP SPU (Stretch Power Uprate) safety analysis TRACE/SNAP model was developed in this research. Then, in order to estimate the safety of Kuosheng NPP under the more severe condition, the SBO (Station Blackout) + LOCA (Loss-of-Coolant Accident) transient analysis of Kuosheng NPP SPU TRACE/SNAP model was performed. Besides, the animation model of Kuosheng NPP was presented using the animation function of SNAP with TRACE/SNAP analysis results.

Keywords: TRACE, safety analysis, BWR/6, severe accident

Procedia PDF Downloads 714
16364 Edmonton Urban Growth Model as a Support Tool for the City Plan Growth Scenarios Development

Authors: Sinisa J. Vukicevic

Abstract:

Edmonton is currently one of the youngest North American cities and has achieved significant growth over the past 40 years. Strong urban shift requires a new approach to how the city is envisioned, planned, and built. This approach is evidence-based scenario development, and an urban growth model was a key support tool in framing Edmonton development strategies, developing urban policies, and assessing policy implications. The urban growth model has been developed using the Metronamica software platform. The Metronamica land use model evaluated the dynamic of land use change under the influence of key development drivers (population and employment), zoning, land suitability, and land and activity accessibility. The model was designed following the Big City Moves ideas: become greener as we grow, develop a rebuildable city, ignite a community of communities, foster a healing city, and create a city of convergence. The Big City Moves were converted to three development scenarios: ‘Strong Central City’, ‘Node City’, and ‘Corridor City’. Each scenario has a narrative story that expressed scenario’s high level goal, scenario’s approach to residential and commercial activities, to transportation vision, and employment and environmental principles. Land use demand was calculated for each scenario according to specific density targets. Spatial policies were analyzed according to their level of importance within the policy set definition for the specific scenario, but also through the policy measures. The model was calibrated on the way to reproduce known historical land use pattern. For the calibration, we used 2006 and 2011 land use data. The validation is done independently, which means we used the data we did not use for the calibration. The model was validated with 2016 data. In general, the modeling process contain three main phases: ‘from qualitative storyline to quantitative modelling’, ‘model development and model run’, and ‘from quantitative modelling to qualitative storyline’. The model also incorporates five spatial indicators: distance from residential to work, distance from residential to recreation, distance to river valley, urban expansion and habitat fragmentation. The major finding of this research could be looked at from two perspectives: the planning perspective and technology perspective. The planning perspective evaluates the model as a tool for scenario development. Using the model, we explored the land use dynamic that is influenced by a different set of policies. The model enables a direct comparison between the three scenarios. We explored the similarities and differences of scenarios and their quantitative indicators: land use change, population change (and spatial allocation), job allocation, density (population, employment, and dwelling unit), habitat connectivity, proximity to objects of interest, etc. From the technology perspective, the model showed one very important characteristic: the model flexibility. The direction for policy testing changed many times during the consultation process and model flexibility in applying all these changes was highly appreciated. The model satisfied our needs as scenario development and evaluation tool, but also as a communication tool during the consultation process.

Keywords: urban growth model, scenario development, spatial indicators, Metronamica

Procedia PDF Downloads 94
16363 A Psychoanalytic Lens: Unmasked Layers of the Self among Post-Graduate Psychology Students in Surviving the COVID-19 Lockdown

Authors: Sharon Sibanda, Benny Motileng

Abstract:

The World Health Organisation (WHO) identified the Sars-Cov-2 (COVID-19) as a pandemic on the 12ᵗʰ of March 2020, with South Africa recording its first case on the 5ᵗʰ of March 2020. The rapidly spreading virus led the South African government to implement one of the strictest nationwide lockdowns globally, resulting in the closing down of all institutions of higher learning effective March 18ᵗʰ 2020. Thus, this qualitative study primarily aimed to explore whether post-graduate psychology students were in a state of a depleted or cohesive self, post the psychological isolation of COVID-19 risk-adjusted level 5 lockdown. Semi-structured interviews from a qualitative interpretive approach comprising N=6 psychology post-graduate students facilitated a rich understanding of their intra-psychic experiences of the self. Thematic analysis of data gathered from the interviews illuminated how students were forced into the self by the emotional isolation of hard lockdown, with the emergence of core psychic conflict often defended against through external self-object experiences. The findings also suggest that lockdown stripped off this sample of psychology post-graduate students’ defensive escape from the inner self through external self-object distractions. The external self was stripped to the core of the internal self by the isolation of hard lockdown, thereby uncovering the psychic function of roles and defenses amalgamated throughout modern cultural consciousness that dictates self-functioning. The study suggests modelling reflexivity skills in the integration of internal and external self-experience dynamics as part of a training model for continued personal and professional development for psychology students.

Keywords: COVID-19, fragmentation, self-object experience, true/false self

Procedia PDF Downloads 57
16362 Developing a Knowledge-Based Lean Six Sigma Model to Improve Healthcare Leadership Performance

Authors: Yousuf N. Al Khamisi, Eduardo M. Hernandez, Khurshid M. Khan

Abstract:

Purpose: This paper presents a model of a Knowledge-Based (KB) using Lean Six Sigma (L6σ) principles to enhance the performance of healthcare leadership. Design/methodology/approach: Using L6σ principles to enhance healthcare leaders’ performance needs a pre-assessment of the healthcare organisation’s capabilities. The model will be developed using a rule-based approach of KB system. Thus, KB system embeds Gauging Absence of Pre-requisite (GAP) for benchmarking and Analytical Hierarchy Process (AHP) for prioritization. A comprehensive literature review will be covered for the main contents of the model with a typical output of GAP analysis and AHP. Findings: The proposed KB system benchmarks the current position of healthcare leadership with the ideal benchmark one (resulting from extensive evaluation by the KB/GAP/AHP system of international leadership concepts in healthcare environments). Research limitations/implications: Future work includes validating the implementation model in healthcare environments around the world. Originality/value: This paper presents a novel application of a hybrid KB combines of GAP and AHP methodology. It implements L6σ principles to enhance healthcare performance. This approach assists healthcare leaders’ decision making to reach performance improvement against a best practice benchmark.

Keywords: Lean Six Sigma (L6σ), Knowledge-Based System (KBS), healthcare leadership, Gauge Absence Prerequisites (GAP), Analytical Hierarchy Process (AHP)

Procedia PDF Downloads 165
16361 Currency Exchange Rate Forecasts Using Quantile Regression

Authors: Yuzhi Cai

Abstract:

In this paper, we discuss a Bayesian approach to quantile autoregressive (QAR) time series model estimation and forecasting. Together with a combining forecasts technique, we then predict USD to GBP currency exchange rates. Combined forecasts contain all the information captured by the fitted QAR models at different quantile levels and are therefore better than those obtained from individual models. Our results show that an unequally weighted combining method performs better than other forecasting methodology. We found that a median AR model can perform well in point forecasting when the predictive density functions are symmetric. However, in practice, using the median AR model alone may involve the loss of information about the data captured by other QAR models. We recommend that combined forecasts should be used whenever possible.

Keywords: combining forecasts, MCMC, predictive density functions, quantile forecasting, quantile modelling

Procedia PDF Downloads 255