Search results for: cluster model approach
26094 A New Bound on the Average Information Ratio of Perfect Secret-Sharing Schemes for Access Structures Based on Bipartite Graphs of Larger Girth
Authors: Hui-Chuan Lu
Abstract:
In a perfect secret-sharing scheme, a dealer distributes a secret among a set of participants in such a way that only qualified subsets of participants can recover the secret and the joint share of the participants in any unqualified subset is statistically independent of the secret. The access structure of the scheme refers to the collection of all qualified subsets. In a graph-based access structures, each vertex of a graph G represents a participant and each edge of G represents a minimal qualified subset. The average information ratio of a perfect secret-sharing scheme realizing a given access structure is the ratio of the average length of the shares given to the participants to the length of the secret. The infimum of the average information ratio of all possible perfect secret-sharing schemes realizing an access structure is called the optimal average information ratio of that access structure. We study the optimal average information ratio of the access structures based on bipartite graphs. Based on some previous results, we give a bound on the optimal average information ratio for all bipartite graphs of girth at least six. This bound is the best possible for some classes of bipartite graphs using our approach.Keywords: secret-sharing scheme, average information ratio, star covering, deduction, core cluster
Procedia PDF Downloads 36226093 Effective Parameter Selection for Audio-Based Music Mood Classification for Christian Kokborok Song: A Regression-Based Approach
Authors: Sanchali Das, Swapan Debbarma
Abstract:
Music mood classification is developing in both the areas of music information retrieval (MIR) and natural language processing (NLP). Some of the Indian languages like Hindi English etc. have considerable exposure in MIR. But research in mood classification in regional language is very less. In this paper, powerful audio based feature for Kokborok Christian song is identified and mood classification task has been performed. Kokborok is an Indo-Burman language especially spoken in the northeastern part of India and also some other countries like Bangladesh, Myanmar etc. For performing audio-based classification task, useful audio features are taken out by jMIR software. There are some standard audio parameters are there for the audio-based task but as known to all that every language has its unique characteristics. So here, the most significant features which are the best fit for the database of Kokborok song is analysed. The regression-based model is used to find out the independent parameters that act as a predictor and predicts the dependencies of parameters and shows how it will impact on overall classification result. For classification WEKA 3.5 is used, and selected parameters create a classification model. And another model is developed by using all the standard audio features that are used by most of the researcher. In this experiment, the essential parameters that are responsible for effective audio based mood classification and parameters that do not significantly change for each of the Christian Kokborok songs are analysed, and a comparison is also shown between the two above model.Keywords: Christian Kokborok song, mood classification, music information retrieval, regression
Procedia PDF Downloads 22226092 Heterogeneous Artifacts Construction for Software Evolution Control
Authors: Mounir Zekkaoui, Abdelhadi Fennan
Abstract:
The software evolution control requires a deep understanding of the changes and their impact on different system heterogeneous artifacts. And an understanding of descriptive knowledge of the developed software artifacts is a prerequisite condition for the success of the evolutionary process. The implementation of an evolutionary process is to make changes more or less important to many heterogeneous software artifacts such as source code, analysis and design models, unit testing, XML deployment descriptors, user guides, and others. These changes can be a source of degradation in functional, qualitative or behavioral terms of modified software. Hence the need for a unified approach for extraction and representation of different heterogeneous artifacts in order to ensure a unified and detailed description of heterogeneous software artifacts, exploitable by several software tools and allowing to responsible for the evolution of carry out the reasoning change concerned.Keywords: heterogeneous software artifacts, software evolution control, unified approach, meta model, software architecture
Procedia PDF Downloads 44526091 Modeling Approach to Better Control Fouling in a Submerged Membrane Bioreactor for Wastewater Treatment: Development of Analytical Expressions in Steady-State Using ASM1
Authors: Benaliouche Hana, Abdessemed Djamal, Meniai Abdessalem, Lesage Geoffroy, Heran Marc
Abstract:
This paper presents a dynamic mathematical model of activated sludge which is able to predict the formation and degradation kinetics of SMP (Soluble microbial products) in membrane bioreactor systems. The model is based on a calibrated version of ASM1 with the theory of production and degradation of SMP. The model was calibrated on the experimental data from MBR (Mathematical modeling Membrane bioreactor) pilot plant. Analytical expressions have been developed, describing the concentrations of the main state variables present in the sludge matrix, with the inclusion of only six additional linear differential equations. The objective is to present a new dynamic mathematical model of activated sludge capable of predicting the formation and degradation kinetics of SMP (UAP and BAP) from the submerged membrane bioreactor (BRMI), operating at low organic load (C / N = 3.5), for two sludge retention times (SRT) fixed at 40 days and 60 days, to study their impact on membrane fouling, The modeling study was carried out under the steady-state condition. Analytical expressions were then validated by comparing their results with those obtained by simulations using GPS-X-Hydromantis software. These equations made it possible, by means of modeling approaches (ASM1), to identify the operating and kinetic parameters and help to predict membrane fouling.Keywords: Activated Sludge Model No. 1 (ASM1), mathematical modeling membrane bioreactor, soluble microbial products, UAP, BAP, Modeling SMP, MBR, heterotrophic biomass
Procedia PDF Downloads 29526090 An Automatic Model Transformation Methodology Based on Semantic and Syntactic Comparisons and the Granularity Issue Involved
Authors: Tiexin Wang, Sebastien Truptil, Frederick Benaben
Abstract:
Model transformation, as a pivotal aspect of Model-driven engineering, attracts more and more attentions both from researchers and practitioners. Many domains (enterprise engineering, software engineering, knowledge engineering, etc.) use model transformation principles and practices to serve to their domain specific problems; furthermore, model transformation could also be used to fulfill the gap between different domains: by sharing and exchanging knowledge. Since model transformation has been widely used, there comes new requirement on it: effectively and efficiently define the transformation process and reduce manual effort that involved in. This paper presents an automatic model transformation methodology based on semantic and syntactic comparisons, and focuses particularly on granularity issue that existed in transformation process. Comparing to the traditional model transformation methodologies, this methodology serves to a general purpose: cross-domain methodology. Semantic and syntactic checking measurements are combined into a refined transformation process, which solves the granularity issue. Moreover, semantic and syntactic comparisons are supported by software tool; manual effort is replaced in this way.Keywords: automatic model transformation, granularity issue, model-driven engineering, semantic and syntactic comparisons
Procedia PDF Downloads 39426089 Prevalence of Elder Abuse and Effects of Social Factors on It
Authors: Ezat Vahidian, Babak Eshrati
Abstract:
Introduction: Elder abuse, a very complex issue with diverse definitions and names, has been very slow to capture the public eye and public policy since it is manifested at many levels. It requires the involvement of different types of professionals. While elder abuse is not a new phenomenon, the speed of population ageing world-wide is likely to lead to an increase in its incidence and prevalence. Elder abuse has devastating consequences for older persons such as poor quality of life, psychological distress, and loss of property and security. It is also associated with increased mortality and morbidity. Elder abuse is a problem that manifests itself in both rich and poor countries and at all levels of society. Purpose: The purpose of this study is to determine the prevalence of elder abuse and effects of social factor on it in Markazi Province. Materials and methods: The society of the study was all of the elders in Markazi Province that were available by geographical address in the table of rural and urban household societies. The study was cross sectional and multi phases in sampling the first one was classification according rural and urban area and the second one was cluster sampling with equal cluster. Estimation of samples were 472 persons and increased by design effect to 1110 persons. Collection data was done by questionnaire and analyzed by SPSS and chi 2 exam. Results: This study showed 70 persons were abused (42/8% male and 57/2% female) mean of ages was 74/7 years. 64% were marred and 31% were widows. There were not any significant meaningful association between elder abuse and area of living (pv=0.299),occupation (p.v=0.104), education (pv=0.358) and age (P.value=0.104) there were significant meaningful association between physical impairment (pv=0.08), and movement impairment (P.value=0.008). Conclusion: Results verify that maltreatment occurred in the aged persons. Analysis of data indicated that elder abuse exist in every socioeconomic group with any context of education in urban area and rural area and in men and women. Prevalence of elder abuse was 6.3% (70 persons) that verify the data of developed countries with limited sample.Keywords: elder abuse, education, occupation, area of living
Procedia PDF Downloads 40326088 LanE-change Path Planning of Autonomous Driving Using Model-Based Optimization, Deep Reinforcement Learning and 5G Vehicle-to-Vehicle Communications
Authors: William Li
Abstract:
Lane-change path planning is a crucial and yet complex task in autonomous driving. The traditional path planning approach based on a system of carefully-crafted rules to cover various driving scenarios becomes unwieldy as more and more rules are added to deal with exceptions and corner cases. This paper proposes to divide the entire path planning to two stages. In the first stage the ego vehicle travels longitudinally in the source lane to reach a safe state. In the second stage the ego vehicle makes lateral lane-change maneuver to the target lane. The paper derives the safe state conditions based on lateral lane-change maneuver calculation to ensure collision free in the second stage. To determine the acceleration sequence that minimizes the time to reach a safe state in the first stage, the paper proposes three schemes, namely, kinetic model based optimization, deep reinforcement learning, and 5G vehicle-to-vehicle (V2V) communications. The paper investigates these schemes via simulation. The model-based optimization is sensitive to the model assumptions. The deep reinforcement learning is more flexible in handling scenarios beyond the model assumed by the optimization. The 5G V2V eliminates uncertainty in predicting future behaviors of surrounding vehicles by sharing driving intents and enabling cooperative driving.Keywords: lane change, path planning, autonomous driving, deep reinforcement learning, 5G, V2V communications, connected vehicles
Procedia PDF Downloads 25226087 Modeling and Controlling Nonlinear Dynamical Effects in Non-Contact Superconducting and Diamagnetic Suspensions
Authors: Sergey Kuznetsov, Yuri Urman
Abstract:
We present an approach to investigate non-linear dynamical effects occurring in the noncontact superconducting and diamagnetic suspensions, when levitated body has finite size. This approach is based on the calculation of interaction energy between spherical finite size superconducting or diamagnetic body with external magnetic field. Effects of small deviations from spherical shape may be also taken into account by introducing small corrections to the energy. This model allows investigating dynamical effects important for practical applications, such as nonlinear resonances, change of vibration plane, coupling of rotational and translational motions etc. We also show how the geometry of suspension affects various dynamical effects and how an inverse problem may be formulated to enforce or diminish various dynamical effects.Keywords: levitation, non-linear dynamics, superconducting, diamagnetic stability
Procedia PDF Downloads 41026086 Partial Differential Equation-Based Modeling of Brain Response to Stimuli
Authors: Razieh Khalafi
Abstract:
The brain is the information processing centre of the human body. Stimuli in the form of information are transferred to the brain and then brain makes the decision on how to respond to them. In this research, we propose a new partial differential equation which analyses the EEG signals and make a relationship between the incoming stimuli and the brain response to them. In order to test the proposed model, a set of external stimuli applied to the model and the model’s outputs were checked versus the real EEG data. The results show that this model can model the EEG signal well. The proposed model is useful not only for modelling of EEG signal in case external stimuli but it can be used for modelling of brain response in case of internal stimuli.Keywords: brain, stimuli, partial differential equation, response, EEG signal
Procedia PDF Downloads 55426085 Enhancing Technical Trading Strategy on the Bitcoin Market using News Headlines and Language Models
Authors: Mohammad Hosein Panahi, Naser Yazdani
Abstract:
we present a technical trading strategy that leverages the FinBERT language model and financial news analysis with a focus on news related to a subset of Nasdaq 100 stocks. Our approach surpasses the baseline Range Break-out strategy in the Bitcoin market, yielding a remarkable 24.8% increase in the win ratio for all Friday trades and an impressive 48.9% surge in short trades specifically on Fridays. Moreover, we conduct rigorous hypothesis testing to establish the statistical significance of these improvements. Our findings underscore considerable potential of our NLP-driven approach in enhancing trading strategies and achieving greater profitability within financial markets.Keywords: quantitative finance, technical analysis, bitcoin market, NLP, language models, FinBERT, technical trading
Procedia PDF Downloads 7526084 MPC of Single Phase Inverter for PV System
Authors: Irtaza M. Syed, Kaamran Raahemifar
Abstract:
This paper presents a model predictive control (MPC) of a utility interactive (UI) single phase inverter (SPI) for a photovoltaic (PV) system at residential/distribution level. The proposed model uses single-phase phase locked loop (PLL) to synchronize SPI with the grid and performs MPC control in a dq reference frame. SPI model consists of boost converter (BC), maximum power point tracking (MPPT) control, and a full bridge (FB) voltage source inverter (VSI). No PI regulators to tune and carrier and modulating waves are required to produce switching sequence. Instead, the operational model of VSI is used to synthesize sinusoidal current and track the reference. Model is validated using a three kW PV system at the input of UI-SPI in Matlab/Simulink. Implementation and results demonstrate simplicity and accuracy, as well as reliability of the model.Keywords: phase locked loop, voltage source inverter, single phase inverter, model predictive control, Matlab/Simulink
Procedia PDF Downloads 53226083 Seismic Behavior of Pile-Supported Bridges Considering Soil-Structure Interaction and Structural Non-Linearity
Authors: Muhammad Tariq A. Chaudhary
Abstract:
Soil-structure interaction (SSI) in bridges under seismic excitation is a complex phenomenon which involves coupling between the non-linear behavior of bridge pier columns and SSI in the soil-foundation part. It is a common practice in the study of SSI to model the bridge piers as linear elastic while treating the soil and foundation with a non-linear or an equivalent linear modeling approach. Consequently, the contribution of soil and foundation to the SSI phenomenon is disproportionately highlighted. The present study considered non-linear behavior of bridge piers in FEM model of a 4-span, pile-supported bridge that was designed for five different soil conditions in a moderate seismic zone. The FEM model of the bridge system was subjected to a suite of 21 actual ground motions representative of three levels of earthquake hazard (i.e. Design Basis Earthquake, Functional Evaluation Earthquake and Maximum Considered Earthquake). Results of the FEM analysis were used to delineate the influence of pier column non-linearity and SSI on critical design parameters of the bridge system. It was found that pier column non-linearity influenced the bridge lateral displacement and base shear more than SSI for majority of the analysis cases for the class of bridge investigated in the study.Keywords: bridge, FEM model, reinforced concrete pier, pile foundation, seismic loading, soil-structure interaction
Procedia PDF Downloads 23226082 An Infinite Mixture Model for Modelling Stutter Ratio in Forensic Data Analysis
Authors: M. A. C. S. Sampath Fernando, James M. Curran, Renate Meyer
Abstract:
Forensic DNA analysis has received much attention over the last three decades, due to its incredible usefulness in human identification. The statistical interpretation of DNA evidence is recognised as one of the most mature fields in forensic science. Peak heights in an Electropherogram (EPG) are approximately proportional to the amount of template DNA in the original sample being tested. A stutter is a minor peak in an EPG, which is not masking as an allele of a potential contributor, and considered as an artefact that is presumed to be arisen due to miscopying or slippage during the PCR. Stutter peaks are mostly analysed in terms of stutter ratio that is calculated relative to the corresponding parent allele height. Analysis of mixture profiles has always been problematic in evidence interpretation, especially with the presence of PCR artefacts like stutters. Unlike binary and semi-continuous models; continuous models assign a probability (as a continuous weight) for each possible genotype combination, and significantly enhances the use of continuous peak height information resulting in more efficient reliable interpretations. Therefore, the presence of a sound methodology to distinguish between stutters and real alleles is essential for the accuracy of the interpretation. Sensibly, any such method has to be able to focus on modelling stutter peaks. Bayesian nonparametric methods provide increased flexibility in applied statistical modelling. Mixture models are frequently employed as fundamental data analysis tools in clustering and classification of data and assume unidentified heterogeneous sources for data. In model-based clustering, each unknown source is reflected by a cluster, and the clusters are modelled using parametric models. Specifying the number of components in finite mixture models, however, is practically difficult even though the calculations are relatively simple. Infinite mixture models, in contrast, do not require the user to specify the number of components. Instead, a Dirichlet process, which is an infinite-dimensional generalization of the Dirichlet distribution, is used to deal with the problem of a number of components. Chinese restaurant process (CRP), Stick-breaking process and Pólya urn scheme are frequently used as Dirichlet priors in Bayesian mixture models. In this study, we illustrate an infinite mixture of simple linear regression models for modelling stutter ratio and introduce some modifications to overcome weaknesses associated with CRP.Keywords: Chinese restaurant process, Dirichlet prior, infinite mixture model, PCR stutter
Procedia PDF Downloads 33026081 Core Competence Development while Carrying out Organizational Changes
Authors: Olga A. Shvetsova
Abstract:
The paper contains the different issues of competence management in industrial companies. The theoretical bases of human resources management and practical issues of innovative enterprises’ competitiveness are considered. The research is focused on the modern industrial enterprise changes management problems; it focuses on the effective personnel management of industrial enterprises on the basis of competence approach. The influence of organizational changes on the competence development is discussed. The need for development of the new technologies is mentioned, proposal is based on competence-based approach in personnel management including in the conditions of carrying out organizational changes; methods of acquisition and development of missing key professional competences are discussed; importance of key competencies in forming competitive advantage of the organization is mentioned.Keywords: competence model, core competencies, development of industrial company, organizational changes, competitiveness
Procedia PDF Downloads 30326080 Explaining E-Learning Systems Usage in Higher Education Institutions: UTAUT Model
Authors: Muneer Abbad
Abstract:
This research explains the e-learning usage in a university in Jordan. Unified theory of acceptance and use of technology (UTAUT) model has been used as a base model to explain the usage. UTAUT is a model of individual acceptance that is compiled mainly from different models of technology acceptance. This research is the initial part from full explanations of the users' acceptance model that use Structural Equation Modelling (SEM) method to explain the users' acceptance of the e-learning systems based on UTAUT model. In this part data has been collected and prepared for further analysis. The main factors of UTAUT model has been tested as different factors using exploratory factor analysis (EFA). The second phase will be confirmatory factor analysis (CFA) and SEM to explain the users' acceptance of e-learning systems.Keywords: e-learning, moodle, adoption, Unified Theory of Acceptance and Use of Technology (UTAUT)
Procedia PDF Downloads 40726079 Levy Model for Commodity Pricing
Authors: V. Benedico, C. Anacleto, A. Bearzi, L. Brice, V. Delahaye
Abstract:
The aim in present paper is to construct an affordable and reliable commodity prices based on a recalculation of its cost through time which allows visualize the potential risks and thus, take more appropriate decisions regarding forecasts. Here attention has been focused on Levy model, more reliable and realistic than classical random Gaussian one as it takes into consideration observed abrupt jumps in case of sudden price variation. In application to Energy Trading sector where it has never been used before, equations corresponding to Levy model have been written for electricity pricing in European market. Parameters have been set in order to predict and simulate the price and its evolution through time to remarkable accuracy. As predicted by Levy model, the results show significant spikes which reach unconventional levels contrary to currently used Brownian model.Keywords: commodity pricing, Lévy Model, price spikes, electricity market
Procedia PDF Downloads 42926078 A Framework for Auditing Multilevel Models Using Explainability Methods
Authors: Debarati Bhaumik, Diptish Dey
Abstract:
Multilevel models, increasingly deployed in industries such as insurance, food production, and entertainment within functions such as marketing and supply chain management, need to be transparent and ethical. Applications usually result in binary classification within groups or hierarchies based on a set of input features. Using open-source datasets, we demonstrate that popular explainability methods, such as SHAP and LIME, consistently underperform inaccuracy when interpreting these models. They fail to predict the order of feature importance, the magnitudes, and occasionally even the nature of the feature contribution (negative versus positive contribution to the outcome). Besides accuracy, the computational intractability of SHAP for binomial classification is a cause of concern. For transparent and ethical applications of these hierarchical statistical models, sound audit frameworks need to be developed. In this paper, we propose an audit framework for technical assessment of multilevel regression models focusing on three aspects: (i) model assumptions & statistical properties, (ii) model transparency using different explainability methods, and (iii) discrimination assessment. To this end, we undertake a quantitative approach and compare intrinsic model methods with SHAP and LIME. The framework comprises a shortlist of KPIs, such as PoCE (Percentage of Correct Explanations) and MDG (Mean Discriminatory Gap) per feature, for each of these three aspects. A traffic light risk assessment method is furthermore coupled to these KPIs. The audit framework will assist regulatory bodies in performing conformity assessments of AI systems using multilevel binomial classification models at businesses. It will also benefit businesses deploying multilevel models to be future-proof and aligned with the European Commission’s proposed Regulation on Artificial Intelligence.Keywords: audit, multilevel model, model transparency, model explainability, discrimination, ethics
Procedia PDF Downloads 9426077 A Multigranular Linguistic ARAS Model in Group Decision Making
Authors: Wiem Daoud Ben Amor, Luis Martínez López, Hela Moalla Frikha
Abstract:
Most of the multi-criteria group decision making (MCGDM) problems dealing with qualitative criteria require consideration of the large background of expert information. It is common that experts have different degrees of knowledge for giving their alternative assessments according to criteria. So, it seems logical that they use different evaluation scales to express their judgment, i.e., multi granular linguistic scales. In this context, we propose the extension of the classical additive ratio assessment (ARAS) method to the case of a hierarchical linguistics term for managing multi granular linguistic scales in uncertain contexts where uncertainty is modeled by means in linguistic information. The proposed approach is called the extended hierarchical linguistics-ARAS method (ARAS-ELH). Within the ARAS-ELH approach, the DM can diagnose the results (the ranking of the alternatives) in a decomposed style, i.e., not only at one level of the hierarchy but also at the intermediate ones. Also, the developed approach allows a feedback transformation i.e the collective final results of all experts able to be transformed at any level of the extended linguistic hierarchy that each expert has previously used. Therefore, the ARAS-ELH technique makes it easier for decision-makers to understand the results. Finally, An MCGDM case study is given to illustrate the proposed approach.Keywords: additive ratio assessment, extended hierarchical linguistic, multi-criteria group decision making problems, multi granular linguistic contexts
Procedia PDF Downloads 20626076 Finite Element Simulation of Embankment Bumps at Bridge Approaches, Comparison Study
Authors: F. A. Hassona, M. D. Hashem, R. I. Melek, B. M. Hakeem
Abstract:
A differential settlement at the end of a bridge near the interface between the abutment and the embankment is a persistent problem for highway agencies. The differential settlement produces the common ‘bump at the end of the bridge’. Reduction in steering response, distraction to the driver, added risk and expense to maintenance operation, and reduction in a transportation agency’s public image are all undesirable effects of these uneven and irregular transitions. This paper attempts to simulate the bump at the end of the bridge using PLAXIS finite element 2D program. PLAXIS was used to simulate a laboratory model called Bridge to Embankment Simulator of Transition (B.E.S.T.) device which was built by others to investigate this problem. A total of six numerical simulations were conducted using hardening- soil model with rational assumptions of missing soil parameters to estimate the bump at the end of the bridge. The results show good agreements between the numerical and the laboratory models. Important factors influencing bumps at bridge ends were also addressed in light of the model results.Keywords: bridge approach slabs, bridge bump, hardening-soil, PLAXIS 2D, settlement
Procedia PDF Downloads 34826075 Organizational Commitment in Islamic Boarding School: The Implementation of Organizational Behavior Integrative Model
Authors: Siswoyo Haryono
Abstract:
Purpose – The fundamental goal of this research is to see if the integrative organizational behavior model can be used effectively in Islamic boarding schools. This paper also seeks to assess the effect of Islamic organizational culture, leadership, and spiritual intelligence on teachers' organizational commitment to Islamic Boarding schools. The goal of the mediation analysis is to see if the Islamic work ethic has a more significant effect on the instructors' organizational commitment than the direct effects of Islamic organizational culture, leadership, and Islamic spiritual intelligence. Design/methodology/approach – A questionnaire survey was used to obtain data from teachers at Islamic Boarding Schools. This study used the AMOS technique for structural equation modeling to evaluate the expected direct effect. To test the hypothesized indirect effect, employed Sobel test. Findings – Islamic organizational culture, Islamic leadership, and Islamic spiritual intelligence significantly affect Islamic work ethic. When it comes to Islamic corporate culture, Islamic leadership, Islamic spiritual intelligence, and Islamic work ethics have a significant impact. The findings of the mediation study reveal that Islamic organizational culture, leadership, and spiritual intelligence influences organizational commitment through Islamic work ethic. The total effect analysis shows that the most effective path to increasing teachers’ organizational commitment is Islamic leadership - Islamic work ethic – organizational commitment. Originality/value – This study evaluates the Integrative Model of Organizational Behavior by Colquitt (2016) applied in Islamic Boarding School. The model consists of contemporary leadership and individual characteristic as the antecedent. The mediating variables of the model consist of individual mechanisms such as trust, justice, and ethic. Individual performance and organizational commitment are the model's outcomes. These variables, on the other hand, do not represent the Islamic viewpoint as a whole. As a result, this study aims to assess the role of Islamic principles in the model. The study employs reliability and validity tests to get reliable and valid measures. The findings revealed that the evaluation model is proven to improve organizational commitment at Islamic Boarding School.Keywords: Islamic leadership, Islamic spiritual intelligence, Islamic work ethic, organizational commitment, Islamic boarding school
Procedia PDF Downloads 16126074 Finding DEA Targets Using Multi-Objective Programming
Authors: Farzad Sharifi, Raziyeh Shamsi
Abstract:
In this paper, we obtain the projection of inefficient units in data envelopment analysis (DEA) in the case of stochastic inputs and outputs using the multi-objective programming (MOP) structure. In some problems, the inputs might be stochastic while the outputs are deterministic, and vice versa. In such cases, we propose molti-objective DEA-R model, because in some cases (e.g., when unnecessary and irrational weights by the BCC model reduces the efficiency score), an efficient DMU is introduced as inefficient by the BCC model, whereas the DMU is considered efficient by the DEA-R model. In some other case, only the ratio of stochastic data may be available (e.g; the ratio of stochastic inputs to stochastic outputs). Thus, we provide multi objective DEA model without explicit outputs and prove that in-put oriented MOP DEA-R model in the invariable return to scale case can be replacing by MOP- DEA model without explicit outputs in the variable return to scale and vice versa. Using the interactive methods for solving the proposed model, yields a projection corresponding to the viewpoint of the DM and the analyst, which is nearer to reality and more practical. Finally, an application is provided.Keywords: DEA, MOLP, STOCHASTIC, DEA-R
Procedia PDF Downloads 39826073 Model Predictive Controller for Pasteurization Process
Authors: Tesfaye Alamirew Dessie
Abstract:
Our study focuses on developing a Model Predictive Controller (MPC) and evaluating it against a traditional PID for a pasteurization process. Utilizing system identification from the experimental data, the dynamics of the pasteurization process were calculated. Using best fit with data validation, residual, and stability analysis, the quality of several model architectures was evaluated. The validation data fit the auto-regressive with exogenous input (ARX322) model of the pasteurization process by roughly 80.37 percent. The ARX322 model structure was used to create MPC and PID control techniques. After comparing controller performance based on settling time, overshoot percentage, and stability analysis, it was found that MPC controllers outperform PID for those parameters.Keywords: MPC, PID, ARX, pasteurization
Procedia PDF Downloads 16326072 Numerical Investigation into Capture Efficiency of Fibrous Filters
Authors: Jayotpaul Chaudhuri, Lutz Goedeke, Torsten Hallenga, Peter Ehrhard
Abstract:
Purification of gases from aerosols or airborne particles via filters is widely applied in the industry and in our daily lives. This separation especially in the micron and submicron size range is a necessary step to protect the environment and human health. Fibrous filters are often employed due to their low cost and high efficiency. For designing any filter the two most important performance parameters are capture efficiency and pressure drop. Since the capture efficiency is directly proportional to the pressure drop which leads to higher operating costs, a detailed investigation of the separation mechanism is required to optimize the filter designing, i.e., to have a high capture efficiency with a lower pressure drop. Therefore a two-dimensional flow simulation around a single fiber using Ansys CFX and Matlab is used to get insight into the separation process. Instead of simulating a solid fiber, the present Ansys CFX model uses a fictitious domain approach for the fiber by implementing a momentum loss model. This approach has been chosen to avoid creating a new mesh for different fiber sizes, thereby saving time and effort for re-meshing. In a first step, only the flow of the continuous fluid around the fiber is simulated in Ansys CFX and the flow field data is extracted and imported into Matlab and the particle trajectory is calculated in a Matlab routine. This calculation is a Lagrangian, one way coupled approach for particles with all relevant forces acting on it. The key parameters for the simulation in both Ansys CFX and Matlab are the porosity ε, the diameter ratio of particle and fiber D, the fluid Reynolds number Re, the Reynolds particle number Rep, the Stokes number St, the Froude number Fr and the density ratio of fluid and particle ρf/ρp. The simulation results were then compared to the single fiber theory from the literature.Keywords: BBO-equation, capture efficiency, CFX, Matlab, fibrous filter, particle trajectory
Procedia PDF Downloads 20626071 Nonlinear Mathematical Model of the Rotor Motion in a Thin Hydrodynamic Gap
Authors: Jaroslav Krutil, Simona Fialová, , František Pochylý
Abstract:
A nonlinear mathematical model of mutual fluid-structure interaction is presented in the work. The model is applicable to the general shape of sealing gaps. An in compressible fluid and turbulent flow is assumed. The shaft carries a rotational and procession motion, the gap is axially flowed through. The achieved results of the additional mass, damping and stiffness matrices may be used in the solution of the rotor dynamics. The usage of this mathematical model is expected particularly in hydraulic machines. The method of control volumes in the ANSYS Fluent was used for the simulation. The obtained results of the pressure and velocity fields are used in the mathematical model of additional effects.Keywords: nonlinear mathematical model, CFD modeling, hydrodynamic sealing gap, matrices of mass, stiffness, damping
Procedia PDF Downloads 53526070 Image Ranking to Assist Object Labeling for Training Detection Models
Authors: Tonislav Ivanov, Oleksii Nedashkivskyi, Denis Babeshko, Vadim Pinskiy, Matthew Putman
Abstract:
Training a machine learning model for object detection that generalizes well is known to benefit from a training dataset with diverse examples. However, training datasets usually contain many repeats of common examples of a class and lack rarely seen examples. This is due to the process commonly used during human annotation where a person would proceed sequentially through a list of images labeling a sufficiently high total number of examples. Instead, the method presented involves an active process where, after the initial labeling of several images is completed, the next subset of images for labeling is selected by an algorithm. This process of algorithmic image selection and manual labeling continues in an iterative fashion. The algorithm used for the image selection is a deep learning algorithm, based on the U-shaped architecture, which quantifies the presence of unseen data in each image in order to find images that contain the most novel examples. Moreover, the location of the unseen data in each image is highlighted, aiding the labeler in spotting these examples. Experiments performed using semiconductor wafer data show that labeling a subset of the data, curated by this algorithm, resulted in a model with a better performance than a model produced from sequentially labeling the same amount of data. Also, similar performance is achieved compared to a model trained on exhaustive labeling of the whole dataset. Overall, the proposed approach results in a dataset that has a diverse set of examples per class as well as more balanced classes, which proves beneficial when training a deep learning model.Keywords: computer vision, deep learning, object detection, semiconductor
Procedia PDF Downloads 13626069 Exploring the Possibility of Islamic Banking as a Viable Alternative to the Conventional Banking Model
Authors: Lavan Vickneson
Abstract:
In today’s modern economy, the conventional banking model is the primary banking system used around the world. A significant problem faced by the conventional banking model is the recurring nature of banking crises. History’s record of the various banking crises, ranging from the Great Depression to the 2008 subprime mortgage crisis, is testament to the fact that banking crises continue to strike despite the preventive measures in place, such as bank’s minimum capital requirements and deposit guarantee schemes. If banking crises continue to occur despite these preventive measures, it necessarily follows that there are inherent flaws with the conventional banking model itself. In light of this, a possible alternative banking model to the conventional banking model is Islamic banking. To date, Islamic banking has been a niche market, predominantly serving Muslim investors. This paper seeks to explore the possibility of Islamic banking being more than just a niche market and playing a greater role in banking sectors around the world, by being a viable alternative to the conventional banking model.Keywords: bank crises, conventional banking model, Islamic banking, niche market
Procedia PDF Downloads 28226068 Solid Particles Transport and Deposition Prediction in a Turbulent Impinging Jet Using the Lattice Boltzmann Method and a Probabilistic Model on GPU
Authors: Ali Abdul Kadhim, Fue Lien
Abstract:
Solid particle distribution on an impingement surface has been simulated utilizing a graphical processing unit (GPU). In-house computational fluid dynamics (CFD) code has been developed to investigate a 3D turbulent impinging jet using the lattice Boltzmann method (LBM) in conjunction with large eddy simulation (LES) and the multiple relaxation time (MRT) models. This paper proposed an improvement in the LBM-cellular automata (LBM-CA) probabilistic method. In the current model, the fluid flow utilizes the D3Q19 lattice, while the particle model employs the D3Q27 lattice. The particle numbers are defined at the same regular LBM nodes, and transport of particles from one node to its neighboring nodes are determined in accordance with the particle bulk density and velocity by considering all the external forces. The previous models distribute particles at each time step without considering the local velocity and the number of particles at each node. The present model overcomes the deficiencies of the previous LBM-CA models and, therefore, can better capture the dynamic interaction between particles and the surrounding turbulent flow field. Despite the increasing popularity of LBM-MRT-CA model in simulating complex multiphase fluid flows, this approach is still expensive in term of memory size and computational time required to perform 3D simulations. To improve the throughput of each simulation, a single GeForce GTX TITAN X GPU is used in the present work. The CUDA parallel programming platform and the CuRAND library are utilized to form an efficient LBM-CA algorithm. The methodology was first validated against a benchmark test case involving particle deposition on a square cylinder confined in a duct. The flow was unsteady and laminar at Re=200 (Re is the Reynolds number), and simulations were conducted for different Stokes numbers. The present LBM solutions agree well with other results available in the open literature. The GPU code was then used to simulate the particle transport and deposition in a turbulent impinging jet at Re=10,000. The simulations were conducted for L/D=2,4 and 6, where L is the nozzle-to-surface distance and D is the jet diameter. The effect of changing the Stokes number on the particle deposition profile was studied at different L/D ratios. For comparative studies, another in-house serial CPU code was also developed, coupling LBM with the classical Lagrangian particle dispersion model. Agreement between results obtained with LBM-CA and LBM-Lagrangian models and the experimental data is generally good. The present GPU approach achieves a speedup ratio of about 350 against the serial code running on a single CPU.Keywords: CUDA, GPU parallel programming, LES, lattice Boltzmann method, MRT, multi-phase flow, probabilistic model
Procedia PDF Downloads 20726067 A System Dynamics Approach to Technological Learning Impact for Cost Estimation of Solar Photovoltaics
Authors: Rong Wang, Sandra Hasanefendic, Elizabeth von Hauff, Bart Bossink
Abstract:
Technological learning and learning curve models have been continuously used to estimate the photovoltaics (PV) cost development over time for the climate mitigation targets. They can integrate a number of technological learning sources which influence the learning process. Yet the accuracy and realistic predictions for cost estimations of PV development are still difficult to achieve. This paper develops four hypothetical-alternative learning curve models by proposing different combinations of technological learning sources, including both local and global technology experience and the knowledge stock. This paper specifically focuses on the non-linear relationship between the costs and technological learning source and their dynamic interaction and uses the system dynamics approach to predict a more accurate PV cost estimation for future development. As the case study, the data from China is gathered and drawn to illustrate that the learning curve model that incorporates both the global and local experience is more accurate and realistic than the other three models for PV cost estimation. Further, absorbing and integrating the global experience into the local industry has a positive impact on PV cost reduction. Although the learning curve model incorporating knowledge stock is not realistic for current PV cost deployment in China, it still plays an effective positive role in future PV cost reduction.Keywords: photovoltaic, system dynamics, technological learning, learning curve
Procedia PDF Downloads 9626066 An E-Government Implementation Model for Peruvian State Companies Based on COBIT 5.0: Definition and Goals of the Model
Authors: M. Bruzza, M. Tupia, F. Rodríguez
Abstract:
As part of the regulatory compliance process and the streamlining of public administration, the Peruvian government has implemented the National E-Government Plan in all state institutions with the aim of providing citizens with solid services based on the use of Information and Communications Technologies (ICT). As part of the regulations, the requisites to be met by public institutions have been submitted. However, the lack of an implementation model was detected, one that can serve as a guide to such institutions in order to materialize the organizational and technological structures needed, which allow them to provide the required digital services. This paper develops an implementation model of electronic government (e-government) for Peru’s state institutions, in compliance with current regulations based on a COBIT 5.0 framework. Furthermore, the paper introduces phase 1 of this model: business and IT goals, the goals cascade and the future model of processes.Keywords: e-government, u-government, COBIT, implementation model
Procedia PDF Downloads 32626065 Adaptive Programming for Indigenous Early Learning: The Early Years Model
Authors: Rachel Buchanan, Rebecca LaRiviere
Abstract:
Context: The ongoing effects of colonialism continue to be experienced through paternalistic policies and funding processes that cause disjuncture between and across Indigenous early childhood programming on-reserve and in urban and Northern settings in Canada. While various educational organizations and social service providers have risen to address these challenges in the short, medium and long term, there continues to be a lack in nation-wide cohesive, culturally grounded, and meaningful early learning programming for Indigenous children in Canada. Indigenous-centered early learning programs tend to face one of two scaling dilemmas: their program goals are too prescriptive to enable the program to be meaningfully replicated in different cultural/ community settings, or their program goals are too broad to be meaningfully adapted to the unique cultural and contextual needs and desires of Indigenous communities (the “franchise approach”). There are over 600 First Nations communities in Canada representing more than 50 Nations and languages. Consequently, Indigenous early learning programming cannot be applied with a universal or “one size fits all” approach. Sustainable and comprehensive programming must be responsive to each community context, building upon existing strengths and assets to avoid program duplication and irrelevance. Thesis: Community-driven and culturally adapted early childhood programming is critical but cannot be achieved on a large scale within traditional program models that are constrained by prescriptive overarching program goals. Principles, rather than goals, are an effective way to navigate and evaluate complex and dynamic systems. Principles guide an intervention to be adaptable, flexible and scalable. The Martin Family Initiative (MFI) ’s Early Years program engages a principles-based approach to programming. As will be discussed in this paper, this approach enables the program to catalyze existing community-based strengths and organizational assets toward bridging gaps across and disjuncture between Indigenous early learning programs, as well as to scale programming in sustainable, context-responsive and dynamic ways. This paper argues that using a principles-driven and adaptive scaling approach, the Early Years model establishes important learnings for culturally adapted Indigenous early learning programming in Canada. Methodology: The Early Years has leveraged this approach to develop an array of programming with partner organizations and communities across the country. The Early Years began as a singular pilot project in one First Nation. In just three years, it has expanded to five different regions and community organizations. In each context, the program supports the partner organization through different means and to different ends, the extent to which is determined in partnership with each community-based organization: in some cases, this means supporting the organization to build home visiting programming from the ground-up; in others, it means offering organization-specific culturally adapted early learning resources to support the programming that already exists in communities. Principles underpin but do not define the practices of the program in each of these relationships. This paper will explore numerous examples of principles-based adaptability with the context of the Early Years, concluding that the program model offers theadaptability and dynamism necessary to respond to unique and ever-evolving community contexts and needs of Indigenous children today.Keywords: culturally adapted programming, indigenous early learning, principles-based approach, program scaling
Procedia PDF Downloads 186