Search results for: unified theory of acceptance and use of technology (UTAUT) model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26043

Search results for: unified theory of acceptance and use of technology (UTAUT) model

23643 Blockchain Technology in Supply Chain Management: A Systematic Review And Meta-Analysis

Authors: Mohammad Yousuf Khan, Bhavya Alankar

Abstract:

Blockchain is a promising technology with its features such as immutability and decentralized database. It has applications in various fields such as pharmaceutical, finance, & the food industry. At the core of its heart lies its feature, traceability which is the most desired key in supply chains. However, supply chains have always been hit rock bottom by scandals and controversies. In this review paper, we have explored the advancement and research gaps of blockchain technology (BT) in supply chain management (SCM). We have used the Prisma framework for systematic literature review (SLR) and included a minuscule amount of grey literature to reduce publication bias. We found that supply chain traceability and transparency is the most researched objective in SCM. There was hardly any research in supply chain resilience. Further, we found that 40 % of the papers were application based. Most articles have focused on the advantages of BT, rather than analyzing it critically. This study will help identify gaps and suitable actions to be followed for an efficient implementation of BT in SCM.

Keywords: blockchain technology, supply chain management, supply chain transparency, supply chain resilience

Procedia PDF Downloads 147
23642 Simulink Library for Reference Current Generation in Active DC Traction Substations

Authors: Mihaela Popescu, Alexandru Bitoleanu

Abstract:

This paper is focused on the reference current calculation in the compensation mode of the active DC traction substations. The so-called p-q theory of the instantaneous reactive power is used as theoretical foundation. The compensation goal of total compensation is taken into consideration for the operation under both sinusoidal and nonsinusoidal voltage conditions, through the two objectives of unity power factor and perfect harmonic cancelation. Four blocks of reference current generation implement the conceived algorithms and they are included in a specific Simulink library, which is useful in a DSP dSPACE-based platform working under Matlab/Simulink. The simulation results validate the correctness of the implementation and fulfillment of the compensation tasks.

Keywords: active power filter, DC traction, p-q theory, Simulink library

Procedia PDF Downloads 656
23641 Efficient Elimination of Common Allergens through the Application of Dry Microfine Steam on Innate Surfaces

Authors: O. Rachinel, C. Recchia, M. Bourel, B. Recchia

Abstract:

Dry microfine steam (DMS) technology, developed by Laurastar, was shown to effectively eliminate a range of pathogens such as Sars-CoV-2, E. coli, S. aureus and C. Albicans. The aim of this study was to investigate the effect of DMS technology on allergens. Therefore, the application of the DMS technology was tested on two common allergens (Dermatophagoides pteronyssinus and cat allergen Fel d 1), on different inert surfaces (e.g., cotton), during 2 to 3 seconds. Quantification of the remaining allergens was performed and the reduction rates reached 100% in 3 seconds for D. pteronyssinus and 97,74% in 2 seconds for cat allergens. In conclusion, DMS showed high efficacy in the elimination of common allergens and could be seen as a natural solution to improve domestic hygiene and reduce allergies.

Keywords: steam, allergens, dust mites, pollens

Procedia PDF Downloads 125
23640 OpenFOAM Based Simulation of High Reynolds Number Separated Flows Using Bridging Method of Turbulence

Authors: Sagar Saroha, Sawan S. Sinha, Sunil Lakshmipathy

Abstract:

Reynolds averaged Navier-Stokes (RANS) model is the popular computational tool for prediction of turbulent flows. Being computationally less expensive as compared to direct numerical simulation (DNS), RANS has received wide acceptance in industry and research community as well. However, for high Reynolds number flows, the traditional RANS approach based on the Boussinesq hypothesis is incapacitated to capture all the essential flow characteristics, and thus, its performance is restricted in high Reynolds number flows of practical interest. RANS performance turns out to be inadequate in regimes like flow over curved surfaces, flows with rapid changes in the mean strain rate, duct flows involving secondary streamlines and three-dimensional separated flows. In the recent decade, partially averaged Navier-Stokes (PANS) methodology has gained acceptability among seamless bridging methods of turbulence- placed between DNS and RANS. PANS methodology, being a scale resolving bridging method, is inherently more suitable than RANS for simulating turbulent flows. The superior ability of PANS method has been demonstrated for some cases like swirling flows, high-speed mixing environment, and high Reynolds number turbulent flows. In our work, we intend to evaluate PANS in case of separated turbulent flows past bluff bodies -which is of broad aerodynamic research and industrial application. PANS equations, being derived from base RANS, continue to inherit the inadequacies from the parent RANS model based on linear eddy-viscosity model (LEVM) closure. To enhance PANS’ capabilities for simulating separated flows, the shortcomings of the LEVM closure need to be addressed. Inabilities of the LEVMs have inspired the development of non-linear eddy viscosity models (NLEVM). To explore the potential improvement in PANS performance, in our study we evaluate the PANS behavior in conjugation with NLEVM. Our work can be categorized into three significant steps: (i) Extraction of PANS version of NLEVM from RANS model, (ii) testing the model in the homogeneous turbulence environment and (iii) application and evaluation of the model in the canonical case of separated non-homogeneous flow field (flow past prismatic bodies and bodies of revolution at high Reynolds number). PANS version of NLEVM shall be derived and implemented in OpenFOAM -an open source solver. Homogeneous flows evaluation will comprise the study of the influence of the PANS’ filter-width control parameter on the turbulent stresses; the homogeneous analysis performed over typical velocity fields and asymptotic analysis of Reynolds stress tensor. Non-homogeneous flow case will include the study of mean integrated quantities and various instantaneous flow field features including wake structures. Performance of PANS + NLEVM shall be compared against the LEVM based PANS and LEVM based RANS. This assessment will contribute to significant improvement of the predictive ability of the computational fluid dynamics (CFD) tools in massively separated turbulent flows past bluff bodies.

Keywords: bridging methods of turbulence, high Re-CFD, non-linear PANS, separated turbulent flows

Procedia PDF Downloads 136
23639 First Principle study of Electronic Structure of Silicene Doped with Galium

Authors: Mauludi Ariesto Pamungkas, Wafa Maftuhin

Abstract:

Gallium with three outer electrons commonly are used as dopants of silicon to make it P type and N type semiconductor respectively. Silicene, one-atom-thick silicon layer is one of emerging two dimension materials after the success of graphene. The effects of Gallium doping on electronic structure of silicine are investigated by using first principle calculation based on Density Functional Theory (DFT) calculation and norm conserving pseudopotential method implemented in ABINIT code. Bandstructure of Pristine silicene is similar to that of graphene. Effect of Ga doping on bandstructure of silicene depend on the position of Ga adatom on silicene

Keywords: silicene, effects of Gallium doping, Density Functional Theory (DFT), graphene

Procedia PDF Downloads 420
23638 The Potential Threat of Cyberterrorism to the National Security: Theoretical Framework

Authors: Abdulrahman S. Alqahtani

Abstract:

The revolution of computing and networks could revolutionise terrorism in the same way that it has brought about changes in other aspects of life. The modern technological era has faced countries with a new set of security challenges. There are many states and potential adversaries who have the potential and capacity in cyberspace, which makes them able to carry out cyber-attacks in the future. Some of them are currently conducting surveillance, gathering and analysis of technical information, and mapping of networks and nodes and infrastructure of opponents, which may be exploited in future conflicts. This poster presents the results of the quantitative study (survey) to test the validity of the proposed theoretical framework for the cyber terrorist threats. This theoretical framework will help to in-depth understand these new digital terrorist threats. It may also be a practical guide for managers and technicians in critical infrastructure, to understand and assess the threats they face. It might also be the foundation for building a national strategy to counter cyberterrorism. In the beginning, it provides basic information about the data. To purify the data, reliability and exploratory factor analysis, as well as confirmatory factor analysis (CFA) were performed. Then, Structural Equation Modelling (SEM) was utilised to test the final model of the theory and to assess the overall goodness-of-fit between the proposed model and the collected data set.

Keywords: cyberterrorism, critical infrastructure, , national security, theoretical framework, terrorism

Procedia PDF Downloads 384
23637 Background Knowledge and Reading Comprehension in ELT Classes: A Pedagogical Perspective

Authors: Davoud Ansari Kejal, Meysam Sabour

Abstract:

For long, there has been a belief that a reader can easily comprehend a text if he is strong enough in vocabulary and grammatical knowledge but there was no account for the ability of understanding different subjects based on readers’ understanding of the surrounding world which is called world background knowledge. This paper attempts to investigate the reading comprehension process applying the schema theory as an influential factor in comprehending texts, in order to prove the important role of background knowledge in reading comprehension. Based on the discussion, some teaching methods are suggested for employing world background knowledge for an elaborated teaching of reading comprehension in an active learning environment in EFL classes.

Keywords: background knowledge, reading comprehension, schema theory, ELT classes

Procedia PDF Downloads 437
23636 Rheological Model for Describing Spunlace Nonwoven Behavior

Authors: Sana Ridene, Soumaya Sayeb, Houda Helali, Mohammed Ben Hassen

Abstract:

Nonwoven structures have a range of applications which include Medical, filtration, geotextile and recently this unconventional fabric is finding a niche in fashion apparel. In this paper, a modified form of Vangheluwe rheological model is used to describe the mechanical behavior of nonwovens fabrics in uniaxial tension. This model is an association in parallel of three Maxwell elements characterized by damping coefficients η1, η2 and η3 and E1, E2, E3 elastic modulus and a nonlinear spring C. The model is verified experimentally with two types of nonwovens (50% viscose /50% Polyester) and (40% viscose/60% Polyester) and a range of three square weights values. Comparative analysis of the theoretical model and the experimental results of tensile test proofs a high correlation between them. The proposed model can fairly well replicate the behavior of nonwoven fabrics during relaxation and sample traction. This allowed us to predict the mechanical behavior in tension and relaxation of fabrics starting only from their technical parameters (composition and weight).

Keywords: mechanical behavior, tensile strength, relaxation, rheological model

Procedia PDF Downloads 391
23635 Thermal Modelling and Experimental Comparison for a Moving Pantograph Strip

Authors: Nicolas Delcey, Philippe Baucour, Didier Chamagne, Geneviève Wimmer, Auditeau Gérard, Bausseron Thomas, Bouger Odile, Blanvillain Gérard

Abstract:

This paper proposes a thermal study of the catenary/pantograph interface for a train in motion. A 2.5D complex model of the pantograph strip has been defined and created by a coupling between a 1D and a 2D model. Experimental and simulation results are presented and with a comparison allow validating the 2.5D model. Some physical phenomena are described and presented with the help of the model such as the stagger motion thermal effect, particular heats and the effect of the material characteristics. Finally it is possible to predict the critical thermal configuration during a train trip.

Keywords: electro-thermal studies, mathematical optimizations, multi-physical approach, numerical model, pantograph strip wear

Procedia PDF Downloads 311
23634 The Influence of Consumer and Brand-Oriented Capabilities on Business Performance in Young Firms: A Quantitative Causal Model Analysis

Authors: Katharina Buttenberg

Abstract:

Customer and brand-oriented capabilities have been identified as key influencing capabilities for business performance. Especially in the early years of the firm, it is crucial to develop and consciously manage these capabilities. In this paper, the results of a quantitative analysis, investigating the causal relationship between customer- and brand-oriented (marketing) capabilities and business performance will be presented. The research displays the dependencies between the constructs and will provide practical implications for young firms in the acquisition and management of these capabilities.

Keywords: brand-oriented capabilities, customer-oriented capabilities, entrepreneurship, resource-based theory, young firms

Procedia PDF Downloads 332
23633 Theory of the Optimum Signal Approximation Clarifying the Importance in the Recognition of Parallel World and Application to Secure Signal Communication with Feedback

Authors: Takuro Kida, Yuichi Kida

Abstract:

In this paper, it is shown a base of the new trend of algorithm mathematically that treats a historical reason of continuous discrimination in the world as well as its solution by introducing new concepts of parallel world that includes an invisible set of errors as its companion. With respect to a matrix operator-filter bank that the matrix operator-analysis-filter bank H and the matrix operator-sampling-filter bank S are given, firstly, we introduce the detail algorithm to derive the optimum matrix operator-synthesis-filter bank Z that minimizes all the worst-case measures of the matrix operator-error-signals E(ω) = F(ω) − Y(ω) between the matrix operator-input-signals F(ω) and the matrix operator-output-signals Y(ω) of the matrix operator-filter bank at the same time. Further, feedback is introduced to the above approximation theory, and it is indicated that introducing conversations with feedback do not superior automatically to the accumulation of existing knowledge of signal prediction. Secondly, the concept of category in the field of mathematics is applied to the above optimum signal approximation and is indicated that the category-based approximation theory is applied to the set-theoretic consideration of the recognition of humans. Based on this discussion, it is shown naturally why the narrow perception that tends to create isolation shows an apparent advantage in the short term and, often, why such narrow thinking becomes intimate with discriminatory action in a human group. Throughout these considerations, it is presented that, in order to abolish easy and intimate discriminatory behavior, it is important to create a parallel world of conception where we share the set of invisible error signals, including the words and the consciousness of both worlds.

Keywords: matrix filterbank, optimum signal approximation, category theory, simultaneous minimization

Procedia PDF Downloads 123
23632 Cellular Automata Using Fractional Integral Model

Authors: Yasser F. Hassan

Abstract:

In this paper, a proposed model of cellular automata is studied by means of fractional integral function. A cellular automaton is a decentralized computing model providing an excellent platform for performing complex computation with the help of only local information. The paper discusses how using fractional integral function for representing cellular automata memory or state. The architecture of computing and learning model will be given and the results of calibrating of approach are also given.

Keywords: fractional integral, cellular automata, memory, learning

Procedia PDF Downloads 400
23631 A Strategic Communication Design Model for Indigenous Knowledge Management

Authors: Dilina Janadith Nawarathne

Abstract:

This article presents the initial development of a communication model (Model_isi) as the means of gathering, preserving and transferring indigenous knowledge in the field of knowledge management. The article first discusses the need for an appropriate complimentary model for indigenous knowledge management which differs from the existing methods and models. Then the paper suggests the newly developed model for indigenous knowledge management which generate as result of blending key aspects of different disciplines, which can be implemented as a complementary approach for the existing scientific method. The paper further presents the effectiveness of the developed method in reflecting upon a pilot demonstration carried out on selected indigenous communities of Sri Lanka.

Keywords: indigenous knowledge management, knowledge transferring, tacit knowledge, research model, asian centric philosophy

Procedia PDF Downloads 464
23630 Molecular Dynamics Simulation for Buckling Analysis at Nanocomposite Beams

Authors: Babak Safaei, A. M. Fattahi

Abstract:

In the present study we have investigated axial buckling characteristics of nanocomposite beams reinforced by single-walled carbon nanotubes (SWCNTs). Various types of beam theories including Euler-Bernoulli beam theory, Timoshenko beam theory and Reddy beam theory were used to analyze the buckling behavior of carbon nanotube-reinforced composite beams. Generalized differential quadrature (GDQ) method was utilized to discretize the governing differential equations along with four commonly used boundary conditions. The material properties of the nanocomposite beams were obtained using molecular dynamic (MD) simulation corresponding to both short-(10,10) SWCNT and long-(10,10) SWCNT composites which were embedded by amorphous polyethylene matrix. Then the results obtained directly from MD simulations were matched with those calculated by the mixture rule to extract appropriate values of carbon nanotube efficiency parameters accounting for the scale-dependent material properties. The selected numerical results were presented to indicate the influences of nanotube volume fractions and end supports on the critical axial buckling loads of nanocomposite beams relevant to long- and short-nanotube composites.

Keywords: nanocomposites, molecular dynamics simulation, axial buckling, generalized differential quadrature (GDQ)

Procedia PDF Downloads 311
23629 Primal Instinct: Formation of Food Aversion

Authors: Zihuan (Dylan) Wang

Abstract:

This paper analyzes the formation of human food aversion from a biological perspective. It points out that this biased behavior is formed through the accumulation of long-term survival and life experiences. By introducing the "Food Chain Energy Pyramid" model and the analogous deduction of the "Human Food Aversion Pyramid," with energy conversion efficiency as the primary reason, it analyzes the underlying reasons for the formation of food preferences. Food industry professionals can gain inspiration from this article to combine the theory presented with their expertise in order to leverage product quality and promote environmentally conscious practices.

Keywords: food aversion, food preference, energy conversion efficiency, food and culture, nutrition, research and development

Procedia PDF Downloads 40
23628 Efficient Frontier: Comparing Different Volatility Estimators

Authors: Tea Poklepović, Zdravka Aljinović, Mario Matković

Abstract:

Modern Portfolio Theory (MPT) according to Markowitz states that investors form mean-variance efficient portfolios which maximizes their utility. Markowitz proposed the standard deviation as a simple measure for portfolio risk and the lower semi-variance as the only risk measure of interest to rational investors. This paper uses a third volatility estimator based on intraday data and compares three efficient frontiers on the Croatian Stock Market. The results show that range-based volatility estimator outperforms both mean-variance and lower semi-variance model.

Keywords: variance, lower semi-variance, range-based volatility, MPT

Procedia PDF Downloads 498
23627 Residual Analysis and Ground Motion Prediction Equation Ranking Metrics for Western Balkan Strong Motion Database

Authors: Manuela Villani, Anila Xhahysa, Christopher Brooks, Marco Pagani

Abstract:

The geological structure of Western Balkans is strongly affected by the collision between Adria microplate and the southwestern Euroasia margin, resulting in a considerably active seismic region. The Harmonization of Seismic Hazard Maps in the Western Balkan Countries Project (BSHAP) (2007-2011, 2012-2015) by NATO supported the preparation of new seismic hazard maps of the Western Balkan, but when inspecting the seismic hazard models produced later by these countries on a national scale, significant differences in design PGA values are observed in the border, for instance, North Albania-Montenegro, South Albania- Greece, etc. Considering the fact that the catalogues were unified and seismic sources were defined within BSHAP framework, obviously, the differences arise from the Ground Motion Prediction Equations selection, which are generally the component with highest impact on the seismic hazard assessment. At the time of the project, a modest database was present, namely 672 three-component records, whereas nowadays, this strong motion database has increased considerably up to 20,939 records with Mw ranging in the interval 3.7-7 and epicentral distance distribution from 0.47km to 490km. Statistical analysis of the strong motion database showed the lack of recordings in the moderate-to-large magnitude and short distance ranges; therefore, there is need to re-evaluate the Ground Motion Prediction Equation in light of the recently updated database and the new generations of GMMs. In some cases, it was observed that some events were more extensively documented in one database than the other, like the 1979 Montenegro earthquake, with a considerably larger number of records in the BSHAP Analogue SM database when compared to ESM23. Therefore, the strong motion flat-file provided from the Harmonization of Seismic Hazard Maps in the Western Balkan Countries Project was merged with the ESM23 database for the polygon studied in this project. After performing the preliminary residual analysis, the candidate GMPE-s were identified. This process was done using the GMPE performance metrics available within the SMT in the OpenQuake Platform. The Likelihood Model and Euclidean Distance Based Ranking (EDR) were used. Finally, for this study, a GMPE logic tree was selected and following the selection of candidate GMPEs, model weights were assigned using the average sample log-likelihood approach of Scherbaum.

Keywords: residual analysis, GMPE, western balkan, strong motion, openquake

Procedia PDF Downloads 62
23626 Guidance for Strengthening Ethics of Entrepreneurs in Information and Communication Technology Professional

Authors: Routsukol Sunalai

Abstract:

The objectives of this paper were to study current problem of ethics of entrepreneurs in information and communication technology professional, and to build their awareness of ethics, which would be useful as guidance for strengthening professional ethics among them. The study employed quantitative research method in order to analyze relationships or differences found in each ethics factor and report in statistics. The sample of this paper was 300 information technology users of Rajabhat Universities in Bangkok. The findings revealed that the ethics factors which gained the highest and high level of opinion included possessing principles of righteousness, having trust in themselves and others, and respecting different opinions of others and accepting the fact that people of different opinions.

Keywords: communication, ethics, information, entrepreneurs

Procedia PDF Downloads 395
23625 Study on the Model Predicting Post-Construction Settlement of Soft Ground

Authors: Pingshan Chen, Zhiliang Dong

Abstract:

In order to estimate the post-construction settlement more objectively, the power-polynomial model is proposed, which can reflect the trend of settlement development based on the observed settlement data. It was demonstrated by an actual case history of an embankment, and during the prediction. Compared with the other three prediction models, the power-polynomial model can estimate the post-construction settlement more accurately with more simple calculation.

Keywords: prediction, model, post-construction settlement, soft ground

Procedia PDF Downloads 415
23624 Transmission Loss Analysis for Panels Laminated with Felt and Film

Authors: Yoshio Kurosawa

Abstract:

To reduce the interior noise of cars in high-frequency region, sound proof materials are laminated with the body panels and the interior trims. Therefore, sound proof properties of the laminates play an important role for the efficient acoustical design. A program code which predicts both sound absorption properties and sound insulation properties of the laminates are developed. This program code is used for transfer matrix method by Biot theory. This report described the outline of this program code, and the calculation results almost agreed with the experimental results.

Keywords: porous media, transmission loss, Biot theory, transfer matrix method

Procedia PDF Downloads 262
23623 A Proof of the Fact that a Finite Morphism is Proper

Authors: Ying Yi Wu

Abstract:

In this paper, we present a proof of the fact that a finite morphism is proper. We show that a finite morphism is universally closed and of finite type, which are the conditions for properness. Our proof is based on the theory of schemes and involves the use of the projection formula and the base change theorem. We first show that a finite morphism is of finite type and then proceed to show that it is universally closed. We use the fact that a finite morphism is also an affine morphism, which allows us to use the theory of coherent sheaves and their modules. We then show that the map induced by a finite morphism is flat and that the module it induces is of finite type. We use these facts to show that a finite morphism is universally closed. Our proof is constructive, and we provide details for each step of the argument.

Keywords: finite, morphism, schemes, projection.

Procedia PDF Downloads 88
23622 Finding Data Envelopment Analysis Targets Using Multi-Objective Programming in DEA-R with Stochastic Data

Authors: R. Shamsi, F. Sharifi

Abstract:

In this paper, we obtain the projection of inefficient units in data envelopment analysis (DEA) in the case of stochastic inputs and outputs using the multi-objective programming (MOP) structure. In some problems, the inputs might be stochastic while the outputs are deterministic, and vice versa. In such cases, we propose a multi-objective DEA-R model because in some cases (e.g., when unnecessary and irrational weights by the BCC model reduce the efficiency score), an efficient decision-making unit (DMU) is introduced as inefficient by the BCC model, whereas the DMU is considered efficient by the DEA-R model. In some other cases, only the ratio of stochastic data may be available (e.g., the ratio of stochastic inputs to stochastic outputs). Thus, we provide a multi-objective DEA model without explicit outputs and prove that the input-oriented MOP DEA-R model in the invariable return to scale case can be replaced by the MOP-DEA model without explicit outputs in the variable return to scale and vice versa. Using the interactive methods for solving the proposed model yields a projection corresponding to the viewpoint of the DM and the analyst, which is nearer to reality and more practical. Finally, an application is provided.

Keywords: DEA-R, multi-objective programming, stochastic data, data envelopment analysis

Procedia PDF Downloads 96
23621 Automated Driving Deep Neural Networks Model Accuracy and Performance Assessment in a Simulated Environment

Authors: David Tena-Gago, Jose M. Alcaraz Calero, Qi Wang

Abstract:

The evolution and integration of automated vehicles have become more and more tangible in recent years. State-of-the-art technological advances in the field of camera-based Artificial Intelligence (AI) and computer vision greatly favor the performance and reliability of the Advanced Driver Assistance System (ADAS), leading to a greater knowledge of vehicular operation and resembling human behavior. However, the exclusive use of this technology still seems insufficient to control vehicular operation at 100%. To reveal the degree of accuracy of the current camera-based automated driving AI modules, this paper studies the structure and behavior of one of the main solutions in a controlled testing environment. The results obtained clearly outline the lack of reliability when using exclusively the AI model in the perception stage, thereby entailing using additional complementary sensors to improve its safety and performance.

Keywords: accuracy assessment, AI-driven mobility, artificial intelligence, automated vehicles

Procedia PDF Downloads 94
23620 Reformulation of Theory of Critical Distances to Predict the Strength of Notched Plain Concrete Beams under Quasi Static Loading

Authors: Radhika V., J. M. Chandra Kishen

Abstract:

The theory of critical distances (TCD), due to its appealing characteristics, has been successfully used in the past to predict the strength of brittle as well as ductile materials, weakened by the presence of stress risers under both static and fatigue loading. By utilising most of the TCD's unique features, this paper summarises an attempt for a reformulation of the point method of the TCD to predict the strength of notched plain concrete beams under mode I quasi-static loading. A zone of micro cracks, which is responsible for the non-linearity of concrete, is taken into account considering the concept of an effective elastic crack. An attempt is also made to correlate the value of the material characteristic length required for the application of TCD with the maximum aggregate size in the concrete mix, eliminating the need for any extensive experimentation prior to the application of TCD. The devised reformulation and the proposed power law based relationship is found to yield satisfactory predictions for static strength of notched plain concrete beams, with geometric dimensions of the beam, tensile strength, and maximum aggregate size of the concrete mix being the only needed input parameters.

Keywords: characteristic length, effective elastic crack, inherent material strength, modeI loading, theory of critical distances

Procedia PDF Downloads 85
23619 Security Model for RFID Systems

Authors: John Ayoade

Abstract:

Radio Frequency Identification (RFID) has gained a lot of popularity in all walks of life due to its usefulness and diverse use of the technology in almost every application. However, there have been some security concerns most especially in regards to how authentic readers and tags can confirm their authenticity before confidential data is exchanged between them. In this paper, Kerberos protocol is adopted for the mutual authentication of RFID system components in order to ensure the secure communication between those components and to realize the authenticity of the communicating components.

Keywords: RFID, security, mutual authentication, Kerberos

Procedia PDF Downloads 449
23618 Application of Generalized Autoregressive Score Model to Stock Returns

Authors: Katleho Daniel Makatjane, Diteboho Lawrence Xaba, Ntebogang Dinah Moroke

Abstract:

The current study investigates the behaviour of time-varying parameters that are based on the score function of the predictive model density at time t. The mechanism to update the parameters over time is the scaled score of the likelihood function. The results revealed that there is high persistence of time-varying, as the location parameter is higher and the skewness parameter implied the departure of scale parameter from the normality with the unconditional parameter as 1.5. The results also revealed that there is a perseverance of the leptokurtic behaviour in stock returns which implies the returns are heavily tailed. Prior to model estimation, the White Neural Network test exposed that the stock price can be modelled by a GAS model. Finally, we proposed further researches specifically to model the existence of time-varying parameters with a more detailed model that encounters the heavy tail distribution of the series and computes the risk measure associated with the returns.

Keywords: generalized autoregressive score model, South Africa, stock returns, time-varying

Procedia PDF Downloads 486
23617 Leadership Styles and Adoption of Risk Governance in Insurance and Energy Industry: A Comparative Case Study

Authors: Ruchi Agarwal

Abstract:

In today’s world, companies are operating in dynamic, uncertain and ambiguous business environments. Globally, more companies are failing due to Environmental, Social and Governance (ESG) factors than ever. Corporate governance and risk management are intertwined in nature. For decades, corporate governance and risk management have been influenced by internal and external factors. Three schools of thought have influenced risk governance for decades: Agency theory, Contingency theory, and Institutional theory. Agency theory argues that agents have interests conflicting with principal interests and the information problem. Contingency theory suggests that risk management adoption is influenced by internal and external factors, while Institutional theory suggests that organizations legitimize risk management with regulators, competitors, and professional bodies. The conflicting objectives of theories have created problems for executives in organizations in the adoption of Risk Governance. So far, there are many studies that discussed risk culture and the role of actors in risk governance, but there are rare studies discussing the role of risk culture in the adoption of risk governance from a leadership style perspective. This study explores the adoption of risk governance in two contrasting industries, such as the Insurance and energy business, to understand whether risk governance is influenced by internal/external factors or whether risk culture is influenced by leaders. We draw empirical evidence by comparing the cases of an Indian insurance company and a renewable energy-based firm in India. We interviewed more than 20 senior executives of companies and collected annual reports, risk management policies, and more than 10 PPTs and other reports from 2017 to 2024. We visited the company for follow-up questions several times. The findings of my research revealed that both companies have used risk governance for strategic renewal of the company. Insurance companies use a transactional leadership style based on performance and reward for improving risk, while energy companies use rather symbolic management to make debt restructuring meaningful for stakeholders. Overall, both companies turned from loss-making to profitable ones in a few years. This comparative study highlights the role of different leadership styles in the adoption of risk governance. The study is also distinct as previous research rarely studied risk governance in two contrasting industries in reference to leadership styles.

Keywords: leadership style, corporate governance, risk management, risk culture, strategic renewal

Procedia PDF Downloads 30
23616 The Maps of Meaning (MoM) Consciousness Theory

Authors: Scott Andersen

Abstract:

Perhaps simply and rather unadornedly, consciousness is having multiple goals for action and the continuously adjudication of such goals to implement action, referred to as the Maps of Meaning (MoM) Consciousness Theory. The MoM theory triangulates through three parallel corollaries, action (behavior), mechanism (morphology/pathophysiology), and goals (teleology). (1) An organism’s consciousness contains a fluid, nested goals. These goals are not intentionality, but intersectionality, embodiment meeting the world. i.e., Darwinian inclusive fitness or randomization, then survival of the fittest. These goals form via gradual descent under inclusive fitness, the goals being the abstraction of a ‘match’ between the evolutionary environment and organism. Human consciousness implements the brain efficiency hypothesis, genetics, epigenetics, and experience crystallize efficiencies, not necessitating best or objective but fitness, i.e., perceived efficiency based on one’s adaptive environment. These efficiencies are objectively arbitrary, but determine the operation and level of one’s consciousness, termed extreme thrownness. Since inclusive fitness drives efficiencies in physiologic mechanism, morphology and behavior (action) and originates one’s goals, embodiment is necessarily entangled to human consciousness as its the intersection of mechanism or action (both necessitating embodiment) occurring in the world that determines fitness. Perception is the operant process of consciousness and is the consciousness’ de facto goal adjudication process. Goal operationalization is fundamentally efficiency-based via one’s unique neuronal mapping as a byproduct of genetics, epigenetics, and experience. Perception involves information intake and information discrimination, equally underpinned by efficiencies of inclusive fitness via extreme thrownness. Perception isn’t a ‘frame rate,’ but Bayesian priors of efficiency based on one’s extreme thrownness. Consciousness and human consciousness is a modular (i.e., a scalar level of richness, which builds up like building blocks) and dimensionalized (i.e., cognitive abilities become possibilities as emergent phenomena at various modularities, like stratified factors in factor analysis). The meta dimensions of human consciousness seemingly include intelligence quotient, personality (five-factor model), richness of perception intake, and richness of perception discrimination, among other potentialities. Future consciousness research should utilize factor analysis to parse modularities and dimensions of human consciousness and animal models.

Keywords: consciousness, perception, prospection, embodiment

Procedia PDF Downloads 26
23615 Reduction of Rotor-Bearing-Support Finite Element Model through Substructuring

Authors: Abdur Rosyid, Mohamed El-Madany, Mohanad Alata

Abstract:

Due to simplicity and low cost, rotordynamic system is often modeled by using lumped parameters. Recently, finite elements have been used to model rotordynamic system as it offers higher accuracy. However, it involves high degrees of freedom. In some applications such as control design, this requires higher cost. For this reason, various model reduction methods have been proposed. This work demonstrates the quality of model reduction of rotor-bearing-support system through substructuring. The quality of the model reduction is evaluated by comparing some first natural frequencies, modal damping ratio, critical speeds and response of both the full system and the reduced system. The simulation shows that the substructuring is proven adequate to reduce finite element rotor model in the frequency range of interest as long as the numbers and the locations of master nodes are determined appropriately. However, the reduction is less accurate in an unstable or nearly-unstable system.

Keywords: rotordynamic, finite element model, timoshenko beam, 3D solid elements, Guyan reduction method

Procedia PDF Downloads 258
23614 A Biophysical Model of CRISPR/Cas9 on- and off-Target Binding for Rational Design of Guide RNAs

Authors: Iman Farasat, Howard M. Salis

Abstract:

The CRISPR/Cas9 system has revolutionized genome engineering by enabling site-directed and high-throughput genome editing, genome insertion, and gene knockdowns in several species, including bacteria, yeast, flies, worms, and human cell lines. This technology has the potential to enable human gene therapy to treat genetic diseases and cancer at the molecular level; however, the current CRISPR/Cas9 system suffers from seemingly sporadic off-target genome mutagenesis that prevents its use in gene therapy. A comprehensive mechanistic model that explains how the CRISPR/Cas9 functions would enable the rational design of the guide-RNAs responsible for target site selection while minimizing unexpected genome mutagenesis. Here, we present the first quantitative model of the CRISPR/Cas9 genome mutagenesis system that predicts how guide-RNA sequences (crRNAs) control target site selection and cleavage activity. We used statistical thermodynamics and law of mass action to develop a five-step biophysical model of cas9 cleavage, and examined it in vivo and in vitro. To predict a crRNA's binding specificities and cleavage rates, we then compiled a nearest neighbor (NN) energy model that accounts for all possible base pairings and mismatches between the crRNA and the possible genomic DNA sites. These calculations correctly predicted crRNA specificity across 5518 sites. Our analysis reveals that cas9 activity and specificity are anti-correlated, and, the trade-off between them is the determining factor in performing an RNA-mediated cleavage with minimal off-targets. To find an optimal solution, we first created a scheme of safe-design criteria for Cas9 target selection by systematic analysis of available high throughput measurements. We then used our biophysical model to determine the optimal Cas9 expression levels and timing that maximizes on-target cleavage and minimizes off-target activity. We successfully applied this approach in bacterial and mammalian cell lines to reduce off-target activity to near background mutagenesis level while maintaining high on-target cleavage rate.

Keywords: biophysical model, CRISPR, Cas9, genome editing

Procedia PDF Downloads 387