Search results for: alignment error
1320 Stability Bound of Ruin Probability in a Reduced Two-Dimensional Risk Model
Authors: Zina Benouaret, Djamil Aissani
Abstract:
In this work, we introduce the qualitative and quantitative concept of the strong stability method in the risk process modeling two lines of business of the same insurance company or an insurance and re-insurance companies that divide between them both claims and premiums with a certain proportion. The approach proposed is based on the identification of the ruin probability associate to the model considered, with a stationary distribution of a Markov random process called a reversed process. Our objective, after clarifying the condition and the perturbation domain of parameters, is to obtain the stability inequality of the ruin probability which is applied to estimate the approximation error of a model with disturbance parameters by the considered model. In the stability bound obtained, all constants are explicitly written.Keywords: Markov chain, risk models, ruin probabilities, strong stability analysis
Procedia PDF Downloads 2471319 Robust Inference with a Skew T Distribution
Authors: M. Qamarul Islam, Ergun Dogan, Mehmet Yazici
Abstract:
There is a growing body of evidence that non-normal data is more prevalent in nature than the normal one. Examples can be quoted from, but not restricted to, the areas of Economics, Finance and Actuarial Science. The non-normality considered here is expressed in terms of fat-tailedness and asymmetry of the relevant distribution. In this study a skew t distribution that can be used to model a data that exhibit inherent non-normal behavior is considered. This distribution has tails fatter than a normal distribution and it also exhibits skewness. Although maximum likelihood estimates can be obtained by solving iteratively the likelihood equations that are non-linear in form, this can be problematic in terms of convergence and in many other respects as well. Therefore, it is preferred to use the method of modified maximum likelihood in which the likelihood estimates are derived by expressing the intractable non-linear likelihood equations in terms of standardized ordered variates and replacing the intractable terms by their linear approximations obtained from the first two terms of a Taylor series expansion about the quantiles of the distribution. These estimates, called modified maximum likelihood estimates, are obtained in closed form. Hence, they are easy to compute and to manipulate analytically. In fact the modified maximum likelihood estimates are equivalent to maximum likelihood estimates, asymptotically. Even in small samples the modified maximum likelihood estimates are found to be approximately the same as maximum likelihood estimates that are obtained iteratively. It is shown in this study that the modified maximum likelihood estimates are not only unbiased but substantially more efficient than the commonly used moment estimates or the least square estimates that are known to be biased and inefficient in such cases. Furthermore, in conventional regression analysis, it is assumed that the error terms are distributed normally and, hence, the well-known least square method is considered to be a suitable and preferred method for making the relevant statistical inferences. However, a number of empirical researches have shown that non-normal errors are more prevalent. Even transforming and/or filtering techniques may not produce normally distributed residuals. Here, a study is done for multiple linear regression models with random error having non-normal pattern. Through an extensive simulation it is shown that the modified maximum likelihood estimates of regression parameters are plausibly robust to the distributional assumptions and to various data anomalies as compared to the widely used least square estimates. Relevant tests of hypothesis are developed and are explored for desirable properties in terms of their size and power. The tests based upon modified maximum likelihood estimates are found to be substantially more powerful than the tests based upon least square estimates. Several examples are provided from the areas of Economics and Finance where such distributions are interpretable in terms of efficient market hypothesis with respect to asset pricing, portfolio selection, risk measurement and capital allocation, etc.Keywords: least square estimates, linear regression, maximum likelihood estimates, modified maximum likelihood method, non-normality, robustness
Procedia PDF Downloads 3961318 Estimation of Population Mean under Random Non-Response in Two-Occasion Successive Sampling
Authors: M. Khalid, G. N. Singh
Abstract:
In this paper, we have considered the problems of estimation for the population mean on current (second) occasion in two-occasion successive sampling under random non-response situations. Some modified exponential type estimators have been proposed and their properties are studied under the assumptions that the number of sampling unit follows a discrete distribution due to random non-response situations. The performances of the proposed estimators are compared with linear combinations of two estimators, (a) sample mean estimator for fresh sample and (b) ratio estimator for matched sample under the complete response situations. Results are demonstrated through empirical studies which present the effectiveness of the proposed estimators. Suitable recommendations have been made to the survey practitioners.Keywords: modified exponential estimator, successive sampling, random non-response, auxiliary variable, bias, mean square error
Procedia PDF Downloads 3491317 Morphological Analysis of English L1-Persian L2 Adult Learners’ Interlanguage: From the Perspective of SLA Variation
Authors: Maassoumeh Bemani Naeini
Abstract:
Studies on interlanguage have long been engaged in describing the phenomenon of variation in SLA. Pursuing the same goal and particularly addressing the role of linguistic features, this study describes the use of Persian morphology in the interlanguage of two adult English-speaking learners of Persian L2. Taking the general approach of a combination of contrastive analysis, error analysis and interlanguage analysis, this study focuses on the identification and prediction of some possible instances of transfer from English L1 to Persian L2 across six elicitation tasks aiming to investigate whether any of contextual features may variably influence the learners’ order of morpheme accuracy in the areas of copula, possessives, articles, demonstratives, plural form, personal pronouns, and genitive cases. Results describe the existence of task variation in the interlanguage system of Persian L2 learners.Keywords: English L1, Interlanguage Analysis, Persian L2, SLA variation
Procedia PDF Downloads 3151316 The Inherent Flaw in the NBA Playoff Structure
Authors: Larry Turkish
Abstract:
Introduction: The NBA is an example of mediocrity and this will be evident in the following paper. The study examines and evaluates the characteristics of the NBA champions. As divisions and playoff teams increase, there is an increase in the probability that the champion originates from the mediocre category. Since it’s inception in 1947, the league has been mediocre and continues to this day. Why does a professional league allow any team with a less than 50% winning percentage into the playoffs? As long as the finances flow into the league, owners will not change the current algorithm. The objective of this paper is to determine if the regular season has meaning in finding an NBA champion. Statistical Analysis: The data originates from the NBA website. The following variables are part of the statistical analysis: Rank, the rank of a team relative to other teams in the league based on the regular season win-loss record; Winning Percentage of a team based on the regular season; Divisions, the number of divisions within the league and Playoff Teams, the number of playoff teams relative to a particular season. The following statistical applications are applied to the data: Pearson Product-Moment Correlation, Analysis of Variance, Factor and Regression analysis. Conclusion: The results indicate that the divisional structure and number of playoff teams results in a negative effect on the winning percentage of playoff teams. It also prevents teams with higher winning percentages from accessing the playoffs. Recommendations: 1. Teams that have a winning percentage greater than 1 standard deviation from the mean from the regular season will have access to playoffs. (Eliminates mediocre teams.) 2. Eliminate Divisions (Eliminates weaker teams from access to playoffs.) 3. Eliminate Conferences (Eliminates weaker teams from access to the playoffs.) 4. Have a balanced regular season schedule, (Reduces the number of regular season games, creates equilibrium, reduces bias) that will reduce the need for load management.Keywords: alignment, mediocrity, regression, z-score
Procedia PDF Downloads 1291315 Development of a Real Time Axial Force Measurement System and IoT-Based Monitoring for Smart Bearing
Authors: Hassam Ahmed, Yuanzhi Liu, Yassine Selami, Wei Tao, Hui Zhao
Abstract:
The purpose of this research is to develop a real time axial force measurement system for a smart bearing through the use of strain-gauges, whereby the data acquisition is performed by an Arduino microcontroller due to its easy manipulation and low-cost. The measured signal is acquired and then discretized using a Wheatstone Bridge and an Analog-Digital Converter (ADC) respectively. For bearing monitoring, a real time monitoring system based on Internet of things (IoT) and Bluetooth were developed. Experimental tests were performed on a bearing within a force range up to 600 kN. The experimental results show that there is a proportional linear relationship between the applied force and the output voltage, and the error R squared is within 0.9878 based on the regression analysis.Keywords: bearing, force measurement, IoT, strain gauge
Procedia PDF Downloads 1391314 The Size Effects of Keyboards (Keycaps) on Computer Typing Tasks
Authors: Chih-Chun Lai, Jun-Yu Wang
Abstract:
The keyboard is the most important equipment for computer tasks. However, improper design of keyboard would cause some symptoms like ulnar and/or radial deviations. The research goal of this study was to investigate the optimal size(s) of keycaps to increase efficiency. As shown in the questionnaire pre-study with 49 participants aged from 20 to 44, the most commonly used keyboards were 101-key standard keyboards. Most of the keycap sizes (W × L) were 1.3 × 1.5 cm and 1.5 × 1.5 cm. The fingertip breadths of most participants were 1.2 cm. Therefore, in the main study with 18 participants, a standard keyboard with each set of the 3-sized (1.2 × 1.4 cm, 1.3 × 1.5 cm, and 1.5 × 1.5 cm) keycaps was used to investigate their typing efficiency, respectively. The results revealed that the differences between the operating times for using 1.3 × 1.5 cm and 1.2 × 1.4 cm keycaps were insignificant while operating times for using 1.5 × 1.5 cm keycaps were significantly longer than for using 1.2 × 1.4 cm or 1.3 × 1.5 cm, respectively. As for the typing error rate, there was no significant difference.Keywords: keyboard, keycap size, typing efficiency, computer tasks
Procedia PDF Downloads 3821313 Efficient Chiller Plant Control Using Modern Reinforcement Learning
Authors: Jingwei Du
Abstract:
The need of optimizing air conditioning systems for existing buildings calls for control methods designed with energy-efficiency as a primary goal. The majority of current control methods boil down to two categories: empirical and model-based. To be effective, the former heavily relies on engineering expertise and the latter requires extensive historical data. Reinforcement Learning (RL), on the other hand, is a model-free approach that explores the environment to obtain an optimal control strategy often referred to as “policy”. This research adopts Proximal Policy Optimization (PPO) to improve chiller plant control, and enable the RL agent to collaborate with experienced engineers. It exploits the fact that while the industry lacks historical data, abundant operational data is available and allows the agent to learn and evolve safely under human supervision. Thanks to the development of language models, renewed interest in RL has led to modern, online, policy-based RL algorithms such as the PPO. This research took inspiration from “alignment”, a process that utilizes human feedback to finetune the pretrained model in case of unsafe content. The methodology can be summarized into three steps. First, an initial policy model is generated based on minimal prior knowledge. Next, the prepared PPO agent is deployed so feedback from both critic model and human experts can be collected for future finetuning. Finally, the agent learns and adapts itself to the specific chiller plant, updates the policy model and is ready for the next iteration. Besides the proposed approach, this study also used traditional RL methods to optimize the same simulated chiller plants for comparison, and it turns out that the proposed method is safe and effective at the same time and needs less to no historical data to start up.Keywords: chiller plant, control methods, energy efficiency, proximal policy optimization, reinforcement learning
Procedia PDF Downloads 261312 The Choosing the Right Projects With Multi-Criteria Decision Making to Ensure the Sustainability of the Projects
Authors: Saniye Çeşmecioğlu
Abstract:
The importance of project sustainability and success has become increasingly significant due to the proliferation of external environmental factors that have decreased project resistance in contemporary times. The primary approach to forestall the failure of projects is to ensure their long-term viability through the strategic selection of projects as creating judicious project selection framework within the organization. Decision-makers require precise decision contexts (models) that conform to the company's business objectives and sustainability expectations during the project selection process. The establishment of a rational model for project selection enables organizations to create a distinctive and objective framework for the selection process. Additionally, for the optimal implementation of this decision-making model, it is crucial to establish a Project Management Office (PMO) team and Project Steering Committee within the organizational structure to oversee the framework. These teams enable updating project selection criteria and weights in response to changing conditions, ensuring alignment with the company's business goals, and facilitating the selection of potentially viable projects. This paper presents a multi-criteria decision model for selecting project sustainability and project success criteria that ensures timely project completion and retention. The model was developed using MACBETH (Measuring Attractiveness by a Categorical Based Evaluation Technique) and was based on broadcaster companies’ expectations. The ultimate results of this study provide a model that endorses the process of selecting the appropriate project objectively by utilizing project selection and sustainability criteria along with their respective weights for organizations. Additionally, the study offers suggestions that may ascertain helpful in future endeavors.Keywords: project portfolio management, project selection, multi-criteria decision making, project sustainability and success criteria, MACBETH
Procedia PDF Downloads 601311 A Fuzzy Linear Regression Model Based on Dissemblance Index
Authors: Shih-Pin Chen, Shih-Syuan You
Abstract:
Fuzzy regression models are useful for investigating the relationship between explanatory variables and responses in fuzzy environments. To overcome the deficiencies of previous models and increase the explanatory power of fuzzy data, the graded mean integration (GMI) representation is applied to determine representative crisp regression coefficients. A fuzzy regression model is constructed based on the modified dissemblance index (MDI), which can precisely measure the actual total error. Compared with previous studies based on the proposed MDI and distance criterion, the results from commonly used test examples show that the proposed fuzzy linear regression model has higher explanatory power and forecasting accuracy.Keywords: dissemblance index, fuzzy linear regression, graded mean integration, mathematical programming
Procedia PDF Downloads 4351310 Target and Equalizer Design for Perpendicular Heat-Assisted Magnetic Recording
Authors: P. Tueku, P. Supnithi, R. Wongsathan
Abstract:
Heat-Assisted Magnetic Recording (HAMR) is one of the leading technologies identified to enable areal density beyond 1 Tb/in2 of magnetic recording systems. A key challenge to HAMR designing is accuracy of positioning, timing of the firing laser, power of the laser, thermo-magnetic head, head-disk interface and cooling system. We study the effect of HAMR parameters on transition center and transition width. The HAMR is model using Thermal Williams-Comstock (TWC) and microtrack model. The target and equalizer are designed by the minimum mean square error (MMSE). The result shows that the unit energy constraint outperforms other constraints.Keywords: heat-assisted magnetic recording, thermal Williams-Comstock equation, microtrack model, equalizer
Procedia PDF Downloads 3471309 Estimating Cyclone Intensity Using INSAT-3D IR Images Based on Convolution Neural Network Model
Authors: Divvela Vishnu Sai Kumar, Deepak Arora, Sheenu Rizvi
Abstract:
Forecasting a cyclone through satellite images consists of the estimation of the intensity of the cyclone and predicting it before a cyclone comes. This research work can help people to take safety measures before the cyclone comes. The prediction of the intensity of a cyclone is very important to save lives and minimize the damage caused by cyclones. These cyclones are very costliest natural disasters that cause a lot of damage globally due to a lot of hazards. Authors have proposed five different CNN (Convolutional Neural Network) models that estimate the intensity of cyclones through INSAT-3D IR images. There are a lot of techniques that are used to estimate the intensity; the best model proposed by authors estimates intensity with a root mean squared error (RMSE) of 10.02 kts.Keywords: estimating cyclone intensity, deep learning, convolution neural network, prediction models
Procedia PDF Downloads 1241308 Application of EEG Wavelet Power to Prediction of Antidepressant Treatment Response
Authors: Dorota Witkowska, Paweł Gosek, Lukasz Swiecicki, Wojciech Jernajczyk, Bruce J. West, Miroslaw Latka
Abstract:
In clinical practice, the selection of an antidepressant often degrades to lengthy trial-and-error. In this work we employ a normalized wavelet power of alpha waves as a biomarker of antidepressant treatment response. This novel EEG metric takes into account both non-stationarity and intersubject variability of alpha waves. We recorded resting, 19-channel EEG (closed eyes) in 22 inpatients suffering from unipolar (UD, n=10) or bipolar (BD, n=12) depression. The EEG measurement was done at the end of the short washout period which followed previously unsuccessful pharmacotherapy. The normalized alpha wavelet power of 11 responders was markedly different than that of 11 nonresponders at several, mostly temporoparietal sites. Using the prediction of treatment response based on the normalized alpha wavelet power, we achieved 81.8% sensitivity and 81.8% specificity for channel T4.Keywords: alpha waves, antidepressant, treatment outcome, wavelet
Procedia PDF Downloads 3121307 Arabic Character Recognition Using Regression Curves with the Expectation Maximization Algorithm
Authors: Abdullah A. AlShaher
Abstract:
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We then proceed by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.Keywords: character recognition, regression curves, handwritten Arabic letters, expectation maximization algorithm
Procedia PDF Downloads 1431306 Projective Lag Synchronization in Drive-Response Dynamical Networks via Hybrid Feedback Control
Authors: Mohd Salmi Md Noorani, Ghada Al-Mahbashi, Sakhinah Abu Bakar
Abstract:
This paper investigates projective lag synchronization (PLS) behavior in drive response dynamical networks (DRDNs) model with identical nodes. A hybrid feedback control method is designed to achieve the PLS with mismatch and without mismatch terms. The stability of the error dynamics is proven theoretically using the Lyapunov stability theory. Finally, analytical results show that the states of the dynamical network with non-delayed coupling can be asymptotically synchronized onto a desired scaling factor under the designed controller. Moreover, the numerical simulations results demonstrate the validity of the proposed method.Keywords: drive-response dynamical network, projective lag synchronization, hybrid feedback control, stability theory
Procedia PDF Downloads 3901305 Additive White Gaussian Noise Filtering from ECG by Wiener Filter and Median Filter: A Comparative Study
Authors: Hossein Javidnia, Salehe Taheri
Abstract:
The Electrocardiogram (ECG) is the recording of the heart’s electrical potential versus time. ECG signals are often contaminated with noise such as baseline wander and muscle noise. As these signals have been widely used in clinical studies to detect heart diseases, it is essential to filter these noises. In this paper we compare performance of Wiener Filtering and Median Filtering methods to filter Additive White Gaussian (AWG) noise with the determined signal to noise ratio (SNR) ranging from 3 to 5 dB applied to long-term ECG recordings samples. Root mean square error (RMSE) and coefficient of determination (R2) between the filtered ECG and original ECG was used as the filter performance indicator. Experimental results show that Wiener filter has better noise filtering performance than Median filter.Keywords: ECG noise filtering, Wiener filtering, median filtering, Gaussian noise, filtering performance
Procedia PDF Downloads 5271304 Discrete Sliding Modes Regulator with Exponential Holder for Non-Linear Systems
Authors: G. Obregon-Pulido , G. C. Solis-Perales, J. A. Meda-Campaña
Abstract:
In this paper, we present a sliding mode controller in discrete time. The design of the controller is based on the theory of regulation for nonlinear systems. In the problem of disturbance rejection and/or output tracking, it is known that in discrete time, a controller that uses the zero-order holder only guarantees tracking at the sampling instances but not between instances. It is shown that using the so-called exponential holder, it is possible to guarantee asymptotic zero output tracking error, also between the sampling instant. For stabilizing the problem of close loop system we introduce the sliding mode approach relaxing the requirements of the existence of a linear stabilizing control law.Keywords: regulation theory, sliding modes, discrete controller, ripple-free tracking
Procedia PDF Downloads 511303 The Interaction between Human and Environment on the Perspective of Environmental Ethics
Authors: Mella Ismelina Farma Rahayu
Abstract:
Environmental problems could not be separated from unethical human perspectives and behaviors toward the environment. There is a fundamental error in the philosophy of people’s perspective about human and nature and their relationship with the environment, which in turn will create an inappropriate behavior in relation to the environment. The aim of this study is to investigate and to understand the ethics of the environment in the context of humans interacting with the environment by using the hermeneutic approach. The related theories and concepts collected from literature review are used as data, which were analyzed by using interpretation, critical evaluation, internal coherence, comparisons, and heuristic techniques. As a result of this study, there will be a picture related to the interaction of human and environment in the perspective of environmental ethics, as well as the problems of the value of ecological justice in the interaction of humans and environment. We suggest that the interaction between humans and environment need to be based on environmental ethics, in a spirit of mutual respect between humans and the natural world.Keywords: environment, environmental ethics, interaction, value
Procedia PDF Downloads 4201302 Survival and Hazard Maximum Likelihood Estimator with Covariate Based on Right Censored Data of Weibull Distribution
Authors: Al Omari Mohammed Ahmed
Abstract:
This paper focuses on Maximum Likelihood Estimator with Covariate. Covariates are incorporated into the Weibull model. Under this regression model with regards to maximum likelihood estimator, the parameters of the covariate, shape parameter, survival function and hazard rate of the Weibull regression distribution with right censored data are estimated. The mean square error (MSE) and absolute bias are used to compare the performance of Weibull regression distribution. For the simulation comparison, the study used various sample sizes and several specific values of the Weibull shape parameter.Keywords: weibull regression distribution, maximum likelihood estimator, survival function, hazard rate, right censoring
Procedia PDF Downloads 4391301 Robust Speed Sensorless Control to Estimated Error for PMa-SynRM
Authors: Kyoung-Jin Joo, In-Gun Kim, Hyun-Seok Hong, Dong-Woo Kang, Ju Lee
Abstract:
Recently, the permanent magnet-assisted synchronous reluctance motor (PMa-SynRM) that can be substituted for the induction motor has been studying because of the needs of the development of the premium high efficiency motor for the minimum energy performance standard (MEPS). PMa-SynRM is required to the speed and position information for motor speed and torque controls. However, to apply the sensors has many problems that are sensor mounting space shortage and additional cost, etc. Therefore, in this paper, speed-sensorless control based on model reference adaptive system (MRAS) is introduced to eliminate the sensor. The sensorless method is constructed in a reference model as standard and an adaptive model as the state observer. The proposed algorithm is verified by the simulation.Keywords: PMa-SynRM, sensorless control, robust estimation, MRAS method
Procedia PDF Downloads 4021300 Neural Network Based Path Loss Prediction for Global System for Mobile Communication in an Urban Environment
Authors: Danladi Ali
Abstract:
In this paper, we measured GSM signal strength in the Dnepropetrovsk city in order to predict path loss in study area using nonlinear autoregressive neural network prediction and we also, used neural network clustering to determine average GSM signal strength receive at the study area. The nonlinear auto-regressive neural network predicted that the GSM signal is attenuated with the mean square error (MSE) of 2.6748dB, this attenuation value is used to modify the COST 231 Hata and the Okumura-Hata models. The neural network clustering revealed that -75dB to -95dB is received more frequently. This means that the signal strength received at the study is mostly weak signalKeywords: one-dimensional multilevel wavelets, path loss, GSM signal strength, propagation, urban environment and model
Procedia PDF Downloads 3811299 Alternating Expectation-Maximization Algorithm for a Bilinear Model in Isoform Quantification from RNA-Seq Data
Authors: Wenjiang Deng, Tian Mou, Yudi Pawitan, Trung Nghia Vu
Abstract:
Estimation of isoform-level gene expression from RNA-seq data depends on simplifying assumptions, such as uniform reads distribution, that are easily violated in real data. Such violations typically lead to biased estimates. Most existing methods provide a bias correction step(s), which is based on biological considerations, such as GC content–and applied in single samples separately. The main problem is that not all biases are known. For example, new technologies such as single-cell RNA-seq (scRNA-seq) may introduce new sources of bias not seen in bulk-cell data. This study introduces a method called XAEM based on a more flexible and robust statistical model. Existing methods are essentially based on a linear model Xβ, where the design matrix X is known and derived based on the simplifying assumptions. In contrast, XAEM considers Xβ as a bilinear model with both X and β unknown. Joint estimation of X and β is made possible by simultaneous analysis of multi-sample RNA-seq data. Compared to existing methods, XAEM automatically performs empirical correction of potentially unknown biases. XAEM implements an alternating expectation-maximization (AEM) algorithm, alternating between estimation of X and β. For speed XAEM utilizes quasi-mapping for read alignment, thus leading to a fast algorithm. Overall XAEM performs favorably compared to other recent advanced methods. For simulated datasets, XAEM obtains higher accuracy for multiple-isoform genes, particularly for paralogs. In a differential-expression analysis of a real scRNA-seq dataset, XAEM achieves substantially greater rediscovery rates in an independent validation set.Keywords: alternating EM algorithm, bias correction, bilinear model, gene expression, RNA-seq
Procedia PDF Downloads 1401298 Required SNR for PPM in Downlink Gamma-Gamma Turbulence Channel
Authors: Selami Şahin
Abstract:
In this paper, in order to achieve sufficient bit error rate (BER) according to zenith angle of the satellite to ground station, SNR requirement is investigated utilizing pulse position modulation (PPM). To realize explicit results, all parameters such as link distance, Rytov variance, scintillation index, wavelength, aperture diameter of the receiver, Fried's parameter and zenith angle have been taken into account. Results indicate that after some parameters are determined since the constraints of the system, to achieve desired BER, required SNR values are in wide range while zenith angle changes from small to large values. Therefore, in order not to utilize high link margin, either SNR should adjust according to zenith angle or link should establish with predetermined intervals of the zenith angle.Keywords: Free-space optical communication, optical downlink channel, atmospheric turbulence, wireless optical communication
Procedia PDF Downloads 3991297 Adopted Method of Information System Strategy for Knowledge Management System: A Literature Review
Authors: Elin Cahyaningsih, Dana Indra Sensuse, Wahyu Catur Wibowo, Sofiyanti Indriasari
Abstract:
Bureaucracy reform program drives Indonesian government to change their management and supporting unit in order to enhance their organization performance. Information technology as one of supporting unit became one of strategic plan that organization tried to improve, because IT can automate and speed up process, reduce business process life cycle become more effective and efficient. Knowledge management system is a technology application for supporting knowledge management implementation in government which is requirement based on problem and potential functionality of each knowledge management process. Define knowledge management that suitable for each organization it is difficult, that why we should make the knowledge management system strategy as an alignment of knowledge management process in the organization. Knowledge management system is one of information system development in people perspective, because this system has high dependency in human interaction and participation. Strategic plan for developing knowledge management system can be determine using some of information system strategic methods. This research conducted to define type of strategic method of information system, stage of activity each method, the strategic method strength and weakness. The author use literature review methods for identify and classify strategic methods of information system for differentiate method type, categorize common activities, strength and weakness. Result of this research are determine and compare six strategic information system methods, there are Balanced Scorecard, Five Force Porter, SWOT analysis, Value Chain Analysis, Risk Analysis and Gap Analysis. Balanced Scorecard and Risk Analysis believe as common strategic method that usually used and have the highest excellence strength.Keywords: knowledge management system, balanced scorecard, five force, risk analysis, gap analysis, value chain analysis, SWOT analysis
Procedia PDF Downloads 4761296 Enzyme Immobilization on Functionalized Polystyrene Nanofibersfor Bioprocessing Applications
Authors: Mailin Misson, Bo Jin, Sheng Dai, Hu Zhang
Abstract:
Advances in biotechnology have witnessed a growing interest in enzyme applications for the development of green and sustainable bio processes. While known as powerful bio catalysts, enzymes are no longer of economic value when extended to large commercialization. Alternatively, immobilization technology allows enzyme recovery and continuous reuse which subsequently compensates high operating costs. Employment of enzymes on nano structured materials has been recognized as a promising approach to enhance enzyme catalytic performances. High porosity, inter connectivity and self-assembling behaviors endow nano fibers as exciting candidate for enzyme carrier in bio reactor systems. In this study, nano fibers were successfully fabricated via electro spinning system by optimizing the polymer concentration (10-30 %, w/v), applied voltage (10-30 kV) and discharge distance (11-26 cm). Microscopic images have confirmed the quality as homogeneous and good fiber alignment. The nano fibers surface was modified using strong oxidizing agent to facilitate bio molecule binding. Bovine serum albumin and β-galactosidase enzyme were employed as model bio catalysts and immobilized onto the oxidized surfaces through covalent binding. Maximum enzyme adsorption capacity of the modified nano fibers was 3000 mg/g, 3-fold higher than the unmodified counterpart (1000 mg/g). The highest immobilization yield was 80% and reached the saturation point at 2 mg/ml of enzyme concentration. The results indicate a significant increase of activity retention by the enzyme-bound modified nano fibers (80%) as compared to the nascent one (60%), signifying excellent enzyme-nano carrier bio compatibility. The immobilized enzyme was further used for the bio conversion of dairy wastes into value-added products. This study demonstrates great potential of acid-modified electrospun polystyrene nano fibers as enzyme carriers.Keywords: immobilization, enzyme, nanocarrier, nanofibers
Procedia PDF Downloads 2901295 Compilation and Statistical Analysis of an Arabic-English Legal Corpus in Sketch Engine
Authors: C. Brierley, H. El-Farahaty, A. Farhan
Abstract:
The Leeds Parallel Corpus of Arabic-English Constitutions is a parallel corpus for the Arabic legal domain. Analysis of legal language via Corpus Linguistics techniques is an important development. In legal proceedings, a corpus-based approach to disambiguating meaning is set to replace the dictionary as an interpretative tool, and legal scholarship in the States is now attuned to the potential for Text Analytics over vast quantities of text-based legal material, following the business and medical industries. This trend is reflected in Europe: the interdisciplinary research group in Computer Assisted Legal Linguistics mines big data collections of legal and non-legal texts to analyse: legal interpretations; legal discourse; the comprehensibility of legal texts; conflict resolution; and linguistic human rights. This paper focuses on ‘dignity’ as an important aspect of the overarching concept of human rights in current constitutions across the Arab world. We have compiled a parallel, Arabic-English raw text corpus (169,861 Arabic words and 205,893 English words) from reputable websites such as the World Intellectual Property Organisation and CONSTITUTE, and uploaded and queried our corpus in Sketch Engine. Our most challenging task was sentence-level alignment of Arabic-English data. This entailed manual intervention to ensure correspondence on a one-to-many basis since Arabic sentences differ from English in length and punctuation. We have searched for morphological variants of ‘dignity’ (رامة ك, karāma) in the Arabic data and inspected their English translation equivalents. The term occurs most frequently in the Sudanese constitution (10 instances), and not at all in the constitution of Palestine. Its most frequent collocate, determined via the logDice statistic in Sketch Engine, is ‘human’ as in ‘human dignity’.Keywords: Arabic constitution, corpus-based legal linguistics, human rights, parallel Arabic-English legal corpora
Procedia PDF Downloads 1811294 Kinetics and Mechanism Study of Photocatalytic Degradation Using Heterojunction Semiconductors
Authors: Ksenija Milošević, Davor Lončarević, Tihana Mudrinić, Jasmina Dostanić
Abstract:
Heterogeneous photocatalytic processes have gained growing interest as an efficient method to generate hydrogen by using clean energy sources and degrading various organic pollutants. The main obstacles that restrict efficient photoactivity are narrow light-response range and high rates of charge carrier recombination. The formation of heterojunction by combining a semiconductor with low VB and a semiconductor with high CB and a suitable band gap was found to be an efficient method to prepare more sensible materials with improved charge separation, appropriate oxidation and reduction ability, and enhanced visible-light harvesting. In our research, various binary heterojunction systems based on the wide-band gap (TiO₂) and narrow bandgap (g-C₃N₄, CuO, and Co₂O₃) photocatalyst were studied. The morphology, optical, and electrochemical properties of the photocatalysts were analyzed by X-ray diffraction (XRD), scanning electron microscopy (FE-SEM), N₂ physisorption, diffuse reflectance measurements (DRS), and Mott-Schottky analysis. The photocatalytic performance of the synthesized catalysts was tested in single and simultaneous systems. The synthesized photocatalysts displayed good adsorption capacity and enhanced visible-light photocatalytic performance. The mutual interactions of pollutants on their adsorption and degradation efficiency were investigated. The interfacial connection between photocatalyst constituents and the mechanism of the transport pathway of photogenerated charge species was discussed. A radical scavenger study revealed the interaction mechanisms of the photocatalyst constituents in single and multiple pollutant systems under solar and visible light irradiation, indicating the type of heterojunction system (Z scheme or type II).Keywords: bandgap alignment, heterojunction, photocatalysis, reaction mechanism
Procedia PDF Downloads 1001293 Influence of a Company’s Dynamic Capabilities on Its Innovation Capabilities
Authors: Lovorka Galetic, Zeljko Vukelic
Abstract:
The advanced concepts of strategic and innovation management in the sphere of company dynamic and innovation capabilities, and achieving their mutual alignment and a synergy effect, are important elements in business today. This paper analyses the theory and empirically investigates the influence of a company’s dynamic capabilities on its innovation capabilities. A new multidimensional model of dynamic capabilities is presented, consisting of five factors appropriate to real time requirements, while innovation capabilities are considered pursuant to the official OECD and Eurostat standards. After examination of dynamic and innovation capabilities indicated their theoretical links, the empirical study testing the model and examining the influence of a company’s dynamic capabilities on its innovation capabilities showed significant results. In the study, a research model was posed to relate company dynamic and innovation capabilities. One side of the model features the variables that are the determinants of dynamic capabilities defined through their factors, while the other side features the determinants of innovation capabilities pursuant to the official standards. With regard to the research model, five hypotheses were set. The study was performed in late 2014 on a representative sample of large and very large Croatian enterprises with a minimum of 250 employees. The research instrument was a questionnaire administered to company top management. For both variables, the position of the company was tested in comparison to industry competitors, on a fivepoint scale. In order to test the hypotheses, correlation tests were performed to determine whether there is a correlation between each individual factor of company dynamic capabilities with the existence of its innovation capabilities, in line with the research model. The results indicate a strong correlation between a company’s possession of dynamic capabilities in terms of their factors, due to the new multi-dimensional model presented in this paper, with its possession of innovation capabilities. Based on the results, all five hypotheses were accepted. Ultimately, it was concluded that there is a strong association between the dynamic and innovation capabilities of a company.Keywords: dynamic capabilities, innovation capabilities, competitive advantage, business results
Procedia PDF Downloads 3031292 Gamipulation: Exploring Covert Manipulation Through Gamification in the Context of Education
Authors: Aguiar-Castillo Lidia, Perez-Jimenez Rafael
Abstract:
The integration of gamification in educational settings aims to enhance student engagement and motivation through game design elements in learning activities. This paper introduces "Gamipulation," the subtle manipulation of students via gamification techniques serving hidden agendas without explicit consent. It highlights the need to distinguish between beneficial and exploitative uses of gamification in education, focusing on its potential to psychologically manipulate students for purposes misaligned with their best interests Through a literature review and expert interviews, this study presents a conceptual framework outlining gamipulation's features. It examines ethical concerns like gradually introducing desired behaviors, using distraction to divert attention from significant learning objectives, immediacy of rewards fostering short-term engagement over long-term learning, infantilization of students, and exploitation of emotional responses over reflective thinking. Additionally, it discusses ethical issues in collecting and utilizing student data within gamified environments. Key findings suggest that while gamification can enhance motivation and engagement, there's a fine line between ethical motivation and unethical manipulation. The study emphasizes the importance of transparency, respect for student autonomy, and alignment with educational values in gamified systems. It calls for educators and designers to be aware of gamification's manipulative potential and strive for ethical implementation that benefits students. In conclusion, this paper provides a framework for educators and researchers to understand and address gamipulation's ethical challenges. It encourages developing ethical guidelines and practices to ensure gamification in education remains a tool for positive engagement and learning rather than covert manipulation.Keywords: gradualness, distraction, immediacy, infantilization, emotion
Procedia PDF Downloads 261291 In Silico Exploration of Quinazoline Derivatives as EGFR Inhibitors for Lung Cancer: A Multi-Modal Approach Integrating QSAR-3D, ADMET, Molecular Docking, and Molecular Dynamics Analyses
Authors: Mohamed Moussaoui
Abstract:
A series of thirty-one potential inhibitors targeting the epidermal growth factor receptor kinase (EGFR), derived from quinazoline, underwent 3D-QSAR analysis using CoMFA and CoMSIA methodologies. The training and test sets of quinazoline derivatives were utilized to construct and validate the QSAR models, respectively, with dataset alignment performed using the lowest energy conformer of the most active compound. The best-performing CoMFA and CoMSIA models demonstrated impressive determination coefficients, with R² values of 0.981 and 0.978, respectively, and Leave One Out cross-validation determination coefficients, Q², of 0.645 and 0.729, respectively. Furthermore, external validation using a test set of five compounds yielded predicted determination coefficients, R² test, of 0.929 and 0.909 for CoMFA and CoMSIA, respectively. Building upon these promising results, eighteen new compounds were designed and assessed for drug likeness and ADMET properties through in silico methods. Additionally, molecular docking studies were conducted to elucidate the binding interactions between the selected compounds and the enzyme. Detailed molecular dynamics simulations were performed to analyze the stability, conformational changes, and binding interactions of the quinazoline derivatives with the EGFR kinase. These simulations provided deeper insights into the dynamic behavior of the compounds within the active site. This comprehensive analysis enhances the understanding of quinazoline derivatives as potential anti-cancer agents and provides valuable insights for lead optimization in the early stages of drug discovery, particularly for developing highly potent anticancer therapeuticsKeywords: 3D-QSAR, CoMFA, CoMSIA, ADMET, molecular docking, quinazoline, molecular dynamic, egfr inhibitors, lung cancer, anticancer
Procedia PDF Downloads 46