Search results for: data interpolating empirical orthogonal function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 29436

Search results for: data interpolating empirical orthogonal function

28956 The Adoption of Process Management for Accounting Information Systems in Thailand

Authors: Manirath Wongsim

Abstract:

Information Quality (IQ) has become a critical, strategic issue in Accounting Information Systems (AIS) adoption. In order to implement AIS adoption successfully, it is important to consider the quality of information use throughout the adoption process, which seriously impacts the effectiveness of AIS adoption practice and the optimization of AIS adoption decisions. There is a growing need for research to provide insights into issues and solutions related to IQ in AIS adoption. The need for an integrated approach to improve IQ in AIS adoption, as well as the unique characteristics of accounting data, demands an AIS adoption specific IQ framework. This research aims to explore ways of managing information quality and AIS adoption to investigate the relationship between the IQ issues and AIS adoption process. This study has led to the development of a framework for understanding IQ management in AIS adoption. This research was done on 44 respondents as ten organisations from manufacturing firms in Thailand. The findings of the research’s empirical evidence suggest that IQ dimensions in AIS adoption to provide assistance in all process of decision making. This research provides empirical evidence that information quality of AIS adoption affect decision making and suggests that these variables should be considered in adopting AIS in order to improve the effectiveness of AIS.

Keywords: information quality, information quality dimensions, accounting information systems, accounting information system adoption

Procedia PDF Downloads 454
28955 Sentiment Classification of Documents

Authors: Swarnadip Ghosh

Abstract:

Sentiment Analysis is the process of detecting the contextual polarity of text. In other words, it determines whether a piece of writing is positive, negative or neutral.Sentiment analysis of documents holds great importance in today's world, when numerous information is stored in databases and in the world wide web. An efficient algorithm to illicit such information, would be beneficial for social, economic as well as medical purposes. In this project, we have developed an algorithm to classify a document into positive or negative. Using our algorithm, we obtained a feature set from the data, and classified the documents based on this feature set. It is important to note that, in the classification, we have not used the independence assumption, which is considered by many procedures like the Naive Bayes. This makes the algorithm more general in scope. Moreover, because of the sparsity and high dimensionality of such data, we did not use empirical distribution for estimation, but developed a method by finding degree of close clustering of the data points. We have applied our algorithm on a movie review data set obtained from IMDb and obtained satisfactory results.

Keywords: sentiment, Run's Test, cross validation, higher dimensional pmf estimation

Procedia PDF Downloads 384
28954 Automatic Slider Design in Injection Moldings

Authors: Alan C. Lin, Tran Anh Son

Abstract:

This study proposes an approach to determine the undercut regions and their releasing directions for slider design of complex parts represented by the file format of STL (STereoLithography). In order to delineate the border of undercut regions, orthogonal cutting planes are firstly employed to automatically find the inner loops of a part model. To discover the facets belonging to undercut regions, attributes are then assigned to the facets of the part model based on the topological relationship of adjacent facets of each inner loop. After that, the undercut regions are separated from other facets in the model. Through the recognized facets of the undercut regions, the concept of 'visibility map (V-map)' is further applied to determine feasible releasing directions for each of the undercut regions. The undercut regions having the same releasing direction are finally grouped to form a slider in the injection mold.

Keywords: solid model, STL data, injection mold design, visibility map

Procedia PDF Downloads 382
28953 Geotechnical Characterization of Residual Soil for Deterministic Landslide Assessment

Authors: Vera Karla S. Caingles, Glen A. Lorenzo

Abstract:

Soil, as the main material of landslides, plays a vital role in landslide assessment. An efficient and accurate method of doing an assessment is significantly important to prevent damage of properties and loss of lives. The study has two phases: to establish an empirical correlation of the residual soil thickness with the slope angle and to investigate the geotechnical characteristics of residual soil. Digital Elevation Model (DEM) in Geographic Information System (GIS) was used to establish the slope map and to program sampling points for field investigation. Physical and index property tests were undertaken on the 20 soil samples obtained from the area with Pliocene-Pleistocene geology and different slope angle in Kibawe, Bukidnon. The regression analysis result shows that the best fitting model that can describe the soil thickness-slope angle relationship is an exponential function. The physical property results revealed that soils contain a high percentage of clay and silts ranges from 41% - 99.52%. Based on the index properties test results, the soil exhibits a high degree of plasticity and expansion but not collapsible. It is deemed that this compendium will serve as primary data for slope stability analysis and deterministic landslide assessment.

Keywords: collapsibility, correlation, expansiveness, landslide, plasticity

Procedia PDF Downloads 146
28952 Agency Cost, Firm Performance, Corporate Governance: Evidence from Indonesia

Authors: Arnold Sanda Layuk

Abstract:

Fraud in the disclosure of financial statements by management shows that agency conflict is an important issue in the company. The conflict has consequences for the agency costs that must be borne and has an impact on the firm's performance. The effect of agency costs on firm performance is investigated in this study, as well as whether several variables such as corporate governance mechanisms can positively moderate the agency cost and firm performance relationship. The agency cost is measured by the asset utilization ratio and discretionary expenditure ratio. The firm's performance is represented by the return on equity. Data was collected from the manufacturing companies listed on the Indonesia Stock Exchange from 2015 to 2019, then regressed on the panel data using the panel corrected standard error model (PCSE). According to the findings, agency costs are negatively related to firm performance, which supports previous empirical research findings. It also found that the agency cost and firm performance relationship is significantly moderated by board size and ownership concentration as the representatives of corporate governance mechanisms. It suggests that corporate governance can become tools to reduce agency costs and increase firm performance as well. The empirical evidence adds to previous research on agency conflict, particularly in emerging markets. These findings are expected to supplement previous research and provide additional information to shareholders in order to control opportunistic management decisions that affect their investments and discretionary operational expenses.

Keywords: agency cost, corporate governance, asset utilization ratio, firm performance

Procedia PDF Downloads 181
28951 A Theory and Empirical Analysis on the Efficency of Chinese Electricity Pricing

Authors: Jianlin Wang, Jiajia Zhao

Abstract:

This paper applies the theory and empirical method to examine the relationship between electricity price and coal price, as well as electricity and industry output, for China during Jan 1999-Dec 2012. Our results indicate that there is no any causality between coal price and electricity price under other factors are controlled. However, we found a bi-directional causality between electricity consumption and industry output. Overall, the electricity price set by China’s NDRC is inefficient, which lead to the electricity supply shortage after 2004. It is time to reform electricity price system for China’s reformers.

Keywords: electricity price, coal price, power supply, China

Procedia PDF Downloads 455
28950 Technological Enhancements in Supply Chain Management Post COVID-19

Authors: Miran Ismail

Abstract:

COVID-19 has caused widespread disruption in all economical sectors and industries around the world. The COVID-19 lockdown measures have resulted in production halts, restrictions on persons and goods movement, border closures, logistical constraints, and a slowdown in trade and economic activity. The main subject of this paper is to leverage technology to manage the supply chain effectively and efficiently through the usage of artificial intelligence. The research methodology is based on empirical data collected through a questionnaire survey. One of the approaches utilized is a case study of industrial organizations that face obstacles such as high operational costs, large inventory levels, a lack of well-established supplier relationships, human behavior, and system issues. The main contribution of this research to the body of knowledge is the empirical insights and on supply chain sustainability performance measurement. The results provide guidelines for the selection of advanced technologies to support supply chain processes and for the design of sustainable performance measurement systems.

Keywords: information technology, artificial intelligence, supply chain management, industrial organizations

Procedia PDF Downloads 113
28949 Replicating Brain’s Resting State Functional Connectivity Network Using a Multi-Factor Hub-Based Model

Authors: B. L. Ho, L. Shi, D. F. Wang, V. C. T. Mok

Abstract:

The brain’s functional connectivity while temporally non-stationary does express consistency at a macro spatial level. The study of stable resting state connectivity patterns hence provides opportunities for identification of diseases if such stability is severely perturbed. A mathematical model replicating the brain’s spatial connections will be useful for understanding brain’s representative geometry and complements the empirical model where it falls short. Empirical computations tend to involve large matrices and become infeasible with fine parcellation. However, the proposed analytical model has no such computational problems. To improve replicability, 92 subject data are obtained from two open sources. The proposed methodology, inspired by financial theory, uses multivariate regression to find relationships of every cortical region of interest (ROI) with some pre-identified hubs. These hubs acted as representatives for the entire cortical surface. A variance-covariance framework of all ROIs is then built based on these relationships to link up all the ROIs. The result is a high level of match between model and empirical correlations in the range of 0.59 to 0.66 after adjusting for sample size; an increase of almost forty percent. More significantly, the model framework provides an intuitive way to delineate between systemic drivers and idiosyncratic noise while reducing dimensions by more than 30 folds, hence, providing a way to conduct attribution analysis. Due to its analytical nature and simple structure, the model is useful as a standalone toolkit for network dependency analysis or as a module for other mathematical models.

Keywords: functional magnetic resonance imaging, multivariate regression, network hubs, resting state functional connectivity

Procedia PDF Downloads 139
28948 Developing Indicators in System Mapping Process Through Science-Based Visual Tools

Authors: Cristian Matti, Valerie Fowles, Eva Enyedi, Piotr Pogorzelski

Abstract:

The system mapping process can be defined as a knowledge service where a team of facilitators, experts and practitioners facilitate a guided conversation, enable the exchange of information and support an iterative curation process. System mapping processes rely on science-based tools to introduce and simplify a variety of components and concepts of socio-technical systems through metaphors while facilitating an interactive dialogue process to enable the design of co-created maps. System maps work then as “artifacts” to provide information and focus the conversation into specific areas around the defined challenge and related decision-making process. Knowledge management facilitates the curation of that data gathered during the system mapping sessions through practices of documentation and subsequent knowledge co-production for which common practices from data science are applied to identify new patterns, hidden insights, recurrent loops and unexpected elements. This study presents empirical evidence on the application of these techniques to explore mechanisms by which visual tools provide guiding principles to portray system components, key variables and types of data through the lens of climate change. In addition, data science facilitates the structuring of elements that allow the analysis of layers of information through affinity and clustering analysis and, therefore, develop simple indicators for supporting the decision-making process. This paper addresses methodological and empirical elements on the horizontal learning process that integrate system mapping through visual tools, interpretation, cognitive transformation and analysis. The process is designed to introduce practitioners to simple iterative and inclusive processes that create actionable knowledge and enable a shared understanding of the system in which they are embedded.

Keywords: indicators, knowledge management, system mapping, visual tools

Procedia PDF Downloads 181
28947 Analyzing the Effects of Real Income and Biomass Energy Consumption on Carbon Dioxide (CO2) Emissions: Empirical Evidence from the Panel of Biomass-Consuming Countries

Authors: Eyup Dogan

Abstract:

This empirical aims to analyze the impacts of real income and biomass energy consumption on the level of emissions in the EKC model for the panel of biomass-consuming countries over the period 1980-2011. Because we detect the presence of cross-sectional dependence and heterogeneity across countries for the analyzed data, we use panel estimation methods robust to cross-sectional dependence and heterogeneity. The CADF and the CIPS panel unit root tests indicate that carbon emissions, real income and biomass energy consumption are stationary at the first-differences. The LM bootstrap panel cointegration test shows that the analyzed variables are cointegrated. Results from the panel group-mean DOLS and the panel group-mean FMOLS estimators show that increase in biomass energy consumption decreases CO2 emissions and the EKC hypothesis is validated. Therefore, countries are advised to boost their production and increase the use of biomass energy for lower level of emissions.

Keywords: biomass energy, CO2 emissions, EKC model, heterogeneity, cross-sectional dependence

Procedia PDF Downloads 282
28946 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second

Authors: P. V. Pramila , V. Mahesh

Abstract:

Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients esulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF 25, PEF,FEF 25-75, FEF50, and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF 25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects). It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.

Keywords: FEV, multivariate adaptive regression splines pulmonary function test, random forest

Procedia PDF Downloads 296
28945 Strengthening of Concrete Slabs with Steel Beams

Authors: Mizam Doğan

Abstract:

In service life; structures can be damaged if they are subjected to dead and live loads which are greater than design values. For preventing this case; possible loads must be correctly calculated, structure must be designed according to determined loads, and structure must not be used out of its function. If loading case of the structure changes when its function changes; it must be reinforced for continuing it is new function. Reinforcement is a process that is made by increasing the existing strengths of structural system elements of the structure as reinforced concrete walls, beams, and slabs. Reinforcement can be done by casting reinforced concrete, placing steel and fiber structural elements. In this paper, reinforcing of columns and slabs of a structure of which function is changed is studied step by step. This reinforcement is made for increasing vertical and lateral load carrying capacity of the building. Not for repairing damaged structural system.

Keywords: strengthening, RC slabs, seismic load, steel beam, structural irregularity

Procedia PDF Downloads 245
28944 Empirical Exploration for the Correlation between Class Object-Oriented Connectivity-Based Cohesion and Coupling

Authors: Jehad Al Dallal

Abstract:

Attributes and methods are the basic contents of an object-oriented class. The connectivity among these class members and the relationship between the class and other classes play an important role in determining the quality of an object-oriented system. Class cohesion evaluates the degree of relatedness of class attributes and methods, whereas class coupling refers to the degree to which a class is related to other classes. Researchers have proposed several class cohesion and class coupling measures. However, the correlation between class coupling and class cohesion measures have not been thoroughly studied. In this paper, using classes of three open-source Java systems, we empirically investigate the correlation between several measures of connectivity-based class cohesion and coupling. Four connectivity-based cohesion measures and eight coupling measures are considered in the empirical study. The empirical study results show that class connectivity-based cohesion and coupling internal quality attributes are inversely correlated. The strength of the correlation depends highly on the cohesion and coupling measurement approaches.

Keywords: object-oriented class, software quality, class cohesion measure, class coupling measure

Procedia PDF Downloads 306
28943 Learning Gains and Constraints Resulting from Haptic Sensory Feedback among Preschoolers' Engagement during Science Experimentation

Authors: Marios Papaevripidou, Yvoni Pavlou, Zacharias Zacharia

Abstract:

Embodied cognition and additional (touch) sensory channel theories indicate that physical manipulation is crucial to learning since it provides, among others, touch sensory input, which is needed for constructing knowledge. Given these theories, the use of Physical Manipulatives (PM) becomes a prerequisite for learning. On the other hand, empirical research on Virtual Manipulatives (VM) (e.g., simulations) learning has provided evidence showing that the use of PM, and thus haptic sensory input, is not always a prerequisite for learning. In order to investigate which means of experimentation, PM or VM, are required for enhancing student science learning at the kindergarten level, an empirical study was conducted that sought to investigate the impact of haptic feedback on the conceptual understanding of pre-school students (n=44, age mean=5,7) in three science domains: beam balance (D1), sinking/floating (D2) and springs (D3). The participants were equally divided in two groups according to the type of manipulatives used (PM: presence of haptic feedback, VM: absence of haptic feedback) during a semi-structured interview for each of the domains. All interviews followed the Predict-Observe-Explain (POE) strategy and consisted of three phases: initial evaluation, experimentation, final evaluation. The data collected through the interviews were analyzed qualitatively (open-coding for identifying students’ ideas in each domain) and quantitatively (use of non-parametric tests). Findings revealed that the haptic feedback enabled students to distinguish heavier to lighter objects when held in hands during experimentation. In D1 the haptic feedback did not differentiate PM and VM students' conceptual understanding of the function of the beam as a mean to compare the mass of objects. In D2 the haptic feedback appeared to have a negative impact on PM students’ learning. Feeling the weight of an object strengthen PM students’ misconception that heavier objects always sink, whereas the scientifically correct idea that the material of an object determines its sinking/floating behavior in the water was found to be significantly higher among the VM students than the PM ones. In D3 the PM students outperformed significantly the VM students with regard to the idea that the heavier an object is the more the spring will expand, indicating that the haptic input experienced by the PM students served as an advantage to their learning. These findings point to the fact that PMs, and thus touch sensory input, might not always be a requirement for science learning and that VMs could be considered, under certain circumstances, as a viable means for experimentation.

Keywords: haptic feedback, physical and virtual manipulatives, pre-school science learning, science experimentation

Procedia PDF Downloads 124
28942 Feasibility Study of Wind Energy Potential in Turkey: Case Study of Catalca District in Istanbul

Authors: Mohammed Wadi, Bedri Kekezoglu, Mustafa Baysal, Mehmet Rida Tur, Abdulfetah Shobole

Abstract:

This paper investigates the technical evaluation of the wind potential for present and future investments in Turkey taking into account the feasibility of sites, installments, operation, and maintenance. This evaluation based on the hourly measured wind speed data for the three years 2008–2010 at 30 m height for Çatalca district. These data were obtained from national meteorology station in Istanbul–Republic of Turkey are analyzed in order to evaluate the feasibility of wind power potential and to assure supreme assortment of wind turbines installing for the area of interest. Furthermore, the data are extrapolated and analyzed at 60 m and 80 m regarding the variability of roughness factor. Weibull bi-parameter probability function is used to approximate monthly and annually wind potential and power density based on three calculation methods namely, the approximated, the graphical and the energy pattern factor methods. The annual mean wind power densities were to be 400.31, 540.08 and 611.02 W/m² for 30, 60, and 80 m heights respectively. Simulation results prove that the analyzed area is an appropriate place for constructing large-scale wind farms.

Keywords: wind potential in Turkey, Weibull bi-parameter probability function, the approximated method, the graphical method, the energy pattern factor method, capacity factor

Procedia PDF Downloads 244
28941 Transforming Data into Knowledge: Mathematical and Statistical Innovations in Data Analytics

Authors: Zahid Ullah, Atlas Khan

Abstract:

The rapid growth of data in various domains has created a pressing need for effective methods to transform this data into meaningful knowledge. In this era of big data, mathematical and statistical innovations play a crucial role in unlocking insights and facilitating informed decision-making in data analytics. This abstract aims to explore the transformative potential of these innovations and their impact on converting raw data into actionable knowledge. Drawing upon a comprehensive review of existing literature, this research investigates the cutting-edge mathematical and statistical techniques that enable the conversion of data into knowledge. By evaluating their underlying principles, strengths, and limitations, we aim to identify the most promising innovations in data analytics. To demonstrate the practical applications of these innovations, real-world datasets will be utilized through case studies or simulations. This empirical approach will showcase how mathematical and statistical innovations can extract patterns, trends, and insights from complex data, enabling evidence-based decision-making across diverse domains. Furthermore, a comparative analysis will be conducted to assess the performance, scalability, interpretability, and adaptability of different innovations. By benchmarking against established techniques, we aim to validate the effectiveness and superiority of the proposed mathematical and statistical innovations in data analytics. Ethical considerations surrounding data analytics, such as privacy, security, bias, and fairness, will be addressed throughout the research. Guidelines and best practices will be developed to ensure the responsible and ethical use of mathematical and statistical innovations in data analytics. The expected contributions of this research include advancements in mathematical and statistical sciences, improved data analysis techniques, enhanced decision-making processes, and practical implications for industries and policymakers. The outcomes will guide the adoption and implementation of mathematical and statistical innovations, empowering stakeholders to transform data into actionable knowledge and drive meaningful outcomes.

Keywords: data analytics, mathematical innovations, knowledge extraction, decision-making

Procedia PDF Downloads 59
28940 Fractal Behaviour of Earthquake Sequences in Himalaya

Authors: Kamal, Adil Ahmad

Abstract:

Earthquakes are among the most versatile natural and dynamic processes, and hence a fractal model is considered to be the best representative of the same. We present a novel method to process and analyse information hidden in earthquake sequences using Fractal Dimensions and Iterative Function Systems (IFS). Spatial and temporal variations in the fractal dimensions of seismicity observed around the Indian peninsula in last 30 years are studied. This was used as a possible precursor before large earthquakes in the region. IFS images for observed seismicity in the Himalayan belt were also obtained. We scan the whole data set and coarse grain of a selected window to reduce it to four bins. A critical analysis of four-cornered chaos-game clearly shows that the spatial variation in earthquake occurrences in Himalayan range is not random. Two subzones of Himalaya have a tendency to follow each other in time.

Keywords: earthquakes, fractals, Himalaya, iterated function systems

Procedia PDF Downloads 285
28939 Organizational Performance and Impact of Social Innovation

Authors: Alfonso Unceta, Javier Castro-Spila

Abstract:

This paper offers a conceptual and empirical exploration between the organizational performance and the impact of social innovation. The paper contributes on the social innovation field in three domains: a) It provides analytical and empirical evidence linking organizational performance to the impact of social innovation; b) it provides a first outline of impact assessment of social innovation when it is developed by a diversity of heterogeneous actors (systemic social innovation); c) it provides a first outline for the development of innovation policies to support social innovations according to a typology of organizations and a typology of impact.

Keywords: absorptive capacity, social innovation impact, organizational performance, RESINDEX, Basque Country

Procedia PDF Downloads 467
28938 Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

Authors: Yehjune Heo

Abstract:

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Keywords: anti-spoofing, CNN, fingerprint recognition, loss function, optimizer

Procedia PDF Downloads 123
28937 Price to Earnings Growth (PEG) Predicting Future Returns Better than the Price to Earnings (PE) Ratio

Authors: Lindrianasari Stefanie, Aminah Khairudin

Abstract:

This study aims to provide empirical evidence regarding the ability of Price to Earnings Ratio and PEG Ratio in predicting future stock returns issuers. The samples used in this study are stocks that go into LQ45. The main contribution is to assign empirical evidence if the PEG Ratio can provide optimum return compared to Price to Earnings Ratio. This study used a sample of the entire company into the group LQ45 with the period of observation. The data used is limited to the financial statements of a company incorporated in LQ45 period July 2013-July 2014, using the financial statements and the position of the company's closing stock price at the end of 2010 as a reference benchmark for the growth of the company's stock price compared to the closing price of 2013. This study found that the method of PEG Ratio can outperform the method of PE ratio in predicting future returns on the stock portfolio of LQ45.

Keywords: price to earnings growth, price to earnings ratio, future returns, stock price

Procedia PDF Downloads 402
28936 Process Parameter Study on Friction Push Plug Welding of AA6061 Alloy

Authors: H. Li, W. Qin, Ben Ye

Abstract:

Friction Push Plug Welding (FPPW) is a solid phase welding suitable for repairing defective welds and filling self-reacting weld keyholes in Friction Stir Welds. In FPPW process, a tapered shaped plug is rotated at high speed and forced into a tapered hole in the substrate. The plug and substrate metal is softened by the increasing temperature generated by friction and material plastic deformation. This paper aims to investigate the effect of process parameters on the quality of the weld. Orthogonal design methods were employed to reduce the amount of experiment. Three values were selected for each process parameter, rotation speed (1500r/min, 2000r/min, 2500r/min), plunge depth (2mm, 3mm, 4mm) and plunge speed (60mm/min, 90mm/min, 120r/min). AA6061aluminum alloy plug and substrate plate was used in the experiment. In a trial test with the plunge depth of 1mm, a noticeable defect appeared due to the short plunge time and insufficient temperature. From the recorded temperature profiles, it was found that the peak temperature increased with the increase of the rotation speed, plunge speed and plunge depth. In the initial stage, the plunge speed was the main factor affecting heat generation, while in the steady state welding stage, the rotation speed played a more important role. The FPPW weld defect includes flash and incomplete penetration in the upper, middle and bottom interface with the substrate. To obtain defect free weld, the higher rotation speed and proper plunge depth were recommended.

Keywords: friction push plug welding, process parameter, weld defect, orthogonal design

Procedia PDF Downloads 135
28935 Understanding Personal Well-Being among Entrepreneurial Breadwinners: Bibliographic and Empirical Analyses of Relative Resource Theory

Authors: E. Fredrick Rice

Abstract:

Over the past three decades, a substantial body of academic literature has asserted that the pressure to maintain household income can negatively affect the personal well-being of breadwinners. Given that scholars have failed to thoroughly explore this phenomenon with breadwinners who are also business owners, theory has been underdeveloped in the entrepreneurial context. To identify the most appropriate theories to apply to entrepreneurs, the current paper utilized two approaches. First, a comprehensive bibliographic analysis was conducted focusing on works at the intersection of breadwinner status and well-being. Co-authorship and journal citation patterns highlighted relative resource theory as a boundary spanning approach with promising applications in the entrepreneurial space. To build upon this theory, regression analysis was performed using data from the Panel Study of Entrepreneurial Dynamics (PSED). Empirical results showed evidence for the effects of breadwinner status and household income on entrepreneurial well-being. Further, the findings suggest that it is not merely income or job status that predicts well-being, but one’s relative financial contribution compared to that of one’s non-breadwinning organizationally employed partner. This paper offers insight into how breadwinner status can be studied in relation to the entrepreneurial personality.

Keywords: breadwinner, entrepreneurship, household income, well-being.

Procedia PDF Downloads 157
28934 Testing the Life Cycle Theory on the Capital Structure Dynamics of Trade-Off and Pecking Order Theories: A Case of Retail, Industrial and Mining Sectors

Authors: Freddy Munzhelele

Abstract:

Setting: the empirical research has shown that the life cycle theory has an impact on the firms’ financing decisions, particularly the dividend pay-outs. Accordingly, the life cycle theory posits that as a firm matures, it gets to a level and capacity where it distributes more cash as dividends. On the other hand, the young firms prioritise investment opportunities sets and their financing; thus, they pay little or no dividends. The research on firms’ financing decisions also demonstrated, among others, the adoption of trade-off and pecking order theories on the dynamics of firms capital structure. The trade-off theory talks to firms holding a favourable position regarding debt structures particularly as to the cost and benefits thereof; and pecking order is concerned with firms preferring a hierarchical order as to choosing financing sources. The case of life cycle hypothesis explaining the financial managers’ decisions as regards the firms’ capital structure dynamics appears to be an interesting link, yet this link has been neglected in corporate finance research. If this link is to be explored as an empirical research, the financial decision-making alternatives will be enhanced immensely, since no conclusive evidence has been found yet as to the dynamics of capital structure. Aim: the aim of this study is to examine the impact of life cycle theory on the capital structure dynamics trade-off and pecking order theories of firms listed in retail, industrial and mining sectors of the JSE. These sectors are among the key contributors to the GDP in the South African economy. Design and methodology: following the postpositivist research paradigm, the study is quantitative in nature and utilises secondary data obtainable from the financial statements of sampled firm for the period 2010 – 2022. The firms’ financial statements will be extracted from the IRESS database. Since the data will be in panel form, a combination of the static and dynamic panel data estimators will used to analyse data. The overall data analyses will be done using STATA program. Value add: this study directly investigates the link between the life cycle theory and the dynamics of capital structure decisions, particularly the trade-off and pecking order theories.

Keywords: life cycle theory, trade-off theory, pecking order theory, capital structure, JSE listed firms

Procedia PDF Downloads 48
28933 Design of Knowledge Management System with Geographic Information System

Authors: Angga Hidayah Ramadhan, Luciana Andrawina, M. Azani Hasibuan

Abstract:

Data will be as a core of the decision if it has a good treatment or process, which is process that data into information, and information into knowledge to make a wisdom or decision. Today, many companies have not realize it include XYZ University Admission Directorate as executor of National Admission called Seleksi Masuk Bersama (SMB) that during the time, the workers only uses their feeling to make a decision. Whereas if it done, then that company can analyze the data to make a right decision to get a pin sales from student candidate or registrant that follow SMB as many as possible. Therefore, needs Knowledge Management System (KMS) with Geographic Information System (GIS) use 5C4C that can process that company data becomes more useful and can help make decisions. This information system can process data into information based on the pin sold data with 5C (Contextualized, Categorize, Calculation, Correction, Condensed) and convert information into knowledge with 4C (Comparing, Consequence, Connection, Conversation) that has been several steps until these data can be useful to make easier to take a decision or wisdom, resolve problems, communicate, and quicker to learn to the employees have not experience and also for ease of viewing/visualization based on spatial data that equipped with GIS functionality that can be used to indicate events in each province with indicator that facilitate in this system. The system also have a function to save the tacit on the system then to be proceed into explicit in expert system based on the problems that will be found from the consequences of information. With the system each team can make a decision with same ways, structured, and the important is based on the actual event/data.

Keywords: 5C4C, data, information, knowledge

Procedia PDF Downloads 446
28932 The Effects of Interest Rates on Islamic Banks in a Dual Banking System: Empirical Evidence from Saudi Arabia

Authors: Mouldi Djelassi, Jamel Boukhatem

Abstract:

Background: A relation has been established between Islamic banks' activities and interest rates. The aim of this study was to explore the impact of interest rates on the deposits and loans held by Islamic and conventional banks in Saudi Arabia. Methods: A time series data was performed over the period 2008Q1-2020Q2 on eight conventional banks and four Islamic banks. The impacts of interest rate shocks on deposits and loans were identified through panel vector autoregressive models. Results: Impulse response function analysis showed that increasing interest rates reduce loans and conventional deposits. For Islamic banks, deposits are more affected by interest rates than lending. Variance decomposition analysis revealed that deposits contribute to 61% of the Islamic financing variation and only 25% of the conventional loans. Conclusion: Interest rates impacted Islamic banks especially through deposits, which is inconsistent with the theoretical framework. Islamic deposits played an important role in Islamic financing variation and may provide to be a channel for the transmission of the monetary policy in a dual banking system. Monetary policy in Saudi Arabia works in part through “credits” (conventional bank credits) as well as through “money” (conventional and Islamic bank deposits).

Keywords: Islamic banking, interest rates, monetary policy transmission, panel VAR

Procedia PDF Downloads 95
28931 TRAC: A Software Based New Track Circuit for Traffic Regulation

Authors: Jérôme de Reffye, Marc Antoni

Abstract:

Following the development of the ERTMS system, we think it is interesting to develop another software-based track circuit system which would fit secondary railway lines with an easy-to-work implementation and a low sensitivity to rail-wheel impedance variations. We called this track circuit 'Track Railway by Automatic Circuits.' To be internationally implemented, this system must not have any mechanical component and must be compatible with existing track circuit systems. For example, the system is independent from the French 'Joints Isolants Collés' that isolate track sections from one another, and it is equally independent from component used in Germany called 'Counting Axles,' in French 'compteur d’essieux.' This track circuit is fully interoperable. Such universality is obtained by replacing the train detection mechanical system with a space-time filtering of train position. The various track sections are defined by the frequency of a continuous signal. The set of frequencies related to the track sections is a set of orthogonal functions in a Hilbert Space. Thus the failure probability of track sections separation is precisely calculated on the basis of signal-to-noise ratio. SNR is a function of the level of traction current conducted by rails. This is the reason why we developed a very powerful algorithm to reject noise and jamming to obtain an SNR compatible with the precision required for the track circuit and SIL 4 level. The SIL 4 level is thus reachable by an adjustment of the set of orthogonal functions. Our major contributions to railway engineering signalling science are i) Train space localization is precisely defined by a calibration system. The operation bypasses the GSM-R radio system of the ERTMS system. Moreover, the track circuit is naturally protected against radio-type jammers. After the calibration operation, the track circuit is autonomous. ii) A mathematical topology adapted to train space localization by following the train through a linear time filtering of the received signal. Track sections are numerically defined and can be modified with a software update. The system was numerically simulated, and results were beyond our expectations. We achieved a precision of one meter. Rail-ground and rail-wheel impedance sensitivity analysis gave excellent results. Results are now complete and ready to be published. This work was initialised as a research project of the French Railways developed by the Pi-Ramses Company under SNCF contract and required five years to obtain the results. This track circuit is already at Level 3 of the ERTMS system, and it will be much cheaper to implement and to work. The traffic regulation is based on variable length track sections. As the traffic growths, the maximum speed is reduced, and the track section lengths are decreasing. It is possible if the elementary track section is correctly defined for the minimum speed and if every track section is able to emit with variable frequencies.

Keywords: track section, track circuits, space-time crossing, adaptive track section, automatic railway signalling

Procedia PDF Downloads 317
28930 Educators’ Adherence to Learning Theories and Their Perceptions on the Advantages and Disadvantages of E-Learning

Authors: Samson T. Obafemi, Seraphin D. Eyono-Obono

Abstract:

Information and Communication Technologies (ICTs) are pervasive nowadays, including in education where they are expected to improve the performance of learners. However, the hope placed in ICTs to find viable solutions to the problem of poor academic performance in schools in the developing world has not yet yielded the expected benefits. This problem serves as a motivation to this study whose aim is to examine the perceptions of educators on the advantages and disadvantages of e-learning. This aim will be subdivided into two types of research objectives. Objectives on the identification and design of theories and models will be achieved using content analysis and literature review. However, the objective on the empirical testing of such theories and models will be achieved through the survey of educators from different schools in the Pinetown District of the South African Kwazulu-Natal province. SPSS is used to quantitatively analyse the data collected by the questionnaire of this survey using descriptive statistics and Pearson correlations after assessing the validity and the reliability of the data. The main hypothesis driving this study is that there is a relationship between the demographics of educators’ and their adherence to learning theories on one side, and their perceptions on the advantages and disadvantages of e-learning on the other side, as argued by existing research; but this research views these learning theories under three perspectives: educators’ adherence to self-regulated learning, to constructivism, and to progressivism. This hypothesis was fully confirmed by the empirical study except for the demographic factor where teachers’ level of education was found to be the only demographic factor affecting the perceptions of educators on the advantages and disadvantages of e-learning.

Keywords: academic performance, e-learning, learning theories, teaching and learning

Procedia PDF Downloads 265
28929 An Evaluation of Cognitive Function Level, Depression, and Quality of Life of Elderly People Living in a Nursing Home

Authors: Ayse Inel Manav, Saliha Bozdogan Yesilot, Pinar Yesil Demirci, Gursel Oztunc

Abstract:

Introduction: This study was conducted with a view to evaluating cognitive function level, depression, and quality of life of elderly people living in a nursing home. Methods: This study, which is cross-sectional and descriptive in nature, was conducted in the Nursing and Rehabilitation Center for the Elderly in Adana/Turkey between 1st of May and 1st of August, 2016. The participants included 118 elderly people who were chosen using simple random sampling method. The data were collected using the Personal Information Form, the Standardized Mini Mental State Exam (SMMSE), the Geriatric Depression Scale (GDS), and the World Health Organization Quality of Life-OLD (WHOQOL-OLD) module. The data were analyzed using IBM SPSS Statistics 22 (IBM, SPSS, Turkey) program. Results: Of all the participants, 36,4% (n=43) were female, 63,6% (n=75) were male, and average age was 74,08 ± 8,23 years. The participants’ SMMSE mean score was found 20,37 ± 7,08, GDS mean score was 14,92 ± 4,29, and WHOQOL-OLD module mean score was 69,76 ± 11,54. There was a negative, significant relationship between SMMSE and GDS scores, a positive relationship between WHOQOL-OLD module total scores and a negative, significant relationship between GDS scores and WHOQOL-OLD module total scores. Discussıon and Conclusion: Results showed that more than half of the elderly people living in the nursing home experienced cognitive deterioration and depression; and cognitive state, depression, and quality of life were found to be significantly related to each other.

Keywords: depression, cognitive function level, quality of life

Procedia PDF Downloads 281
28928 Accurate Calculation of the Penetration Depth of a Bullet Using ANSYS

Authors: Eunsu Jang, Kang Park

Abstract:

In developing an armored ground combat vehicle (AGCV), it is a very important step to analyze the vulnerability (or the survivability) of the AGCV against enemy’s attack. In the vulnerability analysis, the penetration equations are usually used to get the penetration depth and check whether a bullet can penetrate the armor of the AGCV, which causes the damage of internal components or crews. The penetration equations are derived from penetration experiments which require long time and great efforts. However, they usually hold only for the specific material of the target and the specific type of the bullet used in experiments. Thus, penetration simulation using ANSYS can be another option to calculate penetration depth. However, it is very important to model the targets and select the input parameters in order to get an accurate penetration depth. This paper performed a sensitivity analysis of input parameters of ANSYS on the accuracy of the calculated penetration depth. Two conflicting objectives need to be achieved in adopting ANSYS in penetration analysis: maximizing the accuracy of calculation and minimizing the calculation time. To maximize the calculation accuracy, the sensitivity analysis of the input parameters for ANSYS was performed and calculated the RMS error with the experimental data. The input parameters include mesh size, boundary condition, material properties, target diameter are tested and selected to minimize the error between the calculated result from simulation and the experiment data from the papers on the penetration equation. To minimize the calculation time, the parameter values obtained from accuracy analysis are adjusted to get optimized overall performance. As result of analysis, the followings were found: 1) As the mesh size gradually decreases from 0.9 mm to 0.5 mm, both the penetration depth and calculation time increase. 2) As diameters of the target decrease from 250mm to 60 mm, both the penetration depth and calculation time decrease. 3) As the yield stress which is one of the material property of the target decreases, the penetration depth increases. 4) The boundary condition with the fixed side surface of the target gives more penetration depth than that with the fixed side and rear surfaces. By using above finding, the input parameters can be tuned to minimize the error between simulation and experiments. By using simulation tool, ANSYS, with delicately tuned input parameters, penetration analysis can be done on computer without actual experiments. The data of penetration experiments are usually hard to get because of security reasons and only published papers provide them in the limited target material. The next step of this research is to generalize this approach to anticipate the penetration depth by interpolating the known penetration experiments. This result may not be accurate enough to be used to replace the penetration experiments, but those simulations can be used in the early stage of the design process of AGCV in modelling and simulation stage.

Keywords: ANSYS, input parameters, penetration depth, sensitivity analysis

Procedia PDF Downloads 379
28927 An Adaptive Decomposition for the Variability Analysis of Observation Time Series in Geophysics

Authors: Olivier Delage, Thierry Portafaix, Hassan Bencherif, Guillaume Guimbretiere

Abstract:

Most observation data sequences in geophysics can be interpreted as resulting from the interaction of several physical processes at several time and space scales. As a consequence, measurements time series in geophysics have often characteristics of non-linearity and non-stationarity and thereby exhibit strong fluctuations at all time-scales and require a time-frequency representation to analyze their variability. Empirical Mode Decomposition (EMD) is a relatively new technic as part of a more general signal processing method called the Hilbert-Huang transform. This analysis method turns out to be particularly suitable for non-linear and non-stationary signals and consists in decomposing a signal in an auto adaptive way into a sum of oscillating components named IMFs (Intrinsic Mode Functions), and thereby acts as a bank of bandpass filters. The advantages of the EMD technic are to be entirely data driven and to provide the principal variability modes of the dynamics represented by the original time series. However, the main limiting factor is the frequency resolution that may give rise to the mode mixing phenomenon where the spectral contents of some IMFs overlap each other. To overcome this problem, J. Gilles proposed an alternative entitled “Empirical Wavelet Transform” (EWT) which consists in building from the segmentation of the original signal Fourier spectrum, a bank of filters. The method used is based on the idea utilized in the construction of both Littlewood-Paley and Meyer’s wavelets. The heart of the method lies in the segmentation of the Fourier spectrum based on the local maxima detection in order to obtain a set of non-overlapping segments. Because linked to the Fourier spectrum, the frequency resolution provided by EWT is higher than that provided by EMD and therefore allows to overcome the mode-mixing problem. On the other hand, if the EWT technique is able to detect the frequencies involved in the original time series fluctuations, EWT does not allow to associate the detected frequencies to a specific mode of variability as in the EMD technic. Because EMD is closer to the observation of physical phenomena than EWT, we propose here a new technic called EAWD (Empirical Adaptive Wavelet Decomposition) based on the coupling of the EMD and EWT technics by using the IMFs density spectral content to optimize the segmentation of the Fourier spectrum required by EWT. In this study, EMD and EWT technics are described, then EAWD technic is presented. Comparison of results obtained respectively by EMD, EWT and EAWD technics on time series of ozone total columns recorded at Reunion island over [1978-2019] period is discussed. This study was carried out as part of the SOLSTYCE project dedicated to the characterization and modeling of the underlying dynamics of time series issued from complex systems in atmospheric sciences

Keywords: adaptive filtering, empirical mode decomposition, empirical wavelet transform, filter banks, mode-mixing, non-linear and non-stationary time series, wavelet

Procedia PDF Downloads 122