Search results for: empirical mapping
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1087

Search results for: empirical mapping

277 Wind Power Mapping and NPV of Embedded Generation Systems in Nigeria

Authors: Oluseyi O. Ajayi, Ohiose D. Ohijeagbon, Mercy Ogbonnaya, Ameh Attabo

Abstract:

The study assessed the potential and economic viability of stand-alone wind systems for embedded generation, taking into account its benefits to small off-grid rural communities at 40 meteorological sites in Nigeria. A specific electric load profile was developed to accommodate communities consisting of 200 homes, a school and a community health centre. This load profile was incorporated within the distributed generation analysis producing energy in the MW range, while optimally meeting daily load demand for the rural communities. Twenty-four years (1987 to 2010) of wind speed data at a height of 10m utilized for the study were sourced from the Nigeria Meteorological Department, Oshodi. The HOMER® software optimizing tool was engaged for the feasibility study and design. Each site was suited to 3MW wind turbines in sets of five, thus 15MW was designed for each site. This design configuration was adopted in order to easily compare the distributed generation system amongst the sites to determine their relative economic viability in terms of life cycle cost, as well as levelised cost of producing energy. A net present value was estimated in terms of life cycle cost for 25 of the 40 meteorological sites. On the other hand, the remaining sites yielded a net present cost; meaning the installations at these locations were not economically viable when utilizing the present tariff regime for embedded generation in Nigeria.

Keywords: Wind speed, wind power, distributed generation, cost per kilowatt-hour, clean energy, Nigeria.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1400
276 To Know the Way to the Unknown: A Semi-Experimental Study on the Implication of Skills and Knowledge for Creative Processes in Higher Education

Authors: Mikkel Snorre Wilms Boysen

Abstract:

From a theoretical perspective, expertise is generally considered a precondition for creativity. The assumption is that an individual needs to master the common and accepted rules and techniques within a certain knowledge-domain in order to create something new and valuable. However, real life cases, and a limited amount of empirical studies, demonstrate that this assumption may be overly simple. In this article, this question is explored through a number of semi-experimental case studies conducted within the fields of music, technology, and youth culture. The studies indicate that, in various ways, expertise plays an important part in creative processes. However, the case studies also indicate that expertise sometimes leads to an entrenched perspective, in the sense that knowledge and experience may work as a path into the well-known rather than into the unknown. In this article, these issues are explored with reference to different theoretical approaches to creativity and learning, including actor-network theory, the theory of blind variation and selective retention, and Csikszentmihalyi’s system model. Finally, some educational aspects and implications of this are discussed.

Keywords: Creativity, education, expertise, technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 709
275 Small Sample Bootstrap Confidence Intervals for Long-Memory Parameter

Authors: Josu Arteche, Jesus Orbe

Abstract:

The log periodogram regression is widely used in empirical applications because of its simplicity, since only a least squares regression is required to estimate the memory parameter, d, its good asymptotic properties and its robustness to misspecification of the short term behavior of the series. However, the asymptotic distribution is a poor approximation of the (unknown) finite sample distribution if the sample size is small. Here the finite sample performance of different nonparametric residual bootstrap procedures is analyzed when applied to construct confidence intervals. In particular, in addition to the basic residual bootstrap, the local and block bootstrap that might adequately replicate the structure that may arise in the errors of the regression are considered when the series shows weak dependence in addition to the long memory component. Bias correcting bootstrap to adjust the bias caused by that structure is also considered. Finally, the performance of the bootstrap in log periodogram regression based confidence intervals is assessed in different type of models and how its performance changes as sample size increases.

Keywords: bootstrap, confidence interval, log periodogram regression, long memory.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1701
274 Disparities versus Similarities: WHO GPPQCL and ISO/IEC 17025:2017 International Standards for Quality Management Systems in Pharmaceutical Laboratories

Authors: M. A. Okezue, K. L. Clase, S. R. Byrn, P. Shivanand

Abstract:

Medicines regulatory authorities expect pharmaceutical companies and contract research organizations to seek ways to certify that their laboratory control measurements are reliable. Establishing and maintaining laboratory quality standards are essential in ensuring the accuracy of test results. ‘ISO/IEC 17025:2017’ and ‘WHO Good Practices for Pharmaceutical Quality Control Laboratories (GPPQCL)’ are two quality standards commonly employed in developing laboratory quality systems. A review was conducted on the two standards to elaborate on areas on convergence and divergence. The goal was to understand how differences in each standard's requirements may influence laboratories' choices as to which document is easier to adopt for quality systems. A qualitative review method compared similar items in the two standards while mapping out areas where there were specific differences in the requirements of the two documents. The review also provided a detailed description of the clauses and parts covering management and technical requirements in these laboratory standards. The review showed that both documents share requirements for over ten critical areas covering objectives, infrastructure, management systems, and laboratory processes. There were, however, differences in standard expectations where GPPQCL emphasizes system procedures for planning and future budgets that will ensure continuity. Conversely, ISO 17025 was more focused on the risk management approach to establish laboratory quality systems. Elements in the two documents form common standard requirements to assure the validity of laboratory test results that promote mutual recognition. The ISO standard currently has more global patronage than GPPQCL.

Keywords: ISO/IEC 17025:2017, laboratory standards, quality control, WHO GPPQCL

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 996
273 Between Legal Authority and Epistemic Competence: A Case Study of the Brazilian Supreme Court

Authors: Júlia Massadas

Abstract:

The objective of this paper is to analyze the role played by the institute of the public hearings in the Brazilian Supreme Court. The public hearings are regulated since 1999 by the Brazilian Laws nº 9.868, nº 9.882 and by the Intern Regiment of the Brazilian Supreme Court. According to this legislation, the public hearings are supposed to be called when a matter of circumstance of fact must be clarified, what can be done through the hearing of the testimonies of persons with expertise and authority in the theme related to the cause. This work aims to investigate what is the role played by the public hearings and by the experts in the Brazilian Supreme Court. The hypothesis of this research is that: (I) The public hearings in the Brazilian Supreme Court are used to uphold a rhetoric of a democratic legitimacy of the Court`s decisions; (II) The Legislative intentions have been distorted. To test this hypothesis, the adopted methodology involves an empirical study of the Brazilian jurisprudence. As a conclusion, it follows that the public hearings convened by the Brazilian Supreme Court do not correspond, in practice, to the role assigned to them by the Congress since they do not serve properly to epistemic interests. The public hearings not only do not legitimate democratically the decisions, but also, do not properly clarify technical issues.

Keywords: Brazilian Supreme Court, constitutional law, public hearings, epistemic competence, legal authority.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1540
272 An Approach to Measure Snow Depth of Winter Accumulation at Basin Scale Using Satellite Data

Authors: M. Geetha Priya, D. Krishnaveni

Abstract:

Snow depth estimation and monitoring studies have been carried out for decades using empirical relationship or extrapolation of point measurements carried out in field. With the development of advanced satellite based remote sensing techniques, a modified approach is proposed in the present study to estimate the winter accumulated snow depth at basin scale. Assessment of snow depth by differencing Digital Elevation Model (DEM) generated at the beginning and end of winter season can be experimented for the region of interest (Himalayan and polar regions) accounting for winter accumulation (solid precipitation). The proposed approach is based on existing geodetic method that is being used for glacier mass balance estimation. Considering the satellite datasets purely acquired during beginning and end of winter season, it is possible to estimate the change in depth or thickness for the snow that is accumulated during the winter as it takes one year for the snow to get transformed into firn (snow that has survived one summer or one-year old snow).

Keywords: Digital elevation model, snow depth, geodetic method, snow cover.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 666
271 Inter-Organizational Knowledge Transfer Through Malaysia E-government IT Outsourcing: A Theoretical Review

Authors: Nor Aziati Abdul Hamid, Juhana Salim

Abstract:

The main objective of this paper is to contribute the existing knowledge transfer and IT Outsourcing literature specifically in the context of Malaysia by reviewing the current practices of e-government IT outsourcing in Malaysia including the issues and challenges faced by the public agencies in transferring the knowledge during the engagement. This paper discusses various factors and different theoretical model of knowledge transfer starting from the traditional model to the recent model suggested by the scholars. The present paper attempts to align organizational knowledge from the knowledge-based view (KBV) and organizational learning (OL) lens. This review could help shape the direction of both future theoretical and empirical studies on inter-firm knowledge transfer specifically on how KBV and OL perspectives could play significant role in explaining the complex relationships between the client and vendor in inter-firm knowledge transfer and the role of organizational management information system and Transactive Memory System (TMS) to facilitate the organizational knowledge transferring process. Conclusion is drawn and further research is suggested.

Keywords: E-government, IT Outsourcing, Knowledge Management, Knowledge Transfer

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2326
270 Shaft Friction of Bored Pile Socketed in Weathered Limestone in Qatar

Authors: Thanawat Chuleekiat

Abstract:

Socketing of bored piles in rock is always seen as a matter of debate on construction sites between consultants and contractors. The socketing depth normally depends on the type of rock, depth at which the rock is available below the pile cap and load carrying capacity of the pile. In this paper, the review of field load test data of drilled shaft socketed in weathered limestone conducted using conventional static pile load test and dynamic pile load test was made to evaluate a unit shaft friction for the bored piles socketed in weathered limestone (weak rock). The borehole drilling data were also reviewed in conjunction with the pile test result. In addition, the back-calculated unit shaft friction was reviewed against various empirical methods for bored piles socketed in weak rock. The paper concludes with an estimated ultimate unit shaft friction from the case study in Qatar for preliminary design.

Keywords: Piled foundation, weathered limestone, shaft friction, rock socket, pile load test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1046
269 Guidelines for Developing, Supervising, Assessing and Evaluating Capstone Design Project of BSc in Electrical and Electronic Engineering Program

Authors: Muhibul Haque Bhuyan

Abstract:

Inclusion of any design project in an undergraduate electrical and electronic engineering curriculum and producing creative ideas in the final year capstone design projects have received numerous comments at the Board of Accreditation for Engineering and Technical Education (BAETE) several times by the mentors and visiting program evaluator team members at different public and private universities in Bangladesh. To eradicate this deficiency which is needed for getting the program accreditation, a thorough change was required in the Department of Electrical and Electronic Engineering (EEE) for its BSc in EEE program at Southeast University, Dhaka, Bangladesh. We suggested making changes in the course curriculum titles and contents, emphasizing to include capstone design projects, question setting, examining students through other standard methods, selecting and retaining Outcome-Based Education (OBE)-oriented engineering faculty members, improving laboratories through purchasing new equipment and software as well as developing new experiments for each laboratory courses, and engaging the students to practical designs in various courses and final year projects. This paper reports on capstone design project course objectives, course outcomes, mapping with the program outcomes, cognitive domain of learning, assessment schemes, guidelines, suggestions and recommendations for supervision processes, assessment strategy, and rubric setting, etc. It is expected that this will substantially improve the capstone design projects offering, supervision, and assessment in the undergraduate EEE program to fulfill the arduous requirements of BAETE accreditation based on OBE.

Keywords: Course outcome, capstone design project, assessment and evaluation, electrical and electronic engineering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 468
268 Acceleration-Based Motion Model for Visual SLAM

Authors: Daohong Yang, Xiang Zhang, Wanting Zhou, Lei Li

Abstract:

Visual Simultaneous Localization and Mapping (VSLAM) is a technology that gathers information about the surrounding environment to ascertain its own position and create a map. It is widely used in computer vision, robotics, and various other fields. Many visual SLAM systems, such as OBSLAM3, utilize a constant velocity motion model. The utilization of this model facilitates the determination of the initial pose of the current frame, thereby enhancing the efficiency and precision of feature matching. However, it is often difficult to satisfy the constant velocity motion model in actual situations. This can result in a significant deviation between the obtained initial pose and the true value, leading to errors in nonlinear optimization results. Therefore, this paper proposes a motion model based on acceleration that can be applied to most SLAM systems. To provide a more accurate description of the camera pose acceleration, we separate the pose transformation matrix into its rotation matrix and translation vector components. The rotation matrix is now represented by a rotation vector. We assume that, over a short period, the changes in rotating angular velocity and translation vector remain constant. Based on this assumption, the initial pose of the current frame is estimated. In addition, the error of the constant velocity model is analyzed theoretically. Finally, we apply our proposed approach to the ORBSLAM3 system and evaluate two sets of sequences from the TUM datasets. The results show that our proposed method has a more accurate initial pose estimation, resulting in an improvement of 6.61% and 6.46% in the accuracy of the ORBSLAM3 system on the two test sequences, respectively.

Keywords: Error estimation, constant acceleration motion model, pose estimation, visual SLAM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155
267 Modeling and Simulation of Acoustic Link Using Mackenize Propagation Speed Equation

Authors: Christhu Raj M. R., Rajeev Sukumaran

Abstract:

Underwater acoustic networks have attracted great attention in the last few years because of its numerous applications. High data rate can be achieved by efficiently modeling the physical layer in the network protocol stack. In Acoustic medium, propagation speed of the acoustic waves is dependent on many parameters such as temperature, salinity, density, and depth. Acoustic propagation speed cannot be modeled using standard empirical formulas such as Urick and Thorp descriptions. In this paper, we have modeled the acoustic channel using real time data of temperature, salinity, and speed of Bay of Bengal (Indian Coastal Region). We have modeled the acoustic channel by using Mackenzie speed equation and real time data obtained from National Institute of Oceanography and Technology. It is found that acoustic propagation speed varies between 1503 m/s to 1544 m/s as temperature and depth differs. The simulation results show that temperature, salinity, depth plays major role in acoustic propagation and data rate increases with appropriate data sets substituted in the simulated model.

Keywords: Underwater Acoustics, Mackenzie Speed Equation, Temperature, Salinity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2145
266 Underpricing of IPOs during Hot and Cold Market Periods on the South African Stock Exchange (JSE)

Authors: Brownhilder N. Neneh, A. Van Aardt Smit

Abstract:

Underpricing is one anomaly in initial public offerings (IPO) literature that has been widely observed across different stock markets with different trends emerging over different time periods. This study seeks to determine how IPOs on the JSE performed on the first day, first week and first month over the period of 1996-2011. Underpricing trends are documented for both hot and cold market periods in terms of four main sectors (cyclical, defensive, growth stock and interest rate sensitive stocks). Using a sample of 360 listed companies on the JSE, the empirical findings established that IPOs on the JSE are significantly underpriced with an average market adjusted first day return of 62.9%. It is also established that hot market IPOs on the JSE are more underpriced than the cold market IPOs. Also observed is the fact that as the offer price per share increases above the median price for any given period, the level of underpricing decreases substantially. While significant differences exist in the level of underpricing of IPOs in the four different sectors in the hot and cold market periods, interest rates sensitive stocks showed a different trend from the other sectors and thus require further investigation to uncover this pattern.

Keywords: Underpricing, hot and cold markets, South Africa, JSE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4150
265 Trusting Smart Speakers: Analysing the Different Levels of Trust between Technologies

Authors: Alec Wells, Aminu Bello Usman, Justin McKeown

Abstract:

The growing usage of smart speakers raises many privacy and trust concerns compared to other technologies such as smart phones and computers. In this study, a proxy measure of trust is used to gauge users’ opinions on three different technologies based on an empirical study, and to understand which technology most people are most likely to trust. The collected data were analysed using the Kruskal-Wallis H test to determine the statistical differences between the users’ trust level of the three technologies: smart speaker, computer and smart phone. The findings of the study revealed that despite the wide acceptance, ease of use and reputation of smart speakers, people find it difficult to trust smart speakers with their sensitive information via the Direct Voice Input (DVI) and would prefer to use a keyboard or touchscreen offered by computers and smart phones. Findings from this study can inform future work on users’ trust in technology based on perceived ease of use, reputation, perceived credibility and risk of using technologies via DVI.

Keywords: Direct voice input, risk, security, technology and trust.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 536
264 Predicting Global Solar Radiation Using Recurrent Neural Networks and Climatological Parameters

Authors: Rami El-Hajj Mohamad, Mahmoud Skafi, Ali Massoud Haidar

Abstract:

Several meteorological parameters were used for the  prediction of monthly average daily global solar radiation on  horizontal using recurrent neural networks (RNNs). Climatological  data and measures, mainly air temperature, humidity, sunshine  duration, and wind speed between 1995 and 2007 were used to design  and validate a feed forward and recurrent neural network based  prediction systems. In this paper we present our reference system  based on a feed-forward multilayer perceptron (MLP) as well as the  proposed approach based on an RNN model. The obtained results  were promising and comparable to those obtained by other existing  empirical and neural models. The experimental results showed the  advantage of RNNs over simple MLPs when we deal with time series  solar radiation predictions based on daily climatological data.

Keywords: Recurrent Neural Networks, Global Solar Radiation, Multi-layer perceptron, gradient, Root Mean Square Error.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2519
263 Do C-Test and Cloze Procedure Measure what they Purport to be Measuring? A Case of Criterion-Related Validity

Authors: Masoud Saeedi, Mansour Tavakoli, Shirin Rahimi Kazerooni, Vahid Parvaresh

Abstract:

This article investigated the validity of C-test and Cloze test which purport to measure general English proficiency. To provide empirical evidence pertaining to the validity of the interpretations based on the results of these integrative language tests, their criterion-related validity was investigated. In doing so, the test of English as a foreign language (TOEFL) which is an established, standardized, and internationally administered test of general English proficiency was used as the criterion measure. Some 90 Iranian English majors participated in this study. They were seniors studying English at a university in Tehran, Iran. The results of analyses showed that there is a statistically significant correlation among participants- scores on Cloze test, C-test, and the TOEFL. Building on the findings of the study and considering criterion-related validity as the evidential basis of the validity argument, it was cautiously deducted that these tests measure the same underlying trait. However, considering the limitations of using criterion measures to validate tests, no absolute claims can be made as to the construct validity of these integrative tests.

Keywords: Integrative testing, C-test, Cloze test, theTOEFL, Validity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3273
262 Performance Evaluation of Data Mining Techniques for Predicting Software Reliability

Authors: Pradeep Kumar, Abdul Wahid

Abstract:

Accurate software reliability prediction not only enables developers to improve the quality of software but also provides useful information to help them for planning valuable resources. This paper examines the performance of three well-known data mining techniques (CART, TreeNet and Random Forest) for predicting software reliability. We evaluate and compare the performance of proposed models with Cascade Correlation Neural Network (CCNN) using sixteen empirical databases from the Data and Analysis Center for Software. The goal of our study is to help project managers to concentrate their testing efforts to minimize the software failures in order to improve the reliability of the software systems. Two performance measures, Normalized Root Mean Squared Error (NRMSE) and Mean Absolute Errors (MAE), illustrate that CART model is accurate than the models predicted using Random Forest, TreeNet and CCNN in all datasets used in our study. Finally, we conclude that such methods can help in reliability prediction using real-life failure datasets.

Keywords: Classification, Cascade Correlation Neural Network, Random Forest, Software reliability, TreeNet.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1801
261 Convergence Analysis of Training Two-Hidden-Layer Partially Over-Parameterized ReLU Networks via Gradient Descent

Authors: Zhifeng Kong

Abstract:

Over-parameterized neural networks have attracted a great deal of attention in recent deep learning theory research, as they challenge the classic perspective of over-fitting when the model has excessive parameters and have gained empirical success in various settings. While a number of theoretical works have been presented to demystify properties of such models, the convergence properties of such models are still far from being thoroughly understood. In this work, we study the convergence properties of training two-hidden-layer partially over-parameterized fully connected networks with the Rectified Linear Unit activation via gradient descent. To our knowledge, this is the first theoretical work to understand convergence properties of deep over-parameterized networks without the equally-wide-hidden-layer assumption and other unrealistic assumptions. We provide a probabilistic lower bound of the widths of hidden layers and proved linear convergence rate of gradient descent. We also conducted experiments on synthetic and real-world datasets to validate our theory.

Keywords: Over-parameterization, Rectified Linear Units (ReLU), convergence, gradient descent, neural networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 816
260 An Experimental Consideration of the Hybrid Architecture Based on the Situated Action Generator

Authors: Serin Lee, Takashi Kubota, Ichiro Nakatani

Abstract:

The approaches to make an agent generate intelligent actions in the AI field might be roughly categorized into two ways–the classical planning and situated action system. It is well known that each system have its own strength and weakness. However, each system also has its own application field. In particular, most of situated action systems do not directly deal with the logical problem. This paper first briefly mentions the novel action generator to situatedly extract a set of actions, which is likely to help to achieve the goal at the current situation in the relaxed logical space. After performing the action set, the agent should recognize the situation for deciding the next likely action set. However, since the extracted action is an approximation of the action which helps to achieve the goal, the agent could be caught into the deadlock of the problem. This paper proposes the newly developed hybrid architecture to solve the problem, which combines the novel situated action generator with the conventional planner. The empirical result in some planning domains shows that the quality of the resultant path to the goal is mostly acceptable as well as deriving the fast response time, and suggests the correlation between the structure of problems and the organization of each system which generates the action.

Keywords: Situated reasoning, situated action, planning, hybrid architecture

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1091
259 Volatility Switching between Two Regimes

Authors: Josip Visković, Josip Arnerić, Ante Rozga

Abstract:

Based on the fact that volatility is time varying in high frequency data and that periods of high volatility tend to cluster, the most successful and popular models in modeling time varying volatility are GARCH type models. When financial returns exhibit sudden jumps that are due to structural breaks, standard GARCH models show high volatility persistence, i.e. integrated behavior of the conditional variance. In such situations models in which the parameters are allowed to change over time are more appropriate. This paper compares different GARCH models in terms of their ability to describe structural changes in returns caused by financial crisis at stock markets of six selected central and east European countries. The empirical analysis demonstrates that Markov regime switching GARCH model resolves the problem of excessive persistence and outperforms uni-regime GARCH models in forecasting volatility when sudden switching occurs in response to financial crisis.

Keywords: Central and east European countries, financial crisis, Markov switching GARCH model, transition probabilities.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2483
258 Human Resources and Business Result: An Empirical Approach Based On RBV Theory

Authors: XhevrieMamaqi

Abstract:

Organization capacity learning is a process referring to the sum total of individual and collective learning through training programs, experience and experimentation, among others. Today, in-business ongoing training is one of the most important strategies for human capital development and it is crucial to sustain and improve workers’ knowledge and skills. Many organizations, firms and business are adopting a strategy of continuous learning, encouraging employees to learn new skills continually to be innovative and to try new processes and work in order to achieve a competitive advantage and superior business results. This paper uses the Resource Based View and Capacities (RBV) approach to construct a hypothetical relationships model between training and business results. The test of the model is applied on transversal data. A sample of 266 business of Spanish sector service has been selected. A Structural Equation Model (SEM) is used to estimate the relationship between ongoing training, represented by two latent dimension denominated Human and Social Capital resources and economic business results. The coefficients estimated have shown the efficient of some training aspectsexplaining the variation in business results.

Keywords: Business results, Human and Social Capital resources, training, RBV Theory, SEM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1819
257 Cross Project Software Fault Prediction at Design Phase

Authors: Pradeep Singh, Shrish Verma

Abstract:

Software fault prediction models are created by using the source code, processed metrics from the same or previous version of code and related fault data. Some company do not store and keep track of all artifacts which are required for software fault prediction. To construct fault prediction model for such company, the training data from the other projects can be one potential solution. Earlier we predicted the fault the less cost it requires to correct. The training data consists of metrics data and related fault data at function/module level. This paper investigates fault predictions at early stage using the cross-project data focusing on the design metrics. In this study, empirical analysis is carried out to validate design metrics for cross project fault prediction. The machine learning techniques used for evaluation is Naïve Bayes. The design phase metrics of other projects can be used as initial guideline for the projects where no previous fault data is available. We analyze seven datasets from NASA Metrics Data Program which offer design as well as code metrics. Overall, the results of cross project is comparable to the within company data learning.

Keywords: Software Metrics, Fault prediction, Cross project, Within project.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2484
256 Rain Cell Ratio Technique in Path Attenuation for Terrestrial Radio Links

Authors: Peter Odero Akuon

Abstract:

A rain cell ratio model is proposed that computes attenuation of the smallest rain cell which represents the maximum rain rate value i.e. the cell size when rainfall rate is exceeded 0.01% of the time, R0.01 and predicts attenuation for other cells as the ratio with this maximum. This model incorporates the dependence of the path factor r on the ellipsoidal path variation of the Fresnel zone at different frequencies. In addition, the inhomogeneity of rainfall is modeled by a rain drop packing density factor. In order to derive the model, two empirical methods that can be used to find rain cell size distribution Dc are presented. Subsequently, attenuation measurements from different climatic zones for terrestrial radio links with frequencies F in the range 7-38 GHz are used to test the proposed model. Prediction results show that the path factor computed from the rain cell ratio technique has improved reliability when compared with other path factor and effective rain rate models, including the current ITU-R 530-15 model of 2013.

Keywords: Packing density of rain drops, prediction model, rain attenuation, rain cell ratio technique.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 648
255 Specialized Reduced Models of Dynamic Flows in 2-Stroke Engines

Authors: S. Cagin, X. Fischer, E. Delacourt, N. Bourabaa, C. Morin, D. Coutellier, B. Carré, S. Loumé

Abstract:

The complexity of scavenging by ports and its impact on engine efficiency create the need to understand and to model it as realistically as possible. However, there are few empirical scavenging models and these are highly specialized. In a design optimization process, they appear very restricted and their field of use is limited. This paper presents a comparison of two methods to establish and reduce a model of the scavenging process in 2-stroke diesel engines. To solve the lack of scavenging models, a CFD model has been developed and is used as the referent case. However, its large size requires a reduction. Two techniques have been tested depending on their fields of application: The NTF method and neural networks. They both appear highly appropriate drastically reducing the model’s size (over 90% reduction) with a low relative error rate (under 10%). Furthermore, each method produces a reduced model which can be used in distinct specialized fields of application: the distribution of a quantity (mass fraction for example) in the cylinder at each time step (pseudo-dynamic model) or the qualification of scavenging at the end of the process (pseudo-static model).

Keywords: Diesel engine, Design optimization, Model reduction, Neural network, NTF algorithm, Scavenging.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1277
254 Long Term Examination of the Profitability Estimation Focused on Benefits

Authors: Stephan Printz, Kristina Lahl, René Vossen, Sabina Jeschke

Abstract:

Strategic investment decisions are characterized by high innovation potential and long-term effects on the competitiveness of enterprises. Due to the uncertainty and risks involved in this complex decision making process, the need arises for well-structured support activities. A method that considers cost and the long-term added value is the cost-benefit effectiveness estimation. One of those methods is the “profitability estimation focused on benefits – PEFB”-method developed at the Institute of Management Cybernetics at RWTH Aachen University. The method copes with the challenges associated with strategic investment decisions by integrating long-term non-monetary aspects whilst also mapping the chronological sequence of an investment within the organization’s target system. Thus, this method is characterized as a holistic approach for the evaluation of costs and benefits of an investment. This participation-oriented method was applied to business environments in many workshops. The results of the workshops are a library of more than 96 cost aspects, as well as 122 benefit aspects. These aspects are preprocessed and comparatively analyzed with regards to their alignment to a series of risk levels. For the first time, an accumulation and a distribution of cost and benefit aspects regarding their impact and probability of occurrence are given. The results give evidence that the PEFB-method combines precise measures of financial accounting with the incorporation of benefits. Finally, the results constitute the basics for using information technology and data science for decision support when applying within the PEFB-method.

Keywords: Cost-benefit analysis, multi-criteria decision, profitability estimation focused on benefits, risk and uncertainty analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1447
253 Corporate Governance Practices and Audit Quality: An Empirical Study of the Listed Companies in Egypt

Authors: Mohamed Moustafa Soliman, Mohamed Abd Elsalam

Abstract:

Recent financial international scandals around the world have led to a number of investigations into the effectiveness of corporate governance practices and audit quality. Although evidence of corporate governance practices and audit quality exists from developed economies, very scanty studies have been conducted in Egypt where corporate governance is just evolving. Therefore, this study provides evidence on the effectiveness of corporate governance practices and audit quality from a developing country. The data for analysis are gathered from the top 50 most active companies in the Egyptian Stock Exchange, covering the three year period 2007-2009. Logistic regression was used in investigating the questions that were raised in the study. Findings from the study show that board independence; CEO duality and audit committees significantly have relationship with audit quality. The results also, indicate that institutional investor and managerial ownership have no significantly relationship with audit quality. Evidence also exist that size of the company; complexity and business leverage are important factors in audit quality for companies quoted on the Egypt Stock Exchange.

Keywords: Corporate governance, Boards of directors, corporate ownership, Audit Committees, Audit quality, and Egypt.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3877
252 The Grey Relational Analysis of the Influence Factors of Profit in Cartoon-s Character Merchandising Rights

Authors: Min Li, Tao Li

Abstract:

This paper constructs a four factors theoretical model of Chinese small and medium enterprises based on the “cartoon characters- reputation - enterprise marketing and management capabilities – protection of the cartoon image - institutional environment" by literature research, case studies and investigation. The empirical study show that the greatest impact on current merchandising rights income is the institutional environment friendliness, followed by marketing and management capabilities, input of character image protection and Cartoon characters- reputation through the real-time grey relational analysis, and the greatest impact on post-merchandising rights profit is Cartoon characters reputation, followed by the institutional environment friendliness, then marketing and management ability and input of character image protection through the time-delay grey relational analysis.

Keywords: Cartoon characters, merchandising rights, influencefactors, grey relational analysis

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1558
251 Knowledge and Organisational Success: Developing a Scale of Knowledge Framework

Authors: Mohammed Almohammedali, Peter Duncan, David Edgar

Abstract:

The aim of this exploratory research is to understand further how organisations can evaluate their activities, which generate knowledge creation, to meet changing stakeholder expectations. A Scale of Knowledge (SoK) Framework is proposed which links knowledge management and organisational activities to changing stakeholder expectations. The framework was informed by the knowledge management literature, as well as empirical work conducted via a single case study of a multi-site hospital organisation in Saudi Arabia. Eight in-depth semi-structured interviews were conducted with managers from across the organisation regarding current and future stakeholder expectations, organisational strategy/activities and knowledge management. Data were analysed using thematic analysis and a hierarchical value map technique to identify activities that can produce further knowledge and consequently impact on how stakeholder expectations are met. The SoK Framework developed may be useful to practitioners as an analytical aid to determine if current organisational activities produce organisational knowledge which helps them meet (increasingly higher levels of) stakeholder expectations. The limitations of the research and avenues for future development of the proposed framework are discussed.

Keywords: Knowledge creation, knowledge management, organisational knowledge, scale of knowledge, knowledge impact.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1663
250 Zero Carbon & Low Energy Housing; Comparative Analysis of Two Persian Vernacular Architectural Solutions to Increase Energy Efficiency

Authors: N. Poorang

Abstract:

In order to respond the human needs, all regional, social, and economical factors are available to gain residents’ comfort and ideal architecture. There is no doubt the thermal comfort has to satisfy people not only for daily and physical activities but also creating pleasant area for mental activities and relaxing. It costs energy and increases greenhouse gas emissions.

Reducing energy use in buildings is a critical component of meeting carbon reduction commitments. Hence housing design represents a major opportunity to cut energy use and CO2 emissions.

In terms of energy efficiency, it is vital to propose and research modern design methods for buildings however vernacular architecture techniques are proven empirical existing practices which have to be considered. This research tries to compare two architectural solution were proposed by Persian vernacular architecture, to achieve energy efficiency in hot areas.

The aim of this research is to analyze two forms of traditional Persian architecture in different locations in order to develop a systematic research and sustainable technologies on adaptation to contemporary living standards.

Keywords: Comparative Analysis, Persian Vernacular Architecture, Sustainable architecture.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2249
249 Modelling the Photovoltaic Pump Output Using Empirical Data from Local Conditions in the Vhembe District

Authors: C. Matasane, C. Dwarika, R. Naidoo

Abstract:

The mathematical analysis on radiation obtained and the development of the solar photovoltaic (PV) array groundwater pumping is needed in the rural areas of Thohoyandou for sizing and power performance subject to the climate conditions within the area. A simple methodology approach is developed for the directed coupled solar, controller and submersible ground water pump system. The system consists of a PV array, pump controller and submerged pump, battery backup and charger controller. For this reason, the theoretical solar radiation is obtained for optimal predictions and system performance in order to achieve different design and operating parameters. Here the examination of the PV schematic module in a Direct Current (DC) application is used for obtainable maximum solar power energy for water pumping. In this paper, a simple efficient photovoltaic water pumping system is presented with its theoretical studies and mathematical modeling of photovoltaics (PV) system.

Keywords: Renewable energy sources, solar groundwater pumping, theoretical and mathematical analysis of photovoltaic (PV) system, theoretical solar radiation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2478
248 A Set Theory Based Factoring Technique and Its Use for Low Power Logic Design

Authors: Padmanabhan Balasubramanian, Ryuta Arisaka

Abstract:

Factoring Boolean functions is one of the basic operations in algorithmic logic synthesis. A novel algebraic factorization heuristic for single-output combinatorial logic functions is presented in this paper and is developed based on the set theory paradigm. The impact of factoring is analyzed mainly from a low power design perspective for standard cell based digital designs in this paper. The physical implementation of a number of MCNC/IWLS combinational benchmark functions and sub-functions are compared before and after factoring, based on a simple technology mapping procedure utilizing only standard gate primitives (readily available as standard cells in a technology library) and not cells corresponding to optimized complex logic. The power results were obtained at the gate-level by means of an industry-standard power analysis tool from Synopsys, targeting a 130nm (0.13μm) UMC CMOS library, for the typical case. The wire-loads were inserted automatically and the simulations were performed with maximum input activity. The gate-level simulations demonstrate the advantage of the proposed factoring technique in comparison with other existing methods from a low power perspective, for arbitrary examples. Though the benchmarks experimentation reports mixed results, the mean savings in total power and dynamic power for the factored solution over a non-factored solution were 6.11% and 5.85% respectively. In terms of leakage power, the average savings for the factored forms was significant to the tune of 23.48%. The factored solution is expected to better its non-factored counterpart in terms of the power-delay product as it is well-known that factoring, in general, yields a delay-efficient multi-level solution.

Keywords: Factorization, Set theory, Logic function, Standardcell based design, Low power.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1755