Search results for: integrated audit quality model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10355

Search results for: integrated audit quality model

10115 A Hybrid Model of ARIMA and Multiple Polynomial Regression for Uncertainties Modeling of a Serial Production Line

Authors: Amir Azizi, Amir Yazid b. Ali, Loh Wei Ping, Mohsen Mohammadzadeh

Abstract:

Uncertainties of a serial production line affect on the production throughput. The uncertainties cannot be prevented in a real production line. However the uncertain conditions can be controlled by a robust prediction model. Thus, a hybrid model including autoregressive integrated moving average (ARIMA) and multiple polynomial regression, is proposed to model the nonlinear relationship of production uncertainties with throughput. The uncertainties under consideration of this study are demand, breaktime, scrap, and lead-time. The nonlinear relationship of production uncertainties with throughput are examined in the form of quadratic and cubic regression models, where the adjusted R-squared for quadratic and cubic regressions was 98.3% and 98.2%. We optimized the multiple quadratic regression (MQR) by considering the time series trend of the uncertainties using ARIMA model. Finally the hybrid model of ARIMA and MQR is formulated by better adjusted R-squared, which is 98.9%.

Keywords: ARIMA, multiple polynomial regression, production throughput, uncertainties

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2152
10114 Financial Statement Fraud: The Need for a Paradigm Shift to Forensic Accounting

Authors: Ifedapo Francis Awolowo

Abstract:

The unrelenting series of embarrassing audit failures should stimulate a paradigm shift in accounting. And in this age of information revolution, there is need for a constant improvement on the products or services one offers to the market in order to be relevant. This study explores the perceptions of external auditors, forensic accountants and accounting academics on whether a paradigm shift to forensic accounting can reduce financial statement frauds. Through Neo-empiricism/inductive analytical approach, findings reveal that a paradigm shift to forensic accounting might be the right step in the right direction in order to increase the chances of fraud prevention and detection in the financial statement. This research has implication on accounting education on the need to incorporate forensic accounting into present day accounting curriculum. Accounting professional bodies, accounting standard setters and accounting firms all have roles to play in incorporating forensic accounting education into accounting curriculum. Particularly, there is need to alter the ISA 240 to make the prevention and detection of frauds the responsibilities of bot those charged with the management and governance of companies and statutory auditors.

Keywords: Financial statement fraud, forensic accounting, fraud prevention and detection, auditing, audit expectation gap, corporate governance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3571
10113 Empirical Exploration of Correlations between Software Design Measures: A Replication Study

Authors: Jehad Al Dallal

Abstract:

Software engineers apply different measures to quantify the quality of software design. These measures consider artifacts developed at low or high level software design phases. The results are used to point to design weaknesses and to indicate design points that have to be restructured. Understanding the relationship among the quality measures and among the design quality aspects considered by these measures is important to interpreting the impact of a measure for a quality aspect on other potentially related aspects. In addition, exploring the relationship between quality measures helps to explain the impact of different quality measures on external quality aspects, such as reliability and maintainability. In this paper, we report a replication study that empirically explores the correlation between six well known and commonly applied design quality measures. These measures consider several quality aspects, including complexity, cohesion, coupling, and inheritance. The results indicate that inheritance measures are weakly correlated to other measures, whereas complexity, coupling, and cohesion measures are mostly strongly correlated.  

Keywords: Quality attribute, quality measure, software design quality, spearman correlation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 747
10112 Normalizing Logarithms of Realized Volatility in an ARFIMA Model

Authors: G. L. C. Yap

Abstract:

Modelling realized volatility with high-frequency returns is popular as it is an unbiased and efficient estimator of return volatility. A computationally simple model is fitting the logarithms of the realized volatilities with a fractionally integrated long-memory Gaussian process. The Gaussianity assumption simplifies the parameter estimation using the Whittle approximation. Nonetheless, this assumption may not be met in the finite samples and there may be a need to normalize the financial series. Based on the empirical indices S&P500 and DAX, this paper examines the performance of the linear volatility model pre-treated with normalization compared to its existing counterpart. The empirical results show that by including normalization as a pre-treatment procedure, the forecast performance outperforms the existing model in terms of statistical and economic evaluations.

Keywords: Long-memory, Gaussian process, Whittle estimator, normalization, volatility, value-at-risk.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1636
10111 An Extension of Multi-Layer Perceptron Based on Layer-Topology

Authors: Jānis Zuters

Abstract:

There are a lot of extensions made to the classic model of multi-layer perceptron (MLP). A notable amount of them has been designed to hasten the learning process without considering the quality of generalization. The paper proposes a new MLP extension based on exploiting topology of the input layer of the network. Experimental results show the extended model to improve upon generalization capability in certain cases. The new model requires additional computational resources to compare to the classic model, nevertheless the loss in efficiency isn-t regarded to be significant.

Keywords: Learning algorithm, multi-layer perceptron, topology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1468
10110 Investigation of Buoyant Parameters of k-ε Turbulence Model in Gravity Stratified Flows

Authors: A. Majid Bahari, Kourosh Hejazi

Abstract:

Different variants for buoyancy-affected terms in k-ε turbulence model have been utilized to predict the flow parameters more accurately, and investigate applicability of alternative k-ε turbulence buoyant closures in numerical simulation of a horizontal gravity current. The additional non-isotropic turbulent stress due to buoyancy has been considered in production term, based on Algebraic Stress Model (ASM). In order to account for turbulent scalar fluxes, general gradient diffusion hypothesis has been used along with Boussinesq gradient diffusion hypothesis with a variable turbulent Schmidt number and additional empirical constant c3ε.To simulate buoyant flow domain a 2D vertical numerical model (WISE, Width Integrated Stratified Environments), based on Reynolds- Averaged Navier-Stokes (RANS) equations, has been deployed and the model has been further developed for different k-ε turbulence closures. Results are compared against measured laboratory values of a saline gravity current to explore the efficient turbulence model.

Keywords: Buoyant flows, Buoyant k-ε turbulence model, saline gravity current.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3855
10109 A Framework for Successful TQM Implementation and Its Effect on the Organizational Sustainability Development

Authors: Redha Elhuni, M. Munir Ahmad

Abstract:

The main purpose of this research is to construct a generic model for successful implementation of Total Quality Management (TQM) in Oil sector, and to find out the effects of this model on the organizational sustainability development (OSD) performance of Libyan oil and gas companies using the structured equation modeling (SEM) approach. The research approach covers both quantitative and qualitative methods. A questionnaire was developed in order to identify the quality factors that are seen by Libyan oil and gas companies to be critical to the success of TQM implementation. Hypotheses were developed to evaluate the impact of TQM implementation on O SD. Data analysis reveals that there is a significant positive effect of the TQM implementation on OSD. 24 quality factors are found to be critical and absolutely essential for successful TQM implementation. The results generated a structure of the TQMSD implementation framework based on the four major road map constructs (Top management commitment, employee involvement and participation, customer-driven processes, and continuous improvement culture).

Keywords: TQM, CQFs, Oil & Gas, OSD, Libya.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4221
10108 Model of Multi-Criteria Evaluation for Railway Lines

Authors: Juraj Camaj, Martin Kendra, Jaroslav Masek

Abstract:

The paper is focused to the evaluation railway tracks in the Slovakia by using Multi-Criteria method. Evaluation of railway tracks has important impacts for the assessment of investment in technical equipment. Evaluation of railway tracks also has an important impact for the allocation of marshalling yards. Marshalling yards are in transport model as centers for the operation assigned catchment area. This model is one of the effective ways to meet the development strategy of the European Community's railways. By applying this model in practice, a transport company can guarantee a higher quality of service and then expect an increase in performance. The model is also applicable to other rail networks. This model supplements a theoretical problem of train formation problem of new ways of looking at evaluation of factors affecting the organization of wagon flows.

Keywords: Railway track, multi-criteria methods, evaluation, transportation model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2165
10107 A Location-Allocation-Routing Model for a Home Health Care Supply Chain Problem

Authors: Amir Mohammad Fathollahi Fard, Mostafa Hajiaghaei-Keshteli, Mohammad Mahdi Paydar

Abstract:

With increasing life expectancy in developed countries, the role of home care services is highlighted by both academia and industrial contributors in Home Health Care Supply Chain (HHCSC) companies. The main decisions in such supply chain systems are the location of pharmacies, the allocation of patients to these pharmacies and also the routing and scheduling decisions of nurses to visit their patients. In this study, for the first time, an integrated model is proposed to consist of all preliminary and necessary decisions in these companies, namely, location-allocation-routing model. This model is a type of NP-hard one. Therefore, an Imperialist Competitive Algorithm (ICA) is utilized to solve the model, especially in large sizes. Results confirm the efficiency of the developed model for HHCSC companies as well as the performance of employed ICA.

Keywords: Home health care supply chain, location-allocation-routing problem, imperialist competitive algorithm, optimization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 990
10106 Identification of Vessel Class with LSTM using Kinematic Features in Maritime Traffic Control

Authors: Davide Fuscà, Kanan Rahimli, Roberto Leuzzi

Abstract:

Prevent abuse and illegal activities in a given area of the sea is a very difficult and expensive task. Artificial intelligence offers the possibility to implement new methods to identify the vessel class type from the kinematic features of the vessel itself. The task strictly depends on the quality of the data. This paper explores the application of a deep Long Short-Term Memory model by using AIS flow only with a relatively low quality. The proposed model reaches high accuracy on detecting nine vessel classes representing the most common vessel types in the Ionian-Adriatic Sea. The model has been applied during the Adriatic-Ionian trial period of the international EU ANDROMEDA H2020 project to identify vessels performing behaviours far from the expected one, depending on the declared type.

Keywords: maritime surveillance, artificial intelligence, behaviour analysis, LSTM

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1207
10105 Knowledge Management: The Need for a Total Knowledge Transfer Model to Diffuse Innovation of the Public Health Workforce

Authors: Qatawneh H., Yousef S., Shirvani H.

Abstract:

The purpose of this article is to propose a model designed to achieve Total Knowledge Transfer in the public health sector. The Total Knowledge Transfer Model integrated four essential organizational factors which have been under examined in totality in the literature. The research design was inductive in nature and used a case study for accomplishing the research objectives. The researcher investigated the factors that created a base to design a framework for total knowledge transfer in the public health sector. The results of this study are drawn from a fairly large sample in only two hospitals. A further research can be conducted to cover more responses from a wider health sector. The Total Knowledge Transfer Model is essential to improve the transfer and application of total common health knowledge.

Keywords: Health Care, Knowledge Management, Knowledge Transfer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1722
10104 A Methodology for Quality Problems Diagnosis in SMEs

Authors: Humberto N. Teixeira, Isabel S. Lopes, Sérgio D. Sousa

Abstract:

This article proposes a new methodology to be used by SMEs (Small and Medium enterprises) to characterize their performance in quality, highlighting weaknesses and area for improvement. The methodology aims to identify the principal causes of quality problems and help to prioritize improvement initiatives. This is a self-assessment methodology that intends to be easy to implement by companies with low maturity level in quality. The methodology is organized in six different steps which includes gathering information about predetermined processes and subprocesses of quality management, defined based on the well-known Juran-s trilogy for quality management (Quality planning, quality control and quality improvement) and, predetermined results categories, defined based on quality concept. A set of tools for data collecting and analysis, such as interviews, flowcharts, process analysis diagrams and Failure Mode and effects Analysis (FMEA) are used. The article also presents the conclusions obtained in the application of the methodology in two cases studies.

Keywords: Continuous improvement, Diagnosis, Quality Management, Self-assessment, SMEs

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2440
10103 Improving Topic Quality of Scripts by Using Scene Similarity Based Word Co-Occurrence

Authors: Yunseok Noh, Chang-Uk Kwak, Sun-Joong Kim, Seong-Bae Park

Abstract:

Scripts are one of the basic text resources to understand broadcasting contents. Topic modeling is the method to get the summary of the broadcasting contents from its scripts. Generally, scripts represent contents descriptively with directions and speeches, and provide scene segments that can be seen as semantic units. Therefore, a script can be topic modeled by treating a scene segment as a document. Because scene segments consist of speeches mainly, however, relatively small co-occurrences among words in the scene segments are observed. This causes inevitably the bad quality of topics by statistical learning method. To tackle this problem, we propose a method to improve topic quality with additional word co-occurrence information obtained using scene similarities. The main idea of improving topic quality is that the information that two or more texts are topically related can be useful to learn high quality of topics. In addition, more accurate topical representations lead to get information more accurate whether two texts are related or not. In this paper, we regard two scene segments are related if their topical similarity is high enough. We also consider that words are co-occurred if they are in topically related scene segments together. By iteratively inferring topics and determining semantically neighborhood scene segments, we draw a topic space represents broadcasting contents well. In the experiments, we showed the proposed method generates a higher quality of topics from Korean drama scripts than the baselines.

Keywords: Broadcasting contents, generalized P´olya urn model, scripts, text similarity, topic model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1771
10102 Time Series Forecasting Using a Hybrid RBF Neural Network and AR Model Based On Binomial Smoothing

Authors: Fengxia Zheng, Shouming Zhong

Abstract:

ANNARIMA that combines both autoregressive integrated moving average (ARIMA) model and artificial neural network (ANN) model is a valuable tool for modeling and forecasting nonlinear time series, yet the over-fitting problem is more likely to occur in neural network models. This paper provides a hybrid methodology that combines both radial basis function (RBF) neural network and auto regression (AR) model based on binomial smoothing (BS) technique which is efficient in data processing, which is called BSRBFAR. This method is examined by using the data of Canadian Lynx data. Empirical results indicate that the over-fitting problem can be eased using RBF neural network based on binomial smoothing which is called BS-RBF, and the hybrid model–BS-RBFAR can be an effective way to improve forecasting accuracy achieved by BSRBF used separately.

Keywords: Binomial smoothing (BS), hybrid, Canadian Lynx data, forecasting accuracy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3603
10101 Average Switching Thresholds and Average Throughput for Adaptive Modulation using Markov Model

Authors: Essam S. Altubaishi

Abstract:

The motivation for adaptive modulation and coding is to adjust the method of transmission to ensure that the maximum efficiency is achieved over the link at all times. The receiver estimates the channel quality and reports it back to the transmitter. The transmitter then maps the reported quality into a link mode. This mapping however, is not a one-to-one mapping. In this paper we investigate a method for selecting the proper modulation scheme. This method can dynamically adapt the mapping of the Signal-to- Noise Ratio (SNR) into a link mode. It enables the use of the right modulation scheme irrespective of changes in the channel conditions by incorporating errors in the received data. We propose a Markov model for this method, and use it to derive the average switching thresholds and the average throughput. We show that the average throughput of this method outperforms the conventional threshold method.

Keywords: Adaptive modulation and coding, CDMA, Markov model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1680
10100 Influence of IMV on Space Station

Authors: Fu Shiming, Pei Yifei

Abstract:

To study the impact of the inter-module ventilation (IMV) on the space station, the Computational Fluid Dynamic (CFD) model under the influence of IMV, the mathematical model, boundary conditions and calculation method are established and determined to analyze the influence of IMV on cabin air flow characteristics and velocity distribution firstly; and then an integrated overall thermal mathematical model of the space station is used to consider the impact of IMV on thermal management. The results show that: the IMV has a significant influence on the cabin air flow, the flowrate of IMV within a certain range can effectively improve the air velocity distribution in cabin, if too much may lead to its deterioration; IMV can affect the heat deployment of the different modules in space station, thus affecting its thermal management, the use of IMV can effectively maintain the temperature levels of the different modules and help the space station to dissipate the waste heat.

Keywords: CFD, Environment control and life support, Space station, Thermal management, Thermal mathematical model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1994
10099 Compatibility of Integrated Satellite Systems with Another Satellite System Operating in Adjacent Beam

Authors: Hyemi Gam, Dae-Sub Oh, Bon-Jun Ku

Abstract:

This paper addresses the analysis of the interference between complementary ground component (CGC) base station and mobile earth station (MES). In the frequency sharing scenario between CGC base station and MES, the interference from the adjacent beams must be considered. In this paper, we estimated the interference to MES of an integrated satellite system and the result is presented as the carrier to interference ratio(C/I) with respect to the number of CGC base station in the adjacent beam and the ratio of satellite beam center radius to the total beam radius (R1/R). By using these results, we can determine the minimum separation distance between the CGC base stations of adjacent beam and MES for compatibility. This result can be applied to the CGC base station of an integrated satellite system for the effective frequency sharing.

Keywords: Integrated satellite system, interference, frequency reuse pattern, CGC base station, MES, frequency sharing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737
10098 An Improved GA to Address Integrated Formulation of Project Scheduling and Material Ordering with Discount Options

Authors: Babak H. Tabrizi, Seyed Farid Ghaderi

Abstract:

Concurrent planning of the resource constraint project scheduling and material ordering problems have received significant attention within the last decades. Hence, the issue has been investigated here with the aim to minimize total project costs. Furthermore, the presented model considers different discount options in order to approach the real world conditions. The incorporated alternatives consist of all-unit and incremental discount strategies. On the other hand, a modified version of the genetic algorithm is applied in order to solve the model for larger sizes, in particular. Finally, the applicability and efficiency of the given model is tested by different numerical instances.

Keywords: Genetic algorithm, material ordering, project management, project scheduling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1242
10097 Estimation of Vertical Handover Probability in an Integrated UMTS and WLAN Networks

Authors: Diganta Kumar Pathak, Manashjyoti Bhuyan, Vaskar Deka

Abstract:

Vertical Handover(VHO) among different communication technologies ensuring uninterruption and service continuity is one of the most important performance parameter in Heterogenous networks environment. In an integrated Universal Mobile Telecommunicatin System(UMTS) and Wireless Local Area Network(WLAN), WLAN is given an inherent priority over UMTS because of its high data rates with low cost. Therefore mobile users want to be associated with WLAN maximum of the time while roaming, to enjoy best possible services with low cost. That encourages reduction of number of VHO. In this work the reduction of number of VHO with respect to varying number of WLAN Access Points(APs) in an integrated UMTS and WLAN network is investigated through simulation to provide best possible cost effective service to the users. The simulation has been carried out for an area (7800 × 9006)m2 where COST-231 Hata model and 3GPP (TR 101 112 V 3.1.0) specified models are used for WLAN and UMTS path loss models respectively. The handover decision is triggered based on the received signal level as compared to the fade margin. Fade margin gives a probabilistic measure of the reliability of the communication link. A relationship between number of WLAN APs and the number of VHO is also established in this work.

Keywords: VHO, UMTS, WLAN, MT, AP, BS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1984
10096 Forecasting Stock Price Manipulation in Capital Market

Authors: F. Rahnamay Roodposhti, M. Falah Shams, H. Kordlouie

Abstract:

The aim of the article is extending and developing econometrics and network structure based methods which are able to distinguish price manipulation in Tehran stock exchange. The principal goal of the present study is to offer model for approximating price manipulation in Tehran stock exchange. In order to do so by applying separation method a sample consisting of 397 companies accepted at Tehran stock exchange were selected and information related to their price and volume of trades during years 2001 until 2009 were collected and then through performing runs test, skewness test and duration correlative test the selected companies were divided into 2 sets of manipulated and non manipulated companies. In the next stage by investigating cumulative return process and volume of trades in manipulated companies, the date of starting price manipulation was specified and in this way the logit model, artificial neural network, multiple discriminant analysis and by using information related to size of company, clarity of information, ratio of P/E and liquidity of stock one year prior price manipulation; a model for forecasting price manipulation of stocks of companies present in Tehran stock exchange were designed. At the end the power of forecasting models were studied by using data of test set. Whereas the power of forecasting logit model for test set was 92.1%, for artificial neural network was 94.1% and multi audit analysis model was 90.2%; therefore all of the 3 aforesaid models has high power to forecast price manipulation and there is no considerable difference among forecasting power of these 3 models.

Keywords: Price Manipulation, Liquidity, Size of Company, Floating Stock, Information Clarity

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2782
10095 Water Quality from a Mixed Land-Use Catchment in Miri, Sarawak

Authors: Carrie Ho, Darshana J. Kumar

Abstract:

Urbanization has been found to impact stormwater runoff quantity and quality. A study catchment with mixed land use, residential and industrial were investigated and the water quality discharged from the catchment were sampled and tested for four basic water quality parameters; BOD5, NH3-N, NO3-N and P. One dry weather flow and several stormwater runoff were sampled. Results were compared to the USEPA stormwater quality benchmark values and the Interim National Water Quality Standards for Malaysia (INWQS). The concentration of the parameters was found to vary significantly between storms and the pollutant of concern was found to be NO3-N.

Keywords: Mixed land-use, urban runoff, water quality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2668
10094 Analytics Model in a Telehealth Center Based on Cloud Computing and Local Storage

Authors: L. Ramirez, E. Guillén, J. Sánchez

Abstract:

Some of the main goals about telecare such as monitoring, treatment, telediagnostic are deployed with the integration of applications with specific appliances. In order to achieve a coherent model to integrate software, hardware, and healthcare systems, different telehealth models with Internet of Things (IoT), cloud computing, artificial intelligence, etc. have been implemented, and their advantages are still under analysis. In this paper, we propose an integrated model based on IoT architecture and cloud computing telehealth center. Analytics module is presented as a solution to control an ideal diagnostic about some diseases. Specific features are then compared with the recently deployed conventional models in telemedicine. The main advantage of this model is the availability of controlling the security and privacy about patient information and the optimization on processing and acquiring clinical parameters according to technical characteristics.

Keywords: Analytics, telemedicine, internet of things, cloud computing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1497
10093 Integrated Reasoning Approach for Car Faulty Diagnosis

Authors: Diana M.L. Wong

Abstract:

This paper presents an integrated case based and rule based reasoning method for car faulty diagnosis. The reasoning method is done through extracting the past cases from the Proton Service Center while comparing with the preset rules to deduce a diagnosis/solution to a car service case. New cases will be stored to the knowledge base. The test cases examples illustrate the effectiveness of the proposed integrated reasoning. It has proven accuracy of similar reasoning if carried out by a service advisor from the service center.

Keywords: component; case based reasoning (CBR), rule basedreasoning (RBR), decision support systems, diagnosis tool.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1864
10092 Theoretical Model of a Flat Plate Solar Collector Integrated with Phase Change Material

Authors: Mouna Hamed, Ammar B. Brahim

Abstract:

The objective of this work was to develop a theoretical model to study the dynamic thermal behavior of a flat plate solar collector integrated with a phase change material (PCM). The PCM acted as a heat source for the solar system during low intensity solar radiation and night. The energy balance equations for the various components of the collector as well as for the PCM were formulated and numerically solved using Matlab computational program. The effect of natural convection on heat during the melting process was taken into account by using an effective thermal conductivity. The model was used to investigate the effect of inlet water temperature, water mass flow rate, and PCM thickness on the outlet water temperature and the melt fraction during charging and discharging modes. A comparison with a collector without PCM was made. Results showed that charging and discharging processes of PCM have six stages. The adding of PCM caused a decrease in temperature during charge and an increase during discharge. The rise was most enhanced for higher inlet water temperature, PCM thickness and for lower mass flow rate. Analysis indicated that the complete melting time was shorter than the solidification time due to the high heat transfer coefficient during melting. The increases in PCM height and mass flow rate were not linear with the melting and solidification times.

Keywords: Thermal energy storage, phase change material, melting, solidification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2058
10091 Theoretical Model of a Flat Plate Solar Collector Integrated with Phase Change Material

Authors: Mouna Hamed, Ammar B. Brahim

Abstract:

The objective of this work was to develop a theoretical model to study the dynamic thermal behavior of a flat plate solar collector integrated with a phase change material (PCM). The PCM acted as a heat source for the solar system during low intensity solar radiation and night. The energy balance equations for the various components of the collector as well as for the PCM were formulated and numerically solved using MATLAB computational program. The effect of natural convection on heat during the melting process was taken into account by using an effective thermal conductivity. The model was used to investigate the effect of inlet water temperature, water mass flow rate, and PCM thickness on the outlet water temperature and the melt fraction during charging and discharging modes. A comparison with a collector without PCM was made. Results showed that charging and discharging processes of PCM have six stages. The adding of PCM caused a decrease in temperature during charge and an increase during discharge. The rise was most enhanced for higher inlet water temperature, PCM thickness and for lower mass flow rate. Analysis indicated that the complete melting time was shorter than the solidification time due to the high heat transfer coefficient during melting. The increases in PCM height and mass flow rate were not linear with the melting and solidification times.

Keywords: Thermal energy storage, phase change material, melting, solidification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1202
10090 Scale Development for Measuring E-Service Quality in Banking

Authors: Vivek Agrawal, Vikas Tripathi, Nitin Seth

Abstract:

This study examines several critical dimensions of eservice quality overlooked in the existing literature and proposes a model and instrument framework for measuring customer perceived e-service quality in the banking sector. The initial design was derived from a pool of instrument dimensions and their items from the existing literature review by content analysis. Based on focused group discussion, nine dimensions were extracted. An exploratory factor analysis approach was applied to data from a survey of 323 respondents. The instrument has been designed specifically for the banking sector. Research data was collected from bank customers who use electronic banking in a developing economy. A nine-factor instrument has been proposed to measure the e-service quality. The instrument has been checked for reliability. The validity and sample place limited the applicability of the instrument across economies and service categories. Future research must be conducted to check the validity. This instrument can help bankers in developing economies like India to measure the e-service quality and make improvements. The present study offers a systematic procedure that provides insights on to the conceptual and empirical comprehension of customer perceived e-service quality and its constituents.

Keywords: Testing, instrument, e-service quality, factor analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3759
10089 Organizational Management Model based on Knowledge Management, Talent Management and Technology Management Framework “Gomak“

Authors: Nieto Bernal W., Luna Amaya C.

Abstract:

This paper aims to present a framework for the organizational knowledge management, which seeks to deploy a standardized structure for the integrated management of knowledge is a common language based on domains, processes and global indicators inspired by the COBIT framework 5 (ISACA, 2012), which supports the integration of three technologies, enterprise information architecture (EIA), the business process modeling (BPM) and service-oriented architecture (SOA). The Gomak Framework is a management platform that seeks to integrate the information technology infrastructure, the structure of applications, information infrastructure, and business logic and business model to support a sound strategy of organizational knowledge management, low process-based approach and concurrent engineering. Concurrent engineering (CE) is a systematic approach to integrated product development that respond to customer expectations, involving all perspectives in parallel, from the beginning of the product life cycle. (European Space Agency, 2000).

Keywords: Business Process Modeling, Enterprise Information Architecture, Government and Knowledge Management, Service Oriented Architecture, Process Management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1790
10088 Gauteng-s Waste Outlook: A Reflection

Authors: Edison Muzenda

Abstract:

Gauteng, as the province with the greatest industrial and population density, the economic hub of South Africa also generates the greatest amount of waste, both general and hazardous. Therefore the province has a significant need to develop and apply appropriate integrated waste management policies that ensure that waste is recognised as a serious problem and is managed in an effective integrated manner to preserve both the present and future human health and environment. This paper reflects on Gauteng-s waste outlook in particular the province-s General Waste Minimisation Plan and its Integrated Waste Management Policy. The paper also looks at general waste generation, recyclable waste streams as well as recycling and separation at source initiatives in the province. Both the quantity and nature of solid waste differs considerably across the socio-economic spectrum. People in informal settlements generate an average of 0.16 kg per person per day whereas 2 kg per day is not unusual in affluent areas. For example the amount of waste generated in Johannesburg is approximately 1.2 kg per person per day.

Keywords: General waste, generation, integrated, minimisation, recycling, separation

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1442
10087 Evaluating 8D Reports Using Text-Mining

Authors: Benjamin Kuester, Bjoern Eilert, Malte Stonis, Ludger Overmeyer

Abstract:

Increasing quality requirements make reliable and effective quality management indispensable. This includes the complaint handling in which the 8D method is widely used. The 8D report as a written documentation of the 8D method is one of the key quality documents as it internally secures the quality standards and acts as a communication medium to the customer. In practice, however, the 8D report is mostly faulty and of poor quality. There is no quality control of 8D reports today. This paper describes the use of natural language processing for the automated evaluation of 8D reports. Based on semantic analysis and text-mining algorithms the presented system is able to uncover content and formal quality deficiencies and thus increases the quality of the complaint processing in the long term.

Keywords: 8D report, complaint management, evaluation system, text-mining.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 968
10086 The Effects of Software Size on Development Effort and Software Quality

Authors: Zhizhong Jiang, Peter Naudé, Binghua Jiang

Abstract:

Effective evaluation of software development effort is an important issue during project plan. This study provides a model to predict development effort based on the software size estimated with function points. We generalize the average amount of effort spent on each phase of the development, and give the estimates for the effort used in software building, testing, and implementation. Finally, this paper finds a strong correlation between software defects and software size. As the size of software constantly increases, the quality remains to be a matter which requires major concern.

Keywords: Development effort, function points, software quality, software size.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2221