Search results for: Impact measurement model
9442 Systemic Approach to Risk Measurement of Drainage Systems in Urban Areas
Authors: Jadwiga Królikowska, Andrzej Królikowski, Jarosław Bajer
Abstract:
The work delineates the threats of maladjustment of the capacity of rain canals, designed and built in the early 20th century, in connection to heavy rainfall, especially in summer. This is the cause of the so called 'urban floods.' It directly relates to fierce raise of paving in the cities. Resolving this problem requires a change in philosophy of draining the rainfall by wider use of retention, infiltration and usage of rainwater.
In systemic approach to managing the safety of urban drainage systems the risk, which is directly connected to safety failures, has been accepted as a measure. The risk level defines the probability of occurrence of losses greater than the ones forecast for a given time frame. The procedure of risk modelling, enabling its numeric analysis by using appropriate weights, is a significant issue in this paper.
Keywords: Drainage system, urban areas, risk measurement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16949441 Three-Dimensional, Non-Linear Finite Element Analysis of Bullet Penetration through Thin AISI 4340 Steel Target Plate
Authors: Abhishek Soni, A. Kumaraswamy, M. S. Mahesh
Abstract:
Bullet penetration in steel plate is investigated with the help of three-dimensional, non-linear, transient, dynamic, finite elements analysis using explicit time integration code LSDYNA. The effect of large strain, strain-rate and temperature at very high velocity regime was studied from number of simulations of semi-spherical nose shape bullet penetration through single layered circular plate with 2 mm thickness at impact velocities of 500, 1000, and 1500 m/s with the help of Johnson Cook material model. Mie-Gruneisen equation of state is used in conjunction with Johnson Cook material model to determine pressure-volume relationship at various points of interests. Two material models viz. Plastic-Kinematic and Johnson- Cook resulted in different deformation patterns in steel plate. It is observed from the simulation results that the velocity drop and loss of kinetic energy occurred very quickly up to perforation of plate, after that the change in velocity and changes in kinetic energy are negligibly small. The physics behind this kind of behaviour is presented in the paper.Keywords: AISI 4340 steel, ballistic impact simulation, bullet penetration, non-linear FEM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12649440 Study of the Process of Climate Change According to Data Simulation Using LARS-WG Software during 2010-2030: Case Study of Semnan Province
Authors: Leila Rashidian
Abstract:
Temperature rise on Earth has had harmful effects on the Earth's surface and has led to change in precipitation patterns all around the world. The present research was aimed to study the process of climate change according to the data simulation in future and compare these parameters with current situation in the studied stations in Semnan province including Garmsar, Shahrood and Semnan. In this regard, LARS-WG software, HADCM3 model and A2 scenario were used for the 2010-2030 period. In this model, climatic parameters such as maximum and minimum temperature, precipitation and radiation were used daily. The obtained results indicated that there will be a 4.4% increase in precipitation in Semnan province compared with the observed data, and in general, there will be a 1.9% increase in temperature. This temperature rise has significant impact on precipitation patterns. Most of precipitation will be raining (torrential rains in some cases). According to the results, from west to east, the country will experience more temperature rise and will be warmer.
Keywords: Climate change, Semnan province, LARS-WG model, climate parameters, HADCM3 model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11519439 A New Precautionary Method for Measurement and Improvement the Data Quality
Authors: Seyed Mohammad Hossein Moossavizadeh, Mehran Mohsenzadeh, Nasrin Arshadi
Abstract:
the data quality is a kind of complex and unstructured concept, which is concerned by information systems managers. The reason of this attention is the high amount of Expenses for maintenance and cleaning of the inefficient data. Such a data more than its expenses of lack of quality, cause wrong statistics, analysis and decisions in organizations. Therefor the managers intend to improve the quality of their information systems' data. One of the basic subjects of quality improvement is the evaluation of the amount of it. In this paper, we present a precautionary method, which with its application the data of information systems would have a better quality. Our method would cover different dimensions of data quality; therefor it has necessary integrity. The presented method has tested on three dimensions of accuracy, value-added and believability and the results confirm the improvement and integrity of this method.
Keywords: Data quality, precaution, information system, measurement, improvement.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14689438 Application of Spreadsheet and Queuing Network Model to Capacity Optimization in Product Development
Authors: Muhammad Marsudi, Dzuraidah Abdul Wahab, Che Hassan Che Haron
Abstract:
Modeling of a manufacturing system enables one to identify the effects of key design parameters on the system performance and as a result to make correct decision. This paper proposes a manufacturing system modeling approach using a spreadsheet model based on queuing network theory, in which a static capacity planning model and stochastic queuing model are integrated. The model was used to improve the existing system utilization in relation to product design. The model incorporates few parameters such as utilization, cycle time, throughput, and batch size. The study also showed that the validity of developed model is good enough to apply and the maximum value of relative error is 10%, far below the limit value 32%. Therefore, the model developed in this study is a valuable alternative model in evaluating a manufacturing systemKeywords: Manufacturing system, product design, spreadsheet model, utilization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19209437 Bayesian Inference for Phase Unwrapping Using Conjugate Gradient Method in One and Two Dimensions
Authors: Yohei Saika, Hiroki Sakaematsu, Shota Akiyama
Abstract:
We investigated statistical performance of Bayesian inference using maximum entropy and MAP estimation for several models which approximated wave-fronts in remote sensing using SAR interferometry. Using Monte Carlo simulation for a set of wave-fronts generated by assumed true prior, we found that the method of maximum entropy realized the optimal performance around the Bayes-optimal conditions by using model of the true prior and the likelihood representing optical measurement due to the interferometer. Also, we found that the MAP estimation regarded as a deterministic limit of maximum entropy almost achieved the same performance as the Bayes-optimal solution for the set of wave-fronts. Then, we clarified that the MAP estimation perfectly carried out phase unwrapping without using prior information, and also that the MAP estimation realized accurate phase unwrapping using conjugate gradient (CG) method, if we assumed the model of the true prior appropriately.
Keywords: Bayesian inference using maximum entropy, MAP estimation using conjugate gradient method, SAR interferometry.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17519436 Model Parameters Estimating on Lyman–Kutcher–Burman Normal Tissue Complication Probability for Xerostomia on Head and Neck Cancer
Authors: Tsair-Fwu Lee , Hui-Min Ting , Pei-Ju Chao, Jing-Chuan Jiang, Min-Yuan Chao, Wen-Cheng Chen, Long-Chang Chen, Jia-Ming Wu
Abstract:
The purpose of this study is to derive parameters estimating for the Lyman–Kutcher–Burman (LKB) normal tissue complication probability (NTCP) model using analysis of scintigraphy assessments and quality of life (QoL) measurement questionnaires for the parotid gland (xerostomia). In total, 31 patients with head-and-neck (HN) cancer were enrolled. Salivary excretion factor (SEF) and EORTC QLQ-H&N35 questionnaires datasets are used for the NTCP modeling to describe the incidence of grade 4 xerostomia. Assuming that n= 1, NTCP fitted parameters are given as TD50= 43.6 Gy, m= 0.18 in SEF analysis, and as TD50= 44.1 Gy, m= 0.11 in QoL measurements, respectively. SEF and QoL datasets can validate the Quantitative Analyses of Normal Tissue Effects in the Clinic (QUANTEC) guidelines well, resulting in NPV-s of 100% for the both datasets and suggests that the QUANTEC 25/20Gy gland-spared guidelines are suitable for clinical used for the HN cohort to effectively avoid xerostomia.Keywords: HN, NTCP, SEF, QoL, QUANTEC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21019435 Development of Sustainable Building Environmental Model (SBEM) in Hong Kong
Authors: Kwok W. Mui, Ling T. Wong, F. Xiao, Chin T. Cheung, Ho C. Yu
Abstract:
This study addresses a concept of the Sustainable Building Environmental Model (SBEM) developed to optimize energy consumption in air conditioning and ventilation (ACV) systems without any deterioration of indoor environmental quality (IEQ). The SBEM incorporates two main components: an adaptive comfort temperature control module (ACT) and a new carbon dioxide demand control module (nDCV). These two modules take an innovative approach to maintain satisfaction of the Indoor Environmental Quality (IEQ) with optimum energy consumption; they provide a rational basis of effective control. A total of 2133 sets of measurement data of indoor air temperature (Ta), relative humidity (Rh) and carbon dioxide concentration (CO2) were conducted in some Hong Kong offices to investigate the potential of integrating the SBEM. A simulation was used to evaluate the dynamic performance of the energy and air conditioning system with the integration of the SBEM in an air-conditioned building. It allows us make a clear picture of the control strategies and performed any pre-tuned of controllers before utilized in real systems. With the integration of SBEM, it was able to save up to 12.3% in simulation of overall electricity consumption, and maintain the average carbon dioxide concentration within 1000ppm and occupant dissatisfaction in 20%.
Keywords: —Sustainable building environmental model (SBEM), adaptive comfort temperature (ACT), new demand control ventilation (nDCV), energy saving.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18259434 An Experimental Study to Control Single Droplet by Actuating Waveform with Preliminary and Suppressing Vibration
Authors: Oke Oktavianty, Tadayuki Kyoutani, Shigeyuki Haruyama, Ken Kaminishi
Abstract:
For advancing the experiment system standard of Inkjet printer that is being developed, the actual natural period, fire limitation number in droplet weight measurement and observation distance in droplet velocity measurement was investigated. In another side, the study to control the droplet volume in inkjet printer with negative actuating waveform method is still limited. Therefore, the effect of negative waveform with preliminary and suppressing vibration addition on the droplet formation process, droplet shape, volume and velocity were evaluated. The different voltage and print-head temperature were exerted to obtain the optimum preliminary and suppressing vibration. The mechanism of different phenomenon from each waveform was also discussed.
Keywords: Inkjet printer, DoD, waveform, preliminary and suppressing vibration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9469433 Model of High-Speed Train Energy Consumption
Authors: Romain Bosquet, Pierre-Olivier Vandanjon, Alex Coiret, Tristan Lorino
Abstract:
In the hardening energy context, the transport sector which constitutes a large worldwide energy demand has to be improving for decrease energy demand and global warming impacts. In a controversial situation where subsists an increasing demand for long-distance and high-speed travels, high-speed trains offer many advantages, as consuming significantly less energy than road or air transports. At the project phase of new rail infrastructures, it is nowadays important to characterize accurately the energy that will be induced by its operation phase, in addition to other more classical criteria as construction costs and travel time. Current literature consumption models used to estimate railways operation phase are obsolete or not enough accurate for taking into account the newest train or railways technologies. In this paper, an updated model of consumption for high-speed is proposed, based on experimental data obtained from full-scale tests performed on a new high-speed line. The assessment of the model is achieved by identifying train parameters and measured power consumptions for more than one hundred train routes. Perspectives are then discussed to use this updated model for accurately assess the energy impact of future railway infrastructures.Keywords: High-speed train, energy, model, track profile, infrastructure
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 52099432 Comparative Analysis of the Software Effort Estimation Models
Authors: Jaswinder Kaur, Satwinder Singh, Karanjeet Singh Kahlon
Abstract:
Accurate software cost estimates are critical to both developers and customers. They can be used for generating request for proposals, contract negotiations, scheduling, monitoring and control. The exact relationship between the attributes of the effort estimation is difficult to establish. A neural network is good at discovering relationships and pattern in the data. So, in this paper a comparative analysis among existing Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model and Neural Network Based Model is performed. Neural Network has outperformed the other considered models. Hence, we proposed Neural Network system as a soft computing approach to model the effort estimation of the software systems.Keywords: Effort Estimation, Neural Network, Halstead Model, Walston-Felix Model, Bailey-Basili Model, Doty Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22219431 Model Predictive Control of Three Phase Inverter for PV Systems
Authors: Irtaza M. Syed, Kaamran Raahemifar
Abstract:
This paper presents a model predictive control (MPC) of a utility interactive three phase inverter (TPI) for a photovoltaic (PV) system at commercial level. The proposed model uses phase locked loop (PLL) to synchronize the TPI with the power electric grid (PEG) and performs MPC control in a dq reference frame. TPI model consists of a boost converter (BC), maximum power point tracking (MPPT) control, and a three-leg voltage source inverter (VSI). The operational model of VSI is used to synthesize the sinusoidal current and track the reference. The model is validated using a 35.7 kW PV system in Matlab/Simulink. Implementation results show simplicity and accuracy, as well as reliability of the model.Keywords: Model predictive control, three phase voltage source inverter, PV system, Matlab/Simulink.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 37669430 Outsourcing Opportunities for Internet Banking Solutions
Authors: Ondruska Marek, Matustik Ondrej
Abstract:
The main goal of the article is to present new model of application architecture of banking IT solution providing the Internet Banking services that is particularly outsourced. At first, we propose business rationale and a SWOT analysis to explain the reasons for the model in the article. The most important factor for our model is nowadays- big boom around smart phones and tablet devices. As next, we focus on IT architecture viewpoint where we design application, integration and security model. Finally, we propose a generic governance model that serves as a basis for the specialized governance model. The specialized instance of governance model is designed to ensure that the development and the maintenance of different parts of the IT solution are well governed in time.Keywords: governance model, front-end application, Internet Banking, smart phones
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17479429 Accurate Dimensional Measurement of 3D Round Holes Based on Stereo Vision
Authors: Zhiguo Ren, Lilong Cai
Abstract:
This paper present an effective method to accurately reconstruct and measure the 3D curve edges of small industrial parts based on stereo vision. To effectively fit the curve of the measured parts using a series of line segments in the images, a strategy from coarse to fine is employed based on multi-scale curve fitting. After reconstructing the 3D curve of a hole through a curved surface, its axis is adjusted so that it is parallel to the Z axis with least squares error and the dimensions of the hole can be calculated on the XY plane easily. Experimental results show that the presented method can accurately measure the dimensions of round holes through a curved surface.
Keywords: Stereo Vision, 3D Round Hole Measurement, Curve Fitting, 3D Curve Reconstruction, Least Squares Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16279428 Particle Image Velocimetry for Measuring Water Flow Velocity
Authors: King Kuok Kuok, Po Chan Chiu
Abstract:
Floods are natural phenomena, which may turn into disasters causing widespread damage, health problems and even deaths. Nowadays, floods had become more serious and more frequent due to climatic changes. During flooding, discharge measurement still can be taken by standing on the bridge across the river using portable measurement instrument. However, it is too dangerous to get near to the river especially during high flood. Therefore, this study employs Particle Image Velocimetry (PIV) as a tool to measure the surface flow velocity. PIV is a image processing technique to track the movement of water from one point to another. The PIV codes are developed using Matlab. In this study, 18 ping pong balls were scattered over the surface of the drain and images were taken with a digital SLR camera. The images obtained were analyzed using the PIV code. Results show that PIV is able to produce the flow velocity through analyzing the series of images captured.
Keywords: Particle Image Velocimetry, flow velocity, surface flow.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28459427 A Theoretical Hypothesis on Ferris Wheel Model of University Social Responsibility
Authors: Le Kang
Abstract:
According to the nature of the university, as a free and responsible academic community, USR is based on a different foundation —academic responsibility, so the Pyramid and the IC Model of CSR could not fully explain the most distinguished feature of USR. This paper sought to put forward a new model— Ferris Wheel Model, to illustrate the nature of USR and the process of achievement. The Ferris Wheel Model of USR shows the university creates a balanced, fairness and neutrality systemic structure to afford social responsibilities; that makes the organization could obtain a synergistic effect to achieve more extensive interests of stakeholders and wider social responsibilities.Keywords: USR, Achievement model, Ferris wheel model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15649426 Model based Soft-Sensor for Industrial Crystallization: On-line Mass of Crystals and Solubility Measurement
Authors: Cédric Damour, Michel Benne, Brigitte Grondin-Perez, Jean-Pierre Chabriat
Abstract:
Monitoring and control of cane sugar crystallization processes depend on the stability of the supersaturation (σ ) state. The most widely used information to represent σ is the electrical conductivity κ of the solutions. Nevertheless, previous studies point out the shortcomings of this approach: κ may be regarded as inappropriate to guarantee an accurate estimation of σ in impure solutions. To improve the process control efficiency, additional information is necessary. The mass of crystals in the solution ( c m ) and the solubility (mass ratio of sugar to water / s w m m ) are relevant to complete information. Indeed, c m inherently contains information about the mass balance and / s w m m contains information about the supersaturation state of the solution. The main problem is that c m and / s w m m are not available on-line. In this paper, a model based soft-sensor is presented for a final crystallization stage (C sugar). Simulation results obtained on industrial data show the reliability of this approach, c m and the crystal content ( cc ) being estimated with a sufficient accuracy for achieving on-line monitoring in industryKeywords: Soft-sensor, on-line monitoring, cane sugarcrystallization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21369425 Multi-Hazard Risk Assessment and Management in Tourism Industry- A Case Study from the Island of Taiwan
Authors: Chung-Hung Tsai
Abstract:
Global environmental changes lead to increased frequency and scale of natural disaster, Taiwan is under the influence of global warming and extreme weather. Therefore, the vulnerability was increased and variability and complexity of disasters is relatively enhanced. The purpose of this study is to consider the source and magnitude of hazard characteristics on the tourism industry. Using modern risk management concepts, integration of related domestic and international basic research, this goes beyond the Taiwan typhoon disaster risk assessment model and evaluation of loss. This loss evaluation index system considers the impact of extreme weather, in particular heavy rain on the tourism industry in Taiwan. Consider the extreme climate of the compound impact of disaster for the tourism industry; we try to make multi-hazard risk assessment model, strategies and suggestions. Related risk analysis results are expected to provide government department, the tourism industry asset owners, insurance companies and banking include tourist disaster risk necessary information to help its tourism industry for effective natural disaster risk management.
Keywords: Tourism industry, extreme weather, multi-hazard, vulnerability analysis, loss exceeding probability, risk management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 38309424 All-or-None Principle and Weakness of Hodgkin-Huxley Mathematical Model
Authors: S. A. Sadegh Zadeh, C. Kambhampati
Abstract:
Mathematical and computational modellings are the necessary tools for reviewing, analysing, and predicting processes and events in the wide spectrum range of scientific fields. Therefore, in a field as rapidly developing as neuroscience, the combination of these two modellings can have a significant role in helping to guide the direction the field takes. The paper combined mathematical and computational modelling to prove a weakness in a very precious model in neuroscience. This paper is intended to analyse all-or-none principle in Hodgkin-Huxley mathematical model. By implementation the computational model of Hodgkin-Huxley model and applying the concept of all-or-none principle, an investigation on this mathematical model has been performed. The results clearly showed that the mathematical model of Hodgkin-Huxley does not observe this fundamental law in neurophysiology to generating action potentials. This study shows that further mathematical studies on the Hodgkin-Huxley model are needed in order to create a model without this weakness.
Keywords: All-or-none, computational modelling, mathematical model, transmembrane voltage, action potential.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13629423 Zero Truncated Strict Arcsine Model
Authors: Y. N. Phang, E. F. Loh
Abstract:
The zero truncated model is usually used in modeling count data without zero. It is the opposite of zero inflated model. Zero truncated Poisson and zero truncated negative binomial models are discussed and used by some researchers in analyzing the abundance of rare species and hospital stay. Zero truncated models are used as the base in developing hurdle models. In this study, we developed a new model, the zero truncated strict arcsine model, which can be used as an alternative model in modeling count data without zero and with extra variation. Two simulated and one real life data sets are used and fitted into this developed model. The results show that the model provides a good fit to the data. Maximum likelihood estimation method is used in estimating the parameters.
Keywords: Hurdle models, maximum likelihood estimation method, positive count data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18579422 On the Performance of Information Criteria in Latent Segment Models
Authors: Jaime R. S. Fonseca
Abstract:
Nevertheless the widespread application of finite mixture models in segmentation, finite mixture model selection is still an important issue. In fact, the selection of an adequate number of segments is a key issue in deriving latent segments structures and it is desirable that the selection criteria used for this end are effective. In order to select among several information criteria, which may support the selection of the correct number of segments we conduct a simulation study. In particular, this study is intended to determine which information criteria are more appropriate for mixture model selection when considering data sets with only categorical segmentation base variables. The generation of mixtures of multinomial data supports the proposed analysis. As a result, we establish a relationship between the level of measurement of segmentation variables and some (eleven) information criteria-s performance. The criterion AIC3 shows better performance (it indicates the correct number of the simulated segments- structure more often) when referring to mixtures of multinomial segmentation base variables.Keywords: Quantitative Methods, Multivariate Data Analysis, Clustering, Finite Mixture Models, Information Theoretical Criteria, Simulation experiments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15199421 The Impact of Quality Cost on Revenue Sharing in Supply Chain Management
Authors: Fayza Obied-Allah
Abstract:
Customer’ needs, quality, and value creation while reducing costs through supply chain management provides challenges and opportunities for companies and researchers. In the light of these challenges, modern ideas must contribute to counter these challenges and exploit opportunities. Therefore, this paper discusses the impact of the quality cost on revenue sharing as a most important incentive to configure business networks. This paper develops the quality cost approach to align with the modern era. It develops a model to measure quality costs which might enable firms to manage revenue sharing in a supply chain. The developed model includes five categories; besides the well-known four categories (namely prevention costs, appraisal costs, internal failure costs, and external failure costs), a new category has been developed in this research as a new vision of the relationship between quality costs and innovations in industry. This new category is Recycle Cost. This paper also examines whether such quality costs in supply chains influence the revenue sharing between partners. Using the author's quality cost model, the relationship between quality costs and revenue sharing among partners is examined using a case study in an Egyptian manufacturing company which is a part of a supply chain. This paper argues that the revenue-sharing proportion allocated to supplier increases as the recycle cost of supplier increases, and the revenue-sharing proportion allocated to manufacturer increases as the prevention and appraisal costs increase, as well as the failure costs, the recycle costs of manufacturer, and the recycle costs of suppliers decrease. However, the results present surprising findings. The purposes of this study are developing quality cost approach and understanding the relationships between quality costs and revenue sharing in supply chains. Therefore, the present study contributes to theory and practice by explaining how the cost of recycling can be combined in quality cost model to better understanding the revenue sharing among partners in supply chains.Keywords: Quality cost, Recycle cost, Revenue sharing, Supply chain.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19799420 Sound Insulation between Buildings: The Impact Noise Transmission through Different Floor Configurations
Authors: Abdelouahab Bouttout, Mohamed Amara
Abstract:
The present paper examines the impact noise transmission through some floor building assemblies. The Acoubat software numerical simulation has been used to simulate the impact noise transmission through different floor configurations used in Algerian construction mode. The results are compared with the available measurements. We have developed two experimental methods, i) field method, and ii) laboratory method using Brüel and Kjær equipments. The results show that the different cases of floor configurations need some improvement to ensure the acoustic comfort in the receiving apartment. The recommended value of the impact sound level in the receiving room should not exceed 58 dB. The important results obtained in this paper can be used as platform to improve the Algerian building acoustic regulation aimed at the construction of the multi-storey residential building.Keywords: Impact noise, building acoustic, floor insulation, resilient material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27039419 Asynchronous Microcontroller Simulation Model in VHDL
Authors: M. Kovac
Abstract:
This article describes design of the 8-bit asynchronous microcontroller simulation model in VHDL. The model is created in ISE Foundation design tool and simulated in Modelsim tool. This model is a simple application example of asynchronous systems designed in synchronous design tools. The design process of creating asynchronous system with 4-phase bundled-data protocol and with matching delays is described in the article. The model is described in gate-level abstraction. The simulation waveform of the functional construction is the result of this article. Described construction covers only the simulation model. The next step would be creating synthesizable model to FPGA.Keywords: Asynchronous, Microcontroller, VHDL, FPGA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33019418 A General Model for Acquiring Knowledge
Authors: GuoQiang Peng, Yi Sun
Abstract:
In this paper, based on the work in [1], we further give a general model for acquiring knowledge, which first focuses on the research of how and when things involved in problems are made then describes the goals, the energy and the time to give an optimum model to decide how many related things are supposed to be involved in. Finally, we acquire knowledge from this model in which there are the attributes, actions and connections of the things involved at the time when they are born and the time in their life. This model not only improves AI theories, but also surely brings the effectiveness and accuracy for AI system because systems are given more knowledge when reasoning or computing is used to bring about results.Keywords: Time, knowledge, model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10839417 Impact of Solar Energy Based Power Grid for Future Prospective of Pakistan
Authors: Muhammd Usman Sardar, Mazhar Hussain Baloch, Muhammad Shahbaz Ahmad, Zahir Javed Paracha
Abstract:
Shortfall of electrical energy in Pakistan is a challenge adversely affecting its industrial output and social growth. As elsewhere, Pakistan derives its electrical energy from a number of conventional sources. The exhaustion of petroleum and conventional resources, the rising costs coupled with extremely adverse climatic effects are taking its toll especially on the under-developed countries like Pakistan. As alternate, renewable energy sources like hydropower, solar, wind, even bio-energy and a mix of some or all of them could provide a credible alternative to the conventional energy resources that would not only be cleaner but sustainable as well. As a model, solar energy-based power grid for the near future has been attempted to offset the energy shortfalls as a mix with our existing sustainable natural energy resources. An assessment of solar energy potential for electricity generation is being presented for fulfilling the energy demands with higher level of reliability and sustainability. This model is based on the premise that solar energy potential of Pakistan is not only reliable but also sustainable. This research estimates the present & future approaching renewable energy resource specially the impact of solar energy based power grid for mitigating energy shortage in Pakistan.
Keywords: Powergrid network, solar photovoltaic (SPV) setups, solar power generation, solar energy technology (SET).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 34509416 Impact of Liquidity Crunch on Interbank Network
Authors: I. Lucas, N. Schomberg, F-A. Couturier
Abstract:
Most empirical studies have analyzed how liquidity risks faced by individual institutions turn into systemic risk. Recent banking crisis has highlighted the importance of grasping and controlling the systemic risk, and the acceptance by Central Banks to ease their monetary policies for saving default or illiquid banks. This last point shows that banks would pay less attention to liquidity risk which, in turn, can become a new important channel of loss. The financial regulation focuses on the most important and “systemic” banks in the global network. However, to quantify the expected loss associated with liquidity risk, it is worth to analyze sensitivity to this channel for the various elements of the global bank network. A small bank is not considered as potentially systemic; however the interaction of small banks all together can become a systemic element. This paper analyzes the impact of medium and small banks interaction on a set of banks which is considered as the core of the network. The proposed method uses the structure of agent-based model in a two-class environment. In first class, the data from actual balance sheets of 22 large and systemic banks (such as BNP Paribas or Barclays) are collected. In second one, to model a network as closely as possible to actual interbank market, 578 fictitious banks smaller than the ones belonging to first class have been split into two groups of small and medium ones. All banks are active on the European interbank network and have deposit and market activity. A simulation of 12 three month periods representing a midterm time interval three years is projected. In each period, there is a set of behavioral descriptions: repayment of matured loans, liquidation of deposits, income from securities, collection of new deposits, new demands of credit, and securities sale. The last two actions are part of refunding process developed in this paper. To strengthen reliability of proposed model, random parameters dynamics are managed with stochastic equations as rates the variations of which are generated by Vasicek model. The Central Bank is considered as the lender of last resort which allows banks to borrow at REPO rate and some ejection conditions of banks from the system are introduced.
Liquidity crunch due to exogenous crisis is simulated in the first class and the loss impact on other bank classes is analyzed though aggregate values representing the aggregate of loans and/or the aggregate of borrowing between classes. It is mainly shown that the three groups of European interbank network do not have the same response, and that intermediate banks are the most sensitive to liquidity risk.
Keywords: Systemic Risk, Financial Contagion, Liquidity Risk, Interbank Market, Network Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20269415 The Weight of Corporate Social Responsibility Indicators in Measurement Procedure
Authors: Grigoris Giannarakis, Despina Galani, Charitoudi Georgia, Nikolaos Litinas
Abstract:
The Corporate Social Responsibility (CSR) performance has garnered significant interest during the last two decades as numerous methodologies are proposed by Social Responsible Investment (SRI) indexes. The weight of each indicator is a crucial component of the CSR measurement procedures. Based on a previous study, the appropriate weight of each proposed indicator for the Greek telecommunication sector is specified using the rank reciprocal weighting. The Kendall-s Coefficient of Concordance and Spearman Correlation Coefficient non-parametric tests are adopted to determine the level of consensus among the experts concerning the importance rank of indicators. The results show that there is no consensus regarding the rank of indicators in most of stakeholders- domains. The equal weight for all indicators could be proposed as a solution for the lack of consensus among the experts. The study recommends three different equations concerning the adopted weight approach.
Keywords: Corporate Social Responsibility, Indicator, Weight.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23799414 An AK-Chart for the Non-Normal Data
Authors: Chia-Hau Liu, Tai-Yue Wang
Abstract:
Traditional multivariate control charts assume that measurement from manufacturing processes follows a multivariate normal distribution. However, this assumption may not hold or may be difficult to verify because not all the measurement from manufacturing processes are normal distributed in practice. This study develops a new multivariate control chart for monitoring the processes with non-normal data. We propose a mechanism based on integrating the one-class classification method and the adaptive technique. The adaptive technique is used to improve the sensitivity to small shift on one-class classification in statistical process control. In addition, this design provides an easy way to allocate the value of type I error so it is easier to be implemented. Finally, the simulation study and the real data from industry are used to demonstrate the effectiveness of the propose control charts.
Keywords: Multivariate control chart, statistical process control, one-class classification method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22699413 Study on the Deformation Modes of an Axially Crushed Compact Impact Absorption Member
Authors: Shigeyuki Haruyama, Hiroyuki Tanaka, Dai-Heng Chen, Aidil Khaidir Bin Muhamad
Abstract:
In this paper, the deformation modes of a compact impact absorption member subjected to axial compression are investigated using finite element method and experiments. A multiple combination compact impact absorption member, referred to as a 'compress-expand member', is proposed to substitute the conventional thin-walled circular tube. This study found that the proposed compact impact absorption member has stable load increase characteristics and a wider range of high load efficiency (Pave/Pmax) than the thin-walled circular tube. Moreover, the proposed compact impact absorption member can absorb larger loads in a smaller radius than the thin-walled cylindrical tube, as it can maintain its stable deformation in increased wall thicknesses.
Keywords: axial collapse, compact impact absorption member, finite element method, thin-walled cylindrical tube.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1844