Search results for: Gaussian mixture model
4951 Integrating Artificial Neural Network and Taguchi Method on Constructing the Real Estate Appraisal Model
Authors: Mu-Yen Chen, Min-Hsuan Fan, Chia-Chen Chen, Siang-Yu Jhong
Abstract:
In recent years, real estate prediction or valuation has been a topic of discussion in many developed countries. Improper hype created by investors leads to fluctuating prices of real estate, affecting many consumers to purchase their own homes. Therefore, scholars from various countries have conducted research in real estate valuation and prediction. With the back-propagation neural network that has been popular in recent years and the orthogonal array in the Taguchi method, this study aimed to find the optimal parameter combination at different levels of orthogonal array after the system presented different parameter combinations, so that the artificial neural network obtained the most accurate results. The experimental results also demonstrated that the method presented in the study had a better result than traditional machine learning. Finally, it also showed that the model proposed in this study had the optimal predictive effect, and could significantly reduce the cost of time in simulation operation. The best predictive results could be found with a fewer number of experiments more efficiently. Thus users could predict a real estate transaction price that is not far from the current actual prices.
Keywords: Artificial Neural Network, Taguchi Method, Real Estate Valuation Model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30654950 Material Selection for Footwear Insole Using Analytical Hierarchal Process
Authors: Mohammed A. Almomani, Dina W. Al-Qudah
Abstract:
Product performance depends on the type and quality of its building material. Successful product must be made using high quality material, and using the right methods. Many foot problems took place as a result of using poor insole material. Therefore, selecting a proper insole material is crucial to eliminate these problems. In this study, the analytical hierarchy process (AHP) is used to provide a systematic procedure for choosing the best material adequate for this application among three material alternatives (polyurethane, poron, and plastzote). Several comparison criteria are used to build the AHP model including: density, stiffness, durability, energy absorption, and ease of fabrication. Poron was selected as the best choice. Inconsistency testing indicates that the model is reasonable, and the materials alternative ranking is effective.
Keywords: Materials selection, biomedical insole, footwear insole, AHP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23014949 A Multi-objective Fuzzy Optimization Method of Resource Input Based on Genetic Algorithm
Abstract:
With the increasing complexity of engineering problems, the traditional, single-objective and deterministic optimization method can not meet people-s requirements. A multi-objective fuzzy optimization model of resource input is built for M chlor-alkali chemical eco-industrial park in this paper. First, the model is changed into the form that can be solved by genetic algorithm using fuzzy theory. And then, a fitness function is constructed for genetic algorithm. Finally, a numerical example is presented to show that the method compared with traditional single-objective optimization method is more practical and efficient.Keywords: Fitness function, genetic algorithm, multi-objectivefuzzy optimization, satisfaction degree membership function.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13554948 Implementation of Quantum Rotation Gates Using Controlled Non-Adiabatic Evolutions
Authors: Abdelrahman A. H. Abdelrahim, Gharib Subhi Mahmoud, Sherzod Turaev, Azeddine Messikh
Abstract:
Quantum gates are the basic building blocks in the quantum circuits model. These gates can be implemented using adiabatic or non adiabatic processes. Adiabatic models can be controlled using auxiliary qubits, whereas non adiabatic models can be simplified by using one single-shot implementation. In this paper, the controlled adiabatic evolutions is combined with the single-shot implementation to obtain quantum gates with controlled non adiabatic evolutions. This is an important improvement which can speed the implementation of quantum gates and reduce the errors due to the long run in the adiabatic model. The robustness of our scheme to different types of errors is also investigated.Keywords: Adiabatic evolutions, non adiabatic evolutions, controlled adiabatic evolutions, quantum rotation gates, dephasing rates, master equation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11674947 The Importance of Class Attendance and Cumulative GPA for Academic Success in Industrial Engineering Classes
Authors: Suleiman Obeidat, Adnan Bashir, Wisam Abu Jadayil
Abstract:
The affect of the attendance percentage, the overall GPA and the number of credit hours the student is enrolled in at specific semester on the grade attained in specific course has been studied. This study has been performed on three courses offered in industrial engineering department at the Hashemite University in Jordan. Study has revealed that the grade attained by a student is strongly affected by the attendance percentage and his overall GPA with a value of R2 of 52.5%. Another model that has been investigated is the relation between the semester GPA and the attendance percentage, the number of credit hours enrolled in at specific semester, and the overall GPA. This model gave us a strong relationship between the semester GPA and attendance percentage and the overall GPA with a value of R2 of 76.2%.Keywords: Attendance in classes, GPA, Industrial Engineering, Grade
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35904946 Social Business Models: When Profits and Impacts Are Not at Odds
Authors: Elisa Pautasso, Matteo Castagno, Michele Osella
Abstract:
In the last decade the emergence of new social needs as an effect of the economic crisis has stimulated the flourishing of business endeavours characterised by explicit social goals. Social start-ups, social enterprises or Corporate Social Responsibility operations carried out by traditional companies are quintessential examples in this regard. This paper analyses these kinds of initiatives in order to discover the main characteristics of social business models and to provide insights to social entrepreneurs for developing or improving their strategies. The research is conducted through the integration of literature review and case study analysis and, thanks to the recognition of the importance of both profits and social impacts as the key success factors for a social business model, proposes a framework for identifying indicators suitable for measuring the social impacts generated.Keywords: Business model, case study, impacts, social business.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18114945 Zero Dimensional Simulation of Combustion Process of a DI Diesel Engine Fuelled With Biofuels
Authors: Donepudi Jagadish, Ravi Kumar Puli, K. Madhu Murthy
Abstract:
A zero dimensional model has been used to investigate the combustion performance of a single cylinder direct injection diesel engine fueled by biofuels with options like supercharging and exhaust gas recirculation. The numerical simulation was performed at constant speed. The indicated pressure, temperature diagrams are plotted and compared for different fuels. The emissions of soot and nitrous oxide are computed with phenomenological models. The experimental work was also carried out with biodiesel (palm stearin methyl ester) diesel blends, ethanol diesel blends to validate simulation results with experimental results, and observed that the present model is successful in predicting the engine performance with biofuels.Keywords: Biofuels Zero Dimensional Modeling, EnginePerformance, Engine Emissions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42494944 A New Approach for Classifying Large Number of Mixed Variables
Authors: Hashibah Hamid
Abstract:
The issue of classifying objects into one of predefined groups when the measured variables are mixed with different types of variables has been part of interest among statisticians in many years. Some methods for dealing with such situation have been introduced that include parametric, semi-parametric and nonparametric approaches. This paper attempts to discuss on a problem in classifying a data when the number of measured mixed variables is larger than the size of the sample. A propose idea that integrates a dimensionality reduction technique via principal component analysis and a discriminant function based on the location model is discussed. The study aims in offering practitioners another potential tool in a classification problem that is possible to be considered when the observed variables are mixed and too large.Keywords: classification, location model, mixed variables, principal component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15574943 An Algorithm for Autonomous Aerial Navigation using MATLAB® Mapping Tool Box
Authors: Mansoor Ahsan, Suhail Akhtar, Adnan Ali, Farrukh Mazhar, Muddssar Khalid
Abstract:
In the present era of aviation technology, autonomous navigation and control have emerged as a prime area of active research. Owing to the tremendous developments in the field, autonomous controls have led today’s engineers to claim that future of aerospace vehicle is unmanned. Development of guidance and navigation algorithms for an unmanned aerial vehicle (UAV) is an extremely challenging task, which requires efforts to meet strict, and at times, conflicting goals of guidance and control. In this paper, aircraft altitude and heading controllers and an efficient algorithm for self-governing navigation using MATLAB® mapping toolbox is presented which also enables loitering of a fixed wing UAV over a specified area. For this purpose, a nonlinear mathematical model of a UAV is used. The nonlinear model is linearized around a stable trim point and decoupled for controller design. The linear controllers are tested on the nonlinear aircraft model and navigation algorithm is subsequently developed for for autonomous flight of the UAV. The results are presented for trajectory controllers and waypoint based navigation. Our investigation reveals that MATLAB® mapping toolbox can be exploited to successfully deliver an efficient algorithm for autonomous aerial navigation for a UAV.
Keywords: Navigation, trajectory-control, unmanned aerial vehicle, PID-control, MATLAB® mapping toolbox.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43804942 Bioprocess Optimization Based On Relevance Vector Regression Models and Evolutionary Programming Technique
Authors: R. Simutis, V. Galvanauskas, D. Levisauskas, J. Repsyte
Abstract:
This paper proposes a bioprocess optimization procedure based on Relevance Vector Regression models and evolutionary programming technique. Relevance Vector Regression scheme allows developing a compact and stable data-based process model avoiding time-consuming modeling expenses. The model building and process optimization procedure could be done in a half-automated way and repeated after every new cultivation run. The proposed technique was tested in a simulated mammalian cell cultivation process. The obtained results are promising and could be attractive for optimization of industrial bioprocesses.
Keywords: Bioprocess optimization, Evolutionary programming, Relevance Vector Regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21954941 DWM-CDD: Dynamic Weighted Majority Concept Drift Detection for Spam Mail Filtering
Authors: Leili Nosrati, Alireza Nemaney Pour
Abstract:
Although e-mail is the most efficient and popular communication method, unwanted and mass unsolicited e-mails, also called spam mail, endanger the existence of the mail system. This paper proposes a new algorithm called Dynamic Weighted Majority Concept Drift Detection (DWM-CDD) for content-based filtering. The design purposes of DWM-CDD are first to accurate the performance of the previously proposed algorithms, and second to speed up the time to construct the model. The results show that DWM-CDD can detect both sudden and gradual changes quickly and accurately. Moreover, the time needed for model construction is less than previously proposed algorithms.
Keywords: Concept drift, Content-based filtering, E-mail, Spammail.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19624940 Modeling and Optimization of Abrasive Waterjet Parameters using Regression Analysis
Authors: Farhad Kolahan, A. Hamid Khajavi
Abstract:
Abrasive waterjet is a novel machining process capable of processing wide range of hard-to-machine materials. This research addresses modeling and optimization of the process parameters for this machining technique. To model the process a set of experimental data has been used to evaluate the effects of various parameter settings in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. Depth of cut, as one of the most important output characteristics, has been evaluated based on different parameter settings. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. The pairwise effects of process parameters settings on process response outputs are also shown graphically. The proposed model is then embedded into a Simulated Annealing algorithm to optimize the process parameters. The optimization is carried out for any desired values of depth of cut. The objective is to determine proper levels of process parameters in order to obtain a certain level of depth of cut. Computational results demonstrate that the proposed solution procedure is quite effective in solving such multi-variable problems.
Keywords: AWJ cutting, Mathematical modeling, Simulated Annealing, Optimization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21544939 Research of Dynamics Picking Mechanism of Sulzer Projectile Loom
Authors: A. Jomartov, K. Jomartova
Abstract:
One of the main and responsible units of Sulzer projectile loom is picking mechanism. It is specifically designed to accelerate projectile to speed of 25 m / s. Initial speed projectile of Sulzer projectile loom is independent of speed loom and determined the potential energy torsion rod. This paper investigates the dynamics picking mechanism of Sulzer projectile loom during its discharge. A result of calculation model, we obtain the law of motion lever of picking mechanism during its discharge. Construction of dynamic model the picking mechanism of Sulzer projectile loom on software complex SimulationX can make calculations for different thickness of torsion rods taking into account the backlashes in the connections, the dissipative forces and resistance forcesKeywords: Dynamics, loom, picking mechanism, projectile, SimulationX.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35904938 Multi-Layer Multi-Feature Background Subtraction Using Codebook Model Framework
Authors: Yun-Tao Zhang, Jong-Yeop Bae, Whoi-Yul Kim
Abstract:
Background modeling and subtraction in video analysis has been widely used as an effective method for moving objects detection in many computer vision applications. Recently, a large number of approaches have been developed to tackle different types of challenges in this field. However, the dynamic background and illumination variations are the most frequently occurred problems in the practical situation. This paper presents a favorable two-layer model based on codebook algorithm incorporated with local binary pattern (LBP) texture measure, targeted for handling dynamic background and illumination variation problems. More specifically, the first layer is designed by block-based codebook combining with LBP histogram and mean value of each RGB color channel. Because of the invariance of the LBP features with respect to monotonic gray-scale changes, this layer can produce block wise detection results with considerable tolerance of illumination variations. The pixel-based codebook is employed to reinforce the precision from the output of the first layer which is to eliminate false positives further. As a result, the proposed approach can greatly promote the accuracy under the circumstances of dynamic background and illumination changes. Experimental results on several popular background subtraction datasets demonstrate very competitive performance compared to previous models.Keywords: Background subtraction, codebook model, local binary pattern, dynamic background, illumination changes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19654937 Analysis of Surface Hardness, Surface Roughness, and Near Surface Microstructure of AISI 4140 Steel Worked with Turn-Assisted Deep Cold Rolling Process
Authors: P. R. Prabhu, S. M. Kulkarni, S. S. Sharma, K. Jagannath, Achutha Kini U.
Abstract:
In the present study, response surface methodology has been used to optimize turn-assisted deep cold rolling process of AISI 4140 steel. A regression model is developed to predict surface hardness and surface roughness using response surface methodology and central composite design. In the development of predictive model, deep cold rolling force, ball diameter, initial roughness of the workpiece, and number of tool passes are considered as model variables. The rolling force and the ball diameter are the significant factors on the surface hardness and ball diameter and numbers of tool passes are found to be significant for surface roughness. The predicted surface hardness and surface roughness values and the subsequent verification experiments under the optimal operating conditions confirmed the validity of the predicted model. The absolute average error between the experimental and predicted values at the optimal combination of parameter settings for surface hardness and surface roughness is calculated as 0.16% and 1.58% respectively. Using the optimal processing parameters, the surface hardness is improved from 225 to 306 HV, which resulted in an increase in the near surface hardness by about 36% and the surface roughness is improved from 4.84µm to 0.252 µm, which resulted in decrease in the surface roughness by about 95%. The depth of compression is found to be more than 300µm from the microstructure analysis and this is in correlation with the results obtained from the microhardness measurements. Taylor hobson talysurf tester, micro vickers hardness tester, optical microscopy and X-ray diffractometer are used to characterize the modified surface layer.
Keywords: Surface hardness, response surface methodology, microstructure, central composite design, deep cold rolling, surface roughness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18054936 Simulation Study of Radial Heat and Mass Transfer Inside a Fixed Bed Catalytic Reactor
Authors: K. Vakhshouri, M.M. Y. Motamed Hashemi
Abstract:
A rigorous two-dimensional model is developed for simulating the operation of a less-investigated type steam reformer having a considerably lower operating Reynolds number, higher tube diameter, and non-availability of extra steam in the feed compared with conventional steam reformers. Simulation results show that reasonable predictions can only be achieved when certain correlations for wall to fluid heat transfer equations are applied. Due to severe operating conditions, in all cases, strong radial temperature gradients inside the reformer tubes have been found. Furthermore, the results show how a certain catalyst loading profile will affect the operation of the reformer.
Keywords: Steam reforming, direct reduction, heat transfer, two-dimensional model, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36444935 Continuous Feature Adaptation for Non-Native Speech Recognition
Authors: Y. Deng, X. Li, C. Kwan, B. Raj, R. Stern
Abstract:
The current speech interfaces in many military applications may be adequate for native speakers. However, the recognition rate drops quite a lot for non-native speakers (people with foreign accents). This is mainly because the nonnative speakers have large temporal and intra-phoneme variations when they pronounce the same words. This problem is also complicated by the presence of large environmental noise such as tank noise, helicopter noise, etc. In this paper, we proposed a novel continuous acoustic feature adaptation algorithm for on-line accent and environmental adaptation. Implemented by incremental singular value decomposition (SVD), the algorithm captures local acoustic variation and runs in real-time. This feature-based adaptation method is then integrated with conventional model-based maximum likelihood linear regression (MLLR) algorithm. Extensive experiments have been performed on the NATO non-native speech corpus with baseline acoustic model trained on native American English. The proposed feature-based adaptation algorithm improved the average recognition accuracy by 15%, while the MLLR model based adaptation achieved 11% improvement. The corresponding word error rate (WER) reduction was 25.8% and 2.73%, as compared to that without adaptation. The combined adaptation achieved overall recognition accuracy improvement of 29.5%, and WER reduction of 31.8%, as compared to that without adaptation. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32174934 Modelling the Occurrence of Defects and Change Requests during User Acceptance Testing
Authors: Kevin McDaid, Simon P. Wilson
Abstract:
Software developed for a specific customer under contract typically undergoes a period of testing by the customer before acceptance. This is known as user acceptance testing and the process can reveal both defects in the system and requests for changes to the product. This paper uses nonhomogeneous Poisson processes to model a real user acceptance data set from a recently developed system. In particular a split Poisson process is shown to provide an excellent fit to the data. The paper explains how this model can be used to aid the allocation of resources through the accurate prediction of occurrences both during the acceptance testing phase and before this activity begins. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23404933 Using RASCAL Code to Analyze the Postulated UF6 Fire Accident
Authors: J. R. Wang, Y. Chiang, W. S. Hsu, S. H. Chen, J. H. Yang, S. W. Chen, C. Shih, Y. F. Chang, Y. H. Huang, B. R. Shen
Abstract:
In this research, the RASCAL code was used to simulate and analyze the postulated UF6 fire accident which may occur in the Institute of Nuclear Energy Research (INER). There are four main steps in this research. In the first step, the UF6 data of INER were collected. In the second step, the RASCAL analysis methodology and model was established by using these data. Third, this RASCAL model was used to perform the simulation and analysis of the postulated UF6 fire accident. Three cases were simulated and analyzed in this step. Finally, the analysis results of RASCAL were compared with the hazardous levels of the chemicals. According to the compared results of three cases, Case 3 has the maximum danger in human health.
Keywords: RASCAL, UF6, Safety, Hydrogen fluoride.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8714932 Service Architecture for 3rd Party Operator's Participation
Authors: F. Sarabchi, A. H. Darvishan, H. Yeganeh, H. Ahmadian
Abstract:
Next generation networks with the idea of convergence of service and control layer in existing networks (fixed, mobile and data) and with the intention of providing services in an integrated network, has opened new horizon for telecom operators. On the other hand, economic problems have caused operators to look for new source of income including consider new services, subscription of more users and their promotion in using morenetwork resources and easy participation of service providers or 3rd party operators in utilizing networks. With this requirement, an architecture based on next generation objectives for service layer is necessary. In this paper, a new architecture based on IMS model explains participation of 3rd party operators in creation and implementation of services on an integrated telecom network.
Keywords: Service model, IMS, API, Scripting language, JAIN, Parlay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14734931 The Impact of Knowledge Sharing on Innovation Capability in United Arab Emirates Organizations
Authors: S. Abdallah, A. Khalil, A. Divine
Abstract:
The purpose of this study was to explore the relationship between knowledge sharing and innovation capability, by examining the influence of individual, organizational and technological factors on knowledge sharing. The research is based on a survey of 103 employees from different organizations in the United Arab Emirates. The study is based on a model and a questionnaire that was previously tested by Lin [1]. Thus, the study aims at examining the validity of that model in UAE context. The results of the research show varying degrees of correlation between the different variables, with ICT use having the strongest relationship with the innovation capabilities of organizations. The study also revealed little evidence of knowledge collecting and knowledge sharing among UAE employees.Keywords: Knowledge sharing, Organization Innovation, Technology Use, Innovation Capabilities.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25094930 Two New Relative Efficiencies of Linear Weighted Regression
Authors: Shuimiao Wan, Chao Yuan, Baoguang Tian
Abstract:
In statistics parameter theory, usually the parameter estimations have two kinds, one is the least-square estimation (LSE), and the other is the best linear unbiased estimation (BLUE). Due to the determining theorem of minimum variance unbiased estimator (MVUE), the parameter estimation of BLUE in linear model is most ideal. But since the calculations are complicated or the covariance is not given, people are hardly to get the solution. Therefore, people prefer to use LSE rather than BLUE. And this substitution will take some losses. To quantize the losses, many scholars have presented many kinds of different relative efficiencies in different views. For the linear weighted regression model, this paper discusses the relative efficiencies of LSE of β to BLUE of β. It also defines two new relative efficiencies and gives their lower bounds.Keywords: Linear weighted regression, Relative efficiency, Lower bound, Parameter estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21184929 A Fuzzy Nonlinear Regression Model for Interval Type-2 Fuzzy Sets
Authors: O. Poleshchuk, E.Komarov
Abstract:
This paper presents a regression model for interval type-2 fuzzy sets based on the least squares estimation technique. Unknown coefficients are assumed to be triangular fuzzy numbers. The basic idea is to determine aggregation intervals for type-1 fuzzy sets, membership functions of whose are low membership function and upper membership function of interval type-2 fuzzy set. These aggregation intervals were called weighted intervals. Low and upper membership functions of input and output interval type-2 fuzzy sets for developed regression models are considered as piecewise linear functions.
Keywords: Interval type-2 fuzzy sets, fuzzy regression, weighted interval.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22184928 Injection Molding of Inconel718 Parts for Aerospace Application Using Novel Binder System Based On Palm Oil Derivatives
Authors: R. Ibrahim, M. Azmirruddin, M. Jabir, N. Johari, M. Muhamad, A. R. A. Talib
Abstract:
Inconel718 has been widely used as a super alloy in aerospace application due to the high strength at elevated temperatures, satisfactory oxidation resistance and heat corrosion resistance. In this study, the Inconel718 has been fabricated using high technology of Metal Injection Molding (MIM) process due to the cost effective technique for producing small, complex and precision parts in high volume compared with conventional method through machining. Through MIM, the binder system is one of the most important criteria in order to successfully fabricate the Inconel718. Even though, the binder system is a temporary, but failure in the selection and removal of the binder system will affect on the final properties of the sintered parts. Therefore, the binder system based on palm oil derivative which is palm stearin has been formulated and developed to replace the conventional binder system. The rheological studies of the mixture between the powder and binders system have been determined properly in order to be successful during injection into injection molding machine. After molding, the binder holds the particles in place. The binder system has to be removed completely through debinding step. During debinding step, solvent debinding and thermal pyrolysis has been used to remove completely of the binder system. The debound part is then sintered to give the required physical and mechanical properties. The results show that the properties of the final sintered parts fulfill the Standard Metal Powder Industries Federation (MPIF) 35 for MIM parts.
Keywords: Binder system, rheological study, metal injection molding, debinding and sintered parts.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27154927 The Influence of Biofuels on the Permeability of Sand-Bentonite Liners
Authors: Mousa Bani Baker, Maria Elektorowicz, Adel Hanna, Altayeb Qasem
Abstract:
Liners are made to protect the groundwater table from the infiltration of leachate which normally carries different kinds of toxic materials from landfills. Although these liners are engineered to last for long period of time; unfortunately these liners fail; therefore, toxic materials pass to groundwater. This paper focuses on the changes of the hydraulic conductivity of a sand-bentonite liner due to the infiltration of biofuel and ethanol fuel. Series of laboratory tests were conducted in 20-cm-high PVC columns. Several compositions of sand-bentonite liners were tested: 95% sand: 5% bentonite; 90% sand: 10% bentonite; and 100% sand (passed mesh #40). The columns were subjected to extreme pressures of 40 kPa, and 100 kPa to evaluate the transport of alternative fuels (biofuel and ethanol fuel). For comparative studies, similar tests were carried out using water. Results showed that hydraulic conductivity increased due to the infiltration of alternative fuels through the liners. Accordingly, the increase in the hydraulic conductivity showed significant dependency on the type of liner mixture and the characteristics of the liquid. The hydraulic conductivity of a liner (subjected to biofuel infiltration) consisting of 5% bentonite: 95% sand under pressure of 40 kPa and 100 kPa had increased by one fold. In addition, the hydraulic conductivity of a liner consisting of 10% bentonite: 90% sand under pressure of 40 kPa and 100 kPa and infiltrated by biofuel had increased by three folds. On the other hand, the results obtained by water infiltration under 40 kPa showed lower hydraulic conductivities of 1.50×10-5 and 1.37×10-9 cm/s for 5% bentonite: 95% sand, and 10% bentonite: 90% sand, respectively. Similarly, under 100 kPa, the hydraulic conductivities were 2.30×10-5 and 1.90×10-9 cm/s for 5% bentonite: 95% sand, and 10% bentonite: 90% sand, respectively.Keywords: Biofuel, Ethanol; Hydraulic conductivity Landfill, Leakage, Liner failure, Liner performance Fine-grained soils, Particle size, Sand-bentonite.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20074926 A Description Logics Based Approach for Building Multi-Viewpoints Ontologies
Authors: M. Hemam, M. Djezzar, T. Djouad
Abstract:
We are interested in the problem of building an ontology in a heterogeneous organization, by taking into account different viewpoints and different terminologies of communities in the organization. Such ontology, that we call multi-viewpoint ontology, confers to the same universe of discourse, several partial descriptions, where each one is relative to a particular viewpoint. In addition, these partial descriptions share at global level, ontological elements constituent a consensus between the various viewpoints. In order to provide response elements to this problem we define a multi-viewpoints knowledge model based on viewpoint and ontology notions. The multi-viewpoints knowledge model is used to formalize the multi-viewpoints ontology in description logics language.Keywords: Description logic, knowledge engineering, ontology, viewpoint.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10244925 Effect of Progressive Type-I Right Censoring on Bayesian Statistical Inference of Simple Step–Stress Acceleration Life Testing Plan under Weibull Life Distribution
Authors: Saleem Z. Ramadan
Abstract:
This paper discusses the effects of using progressive Type-I right censoring on the design of the Simple Step Accelerated Life testing using Bayesian approach for Weibull life products under the assumption of cumulative exposure model. The optimization criterion used in this paper is to minimize the expected pre-posterior variance of the Pth percentile time of failures. The model variables are the stress changing time and the stress value for the first step. A comparison between the conventional and the progressive Type-I right censoring is provided. The results have shown that the progressive Type-I right censoring reduces the cost of testing on the expense of the test precision when the sample size is small. Moreover, the results have shown that using strong priors or large sample size reduces the sensitivity of the test precision to the censoring proportion. Hence, the progressive Type-I right censoring is recommended in these cases as progressive Type-I right censoring reduces the cost of the test and doesn't affect the precision of the test a lot. Moreover, the results have shown that using direct or indirect priors affects the precision of the test.
Keywords: Reliability, Accelerated life testing, Cumulative exposure model, Bayesian estimation, Progressive Type-I censoring, Weibull distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21614924 River Stage-Discharge Forecasting Based on Multiple-Gauge Strategy Using EEMD-DWT-LSSVM Approach
Authors: Farhad Alizadeh, Alireza Faregh Gharamaleki, Mojtaba Jalilzadeh, Houshang Gholami, Ali Akhoundzadeh
Abstract:
This study presented hybrid pre-processing approach along with a conceptual model to enhance the accuracy of river discharge prediction. In order to achieve this goal, Ensemble Empirical Mode Decomposition algorithm (EEMD), Discrete Wavelet Transform (DWT) and Mutual Information (MI) were employed as a hybrid pre-processing approach conjugated to Least Square Support Vector Machine (LSSVM). A conceptual strategy namely multi-station model was developed to forecast the Souris River discharge more accurately. The strategy used herein was capable of covering uncertainties and complexities of river discharge modeling. DWT and EEMD was coupled, and the feature selection was performed for decomposed sub-series using MI to be employed in multi-station model. In the proposed feature selection method, some useless sub-series were omitted to achieve better performance. Results approved efficiency of the proposed DWT-EEMD-MI approach to improve accuracy of multi-station modeling strategies.Keywords: River stage-discharge process, LSSVM, discrete wavelet transform (DWT), ensemble empirical decomposition mode (EEMD), multi-station modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6644923 Corruption, Economic Growth, and Income Inequality: Evidence from Ten Countries in Asia
Authors: Chiung-Ju Huang
Abstract:
This study utilizes the panel vector error correction model (PVECM) to examine the relationship among corruption, economic growth, and income inequality experienced within ten Asian countries over the 1995 to 2010 period. According to the empirical results, we do not support the common perception that corruption decreases economic growth. On the contrary, we found that corruption increases economic growth. Meanwhile, an increase in economic growth will cause an increase in income inequality, although the effect is insignificant. Similarly, an increase in income inequality will cause an increase in economic growth but a decrease in corruption, although the effect is also insignificant.Keywords: Corruption, economic growth, income inequality, panel vector error correction model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33684922 Solar Radiation Time Series Prediction
Authors: Cameron Hamilton, Walter Potter, Gerrit Hoogenboom, Ronald McClendon, Will Hobbs
Abstract:
A model was constructed to predict the amount of solar radiation that will make contact with the surface of the earth in a given location an hour into the future. This project was supported by the Southern Company to determine at what specific times during a given day of the year solar panels could be relied upon to produce energy in sufficient quantities. Due to their ability as universal function approximators, an artificial neural network was used to estimate the nonlinear pattern of solar radiation, which utilized measurements of weather conditions collected at the Griffin, Georgia weather station as inputs. A number of network configurations and training strategies were utilized, though a multilayer perceptron with a variety of hidden nodes trained with the resilient propagation algorithm consistently yielded the most accurate predictions. In addition, a modeled direct normal irradiance field and adjacent weather station data were used to bolster prediction accuracy. In later trials, the solar radiation field was preprocessed with a discrete wavelet transform with the aim of removing noise from the measurements. The current model provides predictions of solar radiation with a mean square error of 0.0042, though ongoing efforts are being made to further improve the model’s accuracy.
Keywords: Artificial Neural Networks, Resilient Propagation, Solar Radiation, Time Series Forecasting.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2763