Search results for: nonlinear exponential model
4791 X-Ray Intensity Measurement Using Frequency Output Sensor for Computed Tomography
Authors: R. M. Siddiqui, D. Z. Moghaddam, T. R. Turlapati, S. H. Khan, I. Ul Ahad
Abstract:
Quality of 2D and 3D cross-sectional images produce by Computed Tomography primarily depend upon the degree of precision of primary and secondary X-Ray intensity detection. Traditional method of primary intensity detection is apt to errors. Recently the X-Ray intensity measurement system along with smart X-Ray sensors is developed by our group which is able to detect primary X-Ray intensity unerringly. In this study a new smart X-Ray sensor is developed using Light-to-Frequency converter TSL230 from Texas Instruments which has numerous advantages in terms of noiseless data acquisition and transmission. TSL230 construction is based on a silicon photodiode which converts incoming X-Ray radiation into the proportional current signal. A current to frequency converter is attached to this photodiode on a single monolithic CMOS integrated circuit which provides proportional frequency count to incoming current signal in the form of the pulse train. The frequency count is delivered to the center of PICDEM FS USB board with PIC18F4550 microcontroller mounted on it. With highly compact electronic hardware, this Demo Board efficiently read the smart sensor output data. The frequency output approaches overcome nonlinear behavior of sensors with analog output thus un-attenuated X-Ray intensities could be measured precisely and better normalization could be acquired in order to attain high resolution.Keywords: Computed tomography, detector technology, X-Ray intensity measurement
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26094790 The Importance of Class Attendance and Cumulative GPA for Academic Success in Industrial Engineering Classes
Authors: Suleiman Obeidat, Adnan Bashir, Wisam Abu Jadayil
Abstract:
The affect of the attendance percentage, the overall GPA and the number of credit hours the student is enrolled in at specific semester on the grade attained in specific course has been studied. This study has been performed on three courses offered in industrial engineering department at the Hashemite University in Jordan. Study has revealed that the grade attained by a student is strongly affected by the attendance percentage and his overall GPA with a value of R2 of 52.5%. Another model that has been investigated is the relation between the semester GPA and the attendance percentage, the number of credit hours enrolled in at specific semester, and the overall GPA. This model gave us a strong relationship between the semester GPA and attendance percentage and the overall GPA with a value of R2 of 76.2%.Keywords: Attendance in classes, GPA, Industrial Engineering, Grade
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35904789 Social Business Models: When Profits and Impacts Are Not at Odds
Authors: Elisa Pautasso, Matteo Castagno, Michele Osella
Abstract:
In the last decade the emergence of new social needs as an effect of the economic crisis has stimulated the flourishing of business endeavours characterised by explicit social goals. Social start-ups, social enterprises or Corporate Social Responsibility operations carried out by traditional companies are quintessential examples in this regard. This paper analyses these kinds of initiatives in order to discover the main characteristics of social business models and to provide insights to social entrepreneurs for developing or improving their strategies. The research is conducted through the integration of literature review and case study analysis and, thanks to the recognition of the importance of both profits and social impacts as the key success factors for a social business model, proposes a framework for identifying indicators suitable for measuring the social impacts generated.Keywords: Business model, case study, impacts, social business.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18114788 Zero Dimensional Simulation of Combustion Process of a DI Diesel Engine Fuelled With Biofuels
Authors: Donepudi Jagadish, Ravi Kumar Puli, K. Madhu Murthy
Abstract:
A zero dimensional model has been used to investigate the combustion performance of a single cylinder direct injection diesel engine fueled by biofuels with options like supercharging and exhaust gas recirculation. The numerical simulation was performed at constant speed. The indicated pressure, temperature diagrams are plotted and compared for different fuels. The emissions of soot and nitrous oxide are computed with phenomenological models. The experimental work was also carried out with biodiesel (palm stearin methyl ester) diesel blends, ethanol diesel blends to validate simulation results with experimental results, and observed that the present model is successful in predicting the engine performance with biofuels.Keywords: Biofuels Zero Dimensional Modeling, EnginePerformance, Engine Emissions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 42494787 A New Approach for Classifying Large Number of Mixed Variables
Authors: Hashibah Hamid
Abstract:
The issue of classifying objects into one of predefined groups when the measured variables are mixed with different types of variables has been part of interest among statisticians in many years. Some methods for dealing with such situation have been introduced that include parametric, semi-parametric and nonparametric approaches. This paper attempts to discuss on a problem in classifying a data when the number of measured mixed variables is larger than the size of the sample. A propose idea that integrates a dimensionality reduction technique via principal component analysis and a discriminant function based on the location model is discussed. The study aims in offering practitioners another potential tool in a classification problem that is possible to be considered when the observed variables are mixed and too large.Keywords: classification, location model, mixed variables, principal component analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15574786 Bioprocess Optimization Based On Relevance Vector Regression Models and Evolutionary Programming Technique
Authors: R. Simutis, V. Galvanauskas, D. Levisauskas, J. Repsyte
Abstract:
This paper proposes a bioprocess optimization procedure based on Relevance Vector Regression models and evolutionary programming technique. Relevance Vector Regression scheme allows developing a compact and stable data-based process model avoiding time-consuming modeling expenses. The model building and process optimization procedure could be done in a half-automated way and repeated after every new cultivation run. The proposed technique was tested in a simulated mammalian cell cultivation process. The obtained results are promising and could be attractive for optimization of industrial bioprocesses.
Keywords: Bioprocess optimization, Evolutionary programming, Relevance Vector Regression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21954785 DWM-CDD: Dynamic Weighted Majority Concept Drift Detection for Spam Mail Filtering
Authors: Leili Nosrati, Alireza Nemaney Pour
Abstract:
Although e-mail is the most efficient and popular communication method, unwanted and mass unsolicited e-mails, also called spam mail, endanger the existence of the mail system. This paper proposes a new algorithm called Dynamic Weighted Majority Concept Drift Detection (DWM-CDD) for content-based filtering. The design purposes of DWM-CDD are first to accurate the performance of the previously proposed algorithms, and second to speed up the time to construct the model. The results show that DWM-CDD can detect both sudden and gradual changes quickly and accurately. Moreover, the time needed for model construction is less than previously proposed algorithms.
Keywords: Concept drift, Content-based filtering, E-mail, Spammail.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19624784 Modeling and Optimization of Abrasive Waterjet Parameters using Regression Analysis
Authors: Farhad Kolahan, A. Hamid Khajavi
Abstract:
Abrasive waterjet is a novel machining process capable of processing wide range of hard-to-machine materials. This research addresses modeling and optimization of the process parameters for this machining technique. To model the process a set of experimental data has been used to evaluate the effects of various parameter settings in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. Depth of cut, as one of the most important output characteristics, has been evaluated based on different parameter settings. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. The pairwise effects of process parameters settings on process response outputs are also shown graphically. The proposed model is then embedded into a Simulated Annealing algorithm to optimize the process parameters. The optimization is carried out for any desired values of depth of cut. The objective is to determine proper levels of process parameters in order to obtain a certain level of depth of cut. Computational results demonstrate that the proposed solution procedure is quite effective in solving such multi-variable problems.
Keywords: AWJ cutting, Mathematical modeling, Simulated Annealing, Optimization
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21544783 Research of Dynamics Picking Mechanism of Sulzer Projectile Loom
Authors: A. Jomartov, K. Jomartova
Abstract:
One of the main and responsible units of Sulzer projectile loom is picking mechanism. It is specifically designed to accelerate projectile to speed of 25 m / s. Initial speed projectile of Sulzer projectile loom is independent of speed loom and determined the potential energy torsion rod. This paper investigates the dynamics picking mechanism of Sulzer projectile loom during its discharge. A result of calculation model, we obtain the law of motion lever of picking mechanism during its discharge. Construction of dynamic model the picking mechanism of Sulzer projectile loom on software complex SimulationX can make calculations for different thickness of torsion rods taking into account the backlashes in the connections, the dissipative forces and resistance forcesKeywords: Dynamics, loom, picking mechanism, projectile, SimulationX.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 35904782 Multi-Layer Multi-Feature Background Subtraction Using Codebook Model Framework
Authors: Yun-Tao Zhang, Jong-Yeop Bae, Whoi-Yul Kim
Abstract:
Background modeling and subtraction in video analysis has been widely used as an effective method for moving objects detection in many computer vision applications. Recently, a large number of approaches have been developed to tackle different types of challenges in this field. However, the dynamic background and illumination variations are the most frequently occurred problems in the practical situation. This paper presents a favorable two-layer model based on codebook algorithm incorporated with local binary pattern (LBP) texture measure, targeted for handling dynamic background and illumination variation problems. More specifically, the first layer is designed by block-based codebook combining with LBP histogram and mean value of each RGB color channel. Because of the invariance of the LBP features with respect to monotonic gray-scale changes, this layer can produce block wise detection results with considerable tolerance of illumination variations. The pixel-based codebook is employed to reinforce the precision from the output of the first layer which is to eliminate false positives further. As a result, the proposed approach can greatly promote the accuracy under the circumstances of dynamic background and illumination changes. Experimental results on several popular background subtraction datasets demonstrate very competitive performance compared to previous models.Keywords: Background subtraction, codebook model, local binary pattern, dynamic background, illumination changes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19654781 Analysis of Surface Hardness, Surface Roughness, and Near Surface Microstructure of AISI 4140 Steel Worked with Turn-Assisted Deep Cold Rolling Process
Authors: P. R. Prabhu, S. M. Kulkarni, S. S. Sharma, K. Jagannath, Achutha Kini U.
Abstract:
In the present study, response surface methodology has been used to optimize turn-assisted deep cold rolling process of AISI 4140 steel. A regression model is developed to predict surface hardness and surface roughness using response surface methodology and central composite design. In the development of predictive model, deep cold rolling force, ball diameter, initial roughness of the workpiece, and number of tool passes are considered as model variables. The rolling force and the ball diameter are the significant factors on the surface hardness and ball diameter and numbers of tool passes are found to be significant for surface roughness. The predicted surface hardness and surface roughness values and the subsequent verification experiments under the optimal operating conditions confirmed the validity of the predicted model. The absolute average error between the experimental and predicted values at the optimal combination of parameter settings for surface hardness and surface roughness is calculated as 0.16% and 1.58% respectively. Using the optimal processing parameters, the surface hardness is improved from 225 to 306 HV, which resulted in an increase in the near surface hardness by about 36% and the surface roughness is improved from 4.84µm to 0.252 µm, which resulted in decrease in the surface roughness by about 95%. The depth of compression is found to be more than 300µm from the microstructure analysis and this is in correlation with the results obtained from the microhardness measurements. Taylor hobson talysurf tester, micro vickers hardness tester, optical microscopy and X-ray diffractometer are used to characterize the modified surface layer.
Keywords: Surface hardness, response surface methodology, microstructure, central composite design, deep cold rolling, surface roughness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18054780 Simulation Study of Radial Heat and Mass Transfer Inside a Fixed Bed Catalytic Reactor
Authors: K. Vakhshouri, M.M. Y. Motamed Hashemi
Abstract:
A rigorous two-dimensional model is developed for simulating the operation of a less-investigated type steam reformer having a considerably lower operating Reynolds number, higher tube diameter, and non-availability of extra steam in the feed compared with conventional steam reformers. Simulation results show that reasonable predictions can only be achieved when certain correlations for wall to fluid heat transfer equations are applied. Due to severe operating conditions, in all cases, strong radial temperature gradients inside the reformer tubes have been found. Furthermore, the results show how a certain catalyst loading profile will affect the operation of the reformer.
Keywords: Steam reforming, direct reduction, heat transfer, two-dimensional model, simulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36444779 Comparative Dynamic Performance of Load Frequency Control of Nonlinear Interconnected Hydro-Thermal System Using Intelligent Techniques
Authors: Banaja Mohanty, Prakash Kumar Hota
Abstract:
This paper demonstrates dynamic performance evaluation of load frequency control (LFC) with different intelligent techniques. All non-linearities and physical constraints have been considered in simulation studies such as governor dead band (GDB), generation rate constraint (GRC) and boiler dynamics. The conventional integral time absolute error has been considered as objective function. The design problem is formulated as an optimisation problem and particle swarm optimisation (PSO), bacterial foraging optimisation algorithm (BFOA) and differential evolution (DE) are employed to search optimal controller parameters. The superiority of the proposed approach has been shown by comparing the results with published fuzzy logic control (FLC) for the same interconnected power system. The comparison is done using various performance measures like overshoot, undershoot, settling time and standard error criteria of frequency and tie-line power deviation following a step load perturbation (SLP). It is noticed that, the dynamic performance of proposed controller is better than FLC. Further, robustness analysis is carried out by varying the time constants of speed governor, turbine, tie-line power in the range of +40% to -40% to demonstrate the robustness of the proposed DE optimized PID controller.Keywords: Automatic generation control, governor dead band, generation rate constraint, differential evolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10604778 A New Composition Method of Admissible Support Vector Kernel Based on Reproducing Kernel
Authors: Wei Zhang, Xin Zhao, Yi-Fan Zhu, Xin-Jian Zhang
Abstract:
Kernel function, which allows the formulation of nonlinear variants of any algorithm that can be cast in terms of dot products, makes the Support Vector Machines (SVM) have been successfully applied in many fields, e.g. classification and regression. The importance of kernel has motivated many studies on its composition. It-s well-known that reproducing kernel (R.K) is a useful kernel function which possesses many properties, e.g. positive definiteness, reproducing property and composing complex R.K by simple operation. There are two popular ways to compute the R.K with explicit form. One is to construct and solve a specific differential equation with boundary value whose handicap is incapable of obtaining a unified form of R.K. The other is using a piecewise integral of the Green function associated with a differential operator L. The latter benefits the computation of a R.K with a unified explicit form and theoretical analysis, whereas there are relatively later studies and fewer practical computations. In this paper, a new algorithm for computing a R.K is presented. It can obtain the unified explicit form of R.K in general reproducing kernel Hilbert space. It avoids constructing and solving the complex differential equations manually and benefits an automatic, flexible and rigorous computation for more general RKHS. In order to validate that the R.K computed by the algorithm can be used in SVM well, some illustrative examples and a comparison between R.K and Gaussian kernel (RBF) in support vector regression are presented. The result shows that the performance of R.K is close or slightly superior to that of RBF.
Keywords: admissible support vector kernel, reproducing kernel, reproducing kernel Hilbert space, Green function, support vectorregression
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15444777 Shannon-Weaver Biodiversity of Neutrophils in Fractal Networks of Immunofluorescence for Medical Diagnostics
Authors: N.E.Galich
Abstract:
We develop new nonlinear methods of immunofluorescence analysis for a sensitive technology of respiratory burst reaction of DNA fluorescence due to oxidative activity in the peripheral blood neutrophils. Histograms in flow cytometry experiments represent a fluorescence flashes frequency as functions of fluorescence intensity. We used the Shannon-Weaver index for definition of neutrophils- biodiversity and Hurst index for definition of fractal-s correlations in immunofluorescence for different donors, as the basic quantitative criteria for medical diagnostics of health status. We analyze frequencies of flashes, information, Shannon entropies and their fractals in immunofluorescence networks due to reduction of histogram range. We found the number of simplest universal correlations for biodiversity, information and Hurst index in diagnostics and classification of pathologies for wide spectra of diseases. In addition is determined the clear criterion of a common immunity and human health status in a form of yes/no answers type. These answers based on peculiarities of information in immunofluorescence networks and biodiversity of neutrophils. Experimental data analysis has shown the existence of homeostasis for information entropy in oxidative activity of DNA in neutrophil nuclei for all donors.Keywords: blood and cells fluorescence in diagnostics ofdiseases, cytometric histograms, entropy and information in fractalnetworks of oxidative activity of DNA, long-range chromosomalcorrelations in living cells.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17004776 Continuous Feature Adaptation for Non-Native Speech Recognition
Authors: Y. Deng, X. Li, C. Kwan, B. Raj, R. Stern
Abstract:
The current speech interfaces in many military applications may be adequate for native speakers. However, the recognition rate drops quite a lot for non-native speakers (people with foreign accents). This is mainly because the nonnative speakers have large temporal and intra-phoneme variations when they pronounce the same words. This problem is also complicated by the presence of large environmental noise such as tank noise, helicopter noise, etc. In this paper, we proposed a novel continuous acoustic feature adaptation algorithm for on-line accent and environmental adaptation. Implemented by incremental singular value decomposition (SVD), the algorithm captures local acoustic variation and runs in real-time. This feature-based adaptation method is then integrated with conventional model-based maximum likelihood linear regression (MLLR) algorithm. Extensive experiments have been performed on the NATO non-native speech corpus with baseline acoustic model trained on native American English. The proposed feature-based adaptation algorithm improved the average recognition accuracy by 15%, while the MLLR model based adaptation achieved 11% improvement. The corresponding word error rate (WER) reduction was 25.8% and 2.73%, as compared to that without adaptation. The combined adaptation achieved overall recognition accuracy improvement of 29.5%, and WER reduction of 31.8%, as compared to that without adaptation. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32174775 Effect of Confinement on the Bearing Capacity and Settlement of Spread Foundations
Authors: Tahsin Toma Sabbagh, Ihsan Al-Abboodi, Ali Al-Jazaairry
Abstract:
Allowable-bearing capacity is the competency of soil to safely carries the pressure from the superstructure without experiencing a shear failure with accompanying excessive settlements. Ensuring a safe bearing pressure with respect to failure does not tolerate settlement of the foundation will be within acceptable limits. Therefore, settlement analysis should always be performed since most structures are settlement sensitive. When visualising the movement of a soil wedge in the bearing capacity criterion, both vertically and horizontally, it becomes clear that by confining the soil surrounding the foundation, both the bearing capacity and settlement values improve. In this study, two sizes of spread foundation were considered; (2×4) m and (3×5) m. These represent two real problem case studies of an existing building. The foundations were analysed in terms of dimension as well as position with respect to a confining wall (i.e., sheet piles on both sides). Assuming B is the least foundation dimension, the study comprised the analyses of three distances; (0.1 B), (0.5 B), and (0.75 B) between the sheet piles and foundations alongside three depths of confinement (0.5 B), (1 B), and (1.5 B). Nonlinear three-dimensional finite element analysis (ANSYS) was adopted to perform an analytical investigation on the behaviour of the two foundations contained by the case study. Results showed that confinement of foundations reduced the overall stresses near the foundation by 65% and reduced the vertical displacement by 90%. Moreover, the most effective distance between the confinement wall and the foundation was found to be 0.5 B.
Keywords: Bearing capacity, cohesionless soils, spread footings, soil confinement, soil modelling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8904774 Modelling the Occurrence of Defects and Change Requests during User Acceptance Testing
Authors: Kevin McDaid, Simon P. Wilson
Abstract:
Software developed for a specific customer under contract typically undergoes a period of testing by the customer before acceptance. This is known as user acceptance testing and the process can reveal both defects in the system and requests for changes to the product. This paper uses nonhomogeneous Poisson processes to model a real user acceptance data set from a recently developed system. In particular a split Poisson process is shown to provide an excellent fit to the data. The paper explains how this model can be used to aid the allocation of resources through the accurate prediction of occurrences both during the acceptance testing phase and before this activity begins. Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23404773 Using RASCAL Code to Analyze the Postulated UF6 Fire Accident
Authors: J. R. Wang, Y. Chiang, W. S. Hsu, S. H. Chen, J. H. Yang, S. W. Chen, C. Shih, Y. F. Chang, Y. H. Huang, B. R. Shen
Abstract:
In this research, the RASCAL code was used to simulate and analyze the postulated UF6 fire accident which may occur in the Institute of Nuclear Energy Research (INER). There are four main steps in this research. In the first step, the UF6 data of INER were collected. In the second step, the RASCAL analysis methodology and model was established by using these data. Third, this RASCAL model was used to perform the simulation and analysis of the postulated UF6 fire accident. Three cases were simulated and analyzed in this step. Finally, the analysis results of RASCAL were compared with the hazardous levels of the chemicals. According to the compared results of three cases, Case 3 has the maximum danger in human health.
Keywords: RASCAL, UF6, Safety, Hydrogen fluoride.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8714772 Service Architecture for 3rd Party Operator's Participation
Authors: F. Sarabchi, A. H. Darvishan, H. Yeganeh, H. Ahmadian
Abstract:
Next generation networks with the idea of convergence of service and control layer in existing networks (fixed, mobile and data) and with the intention of providing services in an integrated network, has opened new horizon for telecom operators. On the other hand, economic problems have caused operators to look for new source of income including consider new services, subscription of more users and their promotion in using morenetwork resources and easy participation of service providers or 3rd party operators in utilizing networks. With this requirement, an architecture based on next generation objectives for service layer is necessary. In this paper, a new architecture based on IMS model explains participation of 3rd party operators in creation and implementation of services on an integrated telecom network.
Keywords: Service model, IMS, API, Scripting language, JAIN, Parlay.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14734771 The Impact of Knowledge Sharing on Innovation Capability in United Arab Emirates Organizations
Authors: S. Abdallah, A. Khalil, A. Divine
Abstract:
The purpose of this study was to explore the relationship between knowledge sharing and innovation capability, by examining the influence of individual, organizational and technological factors on knowledge sharing. The research is based on a survey of 103 employees from different organizations in the United Arab Emirates. The study is based on a model and a questionnaire that was previously tested by Lin [1]. Thus, the study aims at examining the validity of that model in UAE context. The results of the research show varying degrees of correlation between the different variables, with ICT use having the strongest relationship with the innovation capabilities of organizations. The study also revealed little evidence of knowledge collecting and knowledge sharing among UAE employees.Keywords: Knowledge sharing, Organization Innovation, Technology Use, Innovation Capabilities.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25094770 Two New Relative Efficiencies of Linear Weighted Regression
Authors: Shuimiao Wan, Chao Yuan, Baoguang Tian
Abstract:
In statistics parameter theory, usually the parameter estimations have two kinds, one is the least-square estimation (LSE), and the other is the best linear unbiased estimation (BLUE). Due to the determining theorem of minimum variance unbiased estimator (MVUE), the parameter estimation of BLUE in linear model is most ideal. But since the calculations are complicated or the covariance is not given, people are hardly to get the solution. Therefore, people prefer to use LSE rather than BLUE. And this substitution will take some losses. To quantize the losses, many scholars have presented many kinds of different relative efficiencies in different views. For the linear weighted regression model, this paper discusses the relative efficiencies of LSE of β to BLUE of β. It also defines two new relative efficiencies and gives their lower bounds.Keywords: Linear weighted regression, Relative efficiency, Lower bound, Parameter estimation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21194769 A Description Logics Based Approach for Building Multi-Viewpoints Ontologies
Authors: M. Hemam, M. Djezzar, T. Djouad
Abstract:
We are interested in the problem of building an ontology in a heterogeneous organization, by taking into account different viewpoints and different terminologies of communities in the organization. Such ontology, that we call multi-viewpoint ontology, confers to the same universe of discourse, several partial descriptions, where each one is relative to a particular viewpoint. In addition, these partial descriptions share at global level, ontological elements constituent a consensus between the various viewpoints. In order to provide response elements to this problem we define a multi-viewpoints knowledge model based on viewpoint and ontology notions. The multi-viewpoints knowledge model is used to formalize the multi-viewpoints ontology in description logics language.Keywords: Description logic, knowledge engineering, ontology, viewpoint.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10244768 Effect of Progressive Type-I Right Censoring on Bayesian Statistical Inference of Simple Step–Stress Acceleration Life Testing Plan under Weibull Life Distribution
Authors: Saleem Z. Ramadan
Abstract:
This paper discusses the effects of using progressive Type-I right censoring on the design of the Simple Step Accelerated Life testing using Bayesian approach for Weibull life products under the assumption of cumulative exposure model. The optimization criterion used in this paper is to minimize the expected pre-posterior variance of the Pth percentile time of failures. The model variables are the stress changing time and the stress value for the first step. A comparison between the conventional and the progressive Type-I right censoring is provided. The results have shown that the progressive Type-I right censoring reduces the cost of testing on the expense of the test precision when the sample size is small. Moreover, the results have shown that using strong priors or large sample size reduces the sensitivity of the test precision to the censoring proportion. Hence, the progressive Type-I right censoring is recommended in these cases as progressive Type-I right censoring reduces the cost of the test and doesn't affect the precision of the test a lot. Moreover, the results have shown that using direct or indirect priors affects the precision of the test.
Keywords: Reliability, Accelerated life testing, Cumulative exposure model, Bayesian estimation, Progressive Type-I censoring, Weibull distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21614767 River Stage-Discharge Forecasting Based on Multiple-Gauge Strategy Using EEMD-DWT-LSSVM Approach
Authors: Farhad Alizadeh, Alireza Faregh Gharamaleki, Mojtaba Jalilzadeh, Houshang Gholami, Ali Akhoundzadeh
Abstract:
This study presented hybrid pre-processing approach along with a conceptual model to enhance the accuracy of river discharge prediction. In order to achieve this goal, Ensemble Empirical Mode Decomposition algorithm (EEMD), Discrete Wavelet Transform (DWT) and Mutual Information (MI) were employed as a hybrid pre-processing approach conjugated to Least Square Support Vector Machine (LSSVM). A conceptual strategy namely multi-station model was developed to forecast the Souris River discharge more accurately. The strategy used herein was capable of covering uncertainties and complexities of river discharge modeling. DWT and EEMD was coupled, and the feature selection was performed for decomposed sub-series using MI to be employed in multi-station model. In the proposed feature selection method, some useless sub-series were omitted to achieve better performance. Results approved efficiency of the proposed DWT-EEMD-MI approach to improve accuracy of multi-station modeling strategies.Keywords: River stage-discharge process, LSSVM, discrete wavelet transform (DWT), ensemble empirical decomposition mode (EEMD), multi-station modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6644766 Corruption, Economic Growth, and Income Inequality: Evidence from Ten Countries in Asia
Authors: Chiung-Ju Huang
Abstract:
This study utilizes the panel vector error correction model (PVECM) to examine the relationship among corruption, economic growth, and income inequality experienced within ten Asian countries over the 1995 to 2010 period. According to the empirical results, we do not support the common perception that corruption decreases economic growth. On the contrary, we found that corruption increases economic growth. Meanwhile, an increase in economic growth will cause an increase in income inequality, although the effect is insignificant. Similarly, an increase in income inequality will cause an increase in economic growth but a decrease in corruption, although the effect is also insignificant.Keywords: Corruption, economic growth, income inequality, panel vector error correction model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33684765 A Comparative Study of Standard, Casted and Riveted Eye Design of a Mono Leaf Spring Using CAE Tools
Authors: Gian Bhushan, Vinkel Arora, M.L. Aggarwal
Abstract:
The objective of the present study is to determine better eye end design of a mono leaf spring used in light motor vehicle. A conventional 65Si7 spring steel leaf spring model with standard eye, casted and riveted eye end are considered. The CAD model of the leaf springs is prepared in CATIA and analyzed using ANSYS. The standard eye, casted and riveted eye leaf springs are subjected to similar loading conditions. The CAE analysis of the leaf spring is performed for various parameters like deflection and Von- Mises stress. Mass reduction of 62.9% is achieved in case of riveted eye mono leaf spring as compared to standard eye mono leaf spring for the same loading conditions.
Keywords: CAE, Leaf Spring, 65Si7 spring steel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23814764 An Agent Based Simulation for Network Formation with Heterogeneous Agents
Authors: Hisashi Kojima, Masatora Daito
Abstract:
We investigate an asymmetric connections model with a dynamic network formation process, using an agent based simulation. We permit heterogeneity of agents- value. Valuable persons seem to have many links on real social networks. We focus on this point of view, and examine whether valuable agents change the structures of the terminal networks. Simulation reveals that valuable agents diversify the terminal networks. We can not find evidence that valuable agents increase the possibility that star networks survive the dynamic process. We find that valuable agents disperse the degrees of agents in each terminal network on an average.Keywords: network formation, agent based simulation, connections model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12884763 Structural Assessment of Low-rise Reinforced Concrete Frames under Tsunami Loads
Authors: Hussain Jiffry, Kypros Pilakoutas, Reyes Garcia
Abstract:
This study examines analytically the effect of tsunami loads on reinforced concrete (RC) frame buildings. The impact of tsunami wave loads and waterborne objects are analyzed using a typical substandard full-scale two-story RC frame building tested as part of the EU-funded Ecoleader project. The building was subjected to shake table tests in bare condition, and subsequently strengthened using Carbon Fiber Reinforced Polymers (CFRP) composites and retested. Numerical models of the building in both bare and CFRP-strengthened conditions are calibrated in DRAIN-3DX software to match the test results. To investigate the response of wave loads and impact forces, the numerical models are subjected to nonlinear dynamic analyses using force time-history input records. The analytical results are compared in terms of displacements at the floors and at the “impact point” of a boat. The results show that the roof displacement of the CFRP-strengthened building reduced by 63% when compared to the bare building. The results also indicate that strengthening only the mid-height of the impact column using CFRP is more effective at reducing damage when compared to strengthening other parts of the column. Alternative solutions to mitigate damage due to tsunami loads are suggested.
Keywords: Tsunami loads, hydrodynamic load, impact load, waterborne objects, RC buildings.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19344762 Magneto-Thermo-Mechanical Analysis of Electromagnetic Devices Using the Finite Element Method
Authors: Michael G. Pantelyat
Abstract:
Fundamental basics of pure and applied research in the area of magneto-thermo-mechanical numerical analysis and design of innovative electromagnetic devices (modern induction heaters, novel thermoelastic actuators, rotating electrical machines, induction cookers, electrophysical devices) are elaborated. Thus, mathematical models of magneto-thermo-mechanical processes in electromagnetic devices taking into account main interactions of interrelated phenomena are developed. In addition, graphical representation of coupled (multiphysics) phenomena under consideration is proposed. Besides, numerical techniques for nonlinear problems solution are developed. On this base, effective numerical algorithms for solution of actual problems of practical interest are proposed, validated and implemented in applied 2D and 3D computer codes developed. Many applied problems of practical interest regarding modern electrical engineering devices are numerically solved. Investigations of the influences of various interrelated physical phenomena (temperature dependences of material properties, thermal radiation, conditions of convective heat transfer, contact phenomena, etc.) on the accuracy of the electromagnetic, thermal and structural analyses are conducted. Important practical recommendations on the choice of rational structures, materials and operation modes of electromagnetic devices under consideration are proposed and implemented in industry.Keywords: Electromagnetic devices, multiphysics, numerical analysis, simulation and design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1722