Search results for: variance analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8844

Search results for: variance analysis

8514 Sensitivity Analysis in Power Systems Reliability Evaluation

Authors: A.R Alesaadi, M. Nafar, A.H. Gheisari

Abstract:

In this paper sensitivity analysis is performed for reliability evaluation of power systems. When examining the reliability of a system, it is useful to recognize how results change as component parameters are varied. This knowledge helps engineers to understand the impact of poor data, and gives insight on how reliability can be improved. For these reasons, a sensitivity analysis can be performed. Finally, a real network was used for testing the presented method.

Keywords: sensitivity analysis, reliability evaluation, powersystems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2273
8513 Compositional and Morphological Characteristics of the Tissues of Three Common Dates Grown in Algeria

Authors: H. Amellal-Chibane, Y. Noui, A. Djouab, S. Benamara

Abstract:

Mech-Degla, Degla-Beida and Frezza are the date (Phoenix dactylifera L.) common varieties with a more or less good availability and feeble trade value. Some morphologic and physicochemical factors were determined. Results show that the whole date weight is significantly different (P= 95%) concerning Mech-Degla and Degla-Beida which are more commercialized than Frezza whereas the pulp mass proportion in relation to whole fruits is highest for Frezza (88.28%). Moreover, there is a large variability concerning the weights and densities of constitutive tissues in each variety. The white tissue is dominant in Mech-Degla in opposite to the two other varieties. The variance analyze showed that the difference in weights between brown and white tissues is significant (P = 95%) for all studied varieties. Some other morphologic and chemical proprieties of the whole pulps and their two constitutive parts (brown or pigmented and white) are also investigated. The predominance of phenolics in Mech-Degla (4.01g/100g, w.b) and Frezza (4.96 g/100g, w.b) pulps brown part is the main result revealed in this study.

Keywords: Common dates, phenolics, sugars, tissues.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2275
8512 Efficient Antenna Array Beamforming with Robustness against Random Steering Mismatch

Authors: Ju-Hong Lee, Ching-Wei Liao, Kun-Che Lee

Abstract:

This paper deals with the problem of using antenna sensors for adaptive beamforming in the presence of random steering mismatch. We present an efficient adaptive array beamformer with robustness to deal with the considered problem. The robustness of the proposed beamformer comes from the efficient designation of the steering vector. Using the received array data vector, we construct an appropriate correlation matrix associated with the received array data vector and a correlation matrix associated with signal sources. Then, the eigenvector associated with the largest eigenvalue of the constructed signal correlation matrix is designated as an appropriate estimate of the steering vector. Finally, the adaptive weight vector required for adaptive beamforming is obtained by using the estimated steering vector and the constructed correlation matrix of the array data vector. Simulation results confirm the effectiveness of the proposed method.

Keywords: Adaptive beamforming, antenna array, linearly constrained minimum variance, robustness, steering vector.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 697
8511 Stock Portfolio Selection Using Chemical Reaction Optimization

Authors: Jin Xu, Albert Y.S. Lam, Victor O.K. Li

Abstract:

Stock portfolio selection is a classic problem in finance, and it involves deciding how to allocate an institution-s or an individual-s wealth to a number of stocks, with certain investment objectives (return and risk). In this paper, we adopt the classical Markowitz mean-variance model and consider an additional common realistic constraint, namely, the cardinality constraint. Thus, stock portfolio optimization becomes a mixed-integer quadratic programming problem and it is difficult to be solved by exact optimization algorithms. Chemical Reaction Optimization (CRO), which mimics the molecular interactions in a chemical reaction process, is a population-based metaheuristic method. Two different types of CRO, named canonical CRO and Super Molecule-based CRO (S-CRO), are proposed to solve the stock portfolio selection problem. We test both canonical CRO and S-CRO on a benchmark and compare their performance under two criteria: Markowitz efficient frontier (Pareto frontier) and Sharpe ratio. Computational experiments suggest that S-CRO is promising in handling the stock portfolio optimization problem.

Keywords: Stock portfolio selection, Markowitz model, Chemical Reaction Optimization, Sharpe ratio

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2075
8510 Acoustic Instabilities on Swirling Flames

Authors: T. Parra, R. Z. Szasz, C. Duwig, R. Pérez, V. Mendoza, F. Castro

Abstract:

The POD makes possible to reduce the complete high-dimensional acoustic field to a low-dimensional subspace where different modes are identified and let reconstruct in a simple way a high percentage of the variance of the field.

Rotating modes are instabilities which are commonly observed in swirling flows. Such modes can appear under both cold and reacting conditions but that they have different sources: while the cold flow rotating mode is essentially hydrodynamic and corresponds to the wellknown PVC (precessing vortex core) observed in many swirled unconfined flows, the rotating structure observed for the reacting case inside the combustion chamber might be not hydrodynamically but acoustically controlled. The two transverse acoustic modes of the combustion chamber couple and create a rotating motion of the flame which leads to a self-sustained turning mode which has the features of a classical PVC but a very different source (acoustics and not hydrodynamics).

Keywords: Acoustic field, POD, swirling flames.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2330
8509 Composition Dependent Formation of Sputtered Co-Cu Film on Cr Under-Layer

Authors: Watcharee Rattanasakulthong, Pichai Sirisangsawang, Supree Pinitsoontorn

Abstract:

Sputtered CoxCu100-x films with the different compositions of x = 57.7, 45.8, 25.5, 13.8, 8.8, 7.5 and 1.8 were deposited on Cr under-layer by RF-sputtering. SEM result reveals that the averaged thickness of Co-Cu film and Cr under-layer are 92 nm and 22nm, respectively. All Co-Cu films are composed of Co (FCC) and Cu (FCC) phases in (111) directions on BCC-Cr (110) under-layers. Magnetic properties, surface roughness and morphology of Co-Cu films are dependent on the film composition. The maximum and minimum surface roughness of 3.24 and 1.16nm are observed on the Co7.5Cu92.5 and Co45.8Cu54.2films, respectively. It can be described that the variance of surface roughness of the film because of the difference of the agglomeration rate of Co and Cu atoms on Cr under-layer. The Co57.5Cu42.3, Co45.8Cu54.2 and Co25.5Cu74.5 films shows the ferromagnetic phase whereas the rest of the film exhibits the paramagnetic phase at room temperature. The saturation magnetization, remnant magnetization and coercive field of Co-Cu films on Cr under-layer are slightly increased with increasing the Co composition. It can be concluded that the required magnetic properties and surface roughness of the Co-Cu film can be adapted by the adjustment of the film composition.

Keywords: Co-Cu films, Under-layers, Sputtering, Surface roughness, Magnetic properties, Atomic force microscopy (AFM).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1944
8508 Effects of School Facilities’ Mechanical and Plumbing Characteristics and Conditions on Student Attendance, Academic Performance and Health

Authors: Erica Cochran Hameen, Bobuchi Ken-Opurum, Shalini Priyadarshini, Berangere Lartigue, Sadhana Anath-Pisipati

Abstract:

School districts throughout the United States are constantly seeking measures to improve test scores, reduce school absenteeism and improve indoor environmental quality. It is imperative to identify key building investments which will provide the largest benefits to schools in terms of improving the aforementioned factors. This study uses Analysis of Variance (ANOVA) tests to statistically evaluate the impact of a school building’s mechanical and plumbing characteristics on a child’s educational performance. The educational performance is measured via three indicators, i.e. test scores, suspensions, and absenteeism. The study investigated 125 New York City school facilities to determine the potential correlations between 50 mechanical and plumbing variables and the performance indicators. Key findings from the tests revealed that elementary schools with pneumatic systems in “good” condition have 48.8% lower percentages of students scoring at the minimum English Language Arts (ELA) competency level compared with those with no pneumatic system. Additionally, elementary schools with “unit heaters/cabinet heaters” in “good to fair” conditions have 1.1% higher attendance rates compared to schools with no “unit heaters/cabinet heaters” or those in inferior condition. Furthermore, elementary schools with air conditioning have 0.6% higher attendance rates compared to schools with no air conditioning, and those with interior floor drains in “good” condition have 1.8% higher attendance rates compared to schools with interior drains in inferior condition.

Keywords: Academic attendance and performance, mechanical and plumbing systems, schools, student health.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 648
8507 Data and Spatial Analysis for Economy and Education of 28 E.U. Member-States for 2014

Authors: Alexiou Dimitra, Fragkaki Maria

Abstract:

The objective of the paper is the study of geographic, economic and educational variables and their contribution to determine the position of each member-state among the EU-28 countries based on the values of seven variables as given by Eurostat. The Data Analysis methods of Multiple Factorial Correspondence Analysis (MFCA) Principal Component Analysis and Factor Analysis have been used. The cross tabulation tables of data consist of the values of seven variables for the 28 countries for 2014. The data are manipulated using the CHIC Analysis V 1.1 software package. The results of this program using MFCA and Ascending Hierarchical Classification are given in arithmetic and graphical form. For comparison reasons with the same data the Factor procedure of Statistical package IBM SPSS 20 has been used. The numerical and graphical results presented with tables and graphs, demonstrate the agreement between the two methods. The most important result is the study of the relation between the 28 countries and the position of each country in groups or clouds, which are formed according to the values of the corresponding variables.

Keywords: Multiple factorial correspondence analysis, principal component analysis, factor analysis, E.U.-28 countries, statistical package IBM SPSS 20, CHIC Analysis V 1.1 Software, Eurostat.eu statistics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1085
8506 Statistical Description of Counterpoise Effective Length Based On Regressive Formulas

Authors: Petar Sarajcev, Josip Vasilj, Damir Jakus

Abstract:

This paper presents a novel statistical description of the counterpoise effective length due to lightning surges, where the (impulse) effective length had been obtained by means of regressive formulas applied to the transient simulation results. The effective length is described in terms of a statistical distribution function, from which median, mean, variance, and other parameters of interest could be readily obtained. The influence of lightning current amplitude, lightning front duration, and soil resistivity on the effective length has been accounted for, assuming statistical nature of these parameters. A method for determining the optimal counterpoise length, in terms of the statistical impulse effective length, is also presented. It is based on estimating the number of dangerous events associated with lightning strikes. Proposed statistical description and the associated method provide valuable information which could aid the design engineer in optimising physical lengths of counterpoises in different grounding arrangements and soil resistivity situations.

Keywords: Counterpoise, Grounding conductor, Effective length, Lightning, Monte Carlo method, Statistical distribution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2794
8505 The Establishment of Probabilistic Risk Assessment Analysis Methodology for Dry Storage Concrete Casks Using SAPHIRE 8

Authors: J. R. Wang, W. Y. Cheng, J. S. Yeh, S. W. Chen, Y. M. Ferng, J. H. Yang, W. S. Hsu, C. Shih

Abstract:

To understand the risk for dry storage concrete casks in the cask loading, transfer, and storage phase, the purpose of this research is to establish the probabilistic risk assessment (PRA) analysis methodology for dry storage concrete casks by using SAPHIRE 8 code. This analysis methodology is used to perform the study of Taiwan nuclear power plants (NPPs) dry storage system. The process of research has three steps. First, the data of the concrete casks and Taiwan NPPs are collected. Second, the PRA analysis methodology is developed by using SAPHIRE 8. Third, the PRA analysis is performed by using this methodology. According to the analysis results, the maximum risk is the multipurpose canister (MPC) drop case.

Keywords: PRA, Dry storage, concrete cask, SAPHIRE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 864
8504 Carbon Disulfide Production via Hydrogen Sulfide Methane Reformation

Authors: H. Hosseini, M. Javadi, M. Moghiman, M. H. Ghodsi Rad

Abstract:

Carbon disulfide is widely used for the production of viscose rayon, rubber, and other organic materials and it is a feedstock for the synthesis of sulfuric acid. The objective of this paper is to analyze possibilities for efficient production of CS2 from sour natural gas reformation (H2SMR) (2H2S+CH4 =CS2 +4H2) . Also, the effect of H2S to CH4 feed ratio and reaction temperature on carbon disulfide production is investigated numerically in a reforming reactor. The chemical reaction model is based on an assumed Probability Density Function (PDF) parameterized by the mean and variance of mixture fraction and β-PDF shape. The results show that the major factors influencing CS2 production are reactor temperature. The yield of carbon disulfide increases with increasing H2S to CH4 feed gas ratio (H2S/CH4≤4). Also the yield of C(s) increases with increasing temperature until the temperature reaches to 1000°K, and then due to increase of CS2 production and consumption of C(s), yield of C(s) drops with further increase in the temperature. The predicted CH4 and H2S conversion and yield of carbon disulfide are in good agreement with result of Huang and TRaissi.

Keywords: Carbon disulfide, sour natural gas, H2SMR, probability density function.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5257
8503 A Chaotic Study on Tremor Behavior of Parkinsonian Patients under Deep Brain Stimulation

Authors: M. Sadeghi, A.H. Jafari, S.M.P. Firoozabadi

Abstract:

Deep Brain Stimulation or DBS is a surgical treatment for Parkinson-s Disease with three stimulation parameters: frequency, pulse width, and voltage. The parameters should be selected appropriately to achieve effective treatment. This selection now, performs clinically. The aim of this research is to study chaotic behavior of recorded tremor of patients under DBS in order to present a computational method to recognize stimulation optimum voltage. We obtained some chaotic features of tremor signal, and discovered embedding space of it has an attractor, and its largest Lyapunov exponent is positive, which show tremor signal has chaotic behavior, also we found out, in optimal voltage, entropy and embedding space variance of tremor signal have minimum values in comparison with other voltages. These differences can help neurologists recognize optimal voltage numerically, which leads to reduce patients' role and discomfort in optimizing stimulation parameters and to do treatment with high accuracy.

Keywords: Chaos, Deep Brain Stimulation, Parkinson's Disease, Stimulation Parameters, tremor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1827
8502 Research on the Predict Method of Random Vibration Cumulative Fatigue Damage Life Based on the Finite Element Analysis

Authors: Wang Chengcheng, Li Chuanri, Xu Fei, Guo Ying

Abstract:

Aiming at most of the aviation products are facing the problem of fatigue fracture in vibration environment, we makes use of the testing result of a bracket, analysis for the structure with ANSYS-Workbench, predict the life of the bracket by different ways, and compared with the testing result. With the research on analysis methods, make an organic combination of simulation analysis and testing, Not only ensure the accuracy of simulation analysis and life predict, but also make a dynamic supervision of product life process, promote the application of finite element simulation analysis in engineering practice.

Keywords: Random vibration, finite element simulation, fatigue, frequency domain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4708
8501 Principal Component Analysis using Singular Value Decomposition of Microarray Data

Authors: Dong Hoon Lim

Abstract:

A series of microarray experiments produces observations of differential expression for thousands of genes across multiple conditions. Principal component analysis(PCA) has been widely used in multivariate data analysis to reduce the dimensionality of the data in order to simplify subsequent analysis and allow for summarization of the data in a parsimonious manner. PCA, which can be implemented via a singular value decomposition(SVD), is useful for analysis of microarray data. For application of PCA using SVD we use the DNA microarray data for the small round blue cell tumors(SRBCT) of childhood by Khan et al.(2001). To decide the number of components which account for sufficient amount of information we draw scree plot. Biplot, a graphic display associated with PCA, reveals important features that exhibit relationship between variables and also the relationship of variables with observations.

Keywords: Principal component analysis, singular value decomposition, microarray data, SRBCT

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3250
8500 Quantitative Analysis of PCA, ICA, LDA and SVM in Face Recognition

Authors: Liton Jude Rozario, Mohammad Reduanul Haque, Md. Ziarul Islam, Mohammad Shorif Uddin

Abstract:

Face recognition is a technique to automatically identify or verify individuals. It receives great attention in identification, authentication, security and many more applications. Diverse methods had been proposed for this purpose and also a lot of comparative studies were performed. However, researchers could not reach unified conclusion. In this paper, we are reporting an extensive quantitative accuracy analysis of four most widely used face recognition algorithms: Principal Component Analysis (PCA), Independent Component Analysis (ICA), Linear Discriminant Analysis (LDA) and Support Vector Machine (SVM) using AT&T, Sheffield and Bangladeshi people face databases under diverse situations such as illumination, alignment and pose variations.

Keywords: PCA, ICA, LDA, SVM, face recognition, noise.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2431
8499 Containment/Penetration Analysis for the Protection of Aircraft Engine External Configuration and Nuclear Power Plant Structures

Authors: Dong Wook Lee, Adrian Mistreanu

Abstract:

The authors have studied a method for analyzing containment and penetration using an explicit nonlinear Finite Element Analysis. This method may be used in the stage of concept design for the protection of external configurations or components of aircraft engines and nuclear power plant structures. This paper consists of the modeling method, the results obtained from the method and the comparison of the results with those calculated from simple analytical method. It shows that the containment capability obtained by proposed method matches well with analytically calculated containment capability.

Keywords: Computer Aided Engineering, CAE, containment analysis, Finite Element Analysis, FEA, impact analysis, penetration analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 539
8498 Phelipanche ramosa (L. - Pomel) Control in Field Tomato Crop

Authors: Disciglio G., Lops F., Carlucci A., Gatta G., Tarantino A., Frabboni L., Carriero F., Cibelli F., Raimondo M. L., Tarantino E.

Abstract:

The tomato is a very important crop, whose cultivation in the Mediterranean basin is severely affected by the phytoparasitic weed Phelipanche ramosa. The semiarid regions of the world are considered the main areas where this parasitic weed is established causing heavy infestation as it is able to produce high numbers of seeds (up to 500,000 per plant), which remain viable for extended period (more than 20 years). In this paper the results obtained from eleven treatments in order to control this parasitic weed including chemical, agronomic, biological and biotechnological methods compared with the untreated test under two plowing depths (30 and 50 cm) are reported. The split-plot design with 3 replicates was adopted. In 2014 a trial was performed in Foggia province (southern Italy) on processing tomato (cv Docet) grown in the field infested by Phelipanche ramosa. Tomato seedlings were transplant on May 5, on a clay-loam soil. During the growing cycle of the tomato crop, at 56-78 and 92 days after transplantation, the number of parasitic shoots emerged in each plot was detected. At tomato harvesting, on August 18, the major quantity-quality yield parameters were determined (marketable yield, mean weight, dry matter, pH, soluble solids and color of fruits). All data were subjected to analysis of variance (ANOVA) and the means were compared by Tukey's test. Each treatment studied did not provide complete control against Phelipanche ramosa. However, among the different methods tested, some of them which Fusarium, gliphosate, radicon biostimulant and Red Setter tomato cv (improved genotypes obtained by Tilling technology) under deeper plowing (50 cm depth) proved to mitigate the virulence of the Phelipanche ramose attacks. It is assumed that these effects can be improved combining some of these treatments each other, especially for a gradual and continuing reduction of the “seed bank” of the parasite in the soil.

Keywords: Control methods, Phelipanche ramosa, tomato crop.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2545
8497 Poincaré Plot for Heart Rate Variability

Authors: Mazhar B. Tayel, Eslam I. AlSaba

Abstract:

Heart is the most important part in the body of living organisms. It affects and is affected by any factor in the body. Therefore, it is a good detector for all conditions in the body. Heart signal is a non-stationary signal; thus, it is utmost important to study the variability of heart signal. The Heart Rate Variability (HRV) has attracted considerable attention in psychology, medicine and has become important dependent measure in psychophysiology and behavioral medicine. The standards of measurements, physiological interpretation and clinical use for HRV that are most often used were described in many researcher papers, however, remain complex issues are fraught with pitfalls. This paper presents one of the nonlinear techniques to analyze HRV. It discusses many points like, what Poincaré plot is and how Poincaré plot works; also, Poincaré plot's merits especially in HRV. Besides, it discusses the limitation of Poincaré cause of standard deviation SD1, SD2 and how to overcome this limitation by using complex correlation measure (CCM). The CCM is most sensitive to changes in temporal structure of the Poincaré plot as compared toSD1 and SD2.

Keywords: Heart rate variability, chaotic system, Poincaré, variance, standard deviation, complex correlation measure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7449
8496 Using SNAP and RADTRAD to Establish the Analysis Model for Maanshan PWR Plant

Authors: J. R. Wang, H. C. Chen, C. Shih, S. W. Chen, J. H. Yang, Y. Chiang

Abstract:

In this study, we focus on the establishment of the analysis model for Maanshan PWR nuclear power plant (NPP) by using RADTRAD and SNAP codes with the FSAR, manuals, and other data. In order to evaluate the cumulative dose at the Exclusion Area Boundary (EAB) and Low Population Zone (LPZ) outer boundary, Maanshan NPP RADTRAD/SNAP model was used to perform the analysis of the DBA LOCA case. The analysis results of RADTRAD were similar to FSAR data. These analysis results were lower than the failure criteria of 10 CFR 100.11 (a total radiation dose to the whole body, 250 mSv; a total radiation dose to the thyroid from iodine exposure, 3000 mSv).

Keywords: RADionuclide, transport, removal, and dose estimation, RADTRAD, symbolic nuclear analysis package, SNAP, dose, PWR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1025
8495 Kinematic and Dynamic Analysis of a Lower Limb Exoskeleton

Authors: Tawakal Hasnain Baluch, Adnan Masood, Javaid Iqbal, Umer Izhar, Umar Shahbaz Khan

Abstract:

This paper will provide the kinematic and dynamic analysis of a lower limb exoskeleton. The forward and inverse kinematics of proposed exoskeleton is performed using Denevit and Hartenberg method. The torques required for the actuators will be calculated using Lagrangian formulation technique. This research can be used to design the control of the proposed exoskeleton.

Keywords: Dynamic Analysis, Exoskeleton, Kinematic Analysis, Lower Limb, Rehabilitation Robotics

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4596
8494 Using the Technology Acceptance Model to Examine Seniors’ Attitudes toward Facebook

Authors: Chien-Jen Liu, Shu Ching Yang

Abstract:

Using the technology acceptance model (TAM), this study examined the external variables of technological complexity (TC) to acquire a better understanding of the factors that influence the acceptance of computer application courses by learners at Active Aging Universities. After the learners in this study had completed a 27-hour Facebook course, 44 learners responded to a modified TAM survey. Data were collected to examine the path relationships among the variables that influence the acceptance of Facebook-mediated community learning. The partial least squares (PLS) method was used to test the measurement and the structural model. The study results demonstrated that attitudes toward Facebook use directly influence behavioral intentions (BI) with respect to Facebook use, evincing a high prediction rate of 58.3%. In addition to the perceived usefulness (PU) and perceived ease of use (PEOU) measures that are proposed in the TAM, other external variables, such as TC, also indirectly influence BI. These four variables can explain 88% of the variance in BI and demonstrate a high level of predictive ability. Finally, limitations of this investigation and implications for further research are discussed.

Keywords: Technology acceptance model (TAM), technological complexity, partial least squares (PLS), perceived usefulness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3195
8493 Constructivism Learning Management in Mathematical Analysis Courses

Authors: K. Paisal

Abstract:

The purposes of this research were (1) to create a learning activity for constructivism, (2) study the Mathematical Analysis courses learning achievement, and (3) study students’ attitude toward the learning activity for constructivism. The samples in this study were divided into 2 parts including 3 Mathematical Analysis courses instructors of Suan Sunandha Rajabhat University who provided basic information and attended the seminar and 17 Mathematical Analysis courses students who were studying in the academic and engaging in the learning activity for constructivism. The research instruments were lesson plans constructivism, subjective Mathematical Analysis courses achievement test with reliability index of 0.8119, and an attitude test concerning the students’ attitude toward the Mathematical Analysis courses learning activity for constructivism. The result of the research show that the efficiency of the Mathematical Analysis courses learning activity for constructivism is 73.05/72.16, which is more than expected criteria of 70/70. The research additionally find that the average score of learning achievement of students who engaged in the learning activities for constructivism are equal to 70% and the students’ attitude toward the learning activity for constructivism are at the medium level.

Keywords: Constructivism, learning management, Mathematical Analysis courses.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1867
8492 Revisiting the Concept of Risk Analysis within the Context of Geospatial Database Design: A Collaborative Framework

Authors: J. Grira, Y. Bédard, S. Roche

Abstract:

The aim of this research is to design a collaborative framework that integrates risk analysis activities into the geospatial database design (GDD) process. Risk analysis is rarely undertaken iteratively as part of the present GDD methods in conformance to requirement engineering (RE) guidelines and risk standards. Accordingly, when risk analysis is performed during the GDD, some foreseeable risks may be overlooked and not reach the output specifications especially when user intentions are not systematically collected. This may lead to ill-defined requirements and ultimately in higher risks of geospatial data misuse. The adopted approach consists of 1) reviewing risk analysis process within the scope of RE and GDD, 2) analyzing the challenges of risk analysis within the context of GDD, and 3) presenting the components of a risk-based collaborative framework that improves the collection of the intended/forbidden usages of the data and helps geo-IT experts to discover implicit requirements and risks.

Keywords: Collaborative risk analysis, intention of use, Geospatial database design, Geospatial data misuse.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1666
8491 A Trends Analysis of Dinghy Yacht Simulator

Authors: Jae-Neung Lee, Sung-Bum Pan, Keun-Chang Kwak

Abstract:

This paper describes an analysis of Yacht Simulator international trends and also explains about Yacht. The results are summarized as follows. Attached to the cockpit are sensors that feed -back information on rudder angle, boat heel angle and mainsheet tension to the computer. Energy expenditure of the sailor measure indirectly using expired gas analysis for the measurement of VO2 and VCO2. At sea course configurations and wind conditions can be preset to suit any level of sailor from complete beginner to advanced sailor.

Keywords: Trends Analysis, Yacht Simulator, Sailing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2225
8490 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm

Authors: Ameur Abdelkader, Abed Bouarfa Hafida

Abstract:

Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.

Keywords: Predictive analysis, big data, predictive analysis algorithms. CART algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1075
8489 Image Contrast Enhancement based Sub-histogram Equalization Technique without Over-equalization Noise

Authors: Hyunsup Yoon, Youngjoon Han, Hernsoo Hahn

Abstract:

In order to enhance the contrast in the regions where the pixels have similar intensities, this paper presents a new histogram equalization scheme. Conventional global equalization schemes over-equalizes these regions so that too bright or dark pixels are resulted and local equalization schemes produce unexpected discontinuities at the boundaries of the blocks. The proposed algorithm segments the original histogram into sub-histograms with reference to brightness level and equalizes each sub-histogram with the limited extents of equalization considering its mean and variance. The final image is determined as the weighted sum of the equalized images obtained by using the sub-histogram equalizations. By limiting the maximum and minimum ranges of equalization operations on individual sub-histograms, the over-equalization effect is eliminated. Also the result image does not miss feature information in low density histogram region since the remaining these area is applied separating equalization. This paper includes how to determine the segmentation points in the histogram. The proposed algorithm has been tested with more than 100 images having various contrasts in the images and the results are compared to the conventional approaches to show its superiority.

Keywords: Contrast Enhancement, Histogram Equalization, Histogram Region Equalization, Equalization Noise

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3419
8488 Effect of Biostimulants to Control the Phelipanche ramosa L. Pomel in Processing Tomato Crop

Authors: G. Disciglio, G. Gatta, F. Lops, A. Libutti, A. Tarantino, E. Tarantino

Abstract:

The experimental trial was carried out in open field at Foggia district (Apulia Region, Southern Italy), during the spring-summer season 2014, in order to evaluate the effect of four biostimulant products (RadiconÒ, Viormon plusÒ, LysodinÒ and SiaptonÒ 10L), compared with a control (no biostimulant), on the infestation of processing tomato crop (cv Dres) by the chlorophyll-lacking root parasite Phelipanche ramosa. Biostimulants consist in different categories of products (microbial inoculants, humic and fulvic acids, hydrolyzed proteins and aminoacids, seaweed extracts) which play various roles in plant growing, including the improvement of crop resistance and quali-quantitative characteristics of yield. The experimental trial was arranged according to a complete randomized block design with five treatments, each of one replicated three times. The processing tomato seedlings were transplanted on 5 May 2014. Throughout the crop cycle, P. ramosa infestation was assessed according to the number of emerged shoots (branched plants) counted in each plot, at 66, 78 and 92 day after transplanting. The tomato fruits were harvested at full-stage of maturity on 8 August 2014. From each plot, the marketable yield was measured and the quali-quantitative yield parameters (mean weight, dry matter content, colour coordinate, colour index and soluble solids content of the fruits) were determined. The whole dataset was tested according to the basic assumptions for the analysis of variance (ANOVA) and the differences between the means were determined using Tukey’s tests at the 5% probability level. The results of the study showed that none of the applied biostimulants provided a whole control of Phelipanche, although some positive effects were obtained from their application. To this respect, the RadiconÒ appeared to be the most effective in reducing the infestation of this root-parasite in tomato crop. This treatment also gave the higher tomato yield.

Keywords: Biostimulants, control methods, Phelipanche ramosa, processing tomato crop.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1904
8487 Speckle Reducing Contourlet Transform for Medical Ultrasound Images

Authors: P.S. Hiremath, Prema T. Akkasaligar, Sharan Badiger

Abstract:

Speckle noise affects all coherent imaging systems including medical ultrasound. In medical images, noise suppression is a particularly delicate and difficult task. A tradeoff between noise reduction and the preservation of actual image features has to be made in a way that enhances the diagnostically relevant image content. Even though wavelets have been extensively used for denoising speckle images, we have found that denoising using contourlets gives much better performance in terms of SNR, PSNR, MSE, variance and correlation coefficient. The objective of the paper is to determine the number of levels of Laplacian pyramidal decomposition, the number of directional decompositions to perform on each pyramidal level and thresholding schemes which yields optimal despeckling of medical ultrasound images, in particular. The proposed method consists of the log transformed original ultrasound image being subjected to contourlet transform, to obtain contourlet coefficients. The transformed image is denoised by applying thresholding techniques on individual band pass sub bands using a Bayes shrinkage rule. We quantify the achieved performance improvement.

Keywords: Contourlet transform, Despeckling, Pyramidal directionalfilter bank, Thresholding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2446
8486 Multi-Dimensional Concerns Mining for Web Applications via Concept-Analysis

Authors: Carlo Bellettini, Alessandro Marchetto, Andrea Trentini

Abstract:

Web applications have become very complex and crucial, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering), the scientific community has focused attention to Web applications design, development, analysis, and testing, by studying and proposing methodologies and tools. This paper proposes an approach to automatic multi-dimensional concern mining for Web Applications, based on concepts analysis, impact analysis, and token-based concern identification. This approach lets the user to analyse and traverse Web software relevant to a particular concern (concept, goal, purpose, etc.) via multi-dimensional separation of concerns, to document, understand and test Web applications. This technique was developed in the context of WAAT (Web Applications Analysis and Testing) project. A semi-automatic tool to support this technique is currently under development.

Keywords: Concepts Analysis, Concerns Mining, Multi-Dimensional Separation of Concerns, Impact Analysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1473
8485 The Performance Analysis of Error Saturation Nonlinearity LMS in Impulsive Noise based on Weighted-Energy Conservation

Authors: T Panigrahi, G Panda, Mulgrew

Abstract:

This paper introduces a new approach for the performance analysis of adaptive filter with error saturation nonlinearity in the presence of impulsive noise. The performance analysis of adaptive filters includes both transient analysis which shows that how fast a filter learns and the steady-state analysis gives how well a filter learns. The recursive expressions for mean-square deviation(MSD) and excess mean-square error(EMSE) are derived based on weighted energy conservation arguments which provide the transient behavior of the adaptive algorithm. The steady-state analysis for co-related input regressor data is analyzed, so this approach leads to a new performance results without restricting the input regression data to be white.

Keywords: Error saturation nonlinearity, transient analysis, impulsive noise.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1781