Search results for: null hypothesis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 317

Search results for: null hypothesis

317 Additional Considerations on a Sequential Life Testing Approach using a Weibull Model

Authors: D. I. De Souza, D. R. Fonseca, R. Rocha

Abstract:

In this paper we will develop further the sequential life test approach presented in a previous article by [1] using an underlying two parameter Weibull sampling distribution. The minimum life will be considered equal to zero. We will again provide rules for making one of the three possible decisions as each observation becomes available; that is: accept the null hypothesis H0; reject the null hypothesis H0; or obtain additional information by making another observation. The product being analyzed is a new type of a low alloy-high strength steel product. To estimate the shape and the scale parameters of the underlying Weibull model we will use a maximum likelihood approach for censored failure data. A new example will further develop the proposed sequential life testing approach.

Keywords: Sequential Life Testing, Underlying Weibull Model, Maximum Likelihood Approach, Hypothesis Testing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1334
316 Contributions to Differential Geometry of Pseudo Null Curves in Semi-Euclidean Space

Authors: Melih Turgut, Süha Yılmaz

Abstract:

In this paper, first, a characterization of spherical Pseudo null curves in Semi-Euclidean space is given. Then, to investigate position vector of a pseudo null curve, a system of differential equation whose solution gives the components of the position vector of a pseudo null curve on the Frenet axis is established by means of Frenet equations. Additionally, in view of some special solutions of mentioned system, characterizations of some special pseudo null curves are presented.

Keywords: Semi-Euclidean Space, Pseudo Null Curves, Position Vectors.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1285
315 Further Thoughtson a Sequential Life Testing Approach Using an Inverse Weibull Model

Authors: D. I. De Souza, G. P. Azevedo, D. R. Fonseca

Abstract:

In this paper we will develop further the sequential life test approach presented in a previous article by [1] using an underlying two parameter Inverse Weibull sampling distribution. The location parameter or minimum life will be considered equal to zero. Once again we will provide rules for making one of the three possible decisions as each observation becomes available; that is: accept the null hypothesis H0; reject the null hypothesis H0; or obtain additional information by making another observation. The product being analyzed is a new electronic component. There is little information available about the possible values the parameters of the corresponding Inverse Weibull underlying sampling distribution could have.To estimate the shape and the scale parameters of the underlying Inverse Weibull model we will use a maximum likelihood approach for censored failure data. A new example will further develop the proposed sequential life testing approach.

Keywords: Sequential Life Testing, Inverse Weibull Model, Maximum Likelihood Approach, Hypothesis Testing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1366
314 Position Vector of a Partially Null Curve Derived from a Vector Differential Equation

Authors: Süha Yılmaz, Emin Özyılmaz, Melih Turgut, Şuur Nizamoğlu

Abstract:

In this paper, position vector of a partially null unit speed curve with respect to standard frame of Minkowski space-time is studied. First, it is proven that position vector of every partially null unit speed curve satisfies a vector differential equation of fourth order. In terms of solution of the differential equation, position vector of a partially null unit speed curve is expressed.

Keywords: Frenet Equations, Partially Null Curves, Minkowski Space-time, Vector Differential Equation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1114
313 On Frenet-Serret Invariants of Non-Null Curves in Lorentzian Space L5

Authors: Melih Turgut, José Luis López-Bonilla, Süha Yılmaz

Abstract:

The aim of this paper is to determine Frenet-Serret invariants of non-null curves in Lorentzian 5-space. First, we define a vector product of four vectors, by this way, we present a method to calculate Frenet-Serret invariants of the non-null curves. Additionally, an algebraic example of presented method is illustrated.

Keywords: Lorentzian 5-space, Frenet-Serret Invariants, Nonnull Curves

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1457
312 Performance Evaluation of 2×2 Switched Beam Antennas with Null Locating for Wireless Mesh Networks

Authors: S. Pradittara, M. Uthansakul, P. Uthansakul

Abstract:

A concept of switched beam antennas consisting of 2×2 rectangular array spaced by λ/4 accompanied with a null locating has been proposed in the previous work. In this letter, the performance evaluations of its prototype are presented. The benefits of using proposed system have been clearly measured in term of signal quality, throughput and delays. Also, the impact of position shift which mesh router is not located on the expected beam direction has also been investigated.

Keywords: Antenna array, Beamforming, Null steering, WMNs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1549
311 Distributed Motion Control Real-Time Contouring Algorithm Implementation and Performance Test

Authors: Francisco J. Lopez-Jaquez, Sandra E. Ramirez-Jara

Abstract:

This paper presents an implementation and performance test of a distributed motion control system based on a master-slave configuration used to move a plasma-cutting torch over a predefined trajectory. The master is a general-purpose computer running on an open source operating system platform and software developer. Software running in the master computer generates commands on real time and we measure performance based on a selected set of differences between expected and observed distances. We are testing the null hypothesis that the outcome trajectory is identical to the input against the alternative hypothesis that there is a shift to the right or left of the input one. We used the Wilcoxon signed ranks test method for the hypothesis test.

Keywords: Distributed, motion, control, real-time, contouring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1438
310 The Quality Assessment of Seismic Reflection Survey Data Using Statistical Analysis: A Case Study of Fort Abbas Area, Cholistan Desert, Pakistan

Authors: U. Waqas, M. F. Ahmed, A. Mehmood, M. A. Rashid

Abstract:

In geophysical exploration surveys, the quality of acquired data holds significant importance before executing the data processing and interpretation phases. In this study, 2D seismic reflection survey data of Fort Abbas area, Cholistan Desert, Pakistan was taken as test case in order to assess its quality on statistical bases by using normalized root mean square error (NRMSE), Cronbach’s alpha test (α) and null hypothesis tests (t-test and F-test). The analysis challenged the quality of the acquired data and highlighted the significant errors in the acquired database. It is proven that the study area is plain, tectonically least affected and rich in oil and gas reserves. However, subsurface 3D modeling and contouring by using acquired database revealed high degrees of structural complexities and intense folding. The NRMSE had highest percentage of residuals between the estimated and predicted cases. The outcomes of hypothesis testing also proved the biasness and erraticness of the acquired database. Low estimated value of alpha (α) in Cronbach’s alpha test confirmed poor reliability of acquired database. A very low quality of acquired database needs excessive static correction or in some cases, reacquisition of data is also suggested which is most of the time not feasible on economic grounds. The outcomes of this study could be used to assess the quality of large databases and to further utilize as a guideline to establish database quality assessment models to make much more informed decisions in hydrocarbon exploration field.

Keywords: Data quality, null hypothesis, seismic lines, seismic reflection survey.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 538
309 The Physics of Gravity: A Hypothesis Based on Classical Physics

Authors: I. V. Kuzminov

Abstract:

The alternative hypothesis of the physics of gravitation is put forward in this paper. The hypothesis is constructed on the laws of classical physics. The process of expansion of the Universe explains the physics of gravity. The expansion of the Universe induces the resistance of gyroscopic forces of electron’s rotation. The second component of gravity forces is the resistance arising from the second derivative of linear expansion. This hypothesis does not reject the existing foundation of settlement, particularly as it is empirically constructed. The forces of gravitation and inertia share a common nature, which has been recognized before. The presented hypothesis does not criticize existing theories of gravitation; rather, it explores a separate theme. It is important to acknowledge that the expansion of the Universe exhibits isotropic characteristics. The proposed hypothesis provides a fundamental direction for further research. It is worth noting that this article does not aim to encompass all possible aspects of future investigations.

Keywords: Gyroscopic forces, the unity of the micro- and macrocosm, the expansion of the universe, the second derivative of expansion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 125
308 Variant Polymorphisms of GST and XRCC Genes and the Early Risk of Age Associated Disease in Kazakhstan

Authors: Zeinep A. Berkimbayeva, Almagul T. Mansharipova, Elmira M. Khussainova, Leyla B. Djansugurova

Abstract:

It is believed that DNA damaging toxic metabolites contributes to the development of different pathological conditions. To prevent harmful influence of toxic agents, cells developed number of protecting mechanisms, such as enzymatic reaction of detoxification of reactive metabolites and repair of DNA damage. The aim of the study was to examine the association between polymorphism of GSTT1/GSTM1 and XRCC1/3 genes and coronary artery disease (CAD) incidence. To examine a polymorphism of these genes in CAD susceptibility in patients and controls, PCR based genotyping assay was performed. For GST genes, frequency of GSTM1 null genotype among CAD affected group was significantly increased than in control group (P<0.001). Frequencies of the GSTT1 null and positive alleles are almost equal in both groups (P>0.1). We found that neither XRCC1 Arg399Gln nor XRCC3 Thr241Met were associated with CAD risk. Obtained data suggests that GSTM1 null genotype carriers are more susceptible to CAD development.

Keywords: Cardiovascular disease, DNA reparation, gene polymorphism, risk factors, xenobiotic detoxification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1877
307 Antenna Array Beamforming Using Neural Network

Authors: Maja Sarevska, Abdel-Badeeh M. Salem

Abstract:

This paper considers the problem of Null-Steering beamforming using Neural Network (NN) approach for antenna array system. Two cases are presented. First, unlike the other authors, the estimated Direction Of Arrivals (DOAs) are used for antenna array weights NN-based determination and the imprecise DOAs estimations are taken into account. Second, the blind null-steering beamforming is presented. In this case the antenna array outputs are presented at the input of the NN without DOAs estimation. The results of computer simulations will show much better relative mean error performances of the first NN approach compared to the NNbased blind beamforming.

Keywords: Beamforming, DOAs, neural network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2352
306 A Bootstrap's Reliability Measure on Tests of Hypotheses

Authors: Al Jefferson J. Pabelic, Dennis A. Tarepe

Abstract:

Bootstrapping has gained popularity in different tests of hypotheses as an alternative in using asymptotic distribution if one is not sure of the distribution of the test statistic under a null hypothesis. This method, in general, has two variants – the parametric and the nonparametric approaches. However, issues on reliability of this method always arise in many applications. This paper addresses the issue on reliability by establishing a reliability measure in terms of quantiles with respect to asymptotic distribution, when this is approximately correct. The test of hypotheses used is Ftest. The simulated results show that using nonparametric bootstrapping in F-test gives better reliability than parametric bootstrapping with relatively higher degrees of freedom.

Keywords: F-test, nonparametric bootstrapping, parametric bootstrapping, reliability measure, tests of hypotheses.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1641
305 Difference in Psychological Well-Being Based On Comparison of Religions: A Case Study in Pekan District, Pahang, Malaysia

Authors: Amran Hassan, Fatimah Yusooff, Khadijah Alavi

Abstract:

The psychological well-being of a family is a subjective matter for evaluation, all the more when it involves the element of religions, whether Islam, Christianity, Buddhism or Hinduism. Each of these religions emphasises similar values and morals on family psychological well-being. This comparative study is specifically to determine the role of religion on family psychological well-being in Pekan district, Pahang, Malaysia. The study adopts a quantitative and qualitative mixed method design and considers a total of 412 samples of parents and children for the quantitative study, and 21 samples for the qualitative study. The quantitative study uses simple random sampling, whereas the qualitative sampling is purposive. The instrument for quantitative study is Ryff’s Psychological Well-being Scale and the qualitative study involves the construction of a guidelines protocol for in-depth interviews of respondents. The quantitative study uses the SPSS version .19 with One Way Anova, and the qualitative analysis is manual based on transcripts with specific codes and themes. The results show nonsignificance, that is, no significant difference among religions in all family psychological well-being constructs in the comparison of Islam, Christianity, Buddhism and Hinduism, thereby accepting a null hypothesis and rejecting an alternative hypothesis. The qualitative study supports the quantitative study, that is, all 21 respondents explain that no difference exists in psychological wellbeing in the comparison of teachings in all the religious mentioned. These implications may be used as guidelines for government and non-government bodies in considering religion as an important element in family psychological well-being in the long run. 

Keywords: Psychological well-being, comparison of religions, family, Malaysia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2273
304 Websites for Hypothesis Testing

Authors: František Mošna

Abstract:

E-learning has become an efficient and widespread means of education at all levels of human activities. Statistics is no exception. Unfortunately the main focus in statistics teaching is usually paid to the substitution in formulas. Suitable websites can simplify and automate calculations and provide more attention and time to the basic principles of statistics, mathematization of real-life situations and following interpretation of results. We now introduce our own web-site for hypothesis testing. Its didactic aspects, the technical possibilities of the individual tools, the experience of use and the advantages or disadvantages are discussed in this paper. This web-site is not a substitute for common statistical software but should significantly improve the teaching of statistics at universities.

Keywords: E-learning, hypothesis testing, PHP, websites.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2298
303 Prevalence of Psychological Resistance to Voluntary Counselling and Testing of HIV/AIDS among Students of Tertiary Institutions in Kano State, Nigeria

Authors: A. S. Haruna

Abstract:

The incessant discomfort for Voluntary Counselling and Testing (VCT) exhibited by students in some tertiary institutions in Kano State, Nigeria is capable of causing Psychological Resistance as well as jeopardizing the purpose of HIV intervention. This study investigated the Prevalence of Psychological Resistance to VCT of HIV/AIDS among students of tertiary institutions in the state. Two null hypotheses were postulated and tested. Cross- Sectional Survey Design was employed in which 1512 sample was selected from a student population of 104,841 following Stratified Random Sampling technique. A self-developed 20-item scale whose reliability coefficient is 0.83 was used for data collection. Data analyzed via Chi-square and t-test reveals a prevalence of 38% with males (Mean=0.34; SD=0.475) constituting 60% and females (Mean=0.45; SD=0.498) 40%. Also, the calculated chi-square and ttest were not significant at 0.05 as such the null hypotheses were upheld. Recommendation offered suggests the use of reinforcement and social support for students who patronize HIV/AIDS counselling.

Keywords: HIV/AIDS, Prevalence rate, Psychological Resistance, VCT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2553
302 Design of Non-uniform Circular Antenna Arrays Using Firefly Algorithm for Side Lobe Level Reduction

Authors: Gopi Ram, Durbadal Mandal, Rajib Kar, Sakti Prasad Ghoshal

Abstract:

A design problem of non-uniform circular antenna arrays for maximum reduction of both the side lobe level (SLL) and first null beam width (FNBW) is dealt with. This problem is modeled as a simple optimization problem. The method of Firefly algorithm (FFA) is used to determine an optimal set of current excitation weights and antenna inter-element separations that provide radiation pattern with maximum SLL reduction and much improvement on FNBW as well. Circular array antenna laid on x-y plane is assumed. FFA is applied on circular arrays of 8-, 10-, and 12- elements. Various simulation results are presented and hence performances of side lobe and FNBW are analyzed. Experimental results show considerable reductions of both the SLL and FNBW with respect to those of the uniform case and some standard algorithms GA, PSO and SA applied to the same problem.

Keywords: Circular arrays, First null beam width, Side lobe level, FFA.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3111
301 Blind Channel Estimation for Frequency Hopping System Using Subspace Based Method

Authors: M. M. Qasaymeh, M. A. Khodeir

Abstract:

Subspace channel estimation methods have been studied widely, where the subspace of the covariance matrix is decomposed to separate the signal subspace from noise subspace. The decomposition is normally done by using either the eigenvalue decomposition (EVD) or the singular value decomposition (SVD) of the auto-correlation matrix (ACM). However, the subspace decomposition process is computationally expensive. This paper considers the estimation of the multipath slow frequency hopping (FH) channel using noise space based method. In particular, an efficient method is proposed to estimate the multipath time delays by applying multiple signal classification (MUSIC) algorithm which is based on the null space extracted by the rank revealing LU (RRLU) factorization. As a result, precise information is provided by the RRLU about the numerical null space and the rank, (i.e., important tool in linear algebra). The simulation results demonstrate the effectiveness of the proposed novel method by approximately decreasing the computational complexity to the half as compared with RRQR methods keeping the same performance.

Keywords: Time Delay Estimation, RRLU, RRQR, MUSIC, LS-ESPRIT, LS-ESPRIT, Frequency Hopping.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1981
300 The Effects of Electrical Muscle Stimulation (EMS) towards Male Skeletal Muscle Mass

Authors: Mohd Faridz Ahmad, Amirul Hakim Hasbullah

Abstract:

Electrical Muscle Stimulation (EMS) has been introduced and globally gained increasing attention on its usefulness. Continuous application of EMS may lead to the increment of muscle mass and indirectly will increase the strength. This study can be used as an alternative to help people especially those living a sedentary lifestyle to improve their muscle activity without having to go through a heavy workout session. Therefore, this study intended to investigate the effectiveness of EMS training program in 5 weeks interventions towards male body composition. It was a quasiexperimental design, held at the Impulse Studio Bangsar, which examined the effects of EMS training towards skeletal muscle mass among the subjects. Fifteen subjects (n = 15) were selected to assist in this study. The demographic data showed that, the average age of the subjects was 43.07 years old ± 9.90, height (173.4 cm ± 9.09) and weight was (85.79 kg ± 18.07). Results showed that there was a significant difference on the skeletal muscle mass (p = 0.01 < 0.05), upper body (p = 0.01 < 0.05) and lower body (p = 0.00 < 0.05). Therefore, the null hypothesis has been rejected in this study. As a conclusion, the application of EMS towards body composition can increase the muscle size and strength. This method has been proven to be able to improve athlete strength and thus, may be implemented in the sports science area of knowledge.

Keywords: Body composition, EMS, skeletal muscle mass, strength.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6179
299 A New Brazilian Friction-Resistant Low Alloy High Strength Steel – A Life Testing Approach

Authors: D. I. De Souza, G. P. Azevedo, R. Rocha

Abstract:

In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.

Keywords: Sequential life testing, normal and inverse Weibull models, maximum likelihood approach, truncation mechanism.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1379
298 Strict Stability of Fuzzy Differential Equations by Lyapunov Functions

Authors: Mustafa Bayram Gücen, Coşkun Yakar

Abstract:

In this study, we have investigated the strict stability of fuzzy differential systems and we compare the classical notion of strict stability criteria of ordinary differential equations and the notion of strict stability of fuzzy differential systems. In addition that, we present definitions of stability and strict stability of fuzzy differential equations and also we have some theorems and comparison results. Strict Stability is a different stability definition and this stability type can give us an information about the rate of decay of the solutions. Lyapunov’s second method is a standard technique used in the study of the qualitative behavior of fuzzy differential systems along with a comparison result that allows the prediction of behavior of a fuzzy differential system when the behavior of the null solution of a fuzzy comparison system is known. This method is a usefull for investigating strict stability of fuzzy systems. First of all, we present definitions and necessary background material. Secondly, we discuss and compare the differences between the classical notion of stability and the recent notion of strict stability. And then, we have a comparison result in which the stability properties of the null solution of the comparison system imply the corresponding stability properties of the fuzzy differential system. Consequently, we give the strict stability results and a comparison theorem. We have used Lyapunov second method and we have proved a comparison result with scalar differential equations.

Keywords: Fuzzy systems, fuzzy differential equations, fuzzy stability, strict stability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1036
297 Bootstrap and MLS Methods-based Individual Bioequivalence Assessment

Authors: Kongsheng Zhang, Li Ge

Abstract:

It is a one-sided hypothesis testing process for assessing bioequivalence. Bootstrap and modified large-sample(MLS) methods are considered to study individual bioequivalence(IBE), type I error and power of hypothesis tests are simulated and compared with FDA(2001). The results show that modified large-sample method is equivalent to the method of FDA(2001) .

Keywords: Individual bioequivalence, bootstrap, Bayesian bootstrap, modified large-sample.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1527
296 Linguistic, Pragmatic and Evolutionary Factors in Wason Selection Task

Authors: Olimpia Matarazzo, Fabrizio Ferrara

Abstract:

In two studies we tested the hypothesis that the appropriate linguistic formulation of a deontic rule – i.e. the formulation which clarifies the monadic nature of deontic operators - should produce more correct responses than the conditional formulation in Wason selection task. We tested this assumption by presenting a prescription rule and a prohibition rule in conditional vs. proper deontic formulation. We contrasted this hypothesis with two other hypotheses derived from social contract theory and relevance theory. According to the first theory, a deontic rule expressed in terms of cost-benefit should elicit a cheater detection module, sensible to mental states attributions and thus able to discriminate intentional rule violations from accidental rule violations. We tested this prevision by distinguishing the two types of violations. According to relevance theory, performance in selection task should improve by increasing cognitive effect and decreasing cognitive effort. We tested this prevision by focusing experimental instructions on the rule vs. the action covered by the rule. In study 1, in which 480 undergraduates participated, we tested these predictions through a 2 x 2 x 2 x 2 (type of the rule x rule formulation x type of violation x experimental instructions) between-subjects design. In study 2 – carried out by means of a 2 x 2 (rule formulation x type of violation) between-subjects design - we retested the hypothesis of rule formulation vs. the cheaterdetection hypothesis through a new version of selection task in which intentional vs. accidental rule violations were better discriminated. 240 undergraduates participated in this study. Results corroborate our hypothesis and challenge the contrasting assumptions. However, they show that the conditional formulation of deontic rules produces a lower performance than what is reported in literature.

Keywords: Deontic reasoning; Evolutionary, linguistic, logical, pragmatic factors; Wason selection task

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1547
295 Study of Storms on the Javits Center Green Roof

Authors: A. Cho, H. Sanyal, J. Cataldo

Abstract:

A quantitative analysis of the different variables on both the South and North green roofs of the Jacob K. Javits Convention Center was taken to find mathematical relationships between net radiation and evapotranspiration (ET), average outside temperature, and the lysimeter weight. Groups of datasets were analyzed, and the relationships were plotted on linear and semi-log graphs to find consistent relationships. Antecedent conditions for each rainstorm were also recorded and plotted against the volumetric water difference within the lysimeter. The first relation was the inverse parabolic relationship between the lysimeter weight and the net radiation and ET. The peaks and valleys of the lysimeter weight corresponded to valleys and peaks in the net radiation and ET respectively, with the 8/22/15 and 1/22/16 datasets showing this trend. The U-shaped and inverse U-shaped plots of the two variables coincided, indicating an inverse relationship between the two variables. Cross variable relationships were examined through graphs with lysimeter weight as the dependent variable on the y-axis. 10 out of 16 of the plots of lysimeter weight vs. outside temperature plots had R² values > 0.9. Antecedent conditions were also recorded for rainstorms, categorized by the amount of precipitation accumulating during the storm. Plotted against the change in the volumetric water weight difference within the lysimeter, a logarithmic regression was found with large R² values. The datasets were compared using the Mann Whitney U-test to see if the datasets were statistically different, using a significance level of 5%; all datasets compared showed a U test statistic value, proving the null hypothesis of the datasets being different from being true.

Keywords: Green roof, green infrastructure, Javits Center, evapotranspiration, net radiation, lysimeter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 282
294 Creation and Annihilation of Spacetime Elements

Authors: Dnyanesh P. Mathur, Gregory L. Slater

Abstract:

Gravitation and the expansion of the universe at a large scale are generally regarded as two completely distinct phenomena. Yet, in General theory of Relativity (GR), they both manifest as 'curvature' of spacetime. We propose a hypothesis which treats these two 'curvature-producing' phenomena as aspects of an underlying process. This process treats spacetime itself as composed of discrete units (Plancktons) and is 'dynamic' in the sense that these elements of spacetime are continually being both created and annihilated. It is these two complementary processes of Planckton creation and Planckton annihilation which manifest themselves as - 'cosmic expansion' on the one hand and as 'gravitational attraction’ on the other. The Planckton hypothesis treats spacetime as a perfect fluid in the same manner as the co-moving frame of reference of Friedman equations and the Gullstrand-Painleve metric; i.e., Planckton hypothesis replaces 'curvature' of spacetime by the 'flow' of Plancktons (spacetime). Here we discuss how this perspective may allow a unified description of both cosmological and gravitational acceleration as well as providing a mechanism for inducing an irreducible action at every point associated with the creation and annihilation of Plancktons, which could be identified as the zero point energy.

Keywords: Discrete spacetime, spacetime flow, zero point energy, dark energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 99
293 Semantic Support for Hypothesis-Based Research from Smart Environment Monitoring and Analysis Technologies

Authors: T. S. Myers, J. Trevathan

Abstract:

Improvements in the data fusion and data analysis phase of research are imperative due to the exponential growth of sensed data. Currently, there are developments in the Semantic Sensor Web community to explore efficient methods for reuse, correlation and integration of web-based data sets and live data streams. This paper describes the integration of remotely sensed data with web-available static data for use in observational hypothesis testing and the analysis phase of research. The Semantic Reef system combines semantic technologies (e.g., well-defined ontologies and logic systems) with scientific workflows to enable hypothesis-based research. A framework is presented for how the data fusion concepts from the Semantic Reef architecture map to the Smart Environment Monitoring and Analysis Technologies (SEMAT) intelligent sensor network initiative. The data collected via SEMAT and the inferred knowledge from the Semantic Reef system are ingested to the Tropical Data Hub for data discovery, reuse, curation and publication.

Keywords: Information architecture, Semantic technologies Sensor networks, Ontologies.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1670
292 Change Detection and Non Stationary Signals Tracking by Adaptive Filtering

Authors: Mounira RouaÐùnia, Noureddine Doghmane

Abstract:

In this paper we consider the problem of change detection and non stationary signals tracking. Using parametric estimation of signals based on least square lattice adaptive filters we consider for change detection statistical parametric methods using likelihood ratio and hypothesis tests. In order to track signals dynamics, we introduce a compensation procedure in the adaptive estimation. This will improve the adaptive estimation performances and fasten it-s convergence after changes detection.

Keywords: Change detection, Hypothesis test, likelihood ratioleast square lattice adaptive filters.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1584
291 Integrated Method for Detection of Unknown Steganographic Content

Authors: Magdalena Pejas

Abstract:

This article concerns the presentation of an integrated method for detection of steganographic content embedded by new unknown programs. The method is based on data mining and aggregated hypothesis testing. The article contains the theoretical basics used to deploy the proposed detection system and the description of improvement proposed for the basic system idea. Further main results of experiments and implementation details are collected and described. Finally example results of the tests are presented.

Keywords: Steganography, steganalysis, data embedding, data mining, feature extraction, knowledge base, system learning, hypothesis testing, error estimation, black box program, file structure.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1516
290 Vehicle Detection Method using Haar-like Feature on Real Time System

Authors: Sungji Han, Youngjoon Han, Hernsoo Hahn

Abstract:

This paper presents a robust vehicle detection approach using Haar-like feature. It is possible to get a strong edge feature from this Haar-like feature. Therefore it is very effective to remove the shadow of a vehicle on the road. And we can detect the boundary of vehicles accurately. In the paper, the vehicle detection algorithm can be divided into two main steps. One is hypothesis generation, and the other is hypothesis verification. In the first step, it determines vehicle candidates using features such as a shadow, intensity, and vertical edge. And in the second step, it determines whether the candidate is a vehicle or not by using the symmetry of vehicle edge features. In this research, we can get the detection rate over 15 frames per second on our embedded system.

Keywords: vehicle detection, haar-like feauture, single camera, real time

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3251
289 Tests for Gaussianity of a Stationary Time Series

Authors: Adnan Al-Smadi

Abstract:

One of the primary uses of higher order statistics in signal processing has been for detecting and estimation of non- Gaussian signals in Gaussian noise of unknown covariance. This is motivated by the ability of higher order statistics to suppress additive Gaussian noise. In this paper, several methods to test for non- Gaussianity of a given process are presented. These methods include histogram plot, kurtosis test, and hypothesis testing using cumulants and bispectrum of the available sequence. The hypothesis testing is performed by constructing a statistic to test whether the bispectrum of the given signal is non-zero. A zero bispectrum is not a proof of Gaussianity. Hence, other tests such as the kurtosis test should be employed. Examples are given to demonstrate the performance of the presented methods.

Keywords: Non-Gaussian, bispectrum, kurtosis, hypothesistesting, histogram.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1844
288 Organic Contribution on Particles Formed on Pacific Ocean: From Phytoplankton Blooms to Climate

Authors: Petri Vaattovaara, Luke Cravigan, Zoran Ristovski, Marc Mallet, Ari Laaksonen, Sarah Lawson, Nick Talbot, Gustavo Olivares, Mike Harvey, Cliff Law

Abstract:

These SOAP project Pacific Ocean measurements reveal that phytoplankton blooms with sunny conditions make possible secondary organic contribution to ultrafine particles size and composition, and thus on cloud formation ability, and finally on climate. This is in agreement with other biologically active region observations about the presence of secondary organics even the exact fraction is also depending on the local marine life (e.g. plankton blooms, seaweeds, corals). An organic contribution is clearly needed to add to CLAW hypothesis.

Keywords: Climate, marine aerosols, phytoplankton, secondary organics, CLAW hypothesis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1760