Search results for: stationary measure
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3510

Search results for: stationary measure

3450 On Generalized Cumulative Past Inaccuracy Measure for Marginal and Conditional Lifetimes

Authors: Amit Ghosh, Chanchal Kundu

Abstract:

Recently, the notion of past cumulative inaccuracy (CPI) measure has been proposed in the literature as a generalization of cumulative past entropy (CPE) in univariate as well as bivariate setup. In this paper, we introduce the notion of CPI of order α (alpha) and study the proposed measure for conditionally specified models of two components failed at different time instants called generalized conditional CPI (GCCPI). We provide some bounds using usual stochastic order and investigate several properties of GCCPI. The effect of monotone transformation on this proposed measure has also been examined. Furthermore, we characterize some bivariate distributions under the assumption of conditional proportional reversed hazard rate model. Moreover, the role of GCCPI in reliability modeling has also been investigated for a real-life problem.

Keywords: cumulative past inaccuracy, marginal and conditional past lifetimes, conditional proportional reversed hazard rate model, usual stochastic order

Procedia PDF Downloads 223
3449 Estimating the Power Influence of an Off-Grid Photovoltaic Panel on the Indicting Rate of a Storage System (Batteries)

Authors: Osamede Asowata

Abstract:

The current resurgence of interest in the use of renewable energy is driven by the need to reduce the high environmental impact of fossil-based energy. The aim of this paper is to evaluate the effect of a stationary PV panel on the charging rate of deep-cycle valve regulated lead-acid (DCVRLA) batteries. Stationary PV panels are set to a fixed tilt and orientation angle, which plays a major role in dictating the output power of a PV panel and subsequently on the charging time of a DCVRLA battery. In a basic PV system, an energy storage device that stores the power from the PV panel is necessary due to the fluctuating nature of the PV voltage caused by climatic conditions. The charging and discharging times of a DCVRLA battery were determined for a twelve month period from January through December 2012. Preliminary results, which include regression analysis (R2), conversion-time per week and work-time per day, indicate that a 36 degrees tilt angle produces a good charging rate for a latitude of 26 degrees south throughout the year.

Keywords: tilt and orientation angles, solar chargers, PV panels, storage devices, direct solar radiation.

Procedia PDF Downloads 221
3448 Antioxidant Capacity, Proximate Biomass Composition and Fatty Acid Profile of Five Marine Microalgal Species with Potential as Aquaculture Feed

Authors: Vasilis Andriopoulos, Maria D. Gkioni, Elena Koutra, Savvas G. Mastropetros, Fotini N. Lamari, Sofia Hatziantoniou, Michalis Kornaros

Abstract:

In the present study, the antioxidant activity of aqueous and methanolic extracts of Chlorella minutissima, Dunaliella salina, Isochrysis galbana, Nannochloropsis oculata and Tisohrysis lutea, as well as the proximate composition and fatty acid profile were evaluated, with the aim to select species suitable for co-production of antioxidants and aquaculture feed. Batch cultivation was performed at 25oC in a modified f/2 medium under continuous illumination and aeration with ambient air. Biomass was collected via centrifugation and extracted first with H2O and subsequently with methanol at two growth phases (early and late stationary). Total phenolic content and antioxidant and reducing activity of the extracts were evaluated. The highest phenolic content was found in the methanolic extract of C. minutissima at the early stationary phase (9.04±0.68 mg Gallic Acid Equivalent g-1 dry weight), and the aqueous extract of D. salina at the late stationary phase (8.78±1.49 mg Gallic Acid Equivalent g-1 Dry weight). Antioxidant activity, measured as 1,1-diphenyl-2-picrylhydrazyl radical scavenging activity, and Ferric reducing antioxidant power assay of methanolic extracts were comparable to the literature and correlated to Total phenolic content and Chlorophyll content of the biomass. No such correlation was found in the aqueous extracts. N. oculata and T. lutea were high in protein (39.88±1.72% Dry weight and 43.30±1.33% Dry weight, respectively) and carotenoids (0.64±0.13% and 0.92±0.02%, respectively). Additionally, they presented high eicosapentaenoic acid and docosahexaenoic acid levels (33.74±9.98 mg eicosapentaenoic acid g-1 DW and 31.31±2.92 mg docosahexaenoic acid g-1 dry weight, respectively). N. oculata and T. lutea are promising candidates for the co-production of antioxidants and aquaculture feed, while C. minutissima and D. salina showed promise due to their higher antioxidant content.

Keywords: aquaculture fee, antioxidant activity, fatty acids, microalgae, total phenolic content

Procedia PDF Downloads 138
3447 Study on Concentration and Temperature Measurement with 760 nm Diode Laser in Combustion System Using Tunable Diode Laser Absorption Spectroscopy

Authors: Miyeon Yoo, Sewon Kim, Changyeop Lee

Abstract:

It is important to measure the internal temperature or temperature distribution precisely in combustion system to increase energy efficiency and reduce the pollutants. Especially in case of large combustion systems such as power plant boiler and reheating furnace of steel making process, it is very difficult to measure those physical properties in detail. Tunable diode laser absorption spectroscopy measurement and analysis can be attractive method to overcome the difficulty. In this paper, TDLAS methods are used to measure the oxygen concentration and temperature distribution in various experimental conditions.

Keywords: tunable diode laser absorption Spectroscopy, temperature distribution, gas concentration

Procedia PDF Downloads 357
3446 Measure-Valued Solutions to a Class of Nonlinear Parabolic Equations with Degenerate Coercivity and Singular Initial Data

Authors: Flavia Smarrazzo

Abstract:

Initial-boundary value problems for nonlinear parabolic equations having a Radon measure as initial data have been widely investigated, looking for solutions which for positive times take values in some function space. On the other hand, if the diffusivity degenerates too fast at infinity, it is well known that function-valued solutions may not exist, singularities may persist, and it looks very natural to consider solutions which, roughly speaking, for positive times describe an orbit in the space of the finite Radon measures. In this general framework, our purpose is to introduce a concept of measure-valued solution which is consistent with respect to regularizing and smoothing approximations, in order to develop an existence theory which does not depend neither on the level of degeneracy of diffusivity at infinity nor on the choice of the initial measures. In more detail, we prove existence of suitably defined measure-valued solutions to the homogeneous Dirichlet initial-boundary value problem for a class of nonlinear parabolic equations without strong coerciveness. Moreover, we also discuss some qualitative properties of the constructed solutions concerning the evolution of their singular part, including conditions (depending both on the initial data and on the strength of degeneracy) under which the constructed solutions are in fact unction-valued or not.

Keywords: degenerate parabolic equations, measure-valued solutions, Radon measures, young measures

Procedia PDF Downloads 259
3445 The Use of AI to Measure Gross National Happiness

Authors: Riona Dighe

Abstract:

This research attempts to identify an alternative approach to the measurement of Gross National Happiness (GNH). It uses artificial intelligence (AI), incorporating natural language processing (NLP) and sentiment analysis to measure GNH. We use ‘off the shelf’ NLP models responsible for the sentiment analysis of a sentence as a building block for this research. We constructed an algorithm using NLP models to derive a sentiment analysis score against sentences. This was then tested against a sample of 20 respondents to derive a sentiment analysis score. The scores generated resembled human responses. By utilising the MLP classifier, decision tree, linear model, and K-nearest neighbors, we were able to obtain a test accuracy of 89.97%, 54.63%, 52.13%, and 47.9%, respectively. This gave us the confidence to use the NLP models against sentences in websites to measure the GNH of a country.

Keywords: artificial intelligence, NLP, sentiment analysis, gross national happiness

Procedia PDF Downloads 70
3444 Multi-Granularity Feature Extraction and Optimization for Pathological Speech Intelligibility Evaluation

Authors: Chunying Fang, Haifeng Li, Lin Ma, Mancai Zhang

Abstract:

Speech intelligibility assessment is an important measure to evaluate the functional outcomes of surgical and non-surgical treatment, speech therapy and rehabilitation. The assessment of pathological speech plays an important role in assisting the experts. Pathological speech usually is non-stationary and mutational, in this paper, we describe a multi-granularity combined feature schemes, and which is optimized by hierarchical visual method. First of all, the difference granularity level pathological features are extracted which are BAFS (Basic acoustics feature set), local spectral characteristics MSCC (Mel s-transform cepstrum coefficients) and nonlinear dynamic characteristics based on chaotic analysis. Latterly, radar chart and F-score are proposed to optimize the features by the hierarchical visual fusion. The feature set could be optimized from 526 to 96-dimensions.The experimental results denote that new features by support vector machine (SVM) has the best performance, with a recognition rate of 84.4% on NKI-CCRT corpus. The proposed method is thus approved to be effective and reliable for pathological speech intelligibility evaluation.

Keywords: pathological speech, multi-granularity feature, MSCC (Mel s-transform cepstrum coefficients), F-score, radar chart

Procedia PDF Downloads 259
3443 Optimization of the Measure of Compromise as a Version of Sorites Paradox

Authors: Aleksandar Hatzivelkos

Abstract:

The term ”compromise” is mostly used casually within the social choice theory. It is usually used as a mere result of the social choice function, and this omits its deeper meaning and ramifications. This paper is based on a mathematical model for the description of a compromise as a version of the Sorites paradox. It introduces a formal definition of d-measure of divergence from a compromise and models a notion of compromise that is often used only colloquially. Such a model for vagueness phenomenon, which lies at the core of the notion of compromise enables the introduction of new mathematical structures. In order to maximize compromise, different methods can be used. In this paper, we explore properties of a social welfare function TdM (from Total d-Measure), which is defined as a function which minimizes the total sum of d-measures of divergence over all possible linear orderings. We prove that TdM satisfy strict Pareto principle and behaves well asymptotically. Furthermore, we show that for certain domain restrictions, TdM satisfy positive responsiveness and IIIA (intense independence of irrelevant alternatives) thus being equivalent to Borda count on such domain restriction. This result gives new opportunities in social choice, especially when there is an emphasis on compromise in the decision-making process.

Keywords: borda count, compromise, measure of divergence, minimization

Procedia PDF Downloads 100
3442 Taylor’s Law and Relationship between Life Expectancy at Birth and Variance in Age at Death in Period Life Table

Authors: David A. Swanson, Lucky M. Tedrow

Abstract:

Taylor’s Law is a widely observed empirical pattern that relates variances to means in sets of non-negative measurements via an approximate power function, which has found application to human mortality. This study adds to this research by showing that Taylor’s Law leads to a model that reasonably describes the relationship between life expectancy at birth (e0, which also is equal to mean age at death in a life table) and variance at age of death in seven World Bank regional life tables measured at two points in time, 1970 and 2000. Using as a benchmark a non-random sample of four Japanese female life tables covering the period from 1950 to 2004, the study finds that the simple linear model provides reasonably accurate estimates of variance in age at death in a life table from e0, where the latter range from 60.9 to 85.59 years. Employing 2017 life tables from the Human Mortality Database, the simple linear model is used to provide estimates of variance at age in death for six countries, three of which have high e0 values and three of which have lower e0 values. The paper provides a substantive interpretation of Taylor’s Law relative to e0 and concludes by arguing that reasonably accurate estimates of variance in age at death in a period life table can be calculated using this approach, which also can be used where e0 itself is estimated rather than generated through the construction of a life table, a useful feature of the model.

Keywords: empirical pattern, mean age at death in a life table, mean age of a stationary population, stationary population

Procedia PDF Downloads 296
3441 Behavior of Droplets in Microfluidic System with T-Junction

Authors: A. Guellati, F-M Lounis, N. Guemras, K. Daoud

Abstract:

Micro droplet formation is considered as a growing emerging area of research due to its wide-range application in chemistry as well as biology. The mechanism of micro droplet formation using two immiscible liquids running through a T-junction has been widely studied. We believe that the flow of these two immiscible phases can be of greater important factor that could have an impact on out-flow hydrodynamic behavior, the droplets generated and the size of the droplets. In this study, the type of the capillary tubes used also represents another important factor that can have an impact on the generation of micro droplets. The tygon capillary tubing with hydrophilic inner surface doesn't allow regular out-flows due to the fact that the continuous phase doesn't adhere to the wall of the capillary inner surface. Teflon capillary tubing, presents better wettability than tygon tubing, and allows to obtain steady and regular regimes of out-flow, and the micro droplets are homogeneoussize. The size of the droplets is directly dependent on the flows of the continuous and dispersed phases. Thus, as increasing the flow of the continuous phase, to flow of the dispersed phase stationary, the size of the drops decreases. Inversely, while increasing the flow of the dispersed phase, to flow of the continuous phase stationary, the size of the droplet increases.

Keywords: microfluidic system, micro droplets generation, t-junction, fluids engineering

Procedia PDF Downloads 314
3440 Elucidation of the Sequential Transcriptional Activity in Escherichia coli Using Time-Series RNA-Seq Data

Authors: Pui Shan Wong, Kosuke Tashiro, Satoru Kuhara, Sachiyo Aburatani

Abstract:

Functional genomics and gene regulation inference has readily expanded our knowledge and understanding of gene interactions with regards to expression regulation. With the advancement of transcriptome sequencing in time-series comes the ability to study the sequential changes of the transcriptome. This method presented here works to augment existing regulation networks accumulated in literature with transcriptome data gathered from time-series experiments to construct a sequential representation of transcription factor activity. This method is applied on a time-series RNA-Seq data set from Escherichia coli as it transitions from growth to stationary phase over five hours. Investigations are conducted on the various metabolic activities in gene regulation processes by taking advantage of the correlation between regulatory gene pairs to examine their activity on a dynamic network. Especially, the changes in metabolic activity during phase transition are analyzed with focus on the pagP gene as well as other associated transcription factors. The visualization of the sequential transcriptional activity is used to describe the change in metabolic pathway activity originating from the pagP transcription factor, phoP. The results show a shift from amino acid and nucleic acid metabolism, to energy metabolism during the transition to stationary phase in E. coli.

Keywords: Escherichia coli, gene regulation, network, time-series

Procedia PDF Downloads 341
3439 Gaussian Probability Density for Forest Fire Detection Using Satellite Imagery

Authors: S. Benkraouda, Z. Djelloul-Khedda, B. Yagoubi

Abstract:

we present a method for early detection of forest fires from a thermal infrared satellite image, using the image matrix of the probability of belonging. The principle of the method is to compare a theoretical mathematical model to an experimental model. We considered that each line of the image matrix, as an embodiment of a non-stationary random process. Since the distribution of pixels in the satellite image is statistically dependent, we divided these lines into small stationary and ergodic intervals to characterize the image by an adequate mathematical model. A standard deviation was chosen to generate random variables, so each interval behaves naturally like white Gaussian noise. The latter has been selected as the mathematical model that represents a set of very majority pixels, which we can be considered as the image background. Before modeling the image, we made a few pretreatments, then the parameters of the theoretical Gaussian model were extracted from the modeled image, these settings will be used to calculate the probability of each interval of the modeled image to belong to the theoretical Gaussian model. The high intensities pixels are regarded as foreign elements to it, so they will have a low probability, and the pixels that belong to the background image will have a high probability. Finally, we did present the reverse of the matrix of probabilities of these intervals for a better fire detection.

Keywords: forest fire, forest fire detection, satellite image, normal distribution, theoretical gaussian model, thermal infrared matrix image

Procedia PDF Downloads 112
3438 Optimal Dynamic Regime for CO Oxidation Reaction Discovered by Policy-Gradient Reinforcement Learning Algorithm

Authors: Lifar M. S., Tereshchenko A. A., Bulgakov A. N., Guda S. A., Guda A. A., Soldatov A. V.

Abstract:

Metal nanoparticles are widely used as heterogeneous catalysts to activate adsorbed molecules and reduce the energy barrier of the reaction. Reaction product yield depends on the interplay between elementary processes - adsorption, activation, reaction, and desorption. These processes, in turn, depend on the inlet feed concentrations, temperature, and pressure. At stationary conditions, the active surface sites may be poisoned by reaction byproducts or blocked by thermodynamically adsorbed gaseous reagents. Thus, the yield of reaction products can significantly drop. On the contrary, the dynamic control accounts for the changes in the surface properties and adjusts reaction parameters accordingly. Therefore dynamic control may be more efficient than stationary control. In this work, a reinforcement learning algorithm has been applied to control the simulation of CO oxidation on a catalyst. The policy gradient algorithm is learned to maximize the CO₂ production rate based on the CO and O₂ flows at a given time step. Nonstationary solutions were found for the regime with surface deactivation. The maximal product yield was achieved for periodic variations of the gas flows, ensuring a balance between available adsorption sites and the concentration of activated intermediates. This methodology opens a perspective for the optimization of catalytic reactions under nonstationary conditions.

Keywords: artificial intelligence, catalyst, co oxidation, reinforcement learning, dynamic control

Procedia PDF Downloads 84
3437 A Similarity/Dissimilarity Measure to Biological Sequence Alignment

Authors: Muhammad A. Khan, Waseem Shahzad

Abstract:

Analysis of protein sequences is carried out for the purpose to discover their structural and ancestry relationship. Sequence similarity determines similar protein structures, similar function, and homology detection. Biological sequences composed of amino acid residues or nucleotides provide significant information through sequence alignment. In this paper, we present a new similarity/dissimilarity measure to sequence alignment based on the primary structure of a protein. The approach finds the distance between the two given sequences using the novel sequence alignment algorithm and a mathematical model. The algorithm runs at a time complexity of O(n²). A distance matrix is generated to construct a phylogenetic tree of different species. The new similarity/dissimilarity measure outperforms other existing methods.

Keywords: alignment, distance, homology, mathematical model, phylogenetic tree

Procedia PDF Downloads 152
3436 New Approaches for the Handwritten Digit Image Features Extraction for Recognition

Authors: U. Ravi Babu, Mohd Mastan

Abstract:

The present paper proposes a novel approach for handwritten digit recognition system. The present paper extract digit image features based on distance measure and derives an algorithm to classify the digit images. The distance measure can be performing on the thinned image. Thinning is the one of the preprocessing technique in image processing. The present paper mainly concentrated on an extraction of features from digit image for effective recognition of the numeral. To find the effectiveness of the proposed method tested on MNIST database, CENPARMI, CEDAR, and newly collected data. The proposed method is implemented on more than one lakh digit images and it gets good comparative recognition results. The percentage of the recognition is achieved about 97.32%.

Keywords: handwritten digit recognition, distance measure, MNIST database, image features

Procedia PDF Downloads 435
3435 Market Illiquidity and Pricing Errors in the Term Structure of CDS

Authors: Lidia Sanchis-Marco, Antonio Rubia, Pedro Serrano

Abstract:

This paper studies the informational content of pricing errors in the term structure of sovereign CDS spreads. The residuals from a non-arbitrage model are employed to construct a Price discrepancy estimate, or noise measure. The noise estimate is understood as an indicator of market distress and reflects frictions such as illiquidity. Empirically, the noise measure is computed for an extensive panel of CDS spreads. Our results reveal an important fraction of systematic risk is not priced in default swap contracts. When projecting the noise measure onto a set of financial variables, the panel-data estimates show that greater price discrepancies are systematically related to a higher level of offsetting transactions of CDS contracts. This evidence suggests that arbitrage capital flows exit the marketplace during time of distress, and this consistent with a market segmentation among investors and arbitrageurs where professional arbitrageurs are particularly ineffective at bringing prices to their fundamental values during turbulent periods. Our empirical findings are robust for the most common CDS pricing models employed in the industry.

Keywords: credit default swaps, noise measure, illiquidity, capital arbitrage

Procedia PDF Downloads 544
3434 Analysing the Stability of Electrical Grid for Increased Renewable Energy Penetration by Focussing on LI-Ion Battery Storage Technology

Authors: Hemendra Singh Rathod

Abstract:

Frequency is, among other factors, one of the governing parameters for maintaining electrical grid stability. The quality of an electrical transmission and supply system is mainly described by the stability of the grid frequency. Over the past few decades, energy generation by intermittent sustainable sources like wind and solar has seen a significant increase globally. Consequently, controlling the associated deviations in grid frequency within safe limits has been gaining momentum so that the balance between demand and supply can be maintained. Lithium-ion battery energy storage system (Li-Ion BESS) has been a promising technology to tackle the challenges associated with grid instability. BESS is, therefore, an effective response to the ongoing debate whether it is feasible to have an electrical grid constantly functioning on a hundred percent renewable power in the near future. In recent years, large-scale manufacturing and capital investment into battery production processes have made the Li-ion battery systems cost-effective and increasingly efficient. The Li-ion systems require very low maintenance and are also independent of geographical constraints while being easily scalable. The paper highlights the use of stationary and moving BESS for balancing electrical energy, thereby maintaining grid frequency at a rapid rate. Moving BESS technology, as implemented in the selected railway network in Germany, is here considered as an exemplary concept for demonstrating the same functionality in the electrical grid system. Further, using certain applications of Li-ion batteries, such as self-consumption of wind and solar parks or their ancillary services, wind and solar energy storage during low demand, black start, island operation, residential home storage, etc. offers a solution to effectively integrate the renewables and support Europe’s future smart grid. EMT software tool DIgSILENT PowerFactory has been utilised to model an electrical transmission system with 100% renewable energy penetration. The stability of such a transmission system has been evaluated together with BESS within a defined frequency band. The transmission system operators (TSO) have the superordinate responsibility for system stability and must also coordinate with the other European transmission system operators. Frequency control is implemented by TSO by maintaining a balance between electricity generation and consumption. Li-ion battery systems are here seen as flexible, controllable loads and flexible, controllable generation for balancing energy pools. Thus using Li-ion battery storage solution, frequency-dependent load shedding, i.e., automatic gradual disconnection of loads from the grid, and frequency-dependent electricity generation, i.e., automatic gradual connection of BESS to the grid, is used as a perfect security measure to maintain grid stability in any case scenario. The paper emphasizes the use of stationary and moving Li-ion battery storage for meeting the demands of maintaining grid frequency and stability for near future operations.

Keywords: frequency control, grid stability, li-ion battery storage, smart grid

Procedia PDF Downloads 122
3433 A Systematic Review of Situational Awareness and Cognitive Load Measurement in Driving

Authors: Aly Elshafei, Daniela Romano

Abstract:

With the development of autonomous vehicles, a human-machine interaction (HMI) system is needed for a safe transition of control when a takeover request (TOR) is required. An important part of the HMI system is the ability to monitor the level of situational awareness (SA) of any driver in real-time, in different scenarios, and without any pre-calibration. Presenting state-of-the-art machine learning models used to measure SA is the purpose of this systematic review. Investigating the limitations of each type of sensor, the gaps, and the most suited sensor and computational model that can be used in driving applications. To the author’s best knowledge this is the first literature review identifying online and offline classification methods used to measure SA, explaining which measurements are subject or session-specific, and how many classifications can be done with each classification model. This information can be very useful for researchers measuring SA to identify the most suited model to measure SA for different applications.

Keywords: situational awareness, autonomous driving, gaze metrics, EEG, ECG

Procedia PDF Downloads 92
3432 Numerical Simulation of Two-Dimensional Flow over a Stationary Circular Cylinder Using Feedback Forcing Scheme Based Immersed Boundary Finite Volume Method

Authors: Ranjith Maniyeri, Ahamed C. Saleel

Abstract:

Two-dimensional fluid flow over a stationary circular cylinder is one of the bench mark problem in the field of fluid-structure interaction in computational fluid dynamics (CFD). Motivated by this, in the present work, a two-dimensional computational model is developed using an improved version of immersed boundary method which combines the feedback forcing scheme of the virtual boundary method with Peskin’s regularized delta function approach. Lagrangian coordinates are used to represent the cylinder and Eulerian coordinates are used to describe the fluid flow. A two-dimensional Dirac delta function is used to transfer the quantities between the sold to fluid domain. Further, continuity and momentum equations governing the fluid flow are solved using fractional step based finite volume method on a staggered Cartesian grid system. The developed code is validated by comparing the values of drag coefficient obtained for different Reynolds numbers with that of other researcher’s results. Also, through numerical simulations for different Reynolds numbers flow behavior is well captured. The stability analysis of the improved version of immersed boundary method is tested for different values of feedback forcing coefficients.

Keywords: Feedback Forcing Scheme, Finite Volume Method, Immersed Boundary Method, Navier-Stokes Equations

Procedia PDF Downloads 279
3431 Explaining Listening Comprehension among L2 Learners of English: The Contribution of Vocabulary Knowledge and Working Memory Capacity

Authors: Ahmed Masrai

Abstract:

Listening comprehension constitutes a considerable challenge for the second language (L2) learners, but a little is known about the explanatory power of different variables in explaining variance in listening comprehension. Since research in this area, to the researcher's knowledge, is relatively small in comparison to that focusing on the relationship between reading comprehension and factors such as vocabulary and working memory, there is a need for studies that are seeking to fill the gap in our knowledge about the specific contribution of working memory capacity (WMC), aural vocabulary knowledge and written vocabulary knowledge to explaining listening comprehension. Among 130 English as foreign language learners, the present study examines what proportion of the variance in listening comprehension is explained by aural vocabulary knowledge, written vocabulary knowledge, and WMC. Four measures were used to collect the required data for the study: (1) A-Lex, a measure of aural vocabulary knowledge; (2) XK-Lex, a measure of written vocabulary knowledge; (3) Listening Span Task, a measure of WMC and; (4) IELTS Listening Test, a measure of listening comprehension. The results show that aural vocabulary knowledge is the strongest predictor of listening comprehension, followed by WMC, while written vocabulary knowledge is the weakest predictor. The study discusses implications for the explanatory power of aural vocabulary knowledge and WMC to listening comprehension and pedagogical practice in L2 classrooms.

Keywords: listening comprehension, second language, vocabulary knowledge, working memory

Procedia PDF Downloads 349
3430 Measure of Pleasure of Drug Users

Authors: Vano Tsertsvadze, Marina Chavchanidze, Lali Khurtsia

Abstract:

Problem of drug use is often seen as a combination of psychological and social problems, but this problem can be considered as economically rational decision in the process of buying pleasure (looking after children, reading, harvesting fruits in the fall, sex, eating, etc.). Before the adoption of the decisions people face to a trade-off - when someone chooses a delicious meal, she takes a completely rational decision, that the pleasure of eating has a lot more value than the pleasure which she will experience after two months diet on the summer beach showing off her beautiful body. This argument is also true for alcohol, drugs and cigarettes. Smoking has a negative effect on health, but smokers are not afraid of the threat of a lung cancer after 40 years, more valuable moment is a pleasure from smoking. Our hypothesis - unsatisfied pleasure and frustration, probably determines the risk of dependence on drug abuse. The purpose of research: 1- to determine the relative measure unit of pleasure, which will be used to measure and assess the intensity of various human pleasures. 2- to compare the intensity of the pleasure from different kinds of activity, with pleasures received from drug use. 3- Based on the analysis of data, to identify factors affecting the rational decision making. Research method: Respondents will be asked to recall the greatest pleasure of their life, which will be used as a measure of the other pleasures. The study will use focus groups and structured interviews.

Keywords: drug, drug-user, measurement, satisfaction

Procedia PDF Downloads 299
3429 An Innovative High Energy Density Power Pack for Portable and Off-Grid Power Applications

Authors: Idit Avrahami, Alex Schechter, Lev Zakhvatkin

Abstract:

This research focuses on developing a compact and light Hydrogen Generator (HG), coupled with fuel cells (FC) to provide a High-Energy-Density Power-Pack (HEDPP) solution, which is 10 times Li-Ion batteries. The HEDPP is designed for portable & off-grid power applications such as Drones, UAVs, stationary off-grid power sources, unmanned marine vehicles, and more. Hydrogen gas provided by this device is delivered in the safest way as a chemical powder at room temperature and ambient pressure is activated only when the power is on. Hydrogen generation is based on a stabilized chemical reaction of Sodium Borohydride (SBH) and water. The proposed solution enables a ‘No Storage’ Hydrogen-based Power Pack. Hydrogen is produced and consumed on-the-spot, during operation; therefore, there’s no need for high-pressure hydrogen tanks, which are large, heavy, and unsafe. In addition to its high energy density, ease of use, and safety, the presented power pack has a significant advantage of versatility and deployment in numerous applications and scales. This patented HG was demonstrated using several prototypes in our lab and was proved to be feasible and highly efficient for several applications. For example, in applications where water is available (such as marine vehicles, water and sewage infrastructure, and stationary applications), the Energy Density of the suggested power pack may reach 2700-3000 Wh/kg, which is again more than 10 times higher than conventional lithium-ion batteries. In other applications (e.g., UAV or small vehicles) the energy density may exceed 1000 Wh/kg.

Keywords: hydrogen energy, sodium borohydride, fixed-wing UAV, energy pack

Procedia PDF Downloads 55
3428 A Combinatorial Representation for the Invariant Measure of Diffusion Processes on Metric Graphs

Authors: Michele Aleandri, Matteo Colangeli, Davide Gabrielli

Abstract:

We study a generalization to a continuous setting of the classical Markov chain tree theorem. In particular, we consider an irreducible diffusion process on a metric graph. The unique invariant measure has an atomic component on the vertices and an absolutely continuous part on the edges. We show that the corresponding density at x can be represented by a normalized superposition of the weights associated to metric arborescences oriented toward the point x. A metric arborescence is a metric tree oriented towards its root. The weight of each oriented metric arborescence is obtained by the product of the exponential of integrals of the form ∫a/b², where b is the drift and σ² is the diffusion coefficient, along the oriented edges, for a weight for each node determined by the local orientation of the arborescence around the node and for the inverse of the diffusion coefficient at x. The metric arborescences are obtained by cutting the original metric graph along some edges.

Keywords: diffusion processes, metric graphs, invariant measure, reversibility

Procedia PDF Downloads 134
3427 Macroeconomic Measure of Projectification: An Empirical Study of Pakistani Economy

Authors: Shafaq Rana, Hina Ansar

Abstract:

Projectification is an emerging phenomenon in Western economies. The projects have become the key driver of the economic actions. The impact of projectification is understudy for over a decade. A methodology was developed to measure the degree of projectification at economical level, which was later adapted to measure the degree of projectification in Germany, Norway, and Iceland; and compared the differences in these project societies, considering their industrial structure, organizational size, and the share of project work. Using the same methodology, this study aims to provide empirical evidence of the project work in the context of Pakistan –a developing nation, keeping into consideration the macroeconomic measures, qualitative and quantitative measures of the project i/c GDP, monetary measures, and project success. The research includes a qualitative pre-study to define these macro-measures in the country-specific context and a quantitative study to measure the project work w.r.t hours working in the organizations on projects. The outcome of this study provides the key data on the projectification in a developing economy, which will help industry practitioners and decision-makers to examine the consequences of projectification and strategize, respectively. This study also provides a foundation for further research in individual sectors of the country while exploring different macroeconomic questions, including the effect of projectification on project productivity, income effects, and labor market.

Keywords: developing economy, Pakistan, project work, projectification

Procedia PDF Downloads 95
3426 Prioritization of Mutation Test Generation with Centrality Measure

Authors: Supachai Supmak, Yachai Limpiyakorn

Abstract:

Mutation testing can be applied for the quality assessment of test cases. Prioritization of mutation test generation has been a critical element of the industry practice that would contribute to the evaluation of test cases. The industry generally delivers the product under the condition of time to the market and thus, inevitably sacrifices software testing tasks, even though many test cases are required for software verification. This paper presents an approach of applying a social network centrality measure, PageRank, to prioritize mutation test generation. The source code with the highest values of PageRank will be focused first when developing their test cases as these modules are vulnerable to defects or anomalies which may cause the consequent defects in many other associated modules. Moreover, the approach would help identify the reducible test cases in the test suite, still maintaining the same criteria as the original number of test cases.

Keywords: software testing, mutation test, network centrality measure, test case prioritization

Procedia PDF Downloads 72
3425 Unmanned Aerial System Development for the Remote Reflectance Sensing Using Above-Water Radiometers

Authors: Sunghun Jung, Wonkook Kim

Abstract:

Due to the difficulty of the utilization of satellite and an aircraft, conventional ocean color remote sensing has a disadvantage in that it is difficult to obtain images of desired places at desired times. These disadvantages make it difficult to capture the anomalies such as the occurrence of the red tide which requires immediate observation. It is also difficult to understand the phenomena such as the resuspension-precipitation process of suspended solids and the spread of low-salinity water originating in the coastal areas. For the remote sensing reflectance of seawater, above-water radiometers (AWR) have been used either by carrying portable AWRs on a ship or installing those at fixed observation points on the Ieodo ocean research station, Socheongcho base, and etc. In particular, however, it requires the high cost to measure the remote reflectance in various seawater environments at various times and it is even not possible to measure it at the desired frequency in the desired sea area at the desired time. Also, in case of the stationary observation, it is advantageous that observation data is continuously obtained, but there is the disadvantage that data of various sea areas cannot be obtained. It is possible to instantly capture various marine phenomena occurring on the coast using the unmanned aerial system (UAS) including vertical takeoff and landing (VTOL) type unmanned aerial vehicles (UAV) since it could move and hover at the one location and acquire data of the desired form at a high resolution. To remotely estimate seawater constituents, it is necessary to install an ultra-spectral sensor. Also, to calculate reflected light from the surface of the sea in consideration of the sun’s incident light, a total of three sensors need to be installed on the UAV. The remote sensing reflectance of seawater is the most basic optical property for remotely estimating color components in seawater and we could remotely estimate the chlorophyll concentration, the suspended solids concentration, and the dissolved organic amount. Estimating seawater physics from the remote sensing reflectance requires the algorithm development using the accumulation data of seawater reflectivity under various seawater and atmospheric conditions. The UAS with three AWRs is developed for the remote reflection sensing on the surface of the sea. Throughout the paper, we explain the details of each UAS component, system operation scenarios, and simulation and experiment results. The UAS consists of a UAV, a solar tracker, a transmitter, a ground control station (GCS), three AWRs, and two gimbals.

Keywords: above-water radiometers (AWR), ground control station (GCS), unmanned aerial system (UAS), unmanned aerial vehicle (UAV)

Procedia PDF Downloads 141
3424 Evaluating Emission Reduction Due to a Proposed Light Rail Service: A Micro-Level Analysis

Authors: Saeid Eshghi, Neeraj Saxena, Abdulmajeed Alsultan

Abstract:

Carbon dioxide (CO2) alongside other gas emissions in the atmosphere cause a greenhouse effect, resulting in an increase of the average temperature of the planet. Transportation vehicles are among the main contributors of CO2 emission. Stationary vehicles with initiated motors produce more emissions than mobile ones. Intersections with traffic lights that force the vehicles to become stationary for a period of time produce more CO2 pollution than other parts of the road. This paper focuses on analyzing the CO2 produced by the traffic flow at Anzac Parade Road - Barker Street intersection in Sydney, Australia, before and after the implementation of Light rail transport (LRT). The data are gathered during the construction phase of the LRT by collecting the number of vehicles on each path of the intersection for 15 minutes during the evening rush hour of 1 week (6-7 pm, July 04-31, 2018) and then multiplied by 4 to calculate the flow of vehicles in 1 hour. For analyzing the data, the microscopic simulation software “VISSIM” has been used. Through the analysis, the traffic flow was processed in three stages: before and after implementation of light rail train, and one during the construction phase. Finally, the traffic results were input into another software called “EnViVer”, to calculate the amount of CO2 during 1 h. The results showed that after the implementation of the light rail, CO2 will drop by a minimum of 13%. This finding provides an evidence that light rail is a sustainable mode of transport.

Keywords: carbon dioxide, emission modeling, light rail, microscopic model, traffic flow

Procedia PDF Downloads 112
3423 The Usage of Bridge Estimator for Hegy Seasonal Unit Root Tests

Authors: Huseyin Guler, Cigdem Kosar

Abstract:

The aim of this study is to propose Bridge estimator for seasonal unit root tests. Seasonality is an important factor for many economic time series. Some variables may contain seasonal patterns and forecasts that ignore important seasonal patterns have a high variance. Therefore, it is very important to eliminate seasonality for seasonal macroeconomic data. There are some methods to eliminate the impacts of seasonality in time series. One of them is filtering the data. However, this method leads to undesired consequences in unit root tests, especially if the data is generated by a stochastic seasonal process. Another method to eliminate seasonality is using seasonal dummy variables. Some seasonal patterns may result from stationary seasonal processes, which are modelled using seasonal dummies but if there is a varying and changing seasonal pattern over time, so the seasonal process is non-stationary, deterministic seasonal dummies are inadequate to capture the seasonal process. It is not suitable to use seasonal dummies for modeling such seasonally nonstationary series. Instead of that, it is necessary to take seasonal difference if there are seasonal unit roots in the series. Different alternative methods are proposed in the literature to test seasonal unit roots, such as Dickey, Hazsa, Fuller (DHF) and Hylleberg, Engle, Granger, Yoo (HEGY) tests. HEGY test can be also used to test the seasonal unit root in different frequencies (monthly, quarterly, and semiannual). Another issue in unit root tests is the lag selection. Lagged dependent variables are added to the model in seasonal unit root tests as in the unit root tests to overcome the autocorrelation problem. In this case, it is necessary to choose the lag length and determine any deterministic components (i.e., a constant and trend) first, and then use the proper model to test for seasonal unit roots. However, this two-step procedure might lead size distortions and lack of power in seasonal unit root tests. Recent studies show that Bridge estimators are good in selecting optimal lag length while differentiating nonstationary versus stationary models for nonseasonal data. The advantage of this estimator is the elimination of the two-step nature of conventional unit root tests and this leads a gain in size and power. In this paper, the Bridge estimator is proposed to test seasonal unit roots in a HEGY model. A Monte-Carlo experiment is done to determine the efficiency of this approach and compare the size and power of this method with HEGY test. Since Bridge estimator performs well in model selection, our approach may lead to some gain in terms of size and power over HEGY test.

Keywords: bridge estimators, HEGY test, model selection, seasonal unit root

Procedia PDF Downloads 299
3422 Rapid Separation of Biomolecules and Neutral Analytes with a Cationic Stationary Phase by Capillary Electrochromatography

Authors: A. Aslihan Gokaltun, Ali Tuncel

Abstract:

The unique properties of capillary electrochromatography (CEC) such as high performance, high selectivity, low consumption of both reagents and analytes ensure this technique an attractive one for the separation of biomolecules including nucleosides and nucleotides, peptides, proteins, carbohydrates. Monoliths have become a well-established separation media for CEC in the format that can be compared to a single large 'particle' that does not include interparticular voids. Convective flow through the pores of monolith significantly accelerates the rate of mass transfer and enables a substantial increase in the speed of the separation. In this work, we propose a new approach for the preparation of cationic monolithic stationary phase for capillary electrochromatography. Instead of utilizing a charge bearing monomer during polymerization, the desired charge-bearing group is generated on the capillary monolith after polymerization by using the reactive moiety of the monolithic support via one-pot, simple reaction. Optimized monolithic column compensates the disadvantages of frequently used reversed phases, which are difficult for separation of polar solutes. Rapid separation and high column efficiencies are achieved for the separation of neutral analytes, nucleic acid bases and nucleosides in reversed phase mode. Capillary monolith showed satisfactory hydrodynamic permeability and mechanical stability with relative standard deviation (RSD) values below 2 %. A new promising, reactive support that has a 'ligand selection flexibility' due to its reactive functionality represent a new family of separation media for CEC.

Keywords: biomolecules, capillary electrochromatography, cationic monolith, neutral analytes

Procedia PDF Downloads 188
3421 Is It Important to Measure the Volumetric Mass Density of Nanofluids?

Authors: Z. Haddad, C. Abid, O. Rahli, O. Margeat, W. Dachraoui, A. Mataoui

Abstract:

The present study aims to measure the volumetric mass density of NiPd-heptane nanofluids synthesized using a one-step method known as thermal decomposition of metal-surfactant complexes. The particle concentration is up to 7.55 g/l and the temperature range of the experiment is from 20°C to 50°C. The measured values were compared with the mixture theory and good agreement between the theoretical equation and measurement were obtained. Moreover, the available nanofluids volumetric mass density data in the literature is reviewed.

Keywords: NiPd nanoparticles, nanofluids, volumetric mass density, stability

Procedia PDF Downloads 370