Search results for: the measurement parameters
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10945

Search results for: the measurement parameters

10675 Readiness of Intellectual Capital Measurement: A Review of the Property Development and Investment Industry

Authors: Edward C. W. Chan, Benny C. F. Cheung

Abstract:

In the knowledge economy, the financial indicator is not the unique instrument to gauge the performance of a company. The role of intellectual capital contributing to the company performance is increasing. To measure the company performance due to intellectual capital, the value-added intellectual capital (VAIC) model is adopted to measure the intellectual capital utilisation efficiency of the subject companies. The purpose of this study is to review the readiness of measuring intellectual capital for the Hong Kong listed companies in the property development and property investment industry by using VAIC model. This study covers the financial reports from the representative Hong Kong listed property development companies and property investment companies in the period 2014-2019. The findings from this study indicated the industry is ready for IC measurement employing VAIC framework but not yet ready for using the extended VAIC model.

Keywords: intellectual capital, intellectual capital measurement, property development, property investment, Skandia navigator, VAIC

Procedia PDF Downloads 110
10674 Challenges and Insights by Electrical Characterization of Large Area Graphene Layers

Authors: Marcus Klein, Martina GrießBach, Richard Kupke

Abstract:

The current advances in the research and manufacturing of large area graphene layers are promising towards the introduction of this exciting material in the display industry and other applications that benefit from excellent electrical and optical characteristics. New production technologies in the fabrication of flexible displays, touch screens or printed electronics apply graphene layers on non-metal substrates and bring new challenges to the required metrology. Traditional measurement concepts of layer thickness, sheet resistance, and layer uniformity, are difficult to apply to graphene production processes and are often harmful to the product layer. New non-contact sensor concepts are required to adapt to the challenges and even the foreseeable inline production of large area graphene. Dedicated non-contact measurement sensors are a pioneering method to leverage these issues in a large variety of applications, while significantly lowering the costs of development and process setup. Transferred and printed graphene layers can be characterized with high accuracy in a huge measurement range using a very high resolution. Large area graphene mappings are applied for process optimization and for efficient quality control for transfer, doping, annealing and stacking processes. Examples of doped, defected and excellent Graphene are presented as quality images and implications for manufacturers are explained.

Keywords: graphene, doping and defect testing, non-contact sheet resistance measurement, inline metrology

Procedia PDF Downloads 306
10673 Dissolved Oxygen Prediction Using Support Vector Machine

Authors: Sorayya Malek, Mogeeb Mosleh, Sharifah M. Syed

Abstract:

In this study, Support Vector Machine (SVM) technique was applied to predict the dichotomized value of Dissolved oxygen (DO) from two freshwater lakes namely Chini and Bera Lake (Malaysia). Data sample contained 11 parameters for water quality features from year 2005 until 2009. All data parameters were used to predicate the dissolved oxygen concentration which was dichotomized into 3 different levels (High, Medium, and Low). The input parameters were ranked, and forward selection method was applied to determine the optimum parameters that yield the lowest errors, and highest accuracy. Initial results showed that pH, water temperature, and conductivity are the most important parameters that significantly affect the predication of DO. Then, SVM model was applied using the Anova kernel with those parameters yielded 74% accuracy rate. We concluded that using SVM models to predicate the DO is feasible, and using dichotomized value of DO yields higher prediction accuracy than using precise DO value.

Keywords: dissolved oxygen, water quality, predication DO, support vector machine

Procedia PDF Downloads 284
10672 Examining the Modular End of Line Control Unit Design Criteria for Vehicle Sliding Door System Slide Profile

Authors: Orhan Kurtuluş, Cüneyt Yavuz

Abstract:

The end of the line controls of the finished products in the automotive industry is important. The control that has been conducted with the manual methods for the sliding doors tracks is not sufficient and faulty products cannot be identified. As a result, the customer has the faulty products. In the scope of this study, the design criteria of the PLC integrated modular end of line control unit has been examined, designed and manufactured to make the control of the 10 different track profile to 2 different vehicles with an objective to minimize the salvage costs by obtaining more sensitive, certain and accurate measurement results. In the study that started with literature and patent review, the design inputs have been specified, the technical concept has been developed, computer supported mechanic design, control system and automation design, design review and design improvement have been made. Laser analog sensors at high sensitivity, probes and modular blocks have been used in the unit. The measurement has been conducted in the system and it is observed that measurement results are more sensitive than the previous methods.

Keywords: control unit design, end of line, modular design, sliding door system

Procedia PDF Downloads 441
10671 Approach for the Mathematical Calculation of the Damping Factor of Railway Bridges with Ballasted Track

Authors: Andreas Stollwitzer, Lara Bettinelli, Josef Fink

Abstract:

The expansion of the high-speed rail network over the past decades has resulted in new challenges for engineers, including traffic-induced resonance vibrations of railway bridges. Excessive resonance-induced speed-dependent accelerations of railway bridges during high-speed traffic can lead to negative consequences such as fatigue symptoms, distortion of the track, destabilisation of the ballast bed, and potentially even derailment. A realistic prognosis of bridge vibrations during high-speed traffic must not only rely on the right choice of an adequate calculation model for both bridge and train but first and foremost on the use of dynamic model parameters which reflect reality appropriately. However, comparisons between measured and calculated bridge vibrations are often characterised by considerable discrepancies, whereas dynamic calculations overestimate the actual responses and therefore lead to uneconomical results. This gap between measurement and calculation constitutes a complex research issue and can be traced to several causes. One major cause is found in the dynamic properties of the ballasted track, more specifically in the persisting, substantial uncertainties regarding the consideration of the ballasted track (mechanical model and input parameters) in dynamic calculations. Furthermore, the discrepancy is particularly pronounced concerning the damping values of the bridge, as conservative values have to be used in the calculations due to normative specifications and lack of knowledge. By using a large-scale test facility, the analysis of the dynamic behaviour of ballasted track has been a major research topic at the Institute of Structural Engineering/Steel Construction at TU Wien in recent years. This highly specialised test facility is designed for isolated research of the ballasted track's dynamic stiffness and damping properties – independent of the bearing structure. Several mechanical models for the ballasted track consisting of one or more continuous spring-damper elements were developed based on the knowledge gained. These mechanical models can subsequently be integrated into bridge models for dynamic calculations. Furthermore, based on measurements at the test facility, model-dependent stiffness and damping parameters were determined for these mechanical models. As a result, realistic mechanical models of the railway bridge with different levels of detail and sufficiently precise characteristic values are available for bridge engineers. Besides that, this contribution also presents another practical application of such a bridge model: Based on the bridge model, determination equations for the damping factor (as Lehr's damping factor) can be derived. This approach constitutes a first-time method that makes the damping factor of a railway bridge calculable. A comparison of this mathematical approach with measured dynamic parameters of existing railway bridges illustrates, on the one hand, the apparent deviation between normatively prescribed and in-situ measured damping factors. On the other hand, it is also shown that a new approach, which makes it possible to calculate the damping factor, provides results that are close to reality and thus raises potentials for minimising the discrepancy between measurement and calculation.

Keywords: ballasted track, bridge dynamics, damping, model design, railway bridges

Procedia PDF Downloads 161
10670 A Case Study on the Condition Monitoring of a Critical Machine in a Tyre Manufacturing Plant

Authors: Ramachandra C. G., Amarnath. M., Prashanth Pai M., Nagesh S. N.

Abstract:

The machine's performance level drops down over a period of time due to the wear and tear of its components. The early detection of an emergent fault becomes very vital in order to obtain uninterrupted production in a plant. Maintenance is an activity that helps to keep the machine's performance at an anticipated level, thereby ensuring the availability of the machine to perform its intended function. At present, a number of modern maintenance techniques are available, such as preventive maintenance, predictive maintenance, condition-based maintenance, total productive maintenance, etc. Condition-based maintenance or condition monitoring is one such modern maintenance technique in which the machine's condition or health is checked by the measurement of certain parameters such as sound level, temperature, velocity, displacement, vibration, etc. It can recognize most of the factors restraining the usefulness and efficacy of the total manufacturing unit. This research work is conducted on a Batch Mill in a tire production unit located in the Southern Karnataka region. The health of the mill is assessed using amplitude of vibration as a parameter of measurement. Most commonly, the vibration level is assessed using various points on the machine bearing. The normal or standard level is fixed using reference materials such as manuals or catalogs supplied by the manufacturers and also by referring vibration standards. The Rio-Vibro meter is placed in different locations on the batch-off mill to record the vibration data. The data collected are analyzed to identify the malfunctioning components in the batch off the mill, and corrective measures are suggested.

Keywords: availability, displacement, vibration, rio-vibro, condition monitoring

Procedia PDF Downloads 83
10669 Optimum Er: YAG Laser Parameters for Orthodontic Composite Debonding: An in vitro Study

Authors: Mohammad Zamzam, Wesam Bachir, Imad Asaad

Abstract:

Several studies have produced estimates of Er:YAG laser parameters and specifications but there is still insufficient data for reliable selection of laser parameters. As a consequence, there is a heightened need for ideal specifications of Er:YAG laser to reduce the amount of enamel ablation. The objective of this paper is to investigate the influence of Er:YAG laser parameters, energy level and pulse duration, on orthodontic composite removal after bracket debonding. The sample consisted of 45 cuboids of orthodontic composite made by plastic moulds. The samples were divided into three groups, each was irradiated with Er:YAG laser set at different energy levels and three values for pulse durations (50 µs, 100 µs, and 300 µs). Geometrical parameters (depth and area) of cavities formed by laser irradiation were determined. ANCOVA test showed statistically significant difference (p < 0.0.5) between the groups indicating a potential effect of laser pulse duration on the geometrical parameters after controlling laser energy level. A post-hoc Bonferroni test ranked the 50µ Er:YAG laser pulse as the most influential factor for all geometrical parameters in removing remnant composite from enamel surface. Also, 300 mJ laser pulses caused the largest removal of the composite. The results of the present study demonstrated the efficacy of 50 µs and 300 mJ Er:YAG laser pulse for removal of remnant orthodontic composite.

Keywords: enamel, Er:YAG, geometrical parameters, orthodontic composite, remnant composite

Procedia PDF Downloads 545
10668 Creation and Validation of a Measurement Scale of E-Management: An Exploratory and Confirmatory Study

Authors: Hamadi Khlif

Abstract:

This paper deals with the understanding of the concept of e-management and the development of a measuring instrument adapted to the new problems encountered during the application of this new practice within the modern enterprise. Two principal e-management factors have been isolated in an exploratory study carried out among 260 participants. A confirmatory study applied to a second sample of 270 participants has been established in a cross-validation of the scale of measurement. The study presents the literature review specifically dedicated to e-management and the results of the exploratory and confirmatory phase of the development of this scale, which demonstrates satisfactory psychometric qualities. The e-management has two dimensions: a managerial dimension and a technological dimension.

Keywords: e-management, management, ICT deployment, mode of management

Procedia PDF Downloads 319
10667 Preparation Control Information and Analyzing of Metering Gas System Based of Orifice Plate

Authors: A. Harrouz, A. Benatiallah, O. Harrouz

Abstract:

This paper presents the search for errors in the measurement instruments in a dynamic system of metering liquid or gas and sees the tolerance defined by the international standards and recommendations. We will implement a program on MATLAB/Simulink which is calculated based on the ISO-5167. This program will take the system parameters on considerations such as: the willingness plates, the size of the orifice, the given design conditions, reference conditions, find pressure drop for a given flow, or flow for a loss of given load. The results are considered very good and satisfactory because the errors identified of measuring instruments system are within the margin of error limit by the regulations.

Keywords: analyzing, control, gas, meters system

Procedia PDF Downloads 395
10666 SiC Merged PiN and Schottky (MPS) Power Diodes Electrothermal Modeling in SPICE

Authors: A. Lakrim, D. Tahri

Abstract:

This paper sets out a behavioral macro-model of a Merged PiN and Schottky (MPS) diode based on silicon carbide (SiC). This model holds good for both static and dynamic electrothermal simulations for industrial applications. Its parameters have been worked out from datasheets curves by drawing on the optimization method: Simulated Annealing (SA) for the SiC MPS diodes made available in the industry. The model also adopts the Analog Behavioral Model (ABM) of PSPICE in which it has been implemented. The thermal behavior of the devices was also taken into consideration by making use of Foster’ canonical network as figured out from electro-thermal measurement provided by the manufacturer of the device.

Keywords: SiC MPS diode, electro-thermal, SPICE model, behavioral macro-model

Procedia PDF Downloads 404
10665 Efficiency Measurement of Turkish via the Stochastic Frontier Model

Authors: Yeliz Mert Kantar, İsmail Yeni̇lmez, Ibrahim Arik

Abstract:

In this study, the efficiency measurement of the top fifty Turkish Universities has been conducted. The top fifty Turkish Universities are listed by The Scientific and Technological Research Council of Turkey (TÜBITAK) according to the Entrepreneur and Innovative University Index every year. The index is calculated based on four components since 2018. Four components are scientific and technological research competency, intellectual property pool, cooperation and interaction, and economic and social contribution. The four components consist of twenty-three sub-components. The 2021 list announced in January 2022 is discussed in this study. Efficiency analysis have been carried out using the Stochastic Frontier Model. Statistical significance of the sub-components that make up the index with certain weights has been examined in terms of the efficiency measurement calculated through the Stochastic Frontier Model. The relationship between the efficiency ranking estimated based on the Stochastic Frontier Model and the Entrepreneur and Innovative University Index ranking is discussed in detail.

Keywords: efficiency, entrepreneur and innovative universities, turkish universities, stochastic frontier model, tübi̇tak

Procedia PDF Downloads 87
10664 Blood Glucose Measurement and Analysis: Methodology

Authors: I. M. Abd Rahim, H. Abdul Rahim, R. Ghazali

Abstract:

There is numerous non-invasive blood glucose measurement technique developed by researchers, and near infrared (NIR) is the potential technique nowadays. However, there are some disagreements on the optimal wavelength range that is suitable to be used as the reference of the glucose substance in the blood. This paper focuses on the experimental data collection technique and also the analysis method used to analyze the data gained from the experiment. The selection of suitable linear and non-linear model structure is essential in prediction system, as the system developed need to be conceivably accurate.

Keywords: linear, near-infrared (NIR), non-invasive, non-linear, prediction system

Procedia PDF Downloads 454
10663 Parametric Dependence of the Advection-Diffusion Equation in Two Dimensions

Authors: Matheus Fernando Pereira, Varese Salvador Timoteo

Abstract:

In this work, we have solved the two-dimensional advection-diffusion equation numerically for a spatially dependent solute dispersion along non-uniform flow with a pulse type source in order to make a systematic study on the influence of medium heterogeneity, initial flow velocity, and initial dispersion coefficient parameters on the solutions of the equation. The behavior of the solutions is then investigated as we change the three parameters independently. Our results show that even though the parameters represent different physical features of the system, the effect on their variation is very similar. We also observe that the effects caused by the parameters on the concentration depend on the distance from the source. Finally, our numerical results are in good agreement with the exact solutions for all values of the parameters we used in our analysis.

Keywords: advection-diffusion equation, dispersion, numerical methods, pulse-type source

Procedia PDF Downloads 236
10662 Effect of Soil and Material Characteristics on Safety of Concrete Structures Including SSI

Authors: A. E. Kurtoglu, A. Cevik, M. Bilgehan

Abstract:

In this parametric study, effect of soil and material characteristics on safety of structures is investigated. The soil parameters such as shear strength, unit weight; geometrical parameters of the structure such as foundation depth and height of building; and material properties such as weight of concrete were selected as input parameters. A real accelerogram of 1989 El-Centro earthquake recorded by the USGS in Imperial Valley is used for this study. It is contained in the standard Strong Motion CD-ROM (SMC) format, which can be recognized and interpreted by FEM software used. The soil-structure interaction model subjected to above-mentioned earthquake was analyzed for 729 cases. Effect of input parameters on safety factor of the soil-structure system was then investigated and the interaction between the input and output parameters is presented in graphical form. Findings showed that all input parameters have significant effects on factor of safety results.

Keywords: factor of safety, finite element method, safety of structures, soil structure interaction

Procedia PDF Downloads 502
10661 Measurement of Reverse Flow Generated at Cold Exit of Vortex Tube

Authors: Mohd Hazwan bin Yusof, Hiroshi Katanoda

Abstract:

In order to clarify the structure of the cold flow discharged from the vortex tube (VT), the pressure of the cold flow was measured, and a simple flow visualization technique using a 0.75 mm-diameter needle and an oily paint is made to study the reverse flow at the cold exit. It is clear that a negative pressure and positive pressure region exist at a certain pressure and cold fraction area, and that a reverse flow is observed in the negative pressure region.

Keywords: flow visualization, pressure measurement, reverse flow, vortex tube

Procedia PDF Downloads 514
10660 Measurement Errors and Misclassifications in Covariates in Logistic Regression: Bayesian Adjustment of Main and Interaction Effects and the Sample Size Implications

Authors: Shahadut Hossain

Abstract:

Measurement errors in continuous covariates and/or misclassifications in categorical covariates are common in epidemiological studies. Regression analysis ignoring such mismeasurements seriously biases the estimated main and interaction effects of covariates on the outcome of interest. Thus, adjustments for such mismeasurements are necessary. In this research, we propose a Bayesian parametric framework for eliminating deleterious impacts of covariate mismeasurements in logistic regression. The proposed adjustment method is unified and thus can be applied to any generalized linear and non-linear regression models. Furthermore, adjustment for covariate mismeasurements requires validation data usually in the form of either gold standard measurements or replicates of the mismeasured covariates on a subset of the study population. Initial investigation shows that adequacy of such adjustment depends on the sizes of main and validation samples, especially when prevalences of the categorical covariates are low. Thus, we investigate the impact of main and validation sample sizes on the adjusted estimates, and provide a general guideline about these sample sizes based on simulation studies.

Keywords: measurement errors, misclassification, mismeasurement, validation sample, Bayesian adjustment

Procedia PDF Downloads 404
10659 Liquefaction Susceptibility of Tailing Storage Facility-Comparison of National Centre for Earthquake Engineering Research and Finite Element Methods

Authors: Mehdi Ghatei, Masoomeh Lorestani

Abstract:

Upstream Tailings Storage Facilities (TSFs) may experience slope instabilities due to soil liquefaction, especially in regions known to be seismically active. In this study, liquefaction susceptibility of an upstream-raised TSF in Western Australia was assessed using two different approaches. The first approach assessed liquefaction susceptibility using Cone Penetration Tests with pore pressure measurement (CPTu) as described by the National Centre for Earthquake Engineering Research (NCEER). This assessment was based on the four CPTu tests that were conducted on the perimeter embankment of the TSF. The second approach used the Finite Element (FE) method with application of an equivalent linear model to predict the undrained cyclic behavior, the pore water pressure and the liquefaction of the materials. The tailings parameters were estimated from the CPTu profiles and from the laboratory tests. The cyclic parameters were estimated from the literature where test results of similar material were available. The results showed that there was a good agreement, in the liquefaction susceptibility of the tailings material, between the NCEER and FE methods with equivalent linear model.

Keywords: liquefaction , CPTU, NCEER, finite element method, equivalent linear model

Procedia PDF Downloads 268
10658 Comparison between Bernardi’s Equation and Heat Flux Sensor Measurement as Battery Heat Generation Estimation Method

Authors: Marlon Gallo, Eduardo Miguel, Laura Oca, Eneko Gonzalez, Unai Iraola

Abstract:

The heat generation of an energy storage system is an essential topic when designing a battery pack and its cooling system. Heat generation estimation is used together with thermal models to predict battery temperature in operation and adapt the design of the battery pack and the cooling system to these thermal needs guaranteeing its safety and correct operation. In the present work, a comparison between the use of a heat flux sensor (HFS) for indirect measurement of heat losses in a cell and the widely used and simplified version of Bernardi’s equation for estimation is presented. First, a Li-ion cell is thermally characterized with an HFS to measure the thermal parameters that are used in a first-order lumped thermal model. These parameters are the equivalent thermal capacity and the thermal equivalent resistance of a single Li-ion cell. Static (when no current is flowing through the cell) and dynamic (making current flow through the cell) tests are conducted in which HFS is used to measure heat between the cell and the ambient, so thermal capacity and resistances respectively can be calculated. An experimental platform records current, voltage, ambient temperature, surface temperature, and HFS output voltage. Second, an equivalent circuit model is built in a Matlab-Simulink environment. This allows the comparison between the generated heat predicted by Bernardi’s equation and the HFS measurements. Data post-processing is required to extrapolate the heat generation from the HFS measurements, as the sensor records the heat released to the ambient and not the one generated within the cell. Finally, the cell temperature evolution is estimated with the lumped thermal model (using both HFS and Bernardi’s equation total heat generation) and compared towards experimental temperature data (measured with a T-type thermocouple). At the end of this work, a critical review of the results obtained and the possible mismatch reasons are reported. The results show that indirectly measuring the heat generation with HFS gives a more precise estimation than Bernardi’s simplified equation. On the one hand, when using Bernardi’s simplified equation, estimated heat generation differs from cell temperature measurements during charges at high current rates. Additionally, for low capacity cells where a small change in capacity has a great influence on the terminal voltage, the estimated heat generation shows high dependency on the State of Charge (SoC) estimation, and therefore open circuit voltage calculation (as it is SoC dependent). On the other hand, with indirect measuring the heat generation with HFS, the resulting error is a maximum of 0.28ºC in the temperature prediction, in contrast with 1.38ºC with Bernardi’s simplified equation. This illustrates the limitations of Bernardi’s simplified equation for applications where precise heat monitoring is required. For higher current rates, Bernardi’s equation estimates more heat generation and consequently, a higher predicted temperature. Bernardi´s equation accounts for no losses after cutting the charging or discharging current. However, HFS measurement shows that after cutting the current the cell continues generating heat for some time, increasing the error of Bernardi´s equation.

Keywords: lithium-ion battery, heat flux sensor, heat generation, thermal characterization

Procedia PDF Downloads 382
10657 Vortex Separator for More Accurate Air Dry-Bulb Temperature Measurement

Authors: Ahmed N. Shmroukh, I. M. S. Taha, A. M. Abdel-Ghany, M. Attalla

Abstract:

Fog systems application for cooling and humidification is still limited, although these systems require less initial cost compared with that of other cooling systems such as pad-and-fan systems. The undesirable relative humidity and air temperature inside the space which have been cooled or humidified are the main reasons for its limited use, which results from the poor control of fog systems. Any accurate control system essentially needs air dry bulb temperature as an input parameter. Therefore, the air dry-bulb temperature in the space needs to be measured accurately. The Scope of the present work is the separation of the fog droplets from the air in a fogged space to measure the air dry bulb temperature accurately. The separation is to be done in a small device inside which the sensor of the temperature measuring instrument is positioned. Vortex separator will be designed and used. Another reference device will be used for measuring the air temperature without separation. A comparative study will be performed to reach at the best device which leads to the most accurate measurement of air dry bulb temperature. The results showed that the proposed devices improved the measured air dry bulb temperature toward the correct direction over that of the free junction. Vortex device was the best. It respectively increased the temperature measured by the free junction in the range from around 2 to around 6°C for different fog on-off duration.

Keywords: fog systems, measuring air dry bulb temperature, temperature measurement, vortex separator

Procedia PDF Downloads 292
10656 Counting People Utilizing Space-Time Imagery

Authors: Ahmed Elmarhomy, K. Terada

Abstract:

An automated method for counting passerby has been proposed using virtual-vertical measurement lines. Space-time image is representing the human regions which are treated using the segmentation process. Different color space has been used to perform the template matching. A proper template matching has been achieved to determine direction and speed of passing people. Distinguish one or two passersby has been investigated using a correlation between passerby speed and the human-pixel area. Finally, the effectiveness of the presented method has been experimentally verified.

Keywords: counting people, measurement line, space-time image, segmentation, template matching

Procedia PDF Downloads 449
10655 Fatigue Crack Growth Rate Measurement by Means of Classic Method and Acoustic Emission

Authors: V. Mentl, V. Koula, P. Mazal, J. Volák

Abstract:

Nowadays, the acoustic emission is a widely recognized method of material damage investigation, mainly in cases of cracks initiation and growth observation and evaluation. This is highly important in structures, e.g. pressure vessels, large steam turbine rotors etc., applied both in classic and nuclear power plants. Nevertheless, the acoustic emission signals must be correlated with the real crack progress to be able to evaluate the cracks and their growth by this non-destructive technique alone in real situations and to reach reliable results when the assessment of the structures' safety and reliability is performed and also when the remaining lifetime should be evaluated. The main aim of this study was to propose a methodology for evaluation of the early manifestations of the fatigue cracks and their growth and thus to quantify the material damage by acoustic emission parameters. Specimens made of several steels used in the power producing industry were subjected to fatigue loading in the low- and high-cycle regimes. This study presents results of the crack growth rate measurement obtained by the classic compliance change method and the acoustic emission signal analysis. The experiments were realized in cooperation between laboratories of Brno University of Technology and West Bohemia University in Pilsen within the solution of the project of the Czech Ministry of Industry and Commerce: "A diagnostic complex for the detection of pressure media and material defects in pressure components of nuclear and classic power plants" and the project “New Technologies for Mechanical Engineering”.

Keywords: fatigue, crack growth rate, acoustic emission, material damage

Procedia PDF Downloads 368
10654 Evaluating Contextually Targeted Advertising with Attention Measurement

Authors: John Hawkins, Graham Burton

Abstract:

Contextual targeting is a common strategy for advertising that places marketing messages in media locations that are expected to be aligned with the target audience. There are multiple major challenges to contextual targeting: the ideal categorisation scheme needs to be known, as well as the most appropriate subsections of that scheme for a given campaign or creative. In addition, the campaign reach is typically limited when targeting becomes narrow, so a balance must be struck between requirements. Finally, refinement of the process is limited by the use of evaluation methods that are either rapid but non-specific (click through rates), or reliable but slow and costly (conversions or brand recall studies). In this study we evaluate the use of attention measurement as a technique for understanding the performance of targeting on the basis of specific contextual topics. We perform the analysis using a large scale dataset of impressions categorised using the iAB V2.0 taxonomy. We evaluate multiple levels of the categorisation hierarchy, using categories at different positions within an initial creative specific ranking. The results illustrate that measuring attention time is an affective signal for the performance of a specific creative within a specific context. Performance is sustained across a ranking of categories from one period to another.

Keywords: contextual targeting, digital advertising, attention measurement, marketing performance

Procedia PDF Downloads 103
10653 Optimizing Machine Vision System Setup Accuracy by Six-Sigma DMAIC Approach

Authors: Joseph C. Chen

Abstract:

Machine vision system provides automatic inspection to reduce manufacturing costs considerably. However, only a few principles have been found to optimize machine vision system and help it function more accurately in industrial practice. Mostly, there were complicated and impractical design techniques to improve the accuracy of machine vision system. This paper discusses implementing the Six Sigma Define, Measure, Analyze, Improve, and Control (DMAIC) approach to optimize the setup parameters of machine vision system when it is used as a direct measurement technique. This research follows a case study showing how Six Sigma DMAIC methodology has been put into use.

Keywords: DMAIC, machine vision system, process capability, Taguchi Parameter Design

Procedia PDF Downloads 432
10652 Spin Coherent States Without Squeezing

Authors: A. Dehghani, S. Shirin

Abstract:

We propose in this article a new configuration of quantum states, |α, β> := |α>×|β>. Which are composed of vector products of two different copies of spin coherent states, |α> and |β>. Some mathematical as well as physical properties of such states are discussed. For instance, it has been shown that the cross products of two coherent vectors remain coherent again. They admit a resolution of the identity through positive definite measures on the complex plane. They represent packets similar to the true coherent states, in other words we would not expect to take spin squeezing in any of the field quadratures Lˆx, Lˆy and Lˆz. Depending on the particular choice of parameters in the above scenarios, they can be converted into the so-called Dicke states which minimize the uncertainty relations of each pair of the angular momentum components.

Keywords: vector (Cross-)products, minimum uncertainty, angular momentum, measurement, Dicke states

Procedia PDF Downloads 409
10651 Automatic Censoring in K-Distribution for Multiple Targets Situations

Authors: Naime Boudemagh, Zoheir Hammoudi

Abstract:

The parameters estimation of the K-distribution is an essential part in radar detection. In fact, presence of interfering targets in reference cells causes a decrease in detection performances. In such situation, the estimate of the shape and the scale parameters are far from the actual values. In the order to avoid interfering targets, we propose an Automatic Censoring (AC) algorithm of radar interfering targets in K-distribution. The censoring technique used in this work offers a good discrimination between homogeneous and non-homogeneous environments. The homogeneous population is then used to estimate the unknown parameters by the classical Method of Moment (MOM). The AC algorithm does not need any prior information about the clutter parameters nor does it require both the number and the position of interfering targets. The accuracy of the estimation parameters obtained by this algorithm are validated and compared to various actual values of the shape parameter, using Monte Carlo simulations, this latter show that the probability of censing in multiple target situations are in good agreement.

Keywords: parameters estimation, method of moments, automatic censoring, K distribution

Procedia PDF Downloads 369
10650 Development of an Atmospheric Radioxenon Detection System for Nuclear Explosion Monitoring

Authors: V. Thomas, O. Delaune, W. Hennig, S. Hoover

Abstract:

Measurement of radioactive isotopes of atmospheric xenon is used to detect, locate and identify any confined nuclear tests as part of the Comprehensive Nuclear Test-Ban Treaty (CTBT). In this context, the Alternative Energies and French Atomic Energy Commission (CEA) has developed a fixed device to continuously measure the concentration of these fission products, the SPALAX process. During its atmospheric transport, the radioactive xenon will undergo a significant dilution between the source point and the measurement station. Regarding the distance between fixed stations located all over the globe, the typical volume activities measured are near 1 mBq m⁻³. To avoid the constraints induced by atmospheric dilution, the development of a mobile detection system is in progress; this system will allow on-site measurements in order to confirm or infringe a suspicious measurement detected by a fixed station. Furthermore, this system will use beta/gamma coincidence measurement technique in order to drastically reduce environmental background (which masks such activities). The detector prototype consists of a gas cell surrounded by two large silicon wafers, coupled with two square NaI(Tl) detectors. The gas cell has a sample volume of 30 cm³ and the silicon wafers are 500 µm thick with an active surface area of 3600 mm². In order to minimize leakage current, each wafer has been segmented into four independent silicon pixels. This cell is sandwiched between two low background NaI(Tl) detectors (70x70x40 mm³ crystal). The expected Minimal Detectable Concentration (MDC) for each radio-xenon is in the order of 1-10 mBq m⁻³. Three 4-channels digital acquisition modules (Pixie-NET) are used to process all the signals. Time synchronization is ensured by a dedicated PTP-network, using the IEEE 1588 Precision Time Protocol. We would like to present this system from its simulation to the laboratory tests.

Keywords: beta/gamma coincidence technique, low level measurement, radioxenon, silicon pixels

Procedia PDF Downloads 121
10649 Implementation of Clinical Monitoring System of Physiological Parameters

Authors: Abdesselam Babouri, Ahcène Lemzadmi, M Rahmane, B. Belhadi, N. Abouchi

Abstract:

Medical monitoring aims at monitoring and remotely controlling the vital physiological parameters of the patient. The physiological sensors provide repetitive measurements of these parameters in the form of electrical signals that vary continuously over time. Various measures allow informing us about the health of the person's physiological data (weight, blood pressure, heart rate or specific to a disease), environmental conditions (temperature, humidity, light, noise level) and displacement and movements (physical efforts and the completion of major daily living activities). The collected data will allow monitoring the patient’s condition and alerting in case of modification. They are also used in the diagnosis and decision making on medical treatment and the health of the patient. This work presents the implementation of a monitoring system to be used for the control of physiological parameters.

Keywords: clinical monitoring, physiological parameters, biomedical sensors, personal health

Procedia PDF Downloads 470
10648 Unknown Groundwater Pollution Source Characterization in Contaminated Mine Sites Using Optimal Monitoring Network Design

Authors: H. K. Esfahani, B. Datta

Abstract:

Groundwater is one of the most important natural resources in many parts of the world; however it is widely polluted due to human activities. Currently, effective and reliable groundwater management and remediation strategies are obtained using characterization of groundwater pollution sources, where the measured data in monitoring locations are utilized to estimate the unknown pollutant source location and magnitude. However, accurately identifying characteristics of contaminant sources is a challenging task due to uncertainties in terms of predicting source flux injection, hydro-geological and geo-chemical parameters, and the concentration field measurement. Reactive transport of chemical species in contaminated groundwater systems, especially with multiple species, is a complex and highly non-linear geochemical process. Although sufficient concentration measurement data is essential to accurately identify sources characteristics, available data are often sparse and limited in quantity. Therefore, this inverse problem-solving method for characterizing unknown groundwater pollution sources is often considered ill-posed, complex and non- unique. Different methods have been utilized to identify pollution sources; however, the linked simulation-optimization approach is one effective method to obtain acceptable results under uncertainties in complex real life scenarios. With this approach, the numerical flow and contaminant transport simulation models are externally linked to an optimization algorithm, with the objective of minimizing the difference between measured concentration and estimated pollutant concentration at observation locations. Concentration measurement data are very important to accurately estimate pollution source properties; therefore, optimal design of the monitoring network is essential to gather adequate measured data at desired times and locations. Due to budget and physical restrictions, an efficient and effective approach for groundwater pollutant source characterization is to design an optimal monitoring network, especially when only inadequate and arbitrary concentration measurement data are initially available. In this approach, preliminary concentration observation data are utilized for preliminary source location, magnitude and duration of source activity identification, and these results are utilized for monitoring network design. Further, feedback information from the monitoring network is used as inputs for sequential monitoring network design, to improve the identification of unknown source characteristics. To design an effective monitoring network of observation wells, optimization and interpolation techniques are used. A simulation model should be utilized to accurately describe the aquifer properties in terms of hydro-geochemical parameters and boundary conditions. However, the simulation of the transport processes becomes complex when the pollutants are chemically reactive. Three dimensional transient flow and reactive contaminant transport process is considered. The proposed methodology uses HYDROGEOCHEM 5.0 (HGCH) as the simulation model for flow and transport processes with chemically multiple reactive species. Adaptive Simulated Annealing (ASA) is used as optimization algorithm in linked simulation-optimization methodology to identify the unknown source characteristics. Therefore, the aim of the present study is to develop a methodology to optimally design an effective monitoring network for pollution source characterization with reactive species in polluted aquifers. The performance of the developed methodology will be evaluated for an illustrative polluted aquifer sites, for example an abandoned mine site in Queensland, Australia.

Keywords: monitoring network design, source characterization, chemical reactive transport process, contaminated mine site

Procedia PDF Downloads 228
10647 Analysis of the IEEE 802.15.4 MAC Parameters to Achive Lower Packet Loss Rates

Authors: Imen Bouazzi

Abstract:

The IEEE-802.15.4 standard utilizes the CSMA-CA mechanism to control nodes access to the shared wireless communication medium. It is becoming the popular choice for various applications of surveillance and control used in wireless sensor network (WSN). The benefit of this standard is evaluated regarding of the packet loss probability who depends on the configuration of IEEE 802.15.4 MAC parameters and the traffic load. Our exigency is to evaluate the effects of various configurable MAC parameters on the performance of beaconless IEEE 802.15.4 networks under different traffic loads, static values of IEEE 802.15.4 MAC parameters (macMinBE, macMaxCSMABackoffs, and macMaxFrame Retries) will be evaluated. To performance analysis, we use ns-2[2] network simulator.

Keywords: WSN, packet loss, CSMA/CA, IEEE-802.15.4

Procedia PDF Downloads 336
10646 Shariah Guideline on Value-Based Intermediation Implementation in the Light of Maqasid Shariah Analysis

Authors: Muhammad Izzam Bin Mohd Khazar, Ruqayyah Binti Mohamad Ali, Nurul Atiqah Binti Yusri

Abstract:

Value-based intermediation (VBI) has been introduced by Bank Negara Malaysia (BNM) as the next strategic direction and growth driver for Islamic banking institutions. The aim of VBI is to deliver the intended outcome of Shariah through practices, conducts, and offerings that generate positive and sustainable impact to the economy, community and environment which is aligned to Maqasid Shariah in preserving the common interest of society by preventing harm and maximizing benefit. Hence, upon its implementation, VBI will experiment the current Shariah compliance treatment and revolutionise new policies and systems that can meritoriously entrench and convey the objectives of Shariah. However, discussion revolving VBI in the light of Maqasid analysis is still scarce hence further research needs to be undertaken. The idea of implementation of VBI vision into quantifiable Maqasid Shariah measurement is yet to be explored due to the nature of Maqasid that is variable. The contemporary scholars also have different views on the implementation of VBI. This paper aims to discuss on the importance of Maqasid Shariah in the current Islamic finance transactions by providing Shariah index measurement in the application of VBI. This study also intends to explore basic Shariah guidelines and parameters based on the objectives of Shariah; preservation of the five pillars (religion, life, progeny, intellect and wealth) with further elaboration on preservation of wealth under five headings: rawaj (circulation and marketability); wuduh (transparency); hifz (preservation); thabat (durability and tranquillity); and ‘adl (equity and justice). In alignment with these headings, Islamic finance can be innovated for VBI implementation, particularly in Maybank Islamic being a significant leader in the IFI market.

Keywords: Islamic Financial Institutions, Maqasid Index, Maqasid Shariah, sustainability, value-based intermediation

Procedia PDF Downloads 161