Search results for: covariance matrix estimation
3495 Heavy Metals Estimation in Coastal Areas Using Remote Sensing, Field Sampling and Classical and Robust Statistic
Authors: Elena Castillo-López, Raúl Pereda, Julio Manuel de Luis, Rubén Pérez, Felipe Piña
Abstract:
Sediments are an important source of accumulation of toxic contaminants within the aquatic environment. Bioassays are a powerful tool for the study of sediments in relation to their toxicity, but they can be expensive. This article presents a methodology to estimate the main physical property of intertidal sediments in coastal zones: heavy metals concentration. This study, which was developed in the Bay of Santander (Spain), applies classical and robust statistic to CASI-2 hyperspectral images to estimate heavy metals presence and ecotoxicity (TOC). Simultaneous fieldwork (radiometric and chemical sampling) allowed an appropriate atmospheric correction to CASI-2 images.Keywords: remote sensing, intertidal sediment, airborne sensors, heavy metals, eTOCoxicity, robust statistic, estimation
Procedia PDF Downloads 4193494 Health as a Proxy for Labour Productivity: The Impact on Wages in Egypt’s Private Sector
Authors: Yasmine Ahmed Shemeis
Abstract:
Determining the impact of productivity increases on wage levels is often difficult due to the unavailability of individual-level productivity data. Accordingly, we proxy for productivity using a self-perceived measure of health based on the postulated positive relationship between better health and productivity improvements. Using Egypt’s labour market data for the years 2012 and 2018 and utilizing a Maximum Likelihood Estimation method, we address two issues: the endogeneity of health in the estimation of wages and a sample selection bias. Our findings indicate the great value that better health has in enhancing wage levels in Egypt’s private sector. Also, we find that overlooking the endogeneity of health underestimates its effect on wages. Thus, the improvement of health states is likely to be beneficial in improving labour market outcomes in terms of wages as well as labour productivity in Egypt.Keywords: labour, Productivity, Wages, Endogeneity, Sample Selection
Procedia PDF Downloads 803493 Fatigue Life Estimation of Spiral Welded Waterworks Pipelines
Authors: Suk Woo Hong, Chang Sung Seok, Jae Mean Koo
Abstract:
Recently, the welding is widely used in modern industry for joining the structures. However, the waterworks pipes are exposed to the fatigue load by cars, earthquake and etc because of being buried underground. Moreover, the residual stress exists in weld zone by welding process and it is well known that the fatigue life of welded structures is degraded by residual stress. Due to such reasons, the crack can occur in the weld zone of pipeline. In this case, The ground subsidence or sinkhole can occur, if the soil and sand are washed down by fluid leaked from the crack of water pipe. These problems can lead to property damage and endangering lives. For these reasons, the estimation of fatigue characteristics for water pipeline weld zone is needed. Therefore, in this study, for fatigue characteristics estimation of spiral welded waterworks pipe, ASTM standard specimens and Curved Plate specimens were collected from the spiral welded waterworks pipe and the fatigue tests were performed. The S-N curves of each specimen were estimated, and then the fatigue life of weldment Curved Plate specimen was predicted by theoretical and analytical methods. After that, the weldment Curved Plate specimens were collected from the pipe and verification fatigue tests were performed. Finally, it was verified that the predicted S-N curve of weldment Curved Plate specimen was good agreement with fatigue test data.Keywords: spiral welded pipe, prediction fatigue life, endurance limit modifying factors, residual stress
Procedia PDF Downloads 2973492 Understanding the Classification of Rain Microstructure and Estimation of Z-R Relationship using a Micro Rain Radar in Tropical Region
Authors: Tomiwa, Akinyemi Clement
Abstract:
Tropical regions experience diverse and complex precipitation patterns, posing significant challenges for accurate rainfall estimation and forecasting. This study addresses the problem of effectively classifying tropical rain types and refining the Z-R (Reflectivity-Rain Rate) relationship to enhance rainfall estimation accuracy. Through a combination of remote sensing, meteorological analysis, and machine learning, the research aims to develop an advanced classification framework capable of distinguishing between different types of tropical rain based on their unique characteristics. This involves utilizing high-resolution satellite imagery, radar data, and atmospheric parameters to categorize precipitation events into distinct classes, providing a comprehensive understanding of tropical rain systems. Additionally, the study seeks to improve the Z-R relationship, a crucial aspect of rainfall estimation. One year of rainfall data was analyzed using a Micro Rain Radar (MRR) located at The Federal University of Technology Akure, Nigeria, measuring rainfall parameters from ground level to a height of 4.8 km with a vertical resolution of 0.16 km. Rain rates were classified into low (stratiform) and high (convective) based on various microstructural attributes such as rain rates, liquid water content, Drop Size Distribution (DSD), average fall speed of the drops, and radar reflectivity. By integrating diverse datasets and employing advanced statistical techniques, the study aims to enhance the precision of Z-R models, offering a more reliable means of estimating rainfall rates from radar reflectivity data. This refined Z-R relationship holds significant potential for improving our understanding of tropical rain systems and enhancing forecasting accuracy in regions prone to heavy precipitation.Keywords: remote sensing, precipitation, drop size distribution, micro rain radar
Procedia PDF Downloads 323491 Image-Based UAV Vertical Distance and Velocity Estimation Algorithm during the Vertical Landing Phase Using Low-Resolution Images
Authors: Seyed-Yaser Nabavi-Chashmi, Davood Asadi, Karim Ahmadi, Eren Demir
Abstract:
The landing phase of a UAV is very critical as there are many uncertainties in this phase, which can easily entail a hard landing or even a crash. In this paper, the estimation of relative distance and velocity to the ground, as one of the most important processes during the landing phase, is studied. Using accurate measurement sensors as an alternative approach can be very expensive for sensors like LIDAR, or with a limited operational range, for sensors like ultrasonic sensors. Additionally, absolute positioning systems like GPS or IMU cannot provide distance to the ground independently. The focus of this paper is to determine whether we can measure the relative distance and velocity of UAV and ground in the landing phase using just low-resolution images taken by a monocular camera. The Lucas-Konda feature detection technique is employed to extract the most suitable feature in a series of images taken during the UAV landing. Two different approaches based on Extended Kalman Filters (EKF) have been proposed, and their performance in estimation of the relative distance and velocity are compared. The first approach uses the kinematics of the UAV as the process and the calculated optical flow as the measurement; On the other hand, the second approach uses the feature’s projection on the camera plane (pixel position) as the measurement while employing both the kinematics of the UAV and the dynamics of variation of projected point as the process to estimate both relative distance and relative velocity. To verify the results, a sequence of low-quality images taken by a camera that is moving on a specifically developed testbed has been used to compare the performance of the proposed algorithm. The case studies show that the quality of images results in considerable noise, which reduces the performance of the first approach. On the other hand, using the projected feature position is much less sensitive to the noise and estimates the distance and velocity with relatively high accuracy. This approach also can be used to predict the future projected feature position, which can drastically decrease the computational workload, as an important criterion for real-time applications.Keywords: altitude estimation, drone, image processing, trajectory planning
Procedia PDF Downloads 1093490 Non-Linear Regression Modeling for Composite Distributions
Authors: Mostafa Aminzadeh, Min Deng
Abstract:
Modeling loss data is an important part of actuarial science. Actuaries use models to predict future losses and manage financial risk, which can be beneficial for marketing purposes. In the insurance industry, small claims happen frequently while large claims are rare. Traditional distributions such as Normal, Exponential, and inverse-Gaussian are not suitable for describing insurance data, which often show skewness and fat tails. Several authors have studied classical and Bayesian inference for parameters of composite distributions, such as Exponential-Pareto, Weibull-Pareto, and Inverse Gamma-Pareto. These models separate small to moderate losses from large losses using a threshold parameter. This research introduces a computational approach using a nonlinear regression model for loss data that relies on multiple predictors. Simulation studies were conducted to assess the accuracy of the proposed estimation method. The simulations confirmed that the proposed method provides precise estimates for regression parameters. It's important to note that this approach can be applied to datasets if goodness-of-fit tests confirm that the composite distribution under study fits the data well. To demonstrate the computations, a real data set from the insurance industry is analyzed. A Mathematica code uses the Fisher information algorithm as an iteration method to obtain the maximum likelihood estimation (MLE) of regression parameters.Keywords: maximum likelihood estimation, fisher scoring method, non-linear regression models, composite distributions
Procedia PDF Downloads 313489 Ti-Mo-N Nano-Grains Embedded into Thin MoSₓ-Based Amorphous Matrix: A Novel Structure for Superhardness and Ultra-Low Wear
Authors: Lina Yang, Mao Wen, Jianhong Chen, Kan Zhang
Abstract:
Molybdenum disulfide (MoS₂) represents a highly sought lubricant for reducing friction based on intrinsic layered structure, but for this reason, practical applications have been greatly restricted due to the fact that its low hardness would cause severe wear. Here, a novel TiMoN/MoSₓ composite coatings with TiMoN solid solution grains embedded into MoSₓ-based amorphous matrix has been successfully designed and synthesized, through magnetron co-sputtering technology. Desirably, in virtue of such special microstructure, superhardness and excellent toughness can be well achieved, along with an ultra-low wear rate at ~2×10⁻¹¹ mm³/Nm in the air environment, simultaneously, low friction at ~0.1 is maintained. It should be noted that this wear level is almost two orders of magnitude lower than that of pure TiN coating, and is, as we know, the lowest wear rate in dry sliding. Investigations of tribofilm reveal that it is amorphous MoS₂ in nature, and its formation arises directly from the MoSₓ amorphous matrix. Which contributes to effective lubrication behavior, coupled with excellent mechanical performances of such composite coating, exceptionally low wear can be guaranteed. The findings in this work suggest that the special composite structure makes it possible for the synthesis of super-hard and super-durable lubricative coating, offering guidance to synthesize ultrahigh performance protective coating for industrial application.Keywords: hardness, MoS₂-containing composite coatings, toughness, tribological properties
Procedia PDF Downloads 1503488 Investigation of Minor Actinide-Contained Thorium Fuel Impacts on CANDU-Type Reactor Neutronics Using Computational Method
Authors: S. A. H. Feghhi, Z. Gholamzadeh, Z. Alipoor, C. Tenreiro
Abstract:
Currently, thorium fuel has been especially noticed because of its proliferation resistance than long half-life alpha emitter minor actinides, breeding capability in fast and thermal neutron flux and mono-isotopic naturally abundant. In recent years, efficiency of minor actinide burning up in PWRs has been investigated. Hence, a minor actinide-contained thorium based fuel matrix can confront both proliferation resistance and nuclear waste depletion aims. In the present work, minor actinide depletion rate in a CANDU-type nuclear core modeled using MCNP code has been investigated. The obtained effects of minor actinide load as mixture of thorium fuel matrix on the core neutronics has been studiedwith comparingpresence and non-presence of minor actinide component in the fuel matrix.Depletion rate of minor actinides in the MA-contained fuel has been calculated using different power loads.According to the obtained computational data, minor actinide loading in the modeled core results in more negative reactivity coefficients. The MA-contained fuel achieves less radial peaking factor in the modeled core. The obtained computational results showed 140 kg of 464 kg initial load of minor actinide has been depleted in during a 6-year burn up in 10 MW power.Keywords: minor actinide burning, CANDU-type reactor, MCNPX code, neutronic parameters
Procedia PDF Downloads 4563487 A Semiparametric Approach to Estimate the Mode of Continuous Multivariate Data
Authors: Tiee-Jian Wu, Chih-Yuan Hsu
Abstract:
Mode estimation is an important task, because it has applications to data from a wide variety of sources. We propose a semi-parametric approach to estimate the mode of an unknown continuous multivariate density function. Our approach is based on a weighted average of a parametric density estimate using the Box-Cox transform and a non-parametric kernel density estimate. Our semi-parametric mode estimate improves both the parametric- and non-parametric- mode estimates. Specifically, our mode estimate solves the non-consistency problem of parametric mode estimates (at large sample sizes) and reduces the variability of non-parametric mode estimates (at small sample sizes). The performance of our method at practical sample sizes is demonstrated by simulation examples and two real examples from the fields of climatology and image recognition.Keywords: Box-Cox transform, density estimation, mode seeking, semiparametric method
Procedia PDF Downloads 2833486 Enhanced Thermal Properties of Rigid PVC Foams Using Fly Ash
Authors: Nidal H. Abu-Zahra, Parisa Khoshnoud, Murtatha Jamel, Subhashini Gunashekar
Abstract:
PVC foam-fly ash composites (PVC-FA) are characterized for their structural, morphological, mechanical and thermal properties. The tensile strength of the composites increased modestly with higher fly ash loading, while there was a significant increase in the elastic modulus for the same composites. On the other hand, a decrease in elongation at UTS was observed upon increasing fly ash content due to increased rigidity of the composites. Similarly, the flexural modulus increased as the fly ash loading increased, where the composites containing 25 phr fly ash showed the highest flexural strength. Thermal properties of PVC-fly ash composites were determined by Thermo Gravimetric Analysis (TGA). The micro structural properties were studied by Scanning Electron Microscopy (SEM). SEM results confirm that fly ash particles were mechanically interlocked in PVC matrix with good inter facial interaction with the matrix. Particle agglomeration and debonding was observed in samples containing higher amounts of fly ash.Keywords: PVC foam, polyvinyl chloride, rigid PVC, fly ash composites, polymer composites
Procedia PDF Downloads 3863485 Lipschitz Classifiers Ensembles: Usage for Classification of Target Events in C-OTDR Monitoring Systems
Authors: Andrey V. Timofeev
Abstract:
This paper introduces an original method for guaranteed estimation of the accuracy of an ensemble of Lipschitz classifiers. The solution was obtained as a finite closed set of alternative hypotheses, which contains an object of classification with a probability of not less than the specified value. Thus, the classification is represented by a set of hypothetical classes. In this case, the smaller the cardinality of the discrete set of hypothetical classes is, the higher is the classification accuracy. Experiments have shown that if the cardinality of the classifiers ensemble is increased then the cardinality of this set of hypothetical classes is reduced. The problem of the guaranteed estimation of the accuracy of an ensemble of Lipschitz classifiers is relevant in the multichannel classification of target events in C-OTDR monitoring systems. Results of suggested approach practical usage to accuracy control in C-OTDR monitoring systems are present.Keywords: Lipschitz classifiers, confidence set, C-OTDR monitoring, classifiers accuracy, classifiers ensemble
Procedia PDF Downloads 4903484 Nafion Nanofiber Composite Membrane Fabrication for Fuel Cell Applications
Authors: C. N. Okafor, M. Maaza, T. A. E. Mokrani
Abstract:
A proton exchange membrane has been developed for Direct Methanol Fuel Cell (DMFC). The nanofiber network composite membranes were prepared by interconnected network of Nafion (perfuorosulfonic acid) nanofibers that have been embedded in an uncharged and inert polymer matrix, by electro-spinning. The spinning solution of Nafion with a low concentration (1 wt. % compared to Nafion) of high molecular weight poly(ethylene oxide), as a carrier polymer. The interconnected network of Nafion nanofibers with average fiber diameter in the range of 160-700nm, were used to make the membranes, with the nanofiber occupying up to 85% of the membrane volume. The matrix polymer was cross-linked with Norland Optical Adhesive 63 under UV. The resulting membranes showed proton conductivity of 0.10 S/cm at 25°C and 80% RH; and methanol permeability of 3.6 x 10-6 cm2/s.Keywords: composite membrane, electrospinning, fuel cell, nanofibers
Procedia PDF Downloads 2643483 GIS Application in Surface Runoff Estimation for Upper Klang River Basin, Malaysia
Authors: Suzana Ramli, Wardah Tahir
Abstract:
Estimation of surface runoff depth is a vital part in any rainfall-runoff modeling. It leads to stream flow calculation and later predicts flood occurrences. GIS (Geographic Information System) is an advanced and opposite tool used in simulating hydrological model due to its realistic application on topography. The paper discusses on calculation of surface runoff depth for two selected events by using GIS with Curve Number method for Upper Klang River basin. GIS enables maps intersection between soil type and land use that later produces curve number map. The results show good correlation between simulated and observed values with more than 0.7 of R2. Acceptable performance of statistical measurements namely mean error, absolute mean error, RMSE, and bias are also deduced in the paper.Keywords: surface runoff, geographic information system, curve number method, environment
Procedia PDF Downloads 2793482 Evaluation and Selection of Drilling Technologies: An Application of Portfolio Analysis Matrix in South Azadgan Oilfield
Authors: M. Maleki Sadabad, A. Pointing, N. Marashi
Abstract:
With respect to the role and increasing importance of technology for countries development, in recent decades technology development has paid attention in a systematic form. Nowadays the markets face with highly complicated and competitive conditions in foreign markets, therefore, evaluation and selection of technology effectiveness and also formulating technology strategy have changed into a vital subject for some organizations. The study introduces the standards of empowerment evaluation and technology attractiveness especially strategic technologies which explain the way of technology evaluation, selection and finally formulating suitable technology strategy in the field of drilling in South Azadegan oil field. The study firstly identifies the key challenges of oil fields in order to evaluate the technologies in field of drilling in South Azadegan oil field through an interview with the experts of industry and then they have been prioritised. In the following, the existing and new technologies were identified to solve the challenges of South Azadegan oil field. In order to explore the ability, availability, and attractiveness of every technology, a questionnaire based on Julie indices has been designed and distributed among the industry elites. After determining the score of ability, availability and attractiveness, every technology which has been obtained by the average of expert’s ideas, the technology package has been introduced by Morin’s model. The matrix includes four areas which will follow the especial strategy. Finally, by analysing the above matrix, the technology options have been suggested in order to select and invest.Keywords: technology, technology identification, drilling technologies, technology capability
Procedia PDF Downloads 1413481 Nanomechanical Characterization of Healthy and Tumor Lung Tissues at Cell and Extracellular Matrix Level
Authors: Valeria Panzetta, Ida Musella, Sabato Fusco, Paolo Antonio Netti
Abstract:
The study of the biophysics of living cells drew attention to the pivotal role of the cytoskeleton in many cell functions, such as mechanics, adhesion, proliferation, migration, differentiation and neoplastic transformation. In particular, during the complex process of malignant transformation and invasion cell cytoskeleton devolves from a rigid and organized structure to a more compliant state, which confers to the cancer cells a great ability to migrate and adapt to the extracellular environment. In order to better understand the malignant transformation process from a mechanical point of view, it is necessary to evaluate the direct crosstalk between the cells and their surrounding extracellular matrix (ECM) in a context which is close to in vivo conditions. In this study, human biopsy tissues of lung adenocarcinoma were analyzed in order to define their mechanical phenotype at cell and ECM level, by using particle tracking microrheology (PTM) technique. Polystyrene beads (500 nm) were introduced into the sample slice. The motion of beads was obtained by tracking their displacements across cell cytoskeleton and ECM structures and mean squared displacements (MSDs) were calculated from bead trajectories. It has been already demonstrated that the amplitude of MSD is inversely related to the mechanical properties of intracellular and extracellular microenvironment. For this reason, MSDs of particles introduced in cytoplasm and ECM of healthy and tumor tissues were compared. PTM analyses showed that cancerous transformation compromises mechanical integrity of cells and extracellular matrix. In particular, the MSD amplitudes in cells of adenocarcinoma were greater as compared to cells of normal tissues. The increased motion is probably associated to a less structured cytoskeleton and consequently to an increase of deformability of cells. Further, cancer transformation is also accompanied by extracellular matrix stiffening, as confirmed by the decrease of MSDs of matrix in tumor tissue, a process that promotes tumor proliferation and invasiveness, by activating typical oncogenic signaling pathways. In addition, a clear correlation between MSDs of cells and tumor grade was found. MSDs increase when tumor grade passes from 2 to 3, indicating that cells undergo to a trans-differentiation process during tumor progression. ECM stiffening is not dependent on tumor grade, but the tumor stage resulted to be strictly correlated with both cells and ECM mechanical properties. In fact, a greater stage is assigned to tumor spread to regional lymph nodes and characterized by an up-regulation of different ECM proteins, such as collagen I fibers. These results indicate that PTM can be used to get nanomechanical characterization at different scale levels in an interpretative and diagnostic context.Keywords: cytoskeleton, extracellular matrix, mechanical properties, particle tracking microrheology, tumor
Procedia PDF Downloads 2783480 Nonparametric Sieve Estimation with Dependent Data: Application to Deep Neural Networks
Authors: Chad Brown
Abstract:
This paper establishes general conditions for the convergence rates of nonparametric sieve estimators with dependent data. We present two key results: one for nonstationary data and another for stationary mixing data. Previous theoretical results often lack practical applicability to deep neural networks (DNNs). Using these conditions, we derive convergence rates for DNN sieve estimators in nonparametric regression settings with both nonstationary and stationary mixing data. The DNN architectures considered adhere to current industry standards, featuring fully connected feedforward networks with rectified linear unit activation functions, unbounded weights, and a width and depth that grows with sample size.Keywords: sieve extremum estimates, nonparametric estimation, deep learning, neural networks, rectified linear unit, nonstationary processes
Procedia PDF Downloads 403479 Time of Week Intensity Estimation from Interval Censored Data with Application to Police Patrol Planning
Authors: Jiahao Tian, Michael D. Porter
Abstract:
Law enforcement agencies are tasked with crime prevention and crime reduction under limited resources. Having an accurate temporal estimate of the crime rate would be valuable to achieve such a goal. However, estimation is usually complicated by the interval-censored nature of crime data. We cast the problem of intensity estimation as a Poisson regression using an EM algorithm to estimate the parameters. Two special penalties are added that provide smoothness over the time of day and day of the week. This approach presented here provides accurate intensity estimates and can also uncover day-of-week clusters that share the same intensity patterns. Anticipating where and when crimes might occur is a key element to successful policing strategies. However, this task is complicated by the presence of interval-censored data. The censored data refers to the type of data that the event time is only known to lie within an interval instead of being observed exactly. This type of data is prevailing in the field of criminology because of the absence of victims for certain types of crime. Despite its importance, the research in temporal analysis of crime has lagged behind the spatial component. Inspired by the success of solving crime-related problems with a statistical approach, we propose a statistical model for the temporal intensity estimation of crime with censored data. The model is built on Poisson regression and has special penalty terms added to the likelihood. An EM algorithm was derived to obtain maximum likelihood estimates, and the resulting model shows superior performance to the competing model. Our research is in line with the smart policing initiative (SPI) proposed by the Bureau Justice of Assistance (BJA) as an effort to support law enforcement agencies in building evidence-based, data-driven law enforcement tactics. The goal is to identify strategic approaches that are effective in crime prevention and reduction. In our case, we allow agencies to deploy their resources for a relatively short period of time to achieve the maximum level of crime reduction. By analyzing a particular area within cities where data are available, our proposed approach could not only provide an accurate estimate of intensities for the time unit considered but a time-variation crime incidence pattern. Both will be helpful in the allocation of limited resources by either improving the existing patrol plan with the understanding of the discovery of the day of week cluster or supporting extra resources available.Keywords: cluster detection, EM algorithm, interval censoring, intensity estimation
Procedia PDF Downloads 643478 Tensile and Bond Characterization of Basalt-Fabric Reinforced Alkali Activated Matrix
Authors: S. Candamano, A. Iorfida, F. Crea, A. Macario
Abstract:
Recently, basalt fabric reinforced cementitious composites (FRCM) have attracted great attention because they result to be effective in structural strengthening and cost/environment efficient. In this study, authors investigate their mechanical behavior when an inorganic matrix, belonging to the family of alkali-activated binders, is used. In particular, the matrix has been designed to contain high amounts of industrial by-products and waste, such as Ground Granulated Blast Furnace Slag (GGBFS) and Fly Ash. Fresh state properties, such as workability, mechanical properties and shrinkage behavior of the matrix have been measured, while microstructures and reaction products were analyzed by Scanning Electron Microscopy and X-Ray Diffractometry. Reinforcement is made up of a balanced, coated bidirectional fabric made out of basalt fibres and stainless steel micro-wire, with a mesh size of 8x8 mm and an equivalent design thickness equal to 0.064 mm. Mortars mixes have been prepared by maintaining constant the water/(reactive powders) and sand/(reactive powders) ratios at 0.53 and 2.7 respectively. An appropriate experimental campaign based on direct tensile tests on composite specimens and single-lap shear bond test on brickwork substrate has been thus carried out to investigate their mechanical behavior under tension, the stress-transfer mechanism and failure modes. Tensile tests were carried out on composite specimens of nominal dimensions equal to 500 mm x 50 mm x 10 mm, with 6 embedded rovings in the loading direction. Direct shear tests (DST) were carried out on brickwork substrate using an externally bonded basalt-FRCM composite strip 10 mm thick, 50 mm wide and a bonded length of 300 mm. Mortars exhibit, after 28 days of curing, an average compressive strength of 32 MPa and flexural strength of 5.5 MPa. Main hydration product is a poorly crystalline aluminium-modified calcium silicate hydrate (C-A-S-H) gel. The constitutive behavior of the composite has been identified by means of direct tensile tests, with response curves showing a tri-linear behavior. Test results indicate that the behavior is mainly governed by cracks development (II) and widening (III) up to failure. The ultimate tensile strength and strain were respectively σᵤ = 456 MPa and ɛᵤ= 2.20%. The tensile modulus of elasticity in stage III was EIII= 41 GPa. All single-lap shear test specimens failed due to composite debonding. It occurred at the internal fabric-to-matrix interface, and it was the result of a fracture of the matrix between the fibre bundles. For all specimens, transversal cracks were visible on the external surface of the composite and involved only the external matrix layer. This cracking appears when the interfacial shear stresses increase and slippage of the fabric at the internal matrix layer interface occurs. Since the external matrix layer is bonded to the reinforcement fabric, it translates with the slipped fabric. Average peak load around 945 N, peak stress around 308 MPa and global slip around 6 mm were measured. The preliminary test results allow affirming that Alkali-Activated Materials can be considered a potentially valid alternative to traditional mortars in designing FRCM composites.Keywords: Alkali-activated binders, Basalt-FRCM composites, direct shear tests, structural strengthening
Procedia PDF Downloads 1293477 A Homogenized Mechanical Model of Carbon Nanotubes/Polymer Composite with Interface Debonding
Authors: Wenya Shu, Ilinca Stanciulescu
Abstract:
Carbon nanotubes (CNTs) possess attractive properties, such as high stiffness and strength, and high thermal and electrical conductivities, making them promising filler in multifunctional nanocomposites. Although CNTs can be efficient reinforcements, the expected level of mechanical performance of CNT-polymers is not often reached in practice due to the poor mechanical behavior of the CNT-polymer interfaces. It is believed that the interactions of CNT and polymer mainly result from the Van der Waals force. The interface debonding is a fracture and delamination phenomenon. Thus, the cohesive zone modeling (CZM) is deemed to give good capture of the interface behavior. The detailed, cohesive zone modeling provides an option to consider the CNT-matrix interactions, but brings difficulties in mesh generation and also leads to high computational costs. Homogenized models that smear the fibers in the ground matrix and treat the material as homogeneous are studied in many researches to simplify simulations. But based on the perfect interface assumption, the traditional homogenized model obtained by mixing rules severely overestimates the stiffness of the composite, even comparing with the result of the CZM with artificially very strong interface. A mechanical model that can take into account the interface debonding and achieve comparable accuracy to the CZM is thus essential. The present study first investigates the CNT-matrix interactions by employing cohesive zone modeling. Three different coupled CZM laws, i.e., bilinear, exponential and polynomial, are considered. These studies indicate that the shapes of the CZM constitutive laws chosen do not influence significantly the simulations of interface debonding. Assuming a bilinear traction-separation relationship, the debonding process of single CNT in the matrix is divided into three phases and described by differential equations. The analytical solutions corresponding to these phases are derived. A homogenized model is then developed by introducing a parameter characterizing interface sliding into the mixing theory. The proposed mechanical model is implemented in FEAP8.5 as a user material. The accuracy and limitations of the model are discussed through several numerical examples. The CZM simulations in this study reveal important factors in the modeling of CNT-matrix interactions. The analytical solutions and proposed homogenized model provide alternative methods to efficiently investigate the mechanical behaviors of CNT/polymer composites.Keywords: carbon nanotube, cohesive zone modeling, homogenized model, interface debonding
Procedia PDF Downloads 1293476 Online Authenticity Verification of a Biometric Signature Using Dynamic Time Warping Method and Neural Networks
Authors: Gałka Aleksandra, Jelińska Justyna, Masiak Albert, Walentukiewicz Krzysztof
Abstract:
An offline signature is well-known however not the safest way to verify identity. Nowadays, to ensure proper authentication, i.e. in banking systems, multimodal verification is more widely used. In this paper the online signature analysis based on dynamic time warping (DTW) coupled with machine learning approaches has been presented. In our research signatures made with biometric pens were gathered. Signature features as well as their forgeries have been described. For verification of authenticity various methods were used including convolutional neural networks using DTW matrix and multilayer perceptron using sums of DTW matrix paths. System efficiency has been evaluated on signatures and signature forgeries collected on the same day. Results are presented and discussed in this paper.Keywords: dynamic time warping, handwritten signature verification, feature-based recognition, online signature
Procedia PDF Downloads 1743475 Improvement of the Q-System Using the Rock Engineering System: A Case Study of Water Conveyor Tunnel of Azad Dam
Authors: Sahand Golmohammadi, Sana Hosseini Shirazi
Abstract:
Because the status and mechanical parameters of discontinuities in the rock mass are included in the calculations, various methods of rock engineering classification are often used as a starting point for the design of different types of structures. The Q-system is one of the most frequently used methods for stability analysis and determination of support systems of underground structures in rock, including tunnel. In this method, six main parameters of the rock mass, namely, the rock quality designation (RQD), joint set number (Jn), joint roughness number (Jr), joint alteration number (Ja), joint water parameter (Jw) and stress reduction factor (SRF) are required. In this regard, in order to achieve a reasonable and optimal design, identifying the effective parameters for the stability of the mentioned structures is one of the most important goals and the most necessary actions in rock engineering. Therefore, it is necessary to study the relationships between the parameters of a system and how they interact with each other and, ultimately, the whole system. In this research, it has attempted to determine the most effective parameters (key parameters) from the six parameters of rock mass in the Q-system using the rock engineering system (RES) method to improve the relationships between the parameters in the calculation of the Q value. The RES system is, in fact, a method by which one can determine the degree of cause and effect of a system's parameters by making an interaction matrix. In this research, the geomechanical data collected from the water conveyor tunnel of Azad Dam were used to make the interaction matrix of the Q-system. For this purpose, instead of using the conventional methods that are always accompanied by defects such as uncertainty, the Q-system interaction matrix is coded using a technique that is actually a statistical analysis of the data and determining the correlation coefficient between them. So, the effect of each parameter on the system is evaluated with greater certainty. The results of this study show that the formed interaction matrix provides a reasonable estimate of the effective parameters in the Q-system. Among the six parameters of the Q-system, the SRF and Jr parameters have the maximum and minimum impact on the system, respectively, and also the RQD and Jw parameters have the maximum and minimum impact on the system, respectively. Therefore, by developing this method, we can obtain a more accurate relation to the rock mass classification by weighting the required parameters in the Q-system.Keywords: Q-system, rock engineering system, statistical analysis, rock mass, tunnel
Procedia PDF Downloads 703474 Implementing Quality Function Deployment Tool for a Customer Driven New Product Development in a Kuwait SME
Authors: Asma AlQahtani, Jumana AlHadad, Maryam AlQallaf, Shoug AlHasan
Abstract:
New product development (NPD) is the complete process of bringing a new product to the customer by integrating the two broad divisions; one involving the idea generation, product design and detail engineering; and the other involving market research and marketing analysis. It is a common practice for companies to undertake some of these tasks simultaneously (concurrent engineering) and also consider them as an ongoing process (continuous development). The current study explores the framework and methodology for a new product development process utilizing the Quality Function Deployment (QFD) tool for bringing the customer opinion into the product development process. An elaborate customer survey with focus groups in the region was carried out to ensure that customer requirements are integrated into new products as early as the design stage including identifying the recognition of need for the new product. A QFD Matrix (House of Quality) was prepared that links customer requirements to product engineering requirements and a feasibility study and risk assessment exercise was carried out for a Small and Medium Enterprise (SME) in Kuwait for development of the new product. SMEs in Kuwait, particularly in manufacturing sector are mainly focused on serving the local demand, and often lack of product quality adversely affects the ability of the companies to compete on a regional/global basis. Further, lack of focus on identifying customer requirements often deters SMEs to envisage the idea of a New Product Development. The current study therefore focuses in utilizing QFD Matrix right from the conceptual design to detail design and to some extent, extending the link this to design of the manufacturing system. The outcome of the project resulted in a development of the prototype for a new molded product which can ensure consistency between the customer’s requirements and the measurable characteristics of the product. The Engineering Economics and Cost studies were also undertaken to analyse the viability of the new product, the results of which was also linked to the successful implementation of the initial QFD Matrix.Keywords: Quality Function Deployment, QFD Matrix, new product development, NPD, Kuwait SMEs, prototype development
Procedia PDF Downloads 4143473 Michel Foucault’s Docile Bodies and The Matrix Trilogy: A Close Reading Applied to the Human Pods and Growing Fields in the Films
Authors: Julian Iliev
Abstract:
The recent release of The Matrix Resurrections persuaded many film scholars that The Matrix trilogy had lost its appeal and its concepts were largely outdated. This study examines the human pods and growing fields in the trilogy. Their functionality is compared to Michel Foucault’s concept of docile bodies: linking fictional and contemporary worlds. This paradigm is scrutinized through surveillance literature. The analogy brings to light common elements of hidden surveillance practices in technologies. The comparison illustrates the effects of body manipulation portrayed in the movies and their relevance with contemporary surveillance practices. Many scholars have utilized a close reading methodology in film studies (J.Bizzocchi, J.Tanenbaum, P.Larsen, S. Herbrechter, and Deacon et al.). The use of a particular lens through which media text is examined is an indispensable factor that needs to be incorporated into the methodology. The study spotlights both scenes from the trilogy depicting the human pods and growing fields. The functionality of the pods and the fields compare directly with Foucault’s concept of docile bodies. By utilizing Foucault’s study as a lens, the research will unearth hidden components and insights into the films. Foucault recognizes three disciplines that produce docile bodies: 1) manipulation and the interchangeability of individual bodies, 2) elimination of unnecessary movements and management of time, and 3) command system guaranteeing constant supervision and continuity protection. These disciplines can be found in the pods and growing fields. Each body occupies a single pod aiding easier manipulation and fast interchangeability. The movement of the bodies in the pods is reduced to the absolute minimum. Thus, the body is transformed into the ultimate object of control – minimum movement correlates to maximum energy generation. Supervision is exercised by wiring the body with numerous types of cables. This ultimate supervision of body activity reduces the body’s purpose to mere functioning. If a body does not function as an energy source, then it’s unplugged, ejected, and liquefied. The command system secures the constant supervision and continuity of the process. To Foucault, the disciplines are distinctly different from slavery because they stop short of a total takeover of the bodies. This is a clear difference from the slave system implemented in the films. Even though their system might lack sophistication, it makes up for it in the elevation of functionality. Further, surveillance literature illustrates the connection between the generation of body energy in The Matrix trilogy to the generation of individual data in contemporary society. This study found that the three disciplines producing docile bodies were present in the portrayal of the pods and fields in The Matrix trilogy. The above comparison combined with surveillance literature yields insights into analogous processes and contemporary surveillance practices. Thus, the constant generation of energy in The Matrix trilogy can be equated to the consistent data generation in contemporary society. This essay shows the relevance of the body manipulation concept in the Matrix films with contemporary surveillance practices.Keywords: docile bodies, film trilogies, matrix movies, michel foucault, privacy loss, surveillance
Procedia PDF Downloads 933472 Derivation of Neutrino Mass Parameters from the Study of Neutrinoless Double Beta Decay
Authors: Sabin Stoica
Abstract:
In this paper the theoretical challenges in the study of neutrinoless double beta decay are reviewed. Then, new upper limits of the neutrino mass parameters in the case of three isotopes are derived; 48Ca, 76Ge, and 82Se, assuming two possible mechanisms of occurrence of this nuclear process, namely the exchange of i) light left-handed neutrinos and ii) heavy right-handed neutrinos, between two nucleons inside the nucleus. The derivation is based on accurate calculations of the phase space factors and nuclear matrix elements performed with new high-performance computer codes, which are described in more detail in recent publications. These results are useful both for a better understanding of the scale of neutrino absolute mass and for the planning of future double beta decay experiments.Keywords: double beta decay, neutrino properties, nuclear matrix elements, phase space factors
Procedia PDF Downloads 5983471 Operational Matrix Method for Fuzzy Fractional Reaction Diffusion Equation
Authors: Sachin Kumar
Abstract:
Fuzzy fractional diffusion equation is widely useful to depict different physical processes arising in physics, biology, and hydrology. The motive of this article is to deal with the fuzzy fractional diffusion equation. We study a mathematical model of fuzzy space-time fractional diffusion equation in which unknown function, coefficients, and initial-boundary conditions are fuzzy numbers. First, we find out a fuzzy operational matrix of Legendre polynomial of Caputo type fuzzy fractional derivative having a non-singular Mittag-Leffler kernel. The main advantages of this method are that it reduces the fuzzy fractional partial differential equation (FFPDE) to a system of fuzzy algebraic equations from which we can find the solution of the problem. The feasibility of our approach is shown by some numerical examples. Hence, our method is suitable to deal with FFPDE and has good accuracy.Keywords: fractional PDE, fuzzy valued function, diffusion equation, Legendre polynomial, spectral method
Procedia PDF Downloads 2003470 Acquisition of Overt Pronoun Constraint in L2 Turkish by Adult Korean Speakers
Authors: Oktay Cinar
Abstract:
The aim of this study is to investigate the acquisition of Overt Pronoun Constraint (OPC) by adult Korean L2 Turkish speakers in order to find out how constraints regulating the syntax of null and overt subjects are acquired. OPC is claimed to be a universal feature of all null subject languages restricting the co-indexation between overt embedded pronoun and quantified or wh-question antecedents. However, there is no such restriction when the embedded subject is null or the antecedent is a referential subject. Considered as a principle of Universal Grammar (UG), OPC knowledge of L2 speakers has been widely tested with different language pairs. In the light of previous studies on OPC, it can be argued that L2 learners display early sensitivity to OPC constraints during their interlanguage grammar development. Concerning this, the co-indexation between overt embedded pronoun o (third person pronoun) and referential matrix subject is claimed to be controversial in Turkish, which poses problems with the universality of OPC. However, the current study argues against this claim by providing evidence from advanced Korean speakers that OPC is universal to all null subject languages and OPC knowledge can be accessed with direct access to UG. In other words, the performances of adult Korean speakers on the syntax of null and overt subjects are tested to support this claim. In order to test this, OPC task is used. 15 advanced speakers and a control group of adult native Turkish participants are instructed to determine the co-reference relationship between the subject of embedded clause, either overt pronominal o or null, and the subject of the matrix clause, either quantified pronoun and wh-question or referential antecedent. They are asked to select the interpretation of the embedded subject, either as the same person as in the matrix subject or another person who is not the same person in the matrix subject. These relations are represented with four conditions, and each condition has four questions (16 questions in total). The results claim that both control group and Korean L2 Turkish speakers display sensitivity to all constraints that OPC has, which suggests that OPC works in Turkish as well.Keywords: adult Korean speakers, binding theory, generative second language acquisition, overt pronoun constraint
Procedia PDF Downloads 3093469 Cycle Number Estimation Method on Fatigue Crack Initiation Using Voronoi Tessellation and the Tanaka Mura Model
Authors: Mohammad Ridzwan Bin Abd Rahim, Siegfried Schmauder, Yupiter HP Manurung, Peter Binkele, Meor Iqram B. Meor Ahmad, Kiarash Dogahe
Abstract:
This paper deals with the short crack initiation of the material P91 under cyclic loading at two different temperatures, concluded with the estimation of the short crack initiation Wöhler (S/N) curve. An artificial but representative model microstructure was generated using Voronoi tessellation and the Finite Element Method, and the non-uniform stress distribution was calculated accordingly afterward. The number of cycles needed for crack initiation is estimated on the basis of the stress distribution in the model by applying the physically-based Tanaka-Mura model. Initial results show that the number of cycles to generate crack initiation is strongly correlated with temperature.Keywords: short crack initiation, P91, Wöhler curve, Voronoi tessellation, Tanaka-Mura model
Procedia PDF Downloads 1003468 Residual Lifetime Estimation for Weibull Distribution by Fusing Expert Judgements and Censored Data
Authors: Xiang Jia, Zhijun Cheng
Abstract:
The residual lifetime of a product is the operation time between the current time and the time point when the failure happens. The residual lifetime estimation is rather important in reliability analysis. To predict the residual lifetime, it is necessary to assume or verify a particular distribution that the lifetime of the product follows. And the two-parameter Weibull distribution is frequently adopted to describe the lifetime in reliability engineering. Due to the time constraint and cost reduction, a life testing experiment is usually terminated before all the units have failed. Then the censored data is usually collected. In addition, other information could also be obtained for reliability analysis. The expert judgements are considered as it is common that the experts could present some useful information concerning the reliability. Therefore, the residual lifetime is estimated for Weibull distribution by fusing the censored data and expert judgements in this paper. First, the closed-forms concerning the point estimate and confidence interval for the residual lifetime under the Weibull distribution are both presented. Next, the expert judgements are regarded as the prior information and how to determine the prior distribution of Weibull parameters is developed. For completeness, the cases that there is only one, and there are more than two expert judgements are both focused on. Further, the posterior distribution of Weibull parameters is derived. Considering that it is difficult to derive the posterior distribution of residual lifetime, a sample-based method is proposed to generate the posterior samples of Weibull parameters based on the Monte Carlo Markov Chain (MCMC) method. And these samples are used to obtain the Bayes estimation and credible interval for the residual lifetime. Finally, an illustrative example is discussed to show the application. It demonstrates that the proposed method is rather simple, satisfactory, and robust.Keywords: expert judgements, information fusion, residual lifetime, Weibull distribution
Procedia PDF Downloads 1413467 Effective Dose and Size Specific Dose Estimation with and without Tube Current Modulation for Thoracic Computed Tomography Examinations: A Phantom Study
Authors: S. Gharbi, S. Labidi, M. Mars, M. Chelli, F. Ladeb
Abstract:
The purpose of this study is to reduce radiation dose for chest CT examination by including Tube Current Modulation (TCM) to a standard CT protocol. A scan of an anthropomorphic male Alderson phantom was performed on a 128-slice scanner. The estimation of effective dose (ED) in both scans with and without mAs modulation was done via multiplication of Dose Length Product (DLP) to a conversion factor. Results were compared to those measured with a CT-Expo software. The size specific dose estimation (SSDE) values were obtained by multiplication of the volume CT dose index (CTDIvol) with a conversion size factor related to the phantom’s effective diameter. Objective assessment of image quality was performed with Signal to Noise Ratio (SNR) measurements in phantom. SPSS software was used for data analysis. Results showed including CARE Dose 4D; ED was lowered by 48.35% and 51.51% using DLP and CT-expo, respectively. In addition, ED ranges between 7.01 mSv and 6.6 mSv in case of standard protocol, while it ranges between 3.62 mSv and 3.2 mSv with TCM. Similar results are found for SSDE; dose was higher without TCM of 16.25 mGy and was lower by 48.8% including TCM. The SNR values calculated were significantly different (p=0.03<0.05). The highest one is measured on images acquired with TCM and reconstructed with Filtered back projection (FBP). In conclusion, this study proves the potential of TCM technique in SSDE and ED reduction and in conserving image quality with high diagnostic reference level for thoracic CT examinations.Keywords: anthropomorphic phantom, computed tomography, CT-expo, radiation dose
Procedia PDF Downloads 2193466 Energy Management System Based on Voltage Fluctuations Minimization for Droop-Controlled Islanded Microgrid
Authors: Zahra Majd, Mohsen Kalantar
Abstract:
Power management and voltage regulation is one of the most important issues in microgrid (MG) control and scheduling. This paper proposes a multiobjective scheduling formulation that consists of active power costs, voltage fluctuations summation, and technical constraints of MG. Furthermore, load flow and reserve constraints are considered to achieve proper voltage regulation. A modified Jacobian matrix is presented for calculating voltage variations and Mont Carlo simulation is used for generating and reducing scenarios. To convert the problem to a mixed integer linear program, a linearization procedure for nonlinear equations is presented. The proposed model is applied to a typical low-voltage MG and two different cases are investigated. The results show the effectiveness of the proposed model.Keywords: microgrid, energy management system, voltage fluctuations, modified Jacobian matrix
Procedia PDF Downloads 89