Search results for: function analysis system technique
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 45678

Search results for: function analysis system technique

40908 Sensitivity Analysis of Principal Stresses in Concrete Slab of Rigid Pavement Made From Recycled Materials

Authors: Aleš Florian, Lenka Ševelová

Abstract:

Complex sensitivity analysis of stresses in a concrete slab of the real type of rigid pavement made from recycled materials is performed. The computational model of the pavement is designed as a spatial (3D) model, is based on a nonlinear variant of the finite element method that respects the structural nonlinearity, enables to model different arrangements of joints, and the entire model can be loaded by the thermal load. Interaction of adjacent slabs in joints and contact of the slab and the subsequent layer are modeled with the help of special contact elements. Four concrete slabs separated by transverse and longitudinal joints and the additional structural layers and soil to the depth of about 3m are modeled. The thickness of individual layers, physical and mechanical properties of materials, characteristics of joints, and the temperature of the upper and lower surface of slabs are supposed to be random variables. The modern simulation technique Updated Latin Hypercube Sampling with 20 simulations is used. For sensitivity analysis the sensitivity coefficient based on the Spearman rank correlation coefficient is utilized. As a result, the estimates of influence of random variability of individual input variables on the random variability of principal stresses s1 and s3 in 53 points on the upper and lower surface of the concrete slabs are obtained.

Keywords: concrete, FEM, pavement, sensitivity, simulation

Procedia PDF Downloads 332
40907 The Time-Frequency Domain Reflection Method for Aircraft Cable Defects Localization

Authors: Reza Rezaeipour Honarmandzad

Abstract:

This paper introduces an aircraft cable fault detection and location method in light of TFDR keeping in mind the end goal to recognize the intermittent faults adequately and to adapt to the serial and after-connector issues being hard to be distinguished in time domain reflection. In this strategy, the correlation function of reflected and reference signal is used to recognize and find the airplane fault as per the qualities of reflected and reference signal in time-frequency domain, so the hit rate of distinguishing and finding intermittent faults can be enhanced adequately. In the work process, the reflected signal is interfered by the noise and false caution happens frequently, so the threshold de-noising technique in light of wavelet decomposition is used to diminish the noise interference and lessen the shortcoming alert rate. At that point the time-frequency cross connection capacity of the reference signal and the reflected signal based on Wigner-Ville appropriation is figured so as to find the issue position. Finally, LabVIEW is connected to execute operation and control interface, the primary capacity of which is to connect and control MATLAB and LABSQL. Using the solid computing capacity and the bottomless capacity library of MATLAB, the signal processing turn to be effortlessly acknowledged, in addition LabVIEW help the framework to be more dependable and upgraded effectively.

Keywords: aircraft cable, fault location, TFDR, LabVIEW

Procedia PDF Downloads 480
40906 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization

Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman

Abstract:

In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.

Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization

Procedia PDF Downloads 244
40905 Provisions for Risk in Islamic Banking and Finance in Comparison to the Conventional Banks in Malaysia

Authors: Rashid Masoud Ali Al-Mazrui, Ramadhani Mashaka Shabani

Abstract:

Islamic banks and financial institutions are exposed to the same risks as conventional banking. These risks include the rate return risk, credit or market risk, liquidity risk, and operational risk among others. However, being a financial institution that operates Islamic banking and finance operations, there is additional risk associated with its operations different from conventional finance, such as displacing commercial risk. They face Shari'ah compliance risks because of their failure to follow Shari'ah principles. To have proper mitigation and risk management, banks should have proper risk management policies to mitigate risks. This paper aims to study the risk management taken by Islamic banks in comparison with conventional banks. Also, the study evaluates the provisions for risk management taken by selected Islamic banks and conventional banks. The study employs qualitative analysis using secondary data by applying a content analysis approach with a sample size of 4 Islamic banks and four conventional banks ranging from 2010 to 2020. We find that these banks all use the same technique, except for the associated risk. The extra ways are used, but only for additional risks that are available to Islamic banking and finance.

Keywords: emerging risk, risk management, Islamic banking, conventional bank

Procedia PDF Downloads 88
40904 Thermal Analysis and Experimental Procedure of Integrated Phase Change Material in a Storage Tank

Authors: Chargui Ridha, Agrebi Sameh

Abstract:

The integration of phase change materials (PCM) for the storage of thermal energy during the period of sunshine before being released during the night is a complement of free energy to improve the system formed by a solar collector, tank storage, and a heat exchanger. This paper is dedicated to the design of a thermal storage tank based on a PCM-based heat exchanger. The work is divided into two parts: an experimental part using paraffin as PCM was carried out within the Laboratory of Thermal Processes of Borj Cedria in order to improve the performance of the system formed by the coupling of a flat solar collector and a thermal storage tank and to subsequently determine the influence of PCM on the whole system. This phase is based on the measurement instrumentation, namely, a differential scanning calorimeter (DSC) and the thermal analyzer (hot disk: HOT DISK) in order to determine the physical properties of the paraffin (PCM), which has been chosen. The second phase involves the detailed design of the PCM heat exchanger, which is incorporated into a thermal storage tank and coupled with a solar air collector installed at the Research and Technology Centre of Energy (CRTEn). A numerical part based on the TRANSYS and Fluent software, as well as the finite volume method, was carried out for the storage reservoir systems in order to determine the temperature distribution in each chosen system.

Keywords: phase change materials, storage tank, heat exchanger, flat plate collector

Procedia PDF Downloads 100
40903 Software Architecture Optimization Using Swarm Intelligence Techniques

Authors: Arslan Ellahi, Syed Amjad Hussain, Fawaz Saleem Bokhari

Abstract:

Optimization of software architecture can be done with respect to a quality attributes (QA). In this paper, there is an analysis of multiple research papers from different dimensions that have been used to classify those attributes. We have proposed a technique of swarm intelligence Meta heuristic ant colony optimization algorithm as a contribution to solve this critical optimization problem of software architecture. We have ranked quality attributes and run our algorithm on every QA, and then we will rank those on the basis of accuracy. At the end, we have selected the most accurate quality attributes. Ant colony algorithm is an effective algorithm and will perform best in optimizing the QA’s and ranking them.

Keywords: complexity, rapid evolution, swarm intelligence, dimensions

Procedia PDF Downloads 266
40902 Fusion of MOLA-based DEMs and HiRISE Images for Large-Scale Mars Mapping

Authors: Ahmed F. Elaksher, Islam Omar

Abstract:

In this project, we used MOLA-based DEMs to orthorectify HiRISE optical images. The MOLA data was interpolated using the kriging interpolation technique. Corresponding tie points were then digitized from both datasets. These points were employed in co-registering both datasets using GIS analysis tools. Different transformation models, including the affine and projective transformation models, were used with different sets and distributions of tie points. Additionally, we evaluated the use of the MOLA elevations in co-registering the MOLA and HiRISE datasets. The planimetric RMSEs achieved for each model are reported. Results suggested the use of 3D-2D transformation models.

Keywords: photogrammetry, Mars, MOLA, HiRISE

Procedia PDF Downloads 81
40901 Redox-labeled Electrochemical Aptasensor Array for Single-cell Detection

Authors: Shuo Li, Yannick Coffinier, Chann Lagadec, Fabrizio Cleri, Katsuhiko Nishiguchi, Akira Fujiwara, Soo Hyeon Kim, Nicolas Clément

Abstract:

The need for single cell detection and analysis techniques has increased in the past decades because of the heterogeneity of individual living cells, which increases the complexity of the pathogenesis of malignant tumors. In the search for early cancer detection, high-precision medicine and therapy, the technologies most used today for sensitive detection of target analytes and monitoring the variation of these species are mainly including two types. One is based on the identification of molecular differences at the single-cell level, such as flow cytometry, fluorescence-activated cell sorting, next generation proteomics, lipidomic studies, another is based on capturing or detecting single tumor cells from fresh or fixed primary tumors and metastatic tissues, and rare circulating tumors cells (CTCs) from blood or bone marrow, for example, dielectrophoresis technique, microfluidic based microposts chip, electrochemical (EC) approach. Compared to other methods, EC sensors have the merits of easy operation, high sensitivity, and portability. However, despite various demonstrations of low limits of detection (LOD), including aptamer sensors, arrayed EC sensors for detecting single-cell have not been demonstrated. In this work, a new technique based on 20-nm-thick nanopillars array to support cells and keep them at ideal recognition distance for redox-labeled aptamers grafted on the surface. The key advantages of this technology are not only to suppress the false positive signal arising from the pressure exerted by all (including non-target) cells pushing on the aptamers by downward force but also to stabilize the aptamer at the ideal hairpin configuration thanks to a confinement effect. With the first implementation of this technique, a LOD of 13 cells (with5.4 μL of cell suspension) was estimated. In further, the nanosupported cell technology using redox-labeled aptasensors has been pushed forward and fully integrated into a single-cell electrochemical aptasensor array. To reach this goal, the LOD has been reduced by more than one order of magnitude by suppressing parasitic capacitive electrochemical signals by minimizing the sensor area and localizing the cells. Statistical analysis at the single-cell level is demonstrated for the recognition of cancer cells. The future of this technology is discussed, and the potential for scaling over millions of electrodes, thus pushing further integration at sub-cellular level, is highlighted. Despite several demonstrations of electrochemical devices with LOD of 1 cell/mL, the implementation of single-cell bioelectrochemical sensor arrays has remained elusive due to their challenging implementation at a large scale. Here, the introduced nanopillar array technology combined with redox-labeled aptamers targeting epithelial cell adhesion molecule (EpCAM) is perfectly suited for such implementation. Combining nanopillar arrays with microwells determined for single cell trapping directly on the sensor surface, single target cells are successfully detected and analyzed. This first implementation of a single-cell electrochemical aptasensor array based on Brownian-fluctuating redox species opens new opportunities for large-scale implementation and statistical analysis of early cancer diagnosis and cancer therapy in clinical settings.

Keywords: bioelectrochemistry, aptasensors, single-cell, nanopillars

Procedia PDF Downloads 123
40900 Renewable Energy Interfaced Shunt Active Filter Using a Virtual Flux Direct Power Control

Authors: M. R. Bengourina, M. Rahli, L. Hassaine, S. Saadi

Abstract:

In this study, we present a control method entitled virtual flux direct power control of a grid connected photovoltaic system associated with an active power filter. The virtual flux direct control of power (VF-DPC) is employed for the calculation of reference current generation. In this technique, the switches states of inverter are selected from a table of switching based on the immediate errors between the active and reactive powers and their reference values. The objectives of this paper are the reduction of Total Harmonic Distortion (THD) of source current, compensating reactive power and injecting the maximum active power available from the PV array into the load and/or grid. MATLAB/SIMULINK simulations are provided to demonstrate the performance of the proposed approach.

Keywords: shunt active power filter, VF-DPC, photovoltaic, MPPT

Procedia PDF Downloads 325
40899 Gold Nanoparticle: Synthesis, Characterization, Clinico-Pathological, Pathological and Bio-Distribution Studies in Rabbits

Authors: M. M. Bashandy, A. R. Ahmed, M. El-Gaffary, Sahar S. Abd El-Rahman

Abstract:

This study evaluated the acute toxicity and tissue distribution of intravenously administered gold nanoparticles (AuNPs) in male rabbits. Rabbits were exposed to single dose of AuNPs (300 µg/ kg). Toxic effects were assessed via general behavior, hematological parameters, serum biochemical parameters and histopathological examination of various rabbits’ organs. Tissue distribution of AuNPs was evaluated at a dose of 300 µg/ kg in male rabbit. Inductively coupled plasma–mass spectrometry (ICP-MS) was used to determine gold concentrations in tissue samples collected at predetermined time intervals. After one week, AuNPs exerted no obvious acute toxicity in rabbits. However, inflammatory reactions in lung and liver cells were induced in rabbits treated at the300 µg/ kg dose level. The highest gold levels were found in the spleen, followed by liver, lungs and kidneys. These results indicated that AuNPs could be distributed extensively to various tissues in the body, but primarily in the spleen and liver.

Keywords: gold nanoparticles, toxicity, pathology, hematology, liver function, kidney function

Procedia PDF Downloads 341
40898 Motivational Factors on Non-Academic Staff of Higher Education

Authors: Atya Nur Aisha, Pamoedji Hardjomidjojo, Yassierli

Abstract:

Motivation is an important aspect which affects employee behavior to achieve performance. Working motivation tend to be unstable, it easily changing. This condition could be affected by individual factors, namely working ability, and organizational factors, such as working condition and incentives system. The purpose of this study was to examine the impact of individual and organizational factors on non-academic staff motivation. A questionnaire was designed and distributed to 150 non-academic staff of a university in Indonesia. Regression analysis was used to identify the relationship. Results revealed that individual working ability and incentives system had a positive impact on non-academic staff motivation (sig 0.001). This study provides information about practical implication for university authorities and theoretical implications for researchers who interested in exploring motivational and employee performance in a higher education context. It was proposed to increase productivity and work motivation of non-academic staff, university authorities should maintain equality and feasibility of incentives system and design a human resource development to improve employee ability.

Keywords: motivation, incentives, working ability, non-academic staff

Procedia PDF Downloads 414
40897 An Automatic Speech Recognition Tool for the Filipino Language Using the HTK System

Authors: John Lorenzo Bautista, Yoon-Joong Kim

Abstract:

This paper presents the development of a Filipino speech recognition tool using the HTK System. The system was trained from a subset of the Filipino Speech Corpus developed by the DSP Laboratory of the University of the Philippines-Diliman. The speech corpus was both used in training and testing the system by estimating the parameters for phonetic HMM-based (Hidden-Markov Model) acoustic models. Experiments on different mixture-weights were incorporated in the study. The phoneme-level word-based recognition of a 5-state HMM resulted in an average accuracy rate of 80.13 for a single-Gaussian mixture model, 81.13 after implementing a phoneme-alignment, and 87.19 for the increased Gaussian-mixture weight model. The highest accuracy rate of 88.70% was obtained from a 5-state model with 6 Gaussian mixtures.

Keywords: Filipino language, Hidden Markov Model, HTK system, speech recognition

Procedia PDF Downloads 486
40896 The Relationship of Brand Value and Perceived Brand Quality in the Television Business: A Case Study of Television Viewers in Bangkok

Authors: Natnicha Hasoontree

Abstract:

The purpose of this paper was to study the relationship between brand value and perceived brand quality of television viewers in Bangkok towards the television business in Thailand. The population included television viewers in Bangkok, Thailand. A probability sampling technique was performed to get a sample group that included 500 respondents. Taro Yamane technique was utilized to get a proper sample size. A five Likert scale questionnaire was designed specifically to investigate brand value and perceived brand quality from the perspectives of television viewers in Bangkok. The findings implied that consumers in Bangkok attached a high importance towards the brand equity of television companies that comprised brand ability, brand reputation, brand credibility, and business ethics. Perceived brand quality received high rank in all aspects.

Keywords: brand value, perceived brand quality, television business, television viewers

Procedia PDF Downloads 439
40895 Biotechonomy System Dynamics Modelling: Sustainability of Pellet Production

Authors: Andra Blumberga, Armands Gravelsins, Haralds Vigants, Dagnija Blumberga

Abstract:

The paper discovers biotechonomy development analysis by use of system dynamics modelling. The research is connected with investigations of biomass application for production of bioproducts with higher added value. The most popular bioresource is wood, and therefore, the main question today is about future development and eco-design of products. The paper emphasizes and evaluates energy sector which is open for use of wood logs, wood chips, wood pellets and so on. The main aim for this research study was to build a framework to analyse development perspectives for wood pellet production. To reach the goal, a system dynamics model of energy wood supplies, processing, and consumption is built. Production capacity, energy consumption, changes in energy and technology efficiency, required labour source, prices of wood, energy and labour are taken into account. Validation and verification tests with available data and information have been carried out and indicate that the model constitutes the dynamic hypothesis. It is found that the more is invested into pellets production, the higher the specific profit per production unit compared to wood logs and wood chips. As a result, wood chips production is decreasing dramatically and is replaced by wood pellets. The limiting factor for pellet industry growth is availability of wood sources. This is governed by felling limit set by the government based on sustainable forestry principles.

Keywords: bioenergy, biotechonomy, system dynamics modelling, wood pellets

Procedia PDF Downloads 412
40894 A Study of a Plaque Inhibition Through Stenosed Bifurcation Artery considering a Biomagnetic Blood Flow and Elastic Walls

Authors: M. A. Anwar, K. Iqbal, M. Razzaq

Abstract:

Background and Objectives: This numerical study reflects the magnetic field's effect on the reduction of plaque formation due to stenosis in a stenosed bifurcated artery. The entire arterythe wall is assumed as linearly elastic, and blood flow is modeled as a Newtonian, viscous, steady, incompressible, laminar, biomagnetic fluid. Methods: An Arbitrary Lagrangian-Eulerian (ALE) technique is employed to formulate the hemodynamic flow in a bifurcated artery under the effect of the asymmetric magnetic field by two-way Fluid-structure interaction coupling. A stable P2P1 finite element pair is used to discretize thenonlinear system of partial differential equations. The resulting nonlinear system of algebraic equations is solved by the Newton Raphson method. Results: The numerical results for displacement, velocity magnitude, pressure, and wall shear stresses for Reynolds numbers, Re = 500, 1000, 1500, 2000, in the presence of magnetic fields are presented graphically. Conclusions: The numerical results show that the presence of the magnetic field influences the displacement and flows velocity magnitude considerably. The magnetic field reduces the flow separation, recirculation area adjacent to stenosis and gives rise to wall shear stress.

Keywords: bifurcation, elastic walls, finite element, wall shear stress,

Procedia PDF Downloads 185
40893 Design and Development of Novel Anion Selective Chemosensors Derived from Vitamin B6 Cofactors

Authors: Darshna Sharma, Suban K. Sahoo

Abstract:

The detection of intracellular fluoride in human cancer cell HeLa was achieved by chemosensors derived from vitamin B6 cofactors using fluorescence imaging technique. These sensors were first synthesized by condensation of pyridoxal/pyridoxal phosphate with 2-amino(thio)phenol. The anion recognition ability was explored by experimental (UV-VIS, fluorescence and 1H NMR) and theoretical DFT [(B3LYP/6-31G(d,p)] methods in DMSO and mixed DMSO-H2O system. All the developed sensors showed both naked-eye detectable color change and remarkable fluorescence enhancement in the presence of F- and AcO-. The anion recognition was occurred through the formation of hydrogen bonded complexes between these anions and sensor, followed by the partial deprotonation of sensor. The detection limit of these sensors were down to micro(nano) molar level of F- and AcO-.

Keywords: chemosensors, fluoride, acetate, turn-on, live cells imaging, DFT

Procedia PDF Downloads 405
40892 A Hybrid Normalized Gradient Correlation Based Thermal Image Registration for Morphoea

Authors: L. I. Izhar, T. Stathaki, K. Howell

Abstract:

Analyzing and interpreting of thermograms have been increasingly employed in the diagnosis and monitoring of diseases thanks to its non-invasive, non-harmful nature and low cost. In this paper, a novel system is proposed to improve diagnosis and monitoring of morphoea skin disorder based on integration with the published lines of Blaschko. In the proposed system, image registration based on global and local registration methods are found inevitable. This paper presents a modified normalized gradient cross-correlation (NGC) method to reduce large geometrical differences between two multimodal images that are represented by smooth gray edge maps is proposed for the global registration approach. This method is improved further by incorporating an iterative-based normalized cross-correlation coefficient (NCC) method. It is found that by replacing the final registration part of the NGC method where translational differences are solved in the spatial Fourier domain with the NCC method performed in the spatial domain, the performance and robustness of the NGC method can be greatly improved. It is shown in this paper that the hybrid NGC method not only outperforms phase correlation (PC) method but also improved misregistration due to translation, suffered by the modified NGC method alone for thermograms with ill-defined jawline. This also demonstrates that by using the gradients of the gray edge maps and a hybrid technique, the performance of the PC based image registration method can be greatly improved.

Keywords: Blaschko’s lines, image registration, morphoea, thermal imaging

Procedia PDF Downloads 316
40891 Enhanced Disk-Based Databases towards Improved Hybrid in-Memory Systems

Authors: Samuel Kaspi, Sitalakshmi Venkatraman

Abstract:

In-memory database systems are becoming popular due to the availability and affordability of sufficiently large RAM and processors in modern high-end servers with the capacity to manage large in-memory database transactions. While fast and reliable in-memory systems are still being developed to overcome cache misses, CPU/IO bottlenecks and distributed transaction costs, disk-based data stores still serve as the primary persistence. In addition, with the recent growth in multi-tenancy cloud applications and associated security concerns, many organisations consider the trade-offs and continue to require fast and reliable transaction processing of disk-based database systems as an available choice. For these organizations, the only way of increasing throughput is by improving the performance of disk-based concurrency control. This warrants a hybrid database system with the ability to selectively apply an enhanced disk-based data management within the context of in-memory systems that would help improve overall throughput. The general view is that in-memory systems substantially outperform disk-based systems. We question this assumption and examine how a modified variation of access invariance that we call enhanced memory access, (EMA) can be used to allow very high levels of concurrency in the pre-fetching of data in disk-based systems. We demonstrate how this prefetching in disk-based systems can yield close to in-memory performance, which paves the way for improved hybrid database systems. This paper proposes a novel EMA technique and presents a comparative study between disk-based EMA systems and in-memory systems running on hardware configurations of equivalent power in terms of the number of processors and their speeds. The results of the experiments conducted clearly substantiate that when used in conjunction with all concurrency control mechanisms, EMA can increase the throughput of disk-based systems to levels quite close to those achieved by in-memory system. The promising results of this work show that enhanced disk-based systems facilitate in improving hybrid data management within the broader context of in-memory systems.

Keywords: in-memory database, disk-based system, hybrid database, concurrency control

Procedia PDF Downloads 422
40890 Computer Network Applications, Practical Implementations and Structural Control System Representations

Authors: El Miloudi Djelloul

Abstract:

The computer network play an important position for practical implementations of the differently system. To implement a system into network above all is needed to know all the configurations, which is responsible to be a part of the system, and to give adequate information and solution in realtime. So if want to implement this system for example in the school or relevant institutions, the first step is to analyze the types of model which is needed to be configured and another important step is to organize the works in the context of devices, as a part of the general system. Often before configuration, as important point is descriptions and documentations from all the works into the respective process, and then to organize in the aspect of problem-solving. The computer network as critic infrastructure is very specific so the paper present the effectiveness solutions in the structured aspect viewed from one side, and another side is, than the paper reflect the positive aspect in the context of modeling and block schema presentations as an better alternative to solve the specific problem because of continually distortions of the system from the line of devices, programs and signals or packed collisions, which are in movement from one computer node to another nodes.

Keywords: local area networks, LANs, block schema presentations, computer network system, computer node, critical infrastructure packed collisions, structural control system representations, computer network, implementations, modeling structural representations, companies, computers, context, control systems, internet, software

Procedia PDF Downloads 367
40889 Fusion of Finger Inner Knuckle Print and Hand Geometry Features to Enhance the Performance of Biometric Verification System

Authors: M. L. Anitha, K. A. Radhakrishna Rao

Abstract:

With the advent of modern computing technology, there is an increased demand for developing recognition systems that have the capability of verifying the identity of individuals. Recognition systems are required by several civilian and commercial applications for providing access to secured resources. Traditional recognition systems which are based on physical identities are not sufficiently reliable to satisfy the security requirements due to the use of several advances of forgery and identity impersonation methods. Recognizing individuals based on his/her unique physiological characteristics known as biometric traits is a reliable technique, since these traits are not transferable and they cannot be stolen or lost. Since the performance of biometric based recognition system depends on the particular trait that is utilized, the present work proposes a fusion approach which combines Inner knuckle print (IKP) trait of the middle, ring and index fingers with the geometrical features of hand. The hand image captured from a digital camera is preprocessed to find finger IKP as region of interest (ROI) and hand geometry features. Geometrical features are represented as the distances between different key points and IKP features are extracted by applying local binary pattern descriptor on the IKP ROI. The decision level AND fusion was adopted, which has shown improvement in performance of the combined scheme. The proposed approach is tested on the database collected at our institute. Proposed approach is of significance since both hand geometry and IKP features can be extracted from the palm region of the hand. The fusion of these features yields a false acceptance rate of 0.75%, false rejection rate of 0.86% for verification tests conducted, which is less when compared to the results obtained using individual traits. The results obtained confirm the usefulness of proposed approach and suitability of the selected features for developing biometric based recognition system based on features from palmar region of hand.

Keywords: biometrics, hand geometry features, inner knuckle print, recognition

Procedia PDF Downloads 223
40888 Sustainability Impact Assessment of Construction Ecology to Engineering Systems and Climate Change

Authors: Moustafa Osman Mohammed

Abstract:

Construction industry, as one of the main contributor in depletion of natural resources, influences climate change. This paper discusses incremental and evolutionary development of the proposed models for optimization of a life-cycle analysis to explicit strategy for evaluation systems. The main categories are virtually irresistible for introducing uncertainties, uptake composite structure model (CSM) as environmental management systems (EMSs) in a practice science of evaluation small and medium-sized enterprises (SMEs). The model simplified complex systems to reflect nature systems’ input, output and outcomes mode influence “framework measures” and give a maximum likelihood estimation of how elements are simulated over the composite structure. The traditional knowledge of modeling is based on physical dynamic and static patterns regarding parameters influence environment. It unified methods to demonstrate how construction systems ecology interrelated from management prospective in procedure reflects the effect of the effects of engineering systems to ecology as ultimately unified technologies in extensive range beyond constructions impact so as, - energy systems. Sustainability broadens socioeconomic parameters to practice science that meets recovery performance, engineering reflects the generic control of protective systems. When the environmental model employed properly, management decision process in governments or corporations could address policy for accomplishment strategic plans precisely. The management and engineering limitation focuses on autocatalytic control as a close cellular system to naturally balance anthropogenic insertions or aggregation structure systems to pound equilibrium as steady stable conditions. Thereby, construction systems ecology incorporates engineering and management scheme, as a midpoint stage between biotic and abiotic components to predict constructions impact. The later outcomes’ theory of environmental obligation suggests either a procedures of method or technique that is achieved in sustainability impact of construction system ecology (SICSE), as a relative mitigation measure of deviation control, ultimately.

Keywords: sustainability, environmental impact assessment, environemtal management, construction ecology

Procedia PDF Downloads 398
40887 Thermo-Mechanical Approach to Evaluate Softening Behavior of Polystyrene: Validation and Modeling

Authors: Salah Al-Enezi, Rashed Al-Zufairi, Naseer Ahmad

Abstract:

A Thermo-mechanical technique was developed to determine softening point temperature/glass transition temperature (Tg) of polystyrene exposed to high pressures. The design utilizes the ability of carbon dioxide to lower the glass transition temperature of polymers and acts as plasticizer. In this apparatus, the sorption of carbon dioxide to induce softening of polymers as a function of temperature/pressure is performed and the extent of softening is measured in three-point-flexural-bending mode. The polymer strip was placed in the cell in contact with the linear variable differential transformer (LVDT). CO2 was pumped into the cell from a supply cylinder to reach high pressure. The results clearly showed that full softening point of the samples, accompanied by a large deformation on the polymer strip. The deflection curves are initially relatively flat and then undergo a dramatic increase as the temperature is elevated. It was found that increasing the pressure of CO2 causes the temperature curves to shift from higher to lower by increment of about 45 K, over the pressure range of 0-120 bars. The obtained experimental Tg values were validated with the values reported in the literature. Finally, it is concluded that the defection model fits consistently to the generated experimental results, which attempts to describe in more detail how the central deflection of a thin polymer strip affected by the CO2 diffusions in the polymeric samples.

Keywords: softening, high-pressure, polystyrene, CO₂ diffusions

Procedia PDF Downloads 132
40886 An Agent-Based Modelling Simulation Approach to Calculate Processing Delay of GEO Satellite Payload

Authors: V. Vicente E. Mujica, Gustavo Gonzalez

Abstract:

The global coverage of broadband multimedia and internet-based services in terrestrial-satellite networks demand particular interests for satellite providers in order to enhance services with low latencies and high signal quality to diverse users. In particular, the delay of on-board processing is an inherent source of latency in a satellite communication that sometimes is discarded for the end-to-end delay of the satellite link. The frame work for this paper includes modelling of an on-orbit satellite payload using an agent model that can reproduce the properties of processing delays. In essence, a comparison of different spatial interpolation methods is carried out to evaluate physical data obtained by an GEO satellite in order to define a discretization function for determining that delay. Furthermore, the performance of the proposed agent and the development of a delay discretization function are together validated by simulating an hybrid satellite and terrestrial network. Simulation results show high accuracy according to the characteristics of initial data points of processing delay for Ku bands.

Keywords: terrestrial-satellite networks, latency, on-orbit satellite payload, simulation

Procedia PDF Downloads 277
40885 Signaling Using Phase Shifting in Wi-Fi Backscatter System

Authors: Chang-Bin Ha, Young-Min Ko, Seongjoo Lee, Hyoung-Kyu Song

Abstract:

In this paper, the signaling scheme using phase shifting is proposed for the improved performance of the Wi-Fi backscatter system. Because the communication in the Wi-Fi backscatter system is based on on-off modulation and impedance modulation by unit of packet, the data rate is very low compared to the conventional wireless systems. Also, because the Wi-Fi backscatter system is based on the RF-powered device, the achievement of high reliability is difficult. In order to increase the low data rate, the proposed scheme transmits information of multiple bits during one packet period. Also, in order to increase the reliability, the proposed scheme shifts the phase of signal in according to the transmitting information. The simulation result shows that the proposed scheme has the improved throughput performance.

Keywords: phase shifting, RF-powered device, Wi-Fi backscatter system, IoT

Procedia PDF Downloads 445
40884 Study of Aqueous Solutions: A Dielectric Spectroscopy Approach

Authors: Kumbharkhane Ashok

Abstract:

The time domain dielectric relaxation spectroscopy (TDRS) probes the interaction of a macroscopic sample with a time-dependent electrical field. The resulting complex permittivity spectrum, characterizes amplitude (voltage) and time scale of the charge-density fluctuations within the sample. These fluctuations may arise from the reorientation of the permanent dipole moments of individual molecules or from the rotation of dipolar moieties in flexible molecules, like polymers. The time scale of these fluctuations depends on the sample and its relative relaxation mechanism. Relaxation times range from some picoseconds in low viscosity liquids to hours in glasses, Therefore the DRS technique covers an extensive dynamical process, its corresponding frequency range from 10-4 Hz to 1012 Hz. This inherent ability to monitor the cooperative motion of molecular ensemble distinguishes dielectric relaxation from methods like NMR or Raman spectroscopy which yield information on the motions of individual molecules. An experimental set up for Time Domain Reflectometry (TDR) technique from 10 MHz to 30 GHz has been developed for the aqueous solutions. This technique has been very simple and covers a wide band of frequencies in the single measurement. Dielectric Relaxation Spectroscopy is especially sensitive to intermolecular interactions. The complex permittivity spectra of aqueous solutions have been fitted using Cole-Davidson (CD) model to determine static dielectric constants and relaxation times for entire concentrations. The heterogeneous molecular interactions in aqueous solutions have been discussed through Kirkwood correlation factor and excess properties.

Keywords: liquid, aqueous solutions, time domain reflectometry

Procedia PDF Downloads 448
40883 A Simplified Model of the Control System with PFM

Authors: Bekmurza H. Aitchanov, Sholpan K. Aitchanova, Olimzhon A. Baimuratov, Aitkul N. Aldibekova

Abstract:

This work considers the automated control system (ACS) of milk quality during its magnetic field processing. For achieving high level of quality control methods were applied transformation of complex nonlinear systems in a linearized system with a less complex structure. Presented ACS is adjustable by seven parameters: mass fraction of fat, mass fraction of dry skim milk residues (DSMR), density, mass fraction of added water, temperature, mass fraction of protein, acidity.

Keywords: fluids magnetization, nuclear magnetic resonance, automated control system, dynamic pulse-frequency modulator, PFM, nonlinear systems, structural model

Procedia PDF Downloads 378
40882 Statistical Analysis to Compare between Smart City and Traditional Housing

Authors: Taha Anjamrooz, Sareh Rajabi, Ayman Alzaatreh

Abstract:

Smart cities are playing important roles in real life. Integration and automation between different features of modern cities and information technologies improve smart city efficiency, energy management, human and equipment resource management, life quality and better utilization of resources for the customers. One of difficulties in this path, is use, interface and link between software, hardware, and other IT technologies to develop and optimize processes in various business fields such as construction, supply chain management and transportation in parallel to cost-effective and resource reduction impacts. Also, Smart cities are certainly intended to demonstrate a vital role in offering a sustainable and efficient model for smart houses while mitigating environmental and ecological matters. Energy management is one of the most important matters within smart houses in the smart cities and communities, because of the sensitivity of energy systems, reduction in energy wastage and maximization in utilizing the required energy. Specially, the consumption of energy in the smart houses is important and considerable in the economic balance and energy management in smart city as it causes significant increment in energy-saving and energy-wastage reduction. This research paper develops features and concept of smart city in term of overall efficiency through various effective variables. The selected variables and observations are analyzed through data analysis processes to demonstrate the efficiency of smart city and compare the effectiveness of each variable. There are ten chosen variables in this study to improve overall efficiency of smart city through increasing effectiveness of smart houses using an automated solar photovoltaic system, RFID System, smart meter and other major elements by interfacing between software and hardware devices as well as IT technologies. Secondly to enhance aspect of energy management by energy-saving within smart house through efficient variables. The main objective of smart city and smart houses is to reproduce energy and increase its efficiency through selected variables with a comfortable and harmless atmosphere for the customers within a smart city in combination of control over the energy consumption in smart house using developed IT technologies. Initially the comparison between traditional housing and smart city samples is conducted to indicate more efficient system. Moreover, the main variables involved in measuring overall efficiency of system are analyzed through various processes to identify and prioritize the variables in accordance to their influence over the model. The result analysis of this model can be used as comparison and benchmarking with traditional life style to demonstrate the privileges of smart cities. Furthermore, due to expensive and expected shortage of natural resources in near future, insufficient and developed research study in the region, and available potential due to climate and governmental vision, the result and analysis of this study can be used as key indicator to select most effective variables or devices during construction phase and design

Keywords: smart city, traditional housing, RFID, photovoltaic system, energy efficiency, energy saving

Procedia PDF Downloads 115
40881 An Intelligent Nondestructive Testing System of Ultrasonic Infrared Thermal Imaging Based on Embedded Linux

Authors: Hao Mi, Ming Yang, Tian-yue Yang

Abstract:

Ultrasonic infrared nondestructive testing is a kind of testing method with high speed, accuracy and localization. However, there are still some problems, such as the detection requires manual real-time field judgment, the methods of result storage and viewing are still primitive. An intelligent non-destructive detection system based on embedded linux is put forward in this paper. The hardware part of the detection system is based on the ARM (Advanced Reduced Instruction Set Computer Machine) core and an embedded linux system is built to realize image processing and defect detection of thermal images. The CLAHE algorithm and the Butterworth filter are used to process the thermal image, and then the boa server and CGI (Common Gateway Interface) technology are used to transmit the test results to the display terminal through the network for real-time monitoring and remote monitoring. The system also liberates labor and eliminates the obstacle of manual judgment. According to the experiment result, the system provides a convenient and quick solution for industrial non-destructive testing.

Keywords: remote monitoring, non-destructive testing, embedded Linux system, image processing

Procedia PDF Downloads 230
40880 Platelet Transfusion Thresholds for Pediatrics; A Retrospective Study

Authors: Hessah Alsulami, Majedah Aldosari

Abstract:

Introduction: Platelet threshold of 10x109 /L is recommended for clinically stable thrombocytopenic pediatric patients. Transfusions at a higher level (given the absence of research evidence, as determined by clinical circumstances, generally at threshold of 40x109 /L) may be required for patients with signs of bleeding, high fever, hyper-leukocytosis, rapid fall in platelet count, concomitant coagulation abnormality, critically ill patients, and those with impaired platelet function (including drug induced). Transfusions at a higher level may be also required for patients undergoing invasive procedures. Method: This study is a retrospective observational analysis of platelet transfusion thresholds in a single secondary pediatric hospital in Riyadh. From the blood bank database, the list of the patients who received platelet transfusions in the second half of 2018 was retrieved. Patients were divided into two groups; group A, those belong to the category of high platelet level for transfusion (such as those with bleeding, high fever, rapid fall in platelet count, impaired platelet function or undergoing invasive procedures) and group B, those who were not. Then we looked at the pre and post transfusion platelet levels for each group. The data was analyzed using GraphPad software and the data expressed as Mean ± SD. Result: A total of 112 of transfusion episodes in 61 patients (38% female) were analyzed. The age ranged from 24 days to 8 years. The distribution of platelet transfusion episodes was 64% (n=72) for group A and 36% (n= 40) for group B. The mean pre-transfusion platelet count was 46x103 ± (11x 103) for group A and 28x103 ± (6x103) for group B. the post-transfusion mean platelet count was 61 x 103 ± (14 x 103) and 60 x103 ± (24 x 103) for group A and B respectively. Among the groups the rise in the mean platelet count after transfusion was significant among stable patients (group B) compared to unstable patients (group A) (P < 0.001). Conclusion: The platelet count threshold for transfusion varied with the clinical condition and is higher among unstable patients’ group which is expected. For stable patients the threshold was higher than what it should be which means that the clinicians don’t follow the guidelines in this regard. The rise of platelet count after transfusion was higher among stable patients.

Keywords: platelet, transfusion, threshold, pediatric

Procedia PDF Downloads 75
40879 Architecture Performance-Related Design Based on Graphic Parameterization

Authors: Wenzhe Li, Xiaoyu Ying, Grace Ding

Abstract:

Architecture plane form is an important consideration in the design of green buildings due to its significant impact on energy performance. The most effective method to consider energy performance in the early design stages is parametric modelling. This paper presents a methodology to program plane forms using MATLAB language, generating 16 kinds of plane forms by changing four designed parameters. DesignBuilder (an energy consumption simulation software) was proposed to simulate the energy consumption of the generated planes. A regression mathematical model was established to study the relationship between the plane forms and their energy consumption. The main finding of the study suggested that there was a cubic function relationship between the depth-ratio of U-shaped buildings and energy consumption, and there is also a cubic function relationship between the width-ratio and energy consumption. In the design, the depth-ratio of U-shaped buildings should not be less than 2.5, and the width-ratio should not be less than 2.

Keywords: graphic parameterization, green building design, mathematical model, plane form

Procedia PDF Downloads 157