Search results for: measurement model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18437

Search results for: measurement model

17957 Some Issues of Measurement of Impairment of Non-Financial Assets in the Public Sector

Authors: Mariam Vardiashvili

Abstract:

The economic value of the asset impairment process is quite large. Impairment reflects the reduction of future economic benefits or service potentials itemized in the asset. The assets owned by public sector entities bring economic benefits or are used for delivery of the free-of-charge services. Consequently, they are classified as cash-generating and non-cash-generating assets. IPSAS 21 - Impairment of non-cash-generating assets, and IPSAS 26 - Impairment of cash-generating assets, have been designed considering this specificity.  When measuring impairment of assets, it is important to select the relevant methods. For measurement of the impaired Non-Cash-Generating Assets, IPSAS 21 recommends three methods: Depreciated Replacement Cost Approach, Restoration Cost Approach, and  Service Units Approach. Impairment of Value in Use of Cash-Generating Assets (according to IPSAS 26) is measured by discounted value of the money sources to be received in future. Value in use of the cash-generating asserts (as per IPSAS 26) is measured by the discounted value of the money sources to be received in the future. The article provides classification of the assets in the public sector  as non-cash-generating assets and cash-generating assets and, deals also with the factors which should be considered when evaluating  impairment of assets. An essence of impairment of the non-financial assets and the methods of measurement thereof evaluation are formulated according to IPSAS 21 and IPSAS 26. The main emphasis is put on different methods of measurement of the value in use of the impaired Cash-Generating Assets and Non-Cash-Generation Assets and the methods of their selection. The traditional and the expected cash flow approaches for calculation of the discounted value are reviewed. The article also discusses the issues of recognition of impairment loss and its reflection in the financial reporting. The article concludes that despite a functional purpose of the impaired asset, whichever method is used for measuring the asset, presentation of realistic information regarding the value of the assets should be ensured in the financial reporting. In the theoretical development of the issue, the methods of scientific abstraction, analysis and synthesis were used. The research was carried out with a systemic approach. The research process uses international standards of accounting, theoretical researches and publications of Georgian and foreign scientists.

Keywords: cash-generating assets, non-cash-generating assets, recoverable (usable restorative) value, value of use

Procedia PDF Downloads 125
17956 A New and Simple Method of Plotting Binocular Single Vision Field (BSVF) using the Cervical Range of Motion - CROM - Device

Authors: Mihir Kothari, Heena Khan, Vivek Rathod

Abstract:

Assessment of binocular single vision field (BSVF) is traditionally done using a Goldmann perimeter. The measurement of BSVF is important for the management of incomitant strabismus, viz. orbital fractures, thyroid orbitopathy, oculomotor cranial nerve palsies, Duane syndrome etc. In this paper, we describe a new technique for measuring BSVF using a CROM device. Goldmann perimeter is very bulky and expensive (Euro 5000.00 or more) instrument which is 'almost' obsolete from the contemporary ophthalmology practice. Whereas, CROM can be easily made in the DIY (do it yourself) manner for the fraction of the price of the perimeter (only Euro 15.00). Moreover, CROM is useful for the accurate measurement of ocular torticollis vis. nystagmus, paralytic or incomitant squint etc, and it is highly portable.

Keywords: binocular single vision, perimetry, cervical rgen of motion, visual field, binocular single vision field

Procedia PDF Downloads 54
17955 A New Mathematical Model of Human Olfaction

Authors: H. Namazi, H. T. N. Kuan

Abstract:

It is known that in humans, the adaptation to a given odor occurs within a quite short span of time (typically one minute) after the odor is presented to the brain. Different models of human olfaction have been developed by scientists but none of these models consider the diffusion phenomenon in olfaction. A novel microscopic model of the human olfaction is presented in this paper. We develop this model by incorporating the transient diffusivity. In fact, the mathematical model is written based on diffusion of the odorant within the mucus layer. By the use of the model developed in this paper, it becomes possible to provide quantification of the objective strength of odor.

Keywords: diffusion, microscopic model, mucus layer, olfaction

Procedia PDF Downloads 491
17954 A Two Arm Double Parallel Randomized Controlled Trail of the Effects of Health Education Intervention on Insecticide Treated Nets Use and Its Practices among Pregnant Women Attending Antenatal Clinic: Study Protocol

Authors: Opara Monica, Suriani Ismail, Ahmad Iqmer Nashriq Mohd Nazan

Abstract:

The true magnitude of the mortality and morbidity attributable to malaria worldwide is, at best, a scientific guess, although it is not disputable that the greatest burden is in sub-Saharan Africa. Those at highest risk are children younger than 5 years and pregnant women, particularly primigravidae. Nationally, malaria remains the third leading cause of death and is still considered a major public health problem. Therefore, this study is aimed to assess the effectiveness of health education intervention on insecticide-treated net use and its practices among pregnant women attending antenatal clinics. Materials and Methods: This study will be an intervention study with two arms double parallel randomized controlled trial (blinded) to be conducted in 3 stages. The first stage will develop health belief model (HBM) program, while in the second stage, pregnant women will be recruited, assessed (baseline data), randomized into two arms of the study, and follow-up for six months. The third stage will evaluate the impact of the intervention on HBM and disseminate the findings. Data will be collected with the use of a structured questionnaire which will contain validated tools. The main outcome measurement will be the treatment effect using HBM, while data will be analysed using SPSS, version 22. Discussion: The study will contribute to the existing knowledge on hospital-based care programs for pregnant women in developing countries where the literature is scanty. It will generally give insight into the importance of HBM measurement in interventional studies on malaria and other related infectious diseases in this setting.

Keywords: malaria, health education, insecticide-treated nets, sub-Saharan Africa

Procedia PDF Downloads 104
17953 Application of Fuzzy Analytical Hierarchical Process in Evaluation Supply Chain Performance Measurement

Authors: Riyadh Jamegh, AllaEldin Kassam, Sawsan Sabih

Abstract:

In modern trends of market, organizations face high-pressure environment which is characterized by globalization, high competition, and customer orientation, so it is very crucial to control and know the weak and strong points of the supply chain in order to improve their performance. So the performance measurements presented as an important tool of supply chain management because it's enabled the organizations to control, understand, and improve their efficiency. This paper aims to identify supply chain performance measurement (SCPM) by using Fuzzy Analytical Hierarchical Process (FAHP). In our real application, the performance of organizations estimated based on four parameters these are cost parameter indicator of cost (CPI), inventory turnover parameter indicator of (INPI), raw material parameter (RMPI), and safety stock level parameter indicator (SSPI), these indicators vary in impact on performance depending upon policies and strategies of organization. In this research (FAHP) technique has been used to identify the importance of such parameters, and then first fuzzy inference (FIR1) is applied to identify performance indicator of each factor depending on the importance of the factor and its value. Then, the second fuzzy inference (FIR2) also applied to integrate the effect of these indicators and identify (SCPM) which represent the required output. The developed approach provides an effective tool for evaluation of supply chain performance measurement.

Keywords: fuzzy performance measurements, supply chain, fuzzy logic, key performance indicator

Procedia PDF Downloads 125
17952 An Output Oriented Super-Efficiency Model for Considering Time Lag Effect

Authors: Yanshuang Zhang, Byungho Jeong

Abstract:

There exists some time lag between the consumption of inputs and the production of outputs. This time lag effect should be considered in calculating efficiency of decision making units (DMU). Recently, a couple of DEA models were developed for considering time lag effect in efficiency evaluation of research activities. However, these models can’t discriminate efficient DMUs because of the nature of basic DEA model in which efficiency scores are limited to ‘1’. This problem can be resolved a super-efficiency model. However, a super efficiency model sometimes causes infeasibility problem. This paper suggests an output oriented super-efficiency model for efficiency evaluation under the consideration of time lag effect. A case example using a long term research project is given to compare the suggested model with the MpO model

Keywords: DEA, Super-efficiency, Time Lag, research activities

Procedia PDF Downloads 638
17951 Bio-Hub Ecosystems: Expansion of Traditional Life Cycle Analysis Metrics to Include Zero-Waste Circularity Measures

Authors: Kimberly Samaha

Abstract:

In order to attract new types of investors into the emerging Bio-Economy, a new set of metrics and measurement system is needed to better quantify the environmental, social and economic impacts of circular zero-waste design. The Bio-Hub Ecosystem model was developed to address a critical area of concern within the global energy market regarding the use of biomass as a feedstock for power plants. Lack of an economically-viable business model for bioenergy facilities has resulted in the continuation of idled and decommissioned plants. In particular, the forestry-based plants which have been an invaluable outlet for woody biomass surplus, forest health improvement, timber production enhancement, and especially reduction of wildfire risk. This study looked at repurposing existing biomass-energy plants into Circular Zero-Waste Bio-Hub Ecosystems. A Bio-Hub model that first targets a ‘whole-tree’ approach and then looks at the circular economics of co-hosting diverse industries (wood processing, aquaculture, agriculture) in the vicinity of the Biomass Power Plants facilities. It proposes not only models for integration of forestry, aquaculture, and agriculture in cradle-to-cradle linkages of what have typically been linear systems, but the proposal also allows for the early measurement of the circularity and impact of resource use and investment risk mitigation, for these systems. Typically, life cycle analyses measure environmental impacts of different industrial production stages and are not integrated with indicators of material use circularity. This concept paper proposes the further development of a new set of metrics that would illustrate not only the typical life-cycle analysis (LCA), which shows the reduction in greenhouse gas (GHG) emissions, but also the zero-waste circularity measures of mass balance of the full value chain of the raw material and energy content/caloric value. These new measures quantify key impacts in making hyper-efficient use of natural resources and eliminating waste to landfills. The project utilized traditional LCA using the GREET model where the standalone biomass energy plant case was contrasted with the integration of a jet-fuel biorefinery. The methodology was then expanded to include combinations of co-hosts that optimize the life cycle of woody biomass from tree to energy, CO₂, heat and wood ash both from an energy/caloric value and for mass balance to include reuse of waste streams which are typically landfilled. The major findings of both a formal LCA study resulted in the masterplan for the first Bio-Hub to be built in West Enfield, Maine. Bioenergy facilities are currently at a critical juncture where they have an opportunity to be repurposed into efficient, profitable and socially responsible investments, or be idled and scrapped. If proven as a model, the expedited roll-out of these innovative scenarios can set a new standard for circular zero-waste projects that advance the critical transition from the current ‘take-make-dispose’ paradigm inherent in the energy, forestry and food industries to a more sustainable bio-economy paradigm where waste streams become valuable inputs, supporting local and rural communities in simple, sustainable ways.

Keywords: bio-economy, biomass energy, financing, metrics

Procedia PDF Downloads 142
17950 Validation of the Formal Model of Web Services Applications for Digital Reference Service of Library Information System

Authors: Zainab Magaji Musa, Nordin M. A. Rahman, Julaily Aida Jusoh

Abstract:

The web services applications for digital reference service (WSDRS) of LIS model is an informal model that claims to reduce the problems of digital reference services in libraries. It uses web services technology to provide efficient way of satisfying users’ needs in the reference section of libraries. The formal WSDRS model consists of the Z specifications of all the informal specifications of the model. This paper discusses the formal validation of the Z specifications of WSDRS model. The authors formally verify and thus validate the properties of the model using Z/EVES theorem prover.

Keywords: validation, verification, formal, theorem prover

Procedia PDF Downloads 496
17949 Relative Entropy Used to Determine the Divergence of Cells in Single Cell RNA Sequence Data Analysis

Authors: An Chengrui, Yin Zi, Wu Bingbing, Ma Yuanzhu, Jin Kaixiu, Chen Xiao, Ouyang Hongwei

Abstract:

Single cell RNA sequence (scRNA-seq) is one of the effective tools to study transcriptomics of biological processes. Recently, similarity measurement of cells is Euclidian distance or its derivatives. However, the process of scRNA-seq is a multi-variate Bernoulli event model, thus we hypothesize that it would be more efficient when the divergence between cells is valued with relative entropy than Euclidian distance. In this study, we compared the performances of Euclidian distance, Spearman correlation distance and Relative Entropy using scRNA-seq data of the early, medial and late stage of limb development generated in our lab. Relative Entropy is better than other methods according to cluster potential test. Furthermore, we developed KL-SNE, an algorithm modifying t-SNE whose definition of divergence between cells Euclidian distance to Kullback–Leibler divergence. Results showed that KL-SNE was more effective to dissect cell heterogeneity than t-SNE, indicating the better performance of relative entropy than Euclidian distance. Specifically, the chondrocyte expressing Comp was clustered together with KL-SNE but not with t-SNE. Surprisingly, cells in early stage were surrounded by cells in medial stage in the processing of KL-SNE while medial cells neighbored to late stage with the process of t-SNE. This results parallel to Heatmap which showed cells in medial stage were more heterogenic than cells in other stages. In addition, we also found that results of KL-SNE tend to follow Gaussian distribution compared with those of the t-SNE, which could also be verified with the analysis of scRNA-seq data from another study on human embryo development. Therefore, it is also an effective way to convert non-Gaussian distribution to Gaussian distribution and facilitate the subsequent statistic possesses. Thus, relative entropy is potentially a better way to determine the divergence of cells in scRNA-seq data analysis.

Keywords: Single cell RNA sequence, Similarity measurement, Relative Entropy, KL-SNE, t-SNE

Procedia PDF Downloads 327
17948 Linear MIMO Model Identification Using an Extended Kalman Filter

Authors: Matthew C. Best

Abstract:

Linear Multi-Input Multi-Output (MIMO) dynamic models can be identified, with no a priori knowledge of model structure or order, using a new Generalised Identifying Filter (GIF). Based on an Extended Kalman Filter, the new filter identifies the model iteratively, in a continuous modal canonical form, using only input and output time histories. The filter’s self-propagating state error covariance matrix allows easy determination of convergence and conditioning, and by progressively increasing model order, the best fitting reduced-order model can be identified. The method is shown to be resistant to noise and can easily be extended to identification of smoothly nonlinear systems.

Keywords: system identification, Kalman filter, linear model, MIMO, model order reduction

Procedia PDF Downloads 577
17947 Spectral Linewidth Measurement of Linear Frequency Modulated Continuous Wave Laser with Short Delay within the Coherence Length

Authors: Jongpil La, Jieun Choi

Abstract:

Optical frequency modulation technology for FMCW LiDAR based on Optical Phase Locked Loop(OPLL) configuration is addressed in this paper. The spectral linewidth measurement method of the linear frequency-modulated laser is also described. The single-frequency laser with narrow spectral linewidth is generated using an external cavity diode laser and the excitation frequency of the laser is adjusted by controlling the injection current of the laser. If the injection current of the laser is increased, the lasing frequency is decreased because of the slight increase in the refractive index of the laser gain chip. Dynamic optical frequency change rate is measured by using a Mach-Zehnder interferometer and compared with a proper reference signal. The phase difference between the reference signal and the measured signal using the Mach-Zehnder interferometer is obtained by mixing those two signals. The phase error is used to detect the frequency deviation error from the target value, which is then fed back to the driving current of the laser to compensate for it. The frequency sweep error from the ideal linear frequency waveform will degrade the spectral linewidth of the target spectrum and will degrade the maximum range performance of FMCW LiDAR. Therefore, the spectral linewidth measurement of frequency modulated laser is very important to evaluate the performance of the LiDAR system. However, it is impossible to apply the conventional self-homodyne or self-heterodyne method with a long delay line to evaluate the spectral linewidth of the frequency-modulated laser because the beat frequency generated by the long delay line is too high to measure with a high bandwidth frequency modulated laser. In this article, the spectral linewidth of the frequency-modulated laser is measured by using the newly proposed self-heterodyne method with a short delay line. The theoretical derivation for the proposed linewidth measurement method is provided in this article. The laser's spectral modulation bandwidth and linewidth are measured as 2.91GHz and 287kHz, respectively. LiDAR.

Keywords: FMCW, LiDAR, spectral linewidth, self-heterodyne

Procedia PDF Downloads 24
17946 The Impact of Direct and Indirect Pressure Measuring Systems on the Pressure Mapping for the Medical Compression Garments

Authors: Arash M. Shahidi, Tilak Dias, Gayani K. Nandasiri

Abstract:

While graduated compression is the foundation of treatment and management of many medical complications such as leg ulcer, varicose veins, and lymphedema, monitoring the interface pressure has been conducted using different sensors that operate based on diverse approaches. The variations existed from the pressure readings collected using different interface pressure measurement systems would cause difficulties in taking a decision regarding the compression therapy. It is crucial to acknowledge the differences existing between direct and indirect pressure measurement systems while considering the commercially available systems such as AMI, Picopress and OPM which are under direct measurements systems, and HATRA (BSI), HOSY (RAL-GZ) and FlexiForce which comes under the indirect measurement system. Furthermore, Piezo-resistive sensors (Flexiforce) can measure the changes in resistance corresponding to the applied force on the sensing area. Direct pressure measuring systems are capable of measuring interface pressure on the three-dimensional states, while the indirect pressure measuring systems stretch the fabric in the two-dimensional direction and extrapolate pressure from surface tension measured on the device and neglect the vital factor which is the radius of curvature. In this study, a leg mannequin of known dimensions is selected with a knitted class 3 compression stocking. It has been decided to evaluate the data collected from different available systems (AMI, PicoPress, FlexiForce, and HATRA) and compare the results. The results showed a discrepancy between Hatra, AMI, Picopress, and Flexiforce against the pressure standard used to generate class 3 compression stocking. As predicted a higher pressure value with direct interface measuring systems were monitored against HATRA due to the effect of the radius of curvature.

Keywords: AMI, FlexiForce, graduated compression, HATRA, interface pressure, PicoPress

Procedia PDF Downloads 330
17945 Towards Efficient Reasoning about Families of Class Diagrams Using Union Models

Authors: Tejush Badal, Sanaa Alwidian

Abstract:

Class diagrams are useful tools within the Unified Modelling Language (UML) to model and visualize the relationships between, and properties of objects within a system. As a system evolves over time and space (e.g., products), a series of models with several commonalities and variabilities create what is known as a model family. In circumstances where there are several versions of a model, examining each model individually, becomes expensive in terms of computation resources. To avoid performing redundant operations, this paper proposes an approach for representing a family of class diagrams into Union Models to represent model families using a single generic model. The paper aims to analyze and reason about a family of class diagrams using union models as opposed to individual analysis of each member model in the family. The union algorithm provides a holistic view of the model family, where the latter cannot be otherwise obtained from an individual analysis approach, this in turn, enhances the analysis performed in terms of speeding up the time needed to analyze a family of models together as opposed to analyzing individual models, one model at a time.

Keywords: analysis, class diagram, model family, unified modeling language, union model

Procedia PDF Downloads 56
17944 Increase Productivity by Using Work Measurement Technique

Authors: Mohammed Al Awadh

Abstract:

In order for businesses to take advantage of the opportunities for expanded production and trade that have arisen as a result of globalization and increased levels of competition, productivity growth is required. The number of available sources is decreasing with each passing day, which results in an ever-increasing demand. In response to this, there will be an increased demand placed on firms to improve the efficiency with which they utilise their resources. As a scientific method, work and time research techniques have been employed in all manufacturing and service industries to raise the efficiency of use of the factors of production. These approaches focus on work and time. The goal of this research is to improve the productivity of a manufacturing industry's production system by looking at ways to measure work. The work cycles were broken down into more manageable and quantifiable components. On the observation sheet, these aspects were noted down. The operation has been properly analysed in order to identify value-added and non-value-added components, and observations have been recorded for each of the different trails.

Keywords: time study, work measurement, work study, efficiency

Procedia PDF Downloads 61
17943 A Non-Iterative Shape Reconstruction of an Interface from Boundary Measurement

Authors: Mourad Hrizi

Abstract:

In this paper, we study the inverse problem of reconstructing an interior interface D appearing in the elliptic partial differential equation: Δu+χ(D)u=0 from the knowledge of the boundary measurements. This problem arises from a semiconductor transistor model. We propose a new shape reconstruction procedure that is based on the Kohn-Vogelius formulation and the topological sensitivity method. The inverse problem is formulated as a topology optimization one. A topological sensitivity analysis is derived from a function. The unknown subdomain D is reconstructed using a level-set curve of the topological gradient. Finally, we give several examples to show the viability of our proposed method.

Keywords: inverse problem, topological optimization, topological gradient, Kohn-Vogelius formulation

Procedia PDF Downloads 230
17942 Effect of Filter Paper Technique in Measuring Hydraulic Capacity of Unsaturated Expansive Soil

Authors: Kenechi Kurtis Onochie

Abstract:

This paper shows the use of filter paper technique in the measurement of matric suction of unsaturated expansive soil around the Haspolat region of Lefkosa, North Cyprus in other to establish the soil water characteristics curve (SWCC) or soil water retention curve (SWRC). The dry filter paper approach which is standardized by ASTM, 2003, D 5298-03 in which the filter paper is initially dry was adopted. The whatman No. 42 filter paper was used in the matric suction measurement. The maximum dry density of the soil was obtained as 2.66kg/cm³ and the optimum moisture content as 21%. The soil was discovered to have high air entry value of 1847.46KPa indicating finer particles and 25% hydraulic capacity using filter paper technique. The filter paper technique proved to be very useful for measuring the hydraulic capacity of unsaturated expansive soil.

Keywords: SWCC, matric suction, filter paper, expansive soil

Procedia PDF Downloads 153
17941 Subsurface Elastic Properties Determination for Site Characterization Using Seismic Refraction Tomography at the Pwalugu Dam Area

Authors: Van-Dycke Sarpong Asare, Vincent Adongo

Abstract:

Field measurement of subsurface seismic p-wave velocities was undertaken through seismic refraction tomography. The aim of this work is to obtain a model of the shallow subsurface material elastic properties relevant for geotechnical site characterization. The survey area is at Pwalugu in Northern Ghana, where a multipurpose dam, for electricity generation, irrigation, and potable water delivery, is being planned. A 24-channel seismograph and 24, 10 Hz electromagnetic geophones, deployed 5 m apart constituted the acquisition hardware. Eleven (2-D) seismic refraction profiles, nine of which ran almost perpendicular and two parallel to the White Volta at Pwalugu, were acquired. The refraction tomograms of the thirteen profiles revealed a subsurface model consisting of one minor and one major acoustic impedance boundaries – the top dry/loose sand and the variably weathered sandstone contact, and the overburden-sandstones bedrock contact respectively. The p-wave velocities and by inference, with a priori values of poison ratios, the s-wave velocities, assisted in characterizing the geotechnical conditions of the proposed site and also in evaluating the dynamic properties such as the maximum shear modulus, the bulk modulus, and the Young modulus.

Keywords: tomography, characterization, consolidated, Pwalugu and seismograph

Procedia PDF Downloads 113
17940 Development of an Artificial Ear for Bone-Conducted Objective Occlusion Measurement

Authors: Yu Luan

Abstract:

The bone-conducted objective occlusion effect (OE) is characterized by a discomforting sensation of fullness experienced in an occluded ear. This phenomenon arises from various external stimuli, such as human speech, chewing, and walking, which generate vibrations transmitted through the body to the ear canal walls. The bone-conducted OE occurs due to the pressure build-up inside the occluded ear caused by sound radiating into the ear canal cavity from its walls. In the hearing aid industry, artificial ears are utilized as a tool for developing hearing aids. However, the currently available commercial artificial ears primarily focus on pure acoustics measurements, neglecting the bone-conducted vibration aspect. This research endeavors to develop an artificial ear specifically designed for bone-conducted occlusion measurements. Finite element analysis (FEA) modeling has been employed to gain insights into the behavior of the artificial ear.

Keywords: artificial ear, bone conducted vibration, occlusion measurement, finite element modeling

Procedia PDF Downloads 69
17939 A 3D Cell-Based Biosensor for Real-Time and Non-Invasive Monitoring of 3D Cell Viability and Drug Screening

Authors: Yuxiang Pan, Yong Qiu, Chenlei Gu, Ping Wang

Abstract:

In the past decade, three-dimensional (3D) tumor cell models have attracted increasing interest in the field of drug screening due to their great advantages in simulating more accurately the heterogeneous tumor behavior in vivo. Drug sensitivity testing based on 3D tumor cell models can provide more reliable in vivo efficacy prediction. The gold standard fluorescence staining is hard to achieve the real-time and label-free monitoring of the viability of 3D tumor cell models. In this study, micro-groove impedance sensor (MGIS) was specially developed for dynamic and non-invasive monitoring of 3D cell viability. 3D tumor cells were trapped in the micro-grooves with opposite gold electrodes for the in-situ impedance measurement. The change of live cell number would cause inversely proportional change to the impedance magnitude of the entire cell/matrigel to construct and reflect the proliferation and apoptosis of 3D cells. It was confirmed that 3D cell viability detected by the MGIS platform is highly consistent with the standard live/dead staining. Furthermore, the accuracy of MGIS platform was demonstrated quantitatively using 3D lung cancer model and sophisticated drug sensitivity testing. In addition, the parameters of micro-groove impedance chip processing and measurement experiments were optimized in details. The results demonstrated that the MGIS and 3D cell-based biosensor and would be a promising platform to improve the efficiency and accuracy of cell-based anti-cancer drug screening in vitro.

Keywords: micro-groove impedance sensor, 3D cell-based biosensors, 3D cell viability, micro-electromechanical systems

Procedia PDF Downloads 117
17938 Powder Flow with Normalized Powder Particles Size Distribution and Temperature Analyses in Laser Melting Deposition: Analytical Modelling and Experimental Validation

Authors: Muhammad Arif Mahmood, Andrei C. Popescu, Mihai Oane, Diana Chioibascu, Carmen Ristoscu, Ion N. Mihailescu

Abstract:

Powder flow and temperature distributions are recognized as influencing factors during laser melting deposition (LMD) process, that not only affect the consolidation rate but also characteristics of the deposited layers. Herewith, two simplified analytical models will be presented to simulate the powder flow with the inclusion of powder particles size distribution in Gaussian form, under three powder jet nozzles, and temperature analyses during LMD process. The output of the 1st model will serve as the input in the 2nd model. The models will be validated with experimental data, i.e., weight measurement method for powder particles distribution and infrared imaging for temperature analyses. This study will increase the cost-efficiency of the LMD process by adjustment of the operating parameters for reaching optimal powder debit and energy. This research has received funds under the Marie Sklodowska-Curie grant agreement No. 764935, from the European Union’s Horizon 2020 research and innovation program.

Keywords: laser additive manufacturing, powder particles size distribution in Gaussian form, powder stream distribution, temperature analyses

Procedia PDF Downloads 120
17937 Knowledge Management: Why is So Difficult? From “A Good Idea” to Organizational Contribute

Authors: Lisandro Blas, Héctor Tamanini

Abstract:

From earliest 90 to now, no many companies or organization can “really” implement a knowledge management (KM) system that works (no only viewed from a measurement model, but in this continuity). Which are the reasons of that? Some of the reason maybe could be embedded in how KM is demanded (usefulness, priority, experts, a definition of KM) vs the importance and resources that the organizations afford (budget, responsible of a specific area of KM, intangibility). Many organizations “claim” the importance of Knowledge Management but thhese demands are not reflecting these claims in their future actions. With another’s tools or managements ideas the organizations put the economics and human resources to work. Why it´s not occur in KM? This paper tray to explain some of this reasons and tray to deal with this situations through a survey done in 2011 for a IAPG (Argentinean Institute from Oil & Gas) Congress.

Keywords: knowledge management into organizations, new perspectives, failure in implementation, claim

Procedia PDF Downloads 408
17936 The Social Change Leadership Model for Administrators and Teachers Development in Northeast Thailand

Authors: D. Thawinkarn, S. Wongbutlee

Abstract:

The Social Change Leadership model is strongly aligned with administration’s mission. This research aims to examine the elements of social change leadership, build and develop leadership for social change, and evaluate effectiveness of leadership development model for social change. The research operation has 3 phases: model studies by in-depth interviews and survey research; drafting and creating model which verified by the experts; and trial of model in schools. The results showed that administrators and teachers have the elements of leadership for social change in moderate level. These elements are ranged descending from consciousness of self, common purpose, congruence, collaboration, commitment, citizenship, and controversy with civility. Model of leadership for social change is included the principles, objectives, content, process. Workshop process: Results show that the model of leadership development for social change in administrators and teachers leads to higher score in leadership evaluation prior to administering the operation.

Keywords: leadership, social change model, organization, administrators

Procedia PDF Downloads 401
17935 Modeling of Wind Loads on Heliostats Installed in South Algeria of Various Pylon Height

Authors: Hakim Merarda, Mounir Aksas, Toufik Arrif, Abd Elfateh Belaid, Amor Gama, Reski Khelifi

Abstract:

Knowledge of wind loads is important to develop a heliostat with good performance. These loads can be calculated by mathematical equations based on several parameters: the density, wind velocity, the aspect ratio of the mirror (height/width) and the coefficient of the height of the tower. Measurement data of the wind velocity and the density of the air are used in a numerical simulation of wind profile that was performed on heliostats with different pylon heights, with 1m^2 mirror areas and with aspect ratio of mirror equal to 1. These measurement data are taken from the meteorological station installed in Ghardaia, Algeria. The main aim of this work is to find a mathematical correlation between the wind loads and the height of the tower.

Keywords: heliostat, solar tower power, wind loads simulation, South Algeria

Procedia PDF Downloads 540
17934 Analysis of Urban Flooding in Wazirabad Catchment of Kabul City with Help of Geo-SWMM

Authors: Fazli Rahim Shinwari, Ulrich Dittmer

Abstract:

Like many megacities around the world, Kabul is facing severe problems due to the rising frequency of urban flooding. Since 2001, Kabul is experiencing rapid population growth because of the repatriation of refugees and internal migration. Due to unplanned development, green areas inside city and hilly areas within and around the city are converted into new housing towns that had increased runoff. Trenches along the roadside comprise the unplanned drainage network of the city that drains the combined sewer flow. In rainy season overflow occurs, and after streets become dry, the dust particles contaminate the air which is a major cause of air pollution in Kabul city. In this study, a stormwater management model is introduced as a basis for a systematic approach to urban drainage planning in Kabul. For this purpose, Kabul city is delineated into 8 watersheds with the help of one-meter resolution LIDAR DEM. Storm, water management model, is developed for Wazirabad catchment by using available data and literature values. Due to lack of long term metrological data, the model is only run for hourly rainfall data of a rain event that occurred in April 2016. The rain event from 1st to 3rd April with maximum intensity of 3mm/hr caused huge flooding in Wazirabad Catchment of Kabul City. Model-estimated flooding at some points of the catchment as an actual measurement of flooding was not possible; results were compared with information obtained from local people, Kabul Municipality and Capital Region Independent Development Authority. The model helped to identify areas where flooding occurred because of less capacity of drainage system and areas where the main reason for flooding is due to blockage in the drainage canals. The model was used for further analysis to find a sustainable solution to the problem. The option to construct new canals was analyzed, and two new canals were proposed that will reduce the flooding frequency in Wazirabad catchment of Kabul city. By developing the methodology to develop a stormwater management model from digital data and information, the study had fulfilled the primary objective, and similar methodology can be used for other catchments of Kabul city to prepare an emergency and long-term plan for drainage system of Kabul city.

Keywords: urban hydrology, storm water management, modeling, SWMM, GEO-SWMM, GIS, identification of flood vulnerable areas, urban flooding analysis, sustainable urban drainage

Procedia PDF Downloads 136
17933 Positive Psychology and the Social Emotional Ability Instrument (SEAI)

Authors: Victor William Harris

Abstract:

This research is a validation study of the Social Emotional Ability Inventory (SEAI), a multi-dimensional self-report instrument informed by positive psychology, emotional intelligence, social intelligence, and sociocultural learning theory. Designed for use in tandem with the Social Emotional Development (SEAD) theoretical model, the SEAI provides diagnostic-level guidance for professionals and individuals interested in investigating, identifying, and understanding social, emotional strengths, as well as remediating specific social competency deficiencies. The SEAI was shown to be psychometrically sound, exhibited strong internal reliability, and supported the a priori hypotheses of the SEAD. Additionally, confirmatory factor analysis provided evidence of goodness of fit, convergent and divergent validity, and supported a theoretical model that reflected SEAD expectations. The SEAI and SEAD hold potentially far-reaching and important practical implications for theoretical guidance and diagnostic-level measurement of social, emotional competency across a wide range of domains. Strategies researchers, practitioners, educators, and individuals might use to deploy SEAI in order to improve quality of life outcomes are discussed.

Keywords: emotion, emotional ability, positive psychology-social emotional ability, social emotional ability, social emotional ability instrument

Procedia PDF Downloads 220
17932 Maintaining Experimental Consistency in Geomechanical Studies of Methane Hydrate Bearing Soils

Authors: Lior Rake, Shmulik Pinkert

Abstract:

Methane hydrate has been found in significant quantities in soils offshore within continental margins and in permafrost within arctic regions where low temperature and high pressure are present. The mechanical parameters for geotechnical engineering are commonly evaluated in geomechanical laboratories adapted to simulate the environmental conditions of methane hydrate-bearing sediments (MHBS). Due to the complexity and high cost of natural MHBS sampling, most laboratory investigations are conducted on artificially formed samples. MHBS artificial samples can be formed using different hydrate formation methods in the laboratory, where methane gas and water are supplied into the soil pore space under the methane hydrate phase conditions. The most commonly used formation method is the excess gas method which is considered a relatively simple, time-saving, and repeatable testing method. However, there are several differences in the procedures and techniques used to produce the hydrate using the excess gas method. As a result of the difference between the test facilities and the experimental approaches that were carried out in previous studies, different measurement criteria and analyses were proposed for MHBS geomechanics. The lack of uniformity among the various experimental investigations may adversely impact the reliability of integrating different data sets for unified mechanical model development. In this work, we address some fundamental aspects relevant to reliable MHBS geomechanical investigations, such as hydrate homogeneity in the sample, the hydrate formation duration criterion, the hydrate-saturation evaluation method, and the effect of temperature measurement accuracy. Finally, a set of recommendations for repeatable and reliable MHBS formation will be suggested for future standardization of MHBS geomechanical investigation.

Keywords: experimental study, laboratory investigation, excess gas, hydrate formation, standardization, methane hydrate-bearing sediment

Procedia PDF Downloads 41
17931 Highly Linear and Low Noise AMR Sensor Using Closed Loop and Signal-Chopped Architecture

Authors: N. Hadjigeorgiou, A. C. Tsalikidou, E. Hristoforou, P. P. Sotiriadis

Abstract:

During the last few decades, the continuously increasing demand for accurate and reliable magnetic measurements has paved the way for the development of different types of magnetic sensing systems as well as different measurement techniques. Sensor sensitivity and linearity, signal-to-noise ratio, measurement range, cross-talk between sensors in multi-sensor applications are only some of the aspects that have been examined in the past. In this paper, a fully analog closed loop system in order to optimize the performance of AMR sensors has been developed. The operation of the proposed system has been tested using a Helmholtz coil calibration setup in order to control both the amplitude and direction of magnetic field in the vicinity of the AMR sensor. Experimental testing indicated that improved linearity of sensor response, as well as low noise levels can be achieved, when the system is employed.

Keywords: AMR sensor, closed loop, memory effects, chopper, linearity improvement, sensitivity improvement, magnetic noise, electronic noise

Procedia PDF Downloads 346
17930 Lee-Carter Mortality Forecasting Method with Dynamic Normal Inverse Gaussian Mortality Index

Authors: Funda Kul, İsmail Gür

Abstract:

Pension scheme providers have to price mortality risk by accurate mortality forecasting method. There are many mortality-forecasting methods constructed and used in literature. The Lee-Carter model is the first model to consider stochastic improvement trends in life expectancy. It is still precisely used. Mortality forecasting is done by mortality index in the Lee-Carter model. It is assumed that mortality index fits ARIMA time series model. In this paper, we propose and use dynamic normal inverse gaussian distribution to modeling mortality indes in the Lee-Carter model. Using population mortality data for Italy, France, and Turkey, the model is forecasting capability is investigated, and a comparative analysis with other models is ensured by some well-known benchmarking criterions.

Keywords: mortality, forecasting, lee-carter model, normal inverse gaussian distribution

Procedia PDF Downloads 344
17929 Combining the Dynamic Conditional Correlation and Range-GARCH Models to Improve Covariance Forecasts

Authors: Piotr Fiszeder, Marcin Fałdziński, Peter Molnár

Abstract:

The dynamic conditional correlation model of Engle (2002) is one of the most popular multivariate volatility models. However, this model is based solely on closing prices. It has been documented in the literature that the high and low price of the day can be used in an efficient volatility estimation. We, therefore, suggest a model which incorporates high and low prices into the dynamic conditional correlation framework. Empirical evaluation of this model is conducted on three datasets: currencies, stocks, and commodity exchange-traded funds. The utilisation of realized variances and covariances as proxies for true variances and covariances allows us to reach a strong conclusion that our model outperforms not only the standard dynamic conditional correlation model but also a competing range-based dynamic conditional correlation model.

Keywords: volatility, DCC model, high and low prices, range-based models, covariance forecasting

Procedia PDF Downloads 163
17928 Optimization of Thermopile Sensor Performance of Polycrystalline Silicon Film

Authors: Li Long, Thomas Ortlepp

Abstract:

A theoretical model for the optimization of thermopile sensor performance is developed for thermoelectric-based infrared radiation detection. It is shown that the performance of polycrystalline silicon film thermopile sensor can be optimized according to the thermoelectric quality factor, sensor layer structure factor, and sensor layout geometrical form factor. Based on the properties of electrons, phonons, grain boundaries, and their interactions, the thermoelectric quality factor of polycrystalline silicon is analyzed with the relaxation time approximation of the Boltzmann transport equation. The model includes the effect of grain structure, grain boundary trap properties, and doping concentration. The layer structure factor is analyzed with respect to the infrared absorption coefficient. The optimization of layout design is characterized by the form factor, which is calculated for different sensor designs. A double-layer polycrystalline silicon thermopile infrared sensor on a suspended membrane has been designed and fabricated with a CMOS-compatible process. The theoretical approach is confirmed by measurement results.

Keywords: polycrystalline silicon, relaxation time approximation, specific detectivity, thermal conductivity, thermopile infrared sensor

Procedia PDF Downloads 117