Search results for: measurement problem
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9631

Search results for: measurement problem

8731 Non-Pharmacological Approach to the Improvement and Maintenance of the Convergence Parameter

Authors: Andreas Aceranti, Guido Bighiani, Francesca Crotto, Marco Colorato, Stefania Zaghi, Marino Zanetti, Simonetta Vernocchi

Abstract:

The management of eye parameters such as convergence, accommodation, and miosis is very complex; in fact, both the neurovegetative system and the complex Oculocephalgiria system come into play. We have found the effectiveness of the "highvelocity low amplitude" technique directed on C7-T1 (where the cilio-spinal nucleus of the budge is located) in improving the convergence parameter through the measurement of the point of maximum convergence. With this research, we set out to investigate whether the improvement obtained through the High Velocity Low Amplitude maneuver lasts over time, carrying out a pre-manipulation measurement, one immediately after manipulation and one month after manipulation. We took a population of 30 subjects with both refractive and non-refractive problems. Of the 30 patients tested, 27 gave a positive result after the High Velocity Low Amplitude maneuver, giving an improvement in the point of maximum convergence. After a month, we retested all 27 subjects: some further improved the result, others kept, and three subjects slightly lost the gain obtained. None of the re-tested patients returned to the point of maximum convergence starting pre-manipulation. This result opens the door to a multidisciplinary approach between ophthalmologists and osteopaths with the aim of addressing oculomotricity and convergence deficits that increasingly afflict our society due to the massive use of devices and for the conduct of life in closed and restricted environments.

Keywords: point of maximum convergence, HVLA, improvement in PPC, convergence

Procedia PDF Downloads 77
8730 Scheduling Jobs with Stochastic Processing Times or Due Dates on a Server to Minimize the Number of Tardy Jobs

Authors: H. M. Soroush

Abstract:

The problem of scheduling products and services for on-time deliveries is of paramount importance in today’s competitive environments. It arises in many manufacturing and service organizations where it is desirable to complete jobs (products or services) with different weights (penalties) on or before their due dates. In such environments, schedules should frequently decide whether to schedule a job based on its processing time, due-date, and the penalty for tardy delivery to improve the system performance. For example, it is common to measure the weighted number of late jobs or the percentage of on-time shipments to evaluate the performance of a semiconductor production facility or an automobile assembly line. In this paper, we address the problem of scheduling a set of jobs on a server where processing times or due-dates of jobs are random variables and fixed weights (penalties) are imposed on the jobs’ late deliveries. The goal is to find the schedule that minimizes the expected weighted number of tardy jobs. The problem is NP-hard to solve; however, we explore three scenarios of the problem wherein: (i) both processing times and due-dates are stochastic; (ii) processing times are stochastic and due-dates are deterministic; and (iii) processing times are deterministic and due-dates are stochastic. We prove that special cases of these scenarios are solvable optimally in polynomial time, and introduce efficient heuristic methods for the general cases. Our computational results show that the heuristics perform well in yielding either optimal or near optimal sequences. The results also demonstrate that the stochasticity of processing times or due-dates can affect scheduling decisions. Moreover, the proposed problem is general in the sense that its special cases reduce to some new and some classical stochastic single machine models.

Keywords: number of late jobs, scheduling, single server, stochastic

Procedia PDF Downloads 497
8729 Airport Pavement Crack Measurement Systems and Crack Density for Pavement Evaluation

Authors: Ali Ashtiani, Hamid Shirazi

Abstract:

This paper reviews the status of existing practice and research related to measuring pavement cracking and using crack density as a pavement surface evaluation protocol. Crack density for pavement evaluation is currently not widely used within the airport community and its use by the highway community is limited. However, surface cracking is a distress that is closely monitored by airport staff and significantly influences the development of maintenance, rehabilitation and reconstruction plans for airport pavements. Therefore crack density has the potential to become an important indicator of pavement condition if the type, severity and extent of surface cracking can be accurately measured. A pavement distress survey is an essential component of any pavement assessment. Manual crack surveying has been widely used for decades to measure pavement performance. However, the accuracy and precision of manual surveys can vary depending upon the surveyor and performing surveys may disrupt normal operations. Given the variability of manual surveys, this method has shown inconsistencies in distress classification and measurement. This can potentially impact the planning for pavement maintenance, rehabilitation and reconstruction and the associated funding strategies. A substantial effort has been devoted for the past 20 years to reduce the human intervention and the error associated with it by moving toward automated distress collection methods. The automated methods refer to the systems that identify, classify and quantify pavement distresses through processes that require no or very minimal human intervention. This principally involves the use of a digital recognition software to analyze and characterize pavement distresses. The lack of established protocols for measurement and classification of pavement cracks captured using digital images is a challenge to developing a reliable automated system for distress assessment. Variations in types and severity of distresses, different pavement surface textures and colors and presence of pavement joints and edges all complicate automated image processing and crack measurement and classification. This paper summarizes the commercially available systems and technologies for automated pavement distress evaluation. A comprehensive automated pavement distress survey involves collection, interpretation, and processing of the surface images to identify the type, quantity and severity of the surface distresses. The outputs can be used to quantitatively calculate the crack density. The systems for automated distress survey using digital images reviewed in this paper can assist the airport industry in the development of a pavement evaluation protocol based on crack density. Analysis of automated distress survey data can lead to a crack density index. This index can be used as a means of assessing pavement condition and to predict pavement performance. This can be used by airport owners to determine the type of pavement maintenance and rehabilitation in a more consistent way.

Keywords: airport pavement management, crack density, pavement evaluation, pavement management

Procedia PDF Downloads 185
8728 Determinants of Financial Structure in the Economic Institution

Authors: Abdous Noureddine

Abstract:

The problem of funding in Algeria emerged as a problem you need to study after many Algerians researchers pointed out that the faltering Algerian public economic institution due to the imbalance in the financial structures and lower steering and marketing efficiency, as well as a result of severe expansion of borrowing because of inadequate own resources, and the consequent inability This institution to repay loans and interest payments, in addition to increasing reliance on overdraft so used to finance fixed assets, no doubt that this deterioration requires research and study of the causes and aspects of treatment, which addresses the current study, aside from it.

Keywords: financial structure, financial capital, equity, debt, firm’s value, return, leverage

Procedia PDF Downloads 312
8727 Study on the Non-Contact Sheet Resistance Measuring of Silver Nanowire Coated Film Using Terahertz Wave

Authors: Dong-Hyun Kim, Wan-Ho Chung, Hak-Sung Kim

Abstract:

In this work, non-destructive evaluation was conducted to measure the sheet resistance of silver nanowire coated film and find a damage of that film using terahertz (THz) wave. Pulse type THz instrument was used, and the measurement was performed under transmission and pitch-catch reflection modes with 30 degree of incidence angle. In the transmission mode, the intensity of the THz wave was gradually increased as the conductivity decreased. Meanwhile, the intensity of THz wave was decreased as the conductivity decreased in the pitch-catch reflection mode. To confirm the conductivity of the film, sheet resistance was measured by 4-point probe station. Interaction formula was drawn from a relation between the intensity and the sheet resistance. Through substituting sheet resistance to the formula and comparing the resultant value with measured maximum THz wave intensity, measurement of sheet resistance using THz wave was more suitable than that using 4-point probe station. In addition, the damage on the silver nanowire coated film was detected by applying the THz image system. Therefore, the reliability of the entire film can be also be ensured. In conclusion, real-time monitoring using the THz wave can be applied in the transparent electrodes with detecting the damaged area as well as measuring the sheet resistance.

Keywords: terahertz wave, sheet resistance, non-destructive evaluation, silver nanowire

Procedia PDF Downloads 490
8726 Multi-Objective Variable Neighborhood Search Algorithm to Solving Scheduling Problem with Transportation Times

Authors: Majid Khalili

Abstract:

This paper deals with a bi-objective hybrid no-wait flowshop scheduling problem minimizing the makespan and total weighted tardiness, in which we consider transportation times between stages. Obtaining an optimal solution for this type of complex, large-sized problem in reasonable computational time by using traditional approaches and optimization tools is extremely difficult. This paper presents a new multi-objective variable neighborhood algorithm (MOVNS). A set of experimental instances are carried out to evaluate the algorithm by advanced multi-objective performance measures. The algorithm is carefully evaluated for its performance against available algorithm by means of multi-objective performance measures and statistical tools. The related results show that a variant of our proposed MOVNS provides sound performance comparing with other algorithms.

Keywords: no-wait hybrid flowshop scheduling; multi-objective variable neighborhood algorithm; makespan; total weighted tardiness

Procedia PDF Downloads 418
8725 Density Measurement of Underexpanded Jet Using Stripe Patterned Background Oriented Schlieren Method

Authors: Shinsuke Udagawa, Masato Yamagishi, Masanori Ota

Abstract:

The Schlieren method, which has been conventionally used to visualize high-speed flows, has disadvantages such as the complexity of the experimental setup and the inability to quantitatively analyze the amount of refraction of light. The Background Oriented Schlieren (BOS) method proposed by Meier is one of the measurement methods that solves the problems, as mentioned above. The refraction of light is used for BOS method same as the Schlieren method. The BOS method is characterized using a digital camera to capture the images of the background behind the observation area. The images are later analyzed by a computer to quantitatively detect the amount of shift of the background image. The experimental setup for BOS does not require concave mirrors, pinholes, or color filters, which are necessary in the conventional Schlieren method, thus simplifying the experimental setup. However, the defocusing of the observation results is caused in case of using BOS method. Since the focus of camera on the background image leads to defocusing of the observed object. The defocusing of object becomes greater with increasing the distance between the background and the object. On the other hand, the higher sensitivity can be obtained. Therefore, it is necessary to adjust the distance between the background and the object to be appropriate for the experiment, considering the relation between the defocus and the sensitivity. The purpose of this study is to experimentally clarify the effect of defocus on density field reconstruction. In this study, the visualization experiment of underexpanded jet using BOS measurement system with ronchi ruling as the background that we constructed, have been performed. The reservoir pressure of the jet and the distance between camera and axis of jet is fixed, and the distance between background and axis of jet has been changed as the parameter. The images have been later analyzed by using personal computer to quantitatively detect the amount of shift of the background image from the comparison between the background pattern and the captured image of underexpanded jet. The quantitatively measured amount of shift have been reconstructed into a density flow field using the Abel transformation and the Gradstone-Dale equation. From the experimental results, it is found that the reconstructed density image becomes blurring, and noise becomes decreasing with increasing the distance between background and axis of underexpanded jet. Consequently, it is cralified that the sensitivity constant should be greater than 20, and the circle of confusion diameter should be less than 2.7mm at least in this experimental setup.

Keywords: BOS method, underexpanded jet, abel transformation, density field visualization

Procedia PDF Downloads 78
8724 Enhanced Calibration Map for a Four-Hole Probe for Measuring High Flow Angles

Authors: Jafar Mortadha, Imran Qureshi

Abstract:

This research explains and compares the modern techniques used for measuring the flow angles of a flowing fluid with the traditional technique of using multi-hole pressure probes. In particular, the focus of the study is on four-hole probes, which offer great reliability and benefits in several applications where the use of modern measurement techniques is either inconvenient or impractical. Due to modern advancements in manufacturing, small multi-hole pressure probes can be made with high precision, which eliminates the need for calibrating every manufactured probe. This study aims to improve the range of calibration maps for a four-hole probe to allow high flow angles to be measured accurately. The research methodology comprises a literature review of the successful calibration definitions that have been implemented on five-hole probes. These definitions are then adapted and applied on a four-hole probe using a set of raw pressures data. A comparison of the different definitions will be carried out in Matlab and the results will be analyzed to determine the best calibration definition. Taking simplicity of implementation into account as well as the reliability of flow angles estimation, an adapted technique from a research paper written in 2002 offered the most promising outcome. Consequently, the method is seen as a good enhancement for four-hole probes and it can substitute for the existing calibration definitions that offer less accuracy.

Keywords: calibration definitions, calibration maps, flow measurement techniques, four-hole probes, multi-hole pressure probes

Procedia PDF Downloads 295
8723 Characteristic Function in Estimation of Probability Distribution Moments

Authors: Vladimir S. Timofeev

Abstract:

In this article the problem of distributional moments estimation is considered. The new approach of moments estimation based on usage of the characteristic function is proposed. By statistical simulation technique, author shows that new approach has some robust properties. For calculation of the derivatives of characteristic function there is used numerical differentiation. Obtained results confirmed that author’s idea has a certain working efficiency and it can be recommended for any statistical applications.

Keywords: characteristic function, distributional moments, robustness, outlier, statistical estimation problem, statistical simulation

Procedia PDF Downloads 504
8722 Spherical Harmonic Based Monostatic Anisotropic Point Scatterer Model for RADAR Applications

Authors: Eric Huang, Coleman DeLude, Justin Romberg, Saibal Mukhopadhyay, Madhavan Swaminathan

Abstract:

High performance computing (HPC) based emulators can be used to model the scattering from multiple stationary and moving targets for RADAR applications. These emulators rely on the RADAR Cross Section (RCS) of the targets being available in complex scenarios. Representing the RCS using tables generated from electromagnetic (EM) simulations is often times cumbersome leading to large storage requirement. This paper proposed a spherical harmonic based anisotropic scatterer model to represent the RCS of complex targets. The problem of finding the locations and reflection profiles of all scatterers can be formulated as a linear least square problem with a special sparsity constraint. This paper solves this problem using a modified Orthogonal Matching Pursuit algorithm. The results show that the spherical harmonic based scatterer model can effectively represent the RCS data of complex targets.

Keywords: RADAR, RCS, high performance computing, point scatterer model

Procedia PDF Downloads 191
8721 A Test Methodology to Measure the Open-Loop Voltage Gain of an Operational Amplifier

Authors: Maninder Kaur Gill, Alpana Agarwal

Abstract:

It is practically not feasible to measure the open-loop voltage gain of the operational amplifier in the open loop configuration. It is because the open-loop voltage gain of the operational amplifier is very large. In order to avoid the saturation of the output voltage, a very small input should be given to operational amplifier which is not possible to be measured practically by a digital multimeter. A test circuit for measurement of open loop voltage gain of an operational amplifier has been proposed and verified using simulation tools as well as by experimental methods on breadboard. The main advantage of this test circuit is that it is simple, fast, accurate, cost effective, and easy to handle even on a breadboard. The test circuit requires only the device under test (DUT) along with resistors. This circuit has been tested for measurement of open loop voltage gain for different operational amplifiers. The underlying goal is to design testable circuits for various analog devices that are simple to realize in VLSI systems, giving accurate results and without changing the characteristics of the original system. The DUTs used are LM741CN and UA741CP. For LM741CN, the simulated gain and experimentally measured gain (average) are calculated as 89.71 dB and 87.71 dB, respectively. For UA741CP, the simulated gain and experimentally measured gain (average) are calculated as 101.15 dB and 105.15 dB, respectively. These values are found to be close to the datasheet values.

Keywords: Device Under Test (DUT), open loop voltage gain, operational amplifier, test circuit

Procedia PDF Downloads 447
8720 A Low Cost Non-Destructive Grain Moisture Embedded System for Food Safety and Quality

Authors: Ritula Thakur, Babankumar S. Bansod, Puneet Mehta, S. Chatterji

Abstract:

Moisture plays an important role in storage, harvesting and processing of food grains and related agricultural products. It is an important characteristic of most agricultural products for maintenance of quality. Accurate knowledge of the moisture content can be of significant value in maintaining quality and preventing contamination of cereal grains. The present work reports the design and development of microcontroller based low cost non-destructive moisture meter, which uses complex impedance measurement method for moisture measurement of wheat using parallel plate capacitor arrangement. Moisture can conveniently be sensed by measuring the complex impedance using a small parallel-plate capacitor sensor filled with the kernels in-between the two plates of sensor, exciting the sensor at 30 KHz and 100 KHz frequencies. The effects of density and temperature variations were compensated by providing suitable compensations in the developed algorithm. The results were compared with standard dry oven technique and the developed method was found to be highly accurate with less than 1% error. The developed moisture meter is low cost, highly accurate, non-destructible method for determining the moisture of grains utilizing the fast computing capabilities of microcontroller.

Keywords: complex impedance, moisture content, electrical properties, safety of food

Procedia PDF Downloads 462
8719 Network Connectivity Knowledge Graph Using Dwave Quantum Hybrid Solvers

Authors: Nivedha Rajaram

Abstract:

Hybrid Quantum solvers have been given prime focus in recent days by computation problem-solving domain industrial applications. D’Wave Quantum Computers are one such paragon of systems built using quantum annealing mechanism. Discrete Quadratic Models is a hybrid quantum computing model class supplied by D’Wave Ocean SDK - a real-time software platform for hybrid quantum solvers. These hybrid quantum computing modellers can be employed to solve classic problems. One such problem that we consider in this paper is finding a network connectivity knowledge hub in a huge network of systems. Using this quantum solver, we try to find out the prime system hub, which acts as a supreme connection point for the set of connected computers in a large network. This paper establishes an innovative problem approach to generate a connectivity system hub plot for a set of systems using DWave ocean SDK hybrid quantum solvers.

Keywords: quantum computing, hybrid quantum solver, DWave annealing, network knowledge graph

Procedia PDF Downloads 127
8718 Ultrasonic Evaluation of Periodic Rough Inaccessible Surfaces from Back Side

Authors: Chanh Nghia Nguyen, Yu Kurokawa, Hirotsugu Inoue

Abstract:

The surface roughness is an important parameter for evaluating the quality of material surfaces since it affects functions and performance of industrial components. Although stylus and optical techniques are commonly used for measuring the surface roughness, they are applicable only to accessible surfaces. In practice, surface roughness measurement from the back side is sometimes demanded, for example, in inspection of safety-critical parts such as inner surface of pipes. However, little attention has been paid to the measurement of back surface roughness so far. Since back surface is usually inaccessible by stylus or optical techniques, ultrasonic technique is one of the most effective among others. In this research, an ultrasonic pulse-echo technique is considered for evaluating the pitch and the height of back surface having periodic triangular profile as a very first step. The pitch of the surface profile is measured by applying the diffraction grating theory for oblique incidence; then the height is evaluated by numerical analysis based on the Kirchhoff theory for normal incidence. The validity of the proposed method was verified by both numerical simulation and experiment. It was confirmed that the pitch is accurately measured in most cases. The height was also evaluated with good accuracy when it is smaller than a half of the pitch because of the approximation in the Kirchhoff theory.

Keywords: back side, inaccessible surface, periodic roughness, pulse-echo technique, ultrasonic NDE

Procedia PDF Downloads 275
8717 Vehicle Routing Problem with Mixed Fleet of Conventional and Heterogenous Electric Vehicles and Time Dependent Charging Costs

Authors: Ons Sassi, Wahiba Ramdane Cherif-Khettaf, Ammar Oulamara

Abstract:

In this paper, we consider a new real-life Heterogenous Electric Vehicle Routing Problem with Time Dependant Charging Costs and a Mixed Fleet (HEVRP-TDMF), in which a set of geographically scattered customers have to be served by a mixed fleet of vehicles composed of a heterogenous fleet of Electric Vehicles (EVs), having different battery capacities and operating costs, and Conventional Vehicles (CVs). We include the possibility of charging EVs in the available charging stations during the routes in order to serve all customers. Each charging station offers charging service with a known technology of chargers and time-dependent charging costs. Charging stations are also subject to operating time windows constraints. EVs are not necessarily compatible with all available charging technologies and a partial charging is allowed. Intermittent charging at the depot is also allowed provided that constraints related to the electricity grid are satisfied. The objective is to minimize the number of employed vehicles and then minimize the total travel and charging costs. In this study, we present a Mixed Integer Programming Model and develop a Charging Routing Heuristic and a Local Search Heuristic based on the Inject-Eject routine with three different insertion strategies. All heuristics are tested on real data instances.

Keywords: charging problem, electric vehicle, heuristics, local search, optimization, routing problem

Procedia PDF Downloads 463
8716 Graph Cuts Segmentation Approach Using a Patch-Based Similarity Measure Applied for Interactive CT Lung Image Segmentation

Authors: Aicha Majda, Abdelhamid El Hassani

Abstract:

Lung CT image segmentation is a prerequisite in lung CT image analysis. Most of the conventional methods need a post-processing to deal with the abnormal lung CT scans such as lung nodules or other lesions. The simplest similarity measure in the standard Graph Cuts Algorithm consists of directly comparing the pixel values of the two neighboring regions, which is not accurate because this kind of metrics is extremely sensitive to minor transformations such as noise or other artifacts problems. In this work, we propose an improved version of the standard graph cuts algorithm based on the Patch-Based similarity metric. The boundary penalty term in the graph cut algorithm is defined Based on Patch-Based similarity measurement instead of the simple intensity measurement in the standard method. The weights between each pixel and its neighboring pixels are Based on the obtained new term. The graph is then created using theses weights between its nodes. Finally, the segmentation is completed with the minimum cut/Max-Flow algorithm. Experimental results show that the proposed method is very accurate and efficient, and can directly provide explicit lung regions without any post-processing operations compared to the standard method.

Keywords: graph cuts, lung CT scan, lung parenchyma segmentation, patch-based similarity metric

Procedia PDF Downloads 169
8715 A Two-Dimensional Problem Micropolar Thermoelastic Medium under the Effect of Laser Irradiation and Distributed Sources

Authors: Devinder Singh, Rajneesh Kumar, Arvind Kumar

Abstract:

The present investigation deals with the deformation of micropolar generalized thermoelastic solid subjected to thermo-mechanical loading due to a thermal laser pulse. Laplace transform and Fourier transform techniques are used to solve the problem. Thermo-mechanical laser interactions are taken as distributed sources to describe the application of the approach. The closed form expressions of normal stress, tangential stress, coupled stress and temperature are obtained in the domain. Numerical inversion technique of Laplace transform and Fourier transform has been implied to obtain the resulting quantities in the physical domain after developing a computer program. The normal stress, tangential stress, coupled stress and temperature are depicted graphically to show the effect of relaxation times. Some particular cases of interest are deduced from the present investigation.

Keywords: pulse laser, integral transform, thermoelastic, boundary value problem

Procedia PDF Downloads 616
8714 Elephant Herding Optimization for Service Selection in QoS-Aware Web Service Composition

Authors: Samia Sadouki Chibani, Abdelkamel Tari

Abstract:

Web service composition combines available services to provide new functionality. Given the number of available services with similar functionalities and different non functional aspects (QoS), the problem of finding a QoS-optimal web service composition is considered as an optimization problem belonging to NP-hard class. Thus, an optimal solution cannot be found by exact algorithms within a reasonable time. In this paper, a meta-heuristic bio-inspired is presented to address the QoS aware web service composition; it is based on Elephant Herding Optimization (EHO) algorithm, which is inspired by the herding behavior of elephant group. EHO is characterized by a process of dividing and combining the population to sub populations (clan); this process allows the exchange of information between local searches to move toward a global optimum. However, with Applying others evolutionary algorithms the problem of early stagnancy in a local optimum cannot be avoided. Compared with PSO, the results of experimental evaluation show that our proposition significantly outperforms the existing algorithm with better performance of the fitness value and a fast convergence.

Keywords: bio-inspired algorithms, elephant herding optimization, QoS optimization, web service composition

Procedia PDF Downloads 327
8713 The Effectiveness of Adaptive Difficulty Adjustment in Touch Tablet App on Young Children's Spatial Problem Solving Development

Authors: Chenchen Liu, Jacques Audran

Abstract:

Using tablet apps with a certain educational purpose to promote young children’s cognitive development, is quite common now. Developing an educational app on an Ipad like tablet, especially for a young child (age 3-5) requires an optimal level of challenge to continuously attract children’s attention and obtain an educational effect. Adaptive difficulty adjustment, which could dynamically set the difficulty in the challenge according to children’s performance, seems to be a good solution. Since space concept plays an important role in young children’s cognitive development, we made an experimental comparison in a French kindergarten between one group of 23 children using an educational app ‘Debout Ludo’ with adaptive difficulty settings and another group of 20 children using the previous version of ‘Debout Ludo’ with a classic incremental difficulty adjustment. The experiment results of spatial problem solving indicated that a significantly higher learning outcome was acquired by the young children who used the adaptive version of the app.

Keywords: adaptive difficulty, spatial problem solving, tactile tablet, young children

Procedia PDF Downloads 444
8712 Compressed Sensing of Fetal Electrocardiogram Signals Based on Joint Block Multi-Orthogonal Least Squares Algorithm

Authors: Xiang Jianhong, Wang Cong, Wang Linyu

Abstract:

With the rise of medical IoT technologies, Wireless body area networks (WBANs) can collect fetal electrocardiogram (FECG) signals to support telemedicine analysis. The compressed sensing (CS)-based WBANs system can avoid the sampling of a large amount of redundant information and reduce the complexity and computing time of data processing, but the existing algorithms have poor signal compression and reconstruction performance. In this paper, a Joint block multi-orthogonal least squares (JBMOLS) algorithm is proposed. We apply the FECG signal to the Joint block sparse model (JBSM), and a comparative study of sparse transformation and measurement matrices is carried out. A FECG signal compression transmission mode based on Rbio5.5 wavelet, Bernoulli measurement matrix, and JBMOLS algorithm is proposed to improve the compression and reconstruction performance of FECG signal by CS-based WBANs. Experimental results show that the compression ratio (CR) required for accurate reconstruction of this transmission mode is increased by nearly 10%, and the runtime is saved by about 30%.

Keywords: telemedicine, fetal ECG, compressed sensing, joint sparse reconstruction, block sparse signal

Procedia PDF Downloads 128
8711 Measurement of Natural Radioactivity and Health Hazard Index Evaluation in Major Soils of Tin Mining Areas of Perak

Authors: Habila Nuhu

Abstract:

Natural radionuclides in the environment can significantly contribute to human exposure to ionizing radiation. The knowledge of their levels in an environment can help the radiological protection agencies in policymaking. Measurement of natural radioactivity in major soils in the tin mining state of Perak Malaysia has been conducted using an HPGe detector. Seventy (70) soil samples were collected at widely distributed locations in the state. Six major soil types were sampled, and thirteen districts around the state were covered. The following were the results of the 226Ra (238U), 228Ra (232Th), and 40K activity in the soil samples: 226Ra (238U) has a mean activity concentration of 191.83 Bq kg⁻¹, more than five times the UNSCEAR reference limits of 35 Bq kg⁻¹. The mean activity concentration of 228Ra (232Th) with a value of 232.41 Bq kg⁻¹ is over seven times the UNSCEAR reference values of 30 Bq kg⁻¹. The average concentration of 40K activity was 275.24 Bq kg⁻¹, which was less than the UNSCEAR reference limit of 400 Bq Kg⁻¹. The range of external hazards index (Hₑₓ) values was from 1.03 to 2.05, while the internal hazards index (Hin) was from 1.48 to 3.08. The Hex and Hin should be less than one for minimal external and internal radiation threats as well as secure use of soil material for building construction. The Hₑₓ and Hin results generally indicate that while using the soil types and their derivatives as building materials in the study area, care must be taken.

Keywords: activity concentration, hazard index, soil samples, tin mining

Procedia PDF Downloads 111
8710 Eliminating Arm, Neck and Leg Fatigue of United Asia International Plastics Corporation Workers through Rapid Entire Body Assessment

Authors: John Cheferson R. De Belen, John Paul G. Elizares, Ronald John G. Raz, Janina Elyse A. Reyes, Charie G. Salengua, Aristotle L. Soriano

Abstract:

Plastic is a type of synthetic or man-made polymer that can readily be molded into a variety of products. Its usage over the past century has enabled society to make huge technological advances. The workers of United Asia International Plastics Corporation (UAIPC), a plastic manufacturing company performs manual packaging which causes fatigue and stress on their arm, neck, and legs due to extended periods of standing and repetitive motions. With the use of the Fishbone Diagram, Five-Why Analysis, Rapid Entire Body Assessment (REBA), and Anthropometry, the stressful tasks and activities were identified and analyzed. Given the anthropometric measurements obtained from the workers, improved dimensions for the tables and chairs should be used and provide a new packaging machine. The validation of this proposal shall follow after its implementation. By eliminating fatigue during working hours in the production, the workers will be at ease at performing their work properly; productivity will increase that will lead to more profit. Further areas for study include measurement and comparison of the worker’s anthropometric measurement with the industry standard.

Keywords: anthropometry, fishbone diagram, five-why analysis, rapid entire body assessment

Procedia PDF Downloads 266
8709 Wavelet Method for Numerical Solution of Fourth Order Wave Equation

Authors: A. H. Choudhury

Abstract:

In this paper, a highly accurate numerical method for the solution of one-dimensional fourth-order wave equation is derived. This hyperbolic problem is solved by using semidiscrete approximations. The space direction is discretized by wavelet-Galerkin method, and the time variable is discretized by using Newmark schemes.

Keywords: hyperbolic problem, semidiscrete approximations, stability, Wavelet-Galerkin Method

Procedia PDF Downloads 315
8708 A Mathematical Model for a Two-Stage Assembly Flow-Shop Scheduling Problem with Batch Delivery System

Authors: Saeedeh Ahmadi Basir, Mohammad Mahdavi Mazdeh, Mohammad Namakshenas

Abstract:

Manufacturers often dispatch jobs in batches to reduce delivery costs. However, sending several jobs in batches can have a negative effect on other scheduling-related objective functions such as minimizing the number of tardy jobs which is often used to rate managers’ performance in many manufacturing environments. This paper aims to minimize the number of weighted tardy jobs and the sum of delivery costs of a two-stage assembly flow-shop problem in a batch delivery system. We present a mixed-integer linear programming (MILP) model to solve the problem. As this is an MILP model, the commercial solver (the CPLEX solver) is not guaranteed to find the optimal solution for large-size problems at a reasonable amount of time. We present several numerical examples to confirm the accuracy of the model.

Keywords: scheduling, two-stage assembly flow-shop, tardy jobs, batched delivery system

Procedia PDF Downloads 460
8707 Enhancing Signal Reception in a Mobile Radio Network Using Adaptive Beamforming Antenna Arrays Technology

Authors: Ugwu O. C., Mamah R. O., Awudu W. S.

Abstract:

This work is aimed at enhancing signal reception on a mobile radio network and minimizing outage probability in a mobile radio network using adaptive beamforming antenna arrays. In this research work, an empirical real-time drive measurement was done in a cellular network of Globalcom Nigeria Limited located at Ikeja, the headquarters of Lagos State, Nigeria, with reference base station number KJA 004. The empirical measurement includes Received Signal Strength and Bit Error Rate which were recorded for exact prediction of the signal strength of the network as at the time of carrying out this research work. The Received Signal Strength and Bit Error Rate were measured with a spectrum monitoring Van with the help of a Ray Tracer at an interval of 100 meters up to 700 meters from the transmitting base station. The distance and angular location measurements from the reference network were done with the help Global Positioning System (GPS). The other equipment used were transmitting equipment measurements software (Temsoftware), Laptops and log files, which showed received signal strength with distance from the base station. Results obtained were about 11% from the real-time experiment, which showed that mobile radio networks are prone to signal failure and can be minimized using an Adaptive Beamforming Antenna Array in terms of a significant reduction in Bit Error Rate, which implies improved performance of the mobile radio network. In addition, this work did not only include experiments done through empirical measurement but also enhanced mathematical models that were developed and implemented as a reference model for accurate prediction. The proposed signal models were based on the analysis of continuous time and discrete space, and some other assumptions. These developed (proposed) enhanced models were validated using MATLAB (version 7.6.3.35) program and compared with the conventional antenna for accuracy. These outage models were used to manage the blocked call experience in the mobile radio network. 20% improvement was obtained when the adaptive beamforming antenna arrays were implemented on the wireless mobile radio network.

Keywords: beamforming algorithm, adaptive beamforming, simulink, reception

Procedia PDF Downloads 41
8706 Relationship Between Quetelet Equation and Skin Fold Teckniques in Determining Obesity Among Adolescents in Maiduguri, Borno State, Nigeria

Authors: A. Kaidal, M. M. Abdllahi, O. L. Badaki

Abstract:

The study was conducted to determine the relationship between Quetelet Equation and Skin fold measurement in determining obesity among adolescent male students of University of Maiduguri Demonstration Secondary School, Borno State, Nigeria. A total of 66 students participated in the study, their age ranges from 15-18 years. The ex-post-facto research design was used for this study. Anthropometric measurements were taken at three sites (thigh, abdomen and chest) using accu–measure Skin fold caliper. The values of the three measurements were used to determine the percentage body fat of the participants using the 3-Point Skin Fold Bodyfat calculator of Jackson-Pollock. Body mass index (BMI) was determined using weight (kg) divided by height in (m2). The data obtained was analyzed using descriptive statistics (mean and standard deviation) and inferential statistics of Pearson product moment correlation coefficient to determine the relationship between the two techniques. The result showed a significant positive relationship r=0.673 p<0.05 between body mass index and skin fold measurement techniques. It was however observed that BMI techniques of determining body fat tend to overestimate the actual percent body fat of adolescents studied. Based on this result, it is recommended that the use of BMI as a technique for determining obesity should be used with caution.

Keywords: body max index, skin fold, quetelet, techniques

Procedia PDF Downloads 542
8705 BeamGA Median: A Hybrid Heuristic Search Approach

Authors: Ghada Badr, Manar Hosny, Nuha Bintayyash, Eman Albilali, Souad Larabi Marie-Sainte

Abstract:

The median problem is significantly applied to derive the most reasonable rearrangement phylogenetic tree for many species. More specifically, the problem is concerned with finding a permutation that minimizes the sum of distances between itself and a set of three signed permutations. Genomes with equal number of genes but different order can be represented as permutations. In this paper, an algorithm, namely BeamGA median, is proposed that combines a heuristic search approach (local beam) as an initialization step to generate a number of solutions, and then a Genetic Algorithm (GA) is applied in order to refine the solutions, aiming to achieve a better median with the smallest possible reversal distance from the three original permutations. In this approach, any genome rearrangement distance can be applied. In this paper, we use the reversal distance. To the best of our knowledge, the proposed approach was not applied before for solving the median problem. Our approach considers true biological evolution scenario by applying the concept of common intervals during the GA optimization process. This allows us to imitate a true biological behavior and enhance genetic approach time convergence. We were able to handle permutations with a large number of genes, within an acceptable time performance and with same or better accuracy as compared to existing algorithms.

Keywords: median problem, phylogenetic tree, permutation, genetic algorithm, beam search, genome rearrangement distance

Procedia PDF Downloads 265
8704 Sustainable Balanced Scorecard for Kaizen Evaluation: Comparative Study between Egypt and Japan

Authors: Ola I. S. El Dardery, Ismail Gomaa, Adel R.M. Rayan, Ghada El Khayat, Sara H. Sabry

Abstract:

Continuous improvement activities are becoming a key factor of the success of any organization, those improvement activities include but not limited to kaizen, six sigma, lean projects, and continuous improvement projects. Kaizen is a Japanese philosophy of continuous improvement by making small incremental changes to improve an organization’s performance, reduce costs, reduce delay time, reduce waste in production, etc. This study aims at proposing a new measuring technique for kaizen activities using a Sustainable balanced scorecard structure. A survey questionnaire was developed and introduced to kaizen participants in both Egypt and Japan with the purpose of allocating key performance indicators for both kaizen process (critical success factors) and result (kaizen benefits) into the five perspectives of sustainable balanced scorecard. The study contributes to the literature by presenting a new kaizen measurement of both kaizen process and results, that will illuminate the benefits of using kaizen. Also, the presented measurement can help in the sustainability of kaizen implementation. Determining the combination of the proper kaizen measures could be used by any industry whether service or manufacturing to better measure kaizen activates. The comparison between Japanese measures, as the leaders of kaizen philosophy, and Egyptian measures will help recommending better practices of kaizen in Egypt, and contributing to the 2030 sustainable development goals.

Keywords: continuous improvements, kaizen, performance, sustainable balanced scorecard

Procedia PDF Downloads 148
8703 Analysis of the Inverse Kinematics for 5 DOF Robot Arm Using D-H Parameters

Authors: Apurva Patil, Maithilee Kulkarni, Ashay Aswale

Abstract:

This paper proposes an algorithm to develop the kinematic model of a 5 DOF robot arm. The formulation of the problem is based on finding the D-H parameters of the arm. Brute Force iterative method is employed to solve the system of non linear equations. The focus of the paper is to obtain the accurate solutions by reducing the root mean square error. The result obtained will be implemented to grip the objects. The trajectories followed by the end effector for the required workspace coordinates are plotted. The methodology used here can be used in solving the problem for any other kinematic chain of up to six DOF.

Keywords: 5 DOF robot arm, D-H parameters, inverse kinematics, iterative method, trajectories

Procedia PDF Downloads 203
8702 Effects of Fishbone Creative Thinking Strategy on Problem-Solving Skills of Teaching Personnel in Ogun State, Nigeria

Authors: Olusegun Adeleke Adenuga

Abstract:

The study examined effect of fishbone creative thinking strategy on problem-solving skills of public teachers in Ogun state, Nigeria. A 2x2x2 factorial design was employed for the study which consisted of 80 participants made up of 40 male and 40 female public teachers randomly selected among public teaching personnel from the two local government area headquarters (Ijebu-ode and Ijebu-Igbo) within Ogun East Senatorial District. Each treatment group received 45minutes instructions and training per week for 8weeks. Data was collected from participants with the use of standardized instrument tagged ‘Problem Solving Inventory’ (PSI) developed by the researchers prior to the training to form a pre-test and immediately after eight weeks of training to form a post-test. One hypothesis was tested; the data obtained was analyzed using Analysis of Covariance (ANCOVA) tested at significance level of 0.05. The result of the data analysis shows that there was a significant effect of the fishbone creative thinking technique on the participants (F (2,99) = 12.410; p <.05). Based on the findings, it is therefore recommended that the report of this study be used to effect organizational change and development of teaching service in Nigeria through teachers’ retraining and capacity building.

Keywords: fishbone, creative thinking strategy, and problem-solving skills, public teachers

Procedia PDF Downloads 356