Search results for: input impedance
1607 The Fluid Limit of the Critical Processor Sharing Tandem Queue
Authors: Amal Ezzidani, Abdelghani Ben Tahar, Mohamed Hanini
Abstract:
A sequence of finite tandem queue is considered for this study. Each one has a single server, which operates under the egalitarian processor sharing discipline. External customers arrive at each queue according to a renewal input process and having a general service times distribution. Upon completing service, customers leave the current queue and enter to the next. Under mild assumptions, including critical data, we prove the existence and the uniqueness of the fluid solution. For asymptotic behavior, we provide necessary and sufficient conditions for the invariant state and the convergence to this invariant state. In the end, we establish the convergence of a correctly normalized state process to a fluid limit characterized by a system of algebraic and integral equations.Keywords: fluid limit, fluid model, measure valued process, processor sharing, tandem queue
Procedia PDF Downloads 3231606 Fuzzy Multi-Component DEA with Shared and Undesirable Fuzzy Resources
Authors: Jolly Puri, Shiv Prasad Yadav
Abstract:
Multi-component data envelopment analysis (MC-DEA) is a popular technique for measuring aggregate performance of the decision making units (DMUs) along with their components. However, the conventional MC-DEA is limited to crisp input and output data which may not always be available in exact form. In real life problems, data may be imprecise or fuzzy. Therefore, in this paper, we propose (i) a fuzzy MC-DEA (FMC-DEA) model in which shared and undesirable fuzzy resources are incorporated, (ii) the proposed FMC-DEA model is transformed into a pair of crisp models using cut approach, (iii) fuzzy aggregate performance of a DMU and fuzzy efficiencies of components are defined to be fuzzy numbers, and (iv) a numerical example is illustrated to validate the proposed approach.Keywords: multi-component DEA, fuzzy multi-component DEA, fuzzy resources, decision making units (DMUs)
Procedia PDF Downloads 4071605 Development of an Automatic Computational Machine Learning Pipeline to Process Confocal Fluorescence Images for Virtual Cell Generation
Authors: Miguel Contreras, David Long, Will Bachman
Abstract:
Background: Microscopy plays a central role in cell and developmental biology. In particular, fluorescence microscopy can be used to visualize specific cellular components and subsequently quantify their morphology through development of virtual-cell models for study of effects of mechanical forces on cells. However, there are challenges with these imaging experiments, which can make it difficult to quantify cell morphology: inconsistent results, time-consuming and potentially costly protocols, and limitation on number of labels due to spectral overlap. To address these challenges, the objective of this project is to develop an automatic computational machine learning pipeline to predict cellular components morphology for virtual-cell generation based on fluorescence cell membrane confocal z-stacks. Methods: Registered confocal z-stacks of nuclei and cell membrane of endothelial cells, consisting of 20 images each, were obtained from fluorescence confocal microscopy and normalized through software pipeline for each image to have a mean pixel intensity value of 0.5. An open source machine learning algorithm, originally developed to predict fluorescence labels on unlabeled transmitted light microscopy cell images, was trained using this set of normalized z-stacks on a single CPU machine. Through transfer learning, the algorithm used knowledge acquired from its previous training sessions to learn the new task. Once trained, the algorithm was used to predict morphology of nuclei using normalized cell membrane fluorescence images as input. Predictions were compared to the ground truth fluorescence nuclei images. Results: After one week of training, using one cell membrane z-stack (20 images) and corresponding nuclei label, results showed qualitatively good predictions on training set. The algorithm was able to accurately predict nuclei locations as well as shape when fed only fluorescence membrane images. Similar training sessions with improved membrane image quality, including clear lining and shape of the membrane, clearly showing the boundaries of each cell, proportionally improved nuclei predictions, reducing errors relative to ground truth. Discussion: These results show the potential of pre-trained machine learning algorithms to predict cell morphology using relatively small amounts of data and training time, eliminating the need of using multiple labels in immunofluorescence experiments. With further training, the algorithm is expected to predict different labels (e.g., focal-adhesion sites, cytoskeleton), which can be added to the automatic machine learning pipeline for direct input into Principal Component Analysis (PCA) for generation of virtual-cell mechanical models.Keywords: cell morphology prediction, computational machine learning, fluorescence microscopy, virtual-cell models
Procedia PDF Downloads 2051604 Modelling and Simulation of the Freezing Systems and Heat Pumps Using Unisim® Design
Authors: C. Patrascioiu
Abstract:
The paper describes the modeling and simulation of the heat pumps domain processes. The main objective of the study is the use of the heat pump in propene–propane distillation processes. The modeling and simulation instrument is the Unisim® Design simulator. The paper is structured in three parts: An overview of the compressing gases, the modeling and simulation of the freezing systems, and the modeling and simulation of the heat pumps. For each of these systems, there are presented the Unisim® Design simulation diagrams, the input–output system structure and the numerical results. Future studies will consider modeling and simulation of the propene–propane distillation process with heat pump.Keywords: distillation, heat pump, simulation, unisim design
Procedia PDF Downloads 3631603 Energy Efficient Alternate Hydraulic System Called TejHydroLift
Authors: Tejinder Singh
Abstract:
This paper describes a new more efficient Hydraulic System which uses lesser work to produce more output. Conventional Hydraulic System like Hydraulic Lifts and Rams use lots of water to be pumped to produce output. TejHydroLift will do the equal amount of force with lesser input of water. The paper will show that force applied can be increased manifold without requiring to move smaller force by more distance which used to be required in Conventional Hydraulic Lifts. The paper describes one of the configurations of TejHydroLift System called “Slim Antenna TejHydroLift Configuration”. The TejHydroLift uses lesser water and hence demands lesser work to be performed to move the same load.Keywords: alternate, hydraulic system, efficient, TejHydroLift
Procedia PDF Downloads 2601602 Bayesian Optimization for Reaction Parameter Tuning: An Exploratory Study of Parameter Optimization in Oxidative Desulfurization of Thiophene
Authors: Aman Sharma, Sonali Sengupta
Abstract:
The study explores the utility of Bayesian optimization in tuning the physical and chemical parameters of reactions in an offline experimental setup. A comparative analysis of the influence of the acquisition function on the optimization performance is also studied. For proxy first and second-order reactions, the results are indifferent to the acquisition function used, whereas, while studying the parameters for oxidative desulphurization of thiophene in an offline setup, upper confidence bound (UCB) provides faster convergence along with a marginal trade-off in the maximum conversion achieved. The work also demarcates the critical number of independent parameters and input observations required for both sequential and offline reaction setups to yield tangible results.Keywords: acquisition function, Bayesian optimization, desulfurization, kinetics, thiophene
Procedia PDF Downloads 1821601 Verifiable Secure Computation of Large Scale Two-Point Boundary Value Problems Using Certificate Validation
Authors: Yogita M. Ahire, Nedal M. Mohammed, Ahmed A. Hamoud
Abstract:
Scientific computation outsourcing is gaining popularity because it allows customers with limited computing resources and storage devices to outsource complex computation workloads to more powerful service providers. However, it raises some security and privacy concerns and challenges, such as customer input and output privacy, as well as cloud cheating behaviors. This study was motivated by these concerns and focused on privacy-preserving Two-Point Boundary Value Problems (BVP) as a common and realistic instance for verifiable safe multiparty computing. We'll look at the safe and verifiable schema with correctness guarantees by utilizing standard multiparty approaches to compute the result of a computation and then solely using verifiable ways to check that the result was right.Keywords: verifiable computing, cloud computing, secure and privacy BVP, secure computation outsourcing
Procedia PDF Downloads 971600 Passive Seismic in Hydrogeological Prospecting: The Case Study from Hard Rock and Alluvium Plain
Authors: Prarabdh Tiwari, M. Vidya Sagar, K. Bhima Raju, Joy Choudhury, Subash Chandra, E. Nagaiah, Shakeel Ahmed
Abstract:
Passive seismic, a wavefield interferometric imaging, low cost and rapid tool for subsurface investigation is used for various geotechnical purposes such as hydrocarbon exploration, seismic microzonation, etc. With the recent advancement, its application has also been extended to groundwater exploration by means of finding the bedrock depth. Council of Scientific & Industrial Research (CSIR)-National Geophysical Research Institute (NGRI) has experimented passive seismic studies along with electrical resistivity tomography for groundwater in hard rock (Choutuppal, Hyderabad). Passive Seismic with Electrical Resistivity (ERT) can give more clear 2-D subsurface image for Groundwater Exploration in Hard Rock area. Passive seismic data were collected using a Tromino, a three-component broadband seismometer, to measure background ambient noise and processed using GRILLA software. The passive seismic results are found corroborating with ERT (Electrical Resistivity Tomography) results. For data acquisition purpose, Tromino was kept over 30 locations consist recording of 20 minutes at each station. These location shows strong resonance frequency peak, suggesting good impedance contrast between different subsurface layers (ex. Mica rich Laminated layer, Weathered layer, granite, etc.) This paper presents signature of passive seismic for hard rock terrain. It has been found that passive seismic has potential application for formation characterization and can be used as an alternative tool for delineating litho-stratification in an urban condition where electrical and electromagnetic tools cannot be applied due to high cultural noise. In addition to its general application in combination with electrical and electromagnetic methods can improve the interpreted subsurface model.Keywords: passive seismic, resonant frequency, Tromino, GRILLA
Procedia PDF Downloads 1881599 Efficiency of DMUs in Presence of New Inputs and Outputs in DEA
Authors: Esmat Noroozi, Elahe Sarfi, Farha Hosseinzadeh Lotfi
Abstract:
Examining the impacts of data modification is considered as sensitivity analysis. A lot of studies have considered the data modification of inputs and outputs in DEA. The issues which has not heretofore been considered in DEA sensitivity analysis is modification in the number of inputs and (or) outputs and determining the impacts of this modification in the status of efficiency of DMUs. This paper is going to present systems that show the impacts of adding one or multiple inputs or outputs on the status of efficiency of DMUs and furthermore a model is presented for recognizing the minimum number of inputs and (or) outputs from among specified inputs and outputs which can be added whereas an inefficient DMU will become efficient. Finally the presented systems and model have been utilized for a set of real data and the results have been reported.Keywords: data envelopment analysis, efficiency, sensitivity analysis, input, out put
Procedia PDF Downloads 4501598 Optimization of Process Parameters by Using Taguchi Method for Bainitic Steel Machining
Authors: Vinay Patil, Swapnil Kekade, Ashish Supare, Vinayak Pawar, Shital Jadhav, Rajkumar Singh
Abstract:
In recent days, bainitic steel is used in automobile and non-automobile sectors due to its high strength. Bainitic steel is difficult to machine because of its high hardness, hence in this paper machinability of bainitic steel is studied by using Taguchi design of experiments (DOE) approach. Convectional turning experiments were done by using L16 orthogonal array for three input parameters viz. cutting speed, depth of cut and feed. The Taguchi method is applied to study the performance characteristics of machining parameters with surface roughness (Ra), cutting force and tool wear rate. By using Taguchi analysis, optimized process parameters for best surface finish and minimum cutting forces were analyzed.Keywords: conventional turning, Taguchi method, S/N ratio, bainitic steel machining
Procedia PDF Downloads 3311597 A Neural Approach for the Offline Recognition of the Arabic Handwritten Words of the Algerian Departments
Authors: Salim Ouchtati, Jean Sequeira, Mouldi Bedda
Abstract:
In this work we present an off line system for the recognition of the Arabic handwritten words of the Algerian departments. The study is based mainly on the evaluation of neural network performances, trained with the gradient back propagation algorithm. The used parameters to form the input vector of the neural network are extracted on the binary images of the handwritten word by several methods: the parameters of distribution, the moments centered of the different projections and the Barr features. It should be noted that these methods are applied on segments gotten after the division of the binary image of the word in six segments. The classification is achieved by a multi layers perceptron. Detailed experiments are carried and satisfactory recognition results are reported.Keywords: handwritten word recognition, neural networks, image processing, pattern recognition, features extraction
Procedia PDF Downloads 5131596 The Analysis of TRACE/FRAPTRAN in the Fuel Rods of Maanshan PWR for LBLOCA
Authors: J. R. Wang, W. Y. Li, H. T. Lin, J. H. Yang, C. Shih, S. W. Chen
Abstract:
Fuel rod analysis program transient (FRAPTRAN) code was used to study the fuel rod performance during a postulated large break loss of coolant accident (LBLOCA) in Maanshan nuclear power plant (NPP). Previous transient results from thermal hydraulic code, TRACE, with the same LBLOCA scenario, were used as input boundary conditions for FRAPTRAN. The simulation results showed that the peak cladding temperatures and the fuel center line temperatures were all below the 10CFR50.46 LOCA criteria. In addition, the maximum hoop stress was 18 MPa and the oxide thickness was 0.003 mm for the present simulation cases, which are all within the safety operation ranges. The present study confirms that this analysis method, the FRAPTRAN code combined with TRACE, is an appropriate approach to predict the fuel integrity under LBLOCA with operational ECCS.Keywords: FRAPTRAN, TRACE, LOCA, PWR
Procedia PDF Downloads 5111595 Kernel Parallelization Equation for Identifying Structures under Unknown and Periodic Loads
Authors: Seyed Sadegh Naseralavi
Abstract:
This paper presents a Kernel parallelization equation for damage identification in structures under unknown periodic excitations. Herein, the dynamic differential equation of the motion of structure is viewed as a mapping from displacements to external forces. Utilizing this viewpoint, a new method for damage detection in structures under periodic loads is presented. The developed method requires only two periods of load. The method detects the damages without finding the input loads. The method is based on the fact that structural displacements under free and forced vibrations are associated with two parallel subspaces in the displacement space. Considering the concept, kernel parallelization equation (KPE) is derived for damage detection under unknown periodic loads. The method is verified for a case study under periodic loads.Keywords: Kernel, unknown periodic load, damage detection, Kernel parallelization equation
Procedia PDF Downloads 2841594 Developing a Rational Database Management System (RDBMS) Supporting Product Life Cycle Appications
Authors: Yusri Yusof, Chen Wong Keong
Abstract:
This paper presents the implementation details of a Relational Database Management System of a STEP-technology product model repository. It is able support the implementation of any EXPRESS language schema, although it has been primarily implemented to support mechanical product life cycle applications. This database support the input of STEP part 21 file format from CAD in geometrical and topological data format and support a range of queries for mechanical product life cycle applications. This proposed relational database management system uses entity-to-table method (R1) rather than type-to-table method (R4). The two mapping methods have their own strengths and drawbacks.Keywords: RDBMS, CAD, ISO 10303, part-21 file
Procedia PDF Downloads 5361593 The Effect of Core Training on Physical Fitness Characteristics in Male Volleyball Players
Authors: Sibel Karacaoglu, Fatma Ç. Kayapinar
Abstract:
The aim of the study is to investigate the effect of the core training program on physical fitness characteristics and body composition in male volleyball players. 26 male university volleyball team players aged between 19 to 24 years who had no health problems and injury participated in the study. Subjects were divided into training (TG) and control groups (CG) as randomly. Data from twenty-one players who completed all training sessions were used for statistical analysis (TG,n=11; CG,n=10). A core training program was applied to the training group three days a week for 10 weeks. On the other hand, the control group did not receive any training. Before and after the 10-week training program, pre- and post-testing comprised of body composition measurements (weight, BMI, bioelectrical impedance analysis) and physical fitness measurements including flexibility (sit and reach test), muscle strength (back, leg and grip strength by dynamometer), muscle endurance (sit-ups and push-ups tests), power (one-legged jump and vertical jump tests), speed (20m sprint, 30m sprint) and balance tests (one-legged standing test) were performed. Changes of pre- and post- test values of the groups were determined by using dependent t test. According to the statistical analysis of data, no significant difference was found in terms of body composition in the both groups for pre- and post- test values. In the training group, all physical fitness measurements improved significantly after core training program (p<0.05) except 30m speed and handgrip strength (p>0.05). On the hand, only 20m speed test values improved after post-test period (p<0.05), but the other physical fitness tests values did not differ (p>0.05) between pre- and post- test measurement in the control group. The results of the study suggest that the core training program has positive effect on physical fitness characteristics in male volleyball players.Keywords: body composition, core training, physical fitness, volleyball
Procedia PDF Downloads 3461592 First Investigation on CZTS Electron affinity and Thickness Optimization using SILVACO-Atlas 2D Simulation
Authors: Zeineb Seboui, Samar Dabbabi
Abstract:
In this paper, we study the performance of Cu₂ZnSnS₄ (CZTS) based solar cell. In our knowledge, it is for the first time that the FTO/ZnO:Co/CZTS structure is simulated using the SILVACO-Atlas 2D simulation. Cu₂ZnSnS₄ (CZTS), ZnO:Co and FTO (SnO₂:F) layers have been deposited on glass substrates by the spray pyrolysis technique. The extracted physical properties, such as thickness and optical parameters of CZTS layer, are considered to create a new input data of CZTS based solar cell. The optimization of CZTS electron affinity and thickness is performed to have the best FTO/ZnO: Co/CZTS efficiency. The use of CZTS absorber layer with 3.99 eV electron affinity and 3.2 µm in thickness leads to the higher efficiency of 16.86 %, which is very important in the development of new technologies and new solar cell devices.Keywords: CZTS solar cell, characterization, electron affinity, thickness, SILVACO-atlas 2D simulation
Procedia PDF Downloads 781591 Using Scale Invariant Feature Transform Features to Recognize Characters in Natural Scene Images
Authors: Belaynesh Chekol, Numan Çelebi
Abstract:
The main purpose of this work is to recognize individual characters extracted from natural scene images using scale invariant feature transform (SIFT) features as an input to K-nearest neighbor (KNN); a classification learner algorithm. For this task, 1,068 and 78 images of English alphabet characters taken from Chars74k data set is used to train and test the classifier respectively. For each character image, We have generated describing features by using SIFT algorithm. This set of features is fed to the learner so that it can recognize and label new images of English characters. Two types of KNN (fine KNN and weighted KNN) were trained and the resulted classification accuracy is 56.9% and 56.5% respectively. The training time taken was the same for both fine and weighted KNN.Keywords: character recognition, KNN, natural scene image, SIFT
Procedia PDF Downloads 2811590 Small Scale Batch Anaerobic Digestion of Rice Straw
Authors: V. H. Nguyen, A. Castalone, C. Jamieson, M. Gummert
Abstract:
Rice straw is an abundant biomass resource in Asian countries that can be used for bioenergy. In continuously flooded rice fields, it can be removed without reducing the levels of soil organic matter. One suitable bioenergy technology is anaerobic digestion (AD), but it needs to be further verified using rice straw as a feedstock. For this study, a batch AD system was developed using rice straw and cow dung. It is low cost, farm scale, with the batch capacity ranging from 5 kg to 200 kg of straw mixed with 10% of cow dung. The net energy balance obtained was from 3000 to 4000 MJ per ton of straw input at 15-18% moisture content. Net output energy obtained from biogas and digestate ranged from 4000 to 5000 MJ per ton of straw. This indicates AD as a potential solution for converting rice straw from a waste to a clean fuel, reducing the environmental footprint caused by current disposal practices.Keywords: rice straw, anaerobic digestion, biogas, bioenergy
Procedia PDF Downloads 3521589 Frequency-Dependent and Full Range Tunable Phase Shifter
Authors: Yufu Yin, Tao Lin, Shanghong Zhao, Zihang Zhu, Xuan Li, Wei Jiang, Qiurong Zheng, Hui Wang
Abstract:
In this paper, a frequency-dependent and tunable phase shifter is proposed and numerically analyzed. The key devices are the dual-polarization binary phase shift keying modulator (DP-BPSK) and the fiber Bragg grating (FBG). The phase-frequency response of the FBG is employed to determine the frequency-dependent phase shift. The simulation results show that a linear phase shift of the recovered output microwave signal which depends on the frequency of the input RF signal is achieved. In addition, by adjusting the power of the RF signal, the full range phase shift from 0° to 360° can be realized. This structure shows the spurious free dynamic range (SFDR) of 70.90 dB·Hz2/3 and 72.11 dB·Hz2/3 under different RF powers.Keywords: microwave photonics, phase shifter, spurious free dynamic range, frequency-dependent
Procedia PDF Downloads 2961588 Microwave Plasma Dry Reforming of Methane at High CO2/CH4 Feed Ratio
Authors: Nabil Majd Alawi, Gia Hung Pham, Ahmed Barifcani
Abstract:
Dry reforming of methane that converts two greenhouses gases (CH4 and CO2) to synthesis gas (a mixture of H2 and CO) was studied in a commercial bench scale microwave (MW) plasma reactor system at atmospheric pressure. The CO2, CH4 and N2 conversions; H2, CO selectivities and yields, and syngas ratio (H2/CO) were investigated in a wide range of total feed flow rate (0.45 – 2.1 L/min), MW power (700 – 1200 watt) and CO2/CH4 molar ratio (2 – 5). At the feed flow rates of CH4, CO2 and N2 of 0.2, 0.4 and 1.5 L/min respectively, and the MWs input power of 700 W, the highest conversions of CH4 and CO2, selectivity and yield of H2, CO and H2/CO ratio of 79.35%, 44.82%, 50.12, 58.42, 39.77%, 32.89%, and 0.86, respectively, were achieved. The results of this work show that the product ratio increases slightly with the increasing total feed flow rate, but it decreases significantly with the increasing MW power and feeds CO2/CH4 ratio.Keywords: dry reforming of methane, microwave discharge, plasma technology, synthesis gas production
Procedia PDF Downloads 2741587 Speaker Recognition Using LIRA Neural Networks
Authors: Nestor A. Garcia Fragoso, Tetyana Baydyk, Ernst Kussul
Abstract:
This article contains information from our investigation in the field of voice recognition. For this purpose, we created a voice database that contains different phrases in two languages, English and Spanish, for men and women. As a classifier, the LIRA (Limited Receptive Area) grayscale neural classifier was selected. The LIRA grayscale neural classifier was developed for image recognition tasks and demonstrated good results. Therefore, we decided to develop a recognition system using this classifier for voice recognition. From a specific set of speakers, we can recognize the speaker’s voice. For this purpose, the system uses spectrograms of the voice signals as input to the system, extracts the characteristics and identifies the speaker. The results are described and analyzed in this article. The classifier can be used for speaker identification in security system or smart buildings for different types of intelligent devices.Keywords: extreme learning, LIRA neural classifier, speaker identification, voice recognition
Procedia PDF Downloads 1771586 Hardware Implementation on Field Programmable Gate Array of Two-Stage Algorithm for Rough Set Reduct Generation
Authors: Tomasz Grzes, Maciej Kopczynski, Jaroslaw Stepaniuk
Abstract:
The rough sets theory developed by Prof. Z. Pawlak is one of the tools that can be used in the intelligent systems for data analysis and processing. Banking, medicine, image recognition and security are among the possible fields of utilization. In all these fields, the amount of the collected data is increasing quickly, but with the increase of the data, the computation speed becomes the critical factor. Data reduction is one of the solutions to this problem. Removing the redundancy in the rough sets can be achieved with the reduct. A lot of algorithms of generating the reduct were developed, but most of them are only software implementations, therefore have many limitations. Microprocessor uses the fixed word length, consumes a lot of time for either fetching as well as processing of the instruction and data; consequently, the software based implementations are relatively slow. Hardware systems don’t have these limitations and can process the data faster than a software. Reduct is the subset of the decision attributes that provides the discernibility of the objects. For the given decision table there can be more than one reduct. Core is the set of all indispensable condition attributes. None of its elements can be removed without affecting the classification power of all condition attributes. Moreover, every reduct consists of all the attributes from the core. In this paper, the hardware implementation of the two-stage greedy algorithm to find the one reduct is presented. The decision table is used as an input. Output of the algorithm is the superreduct which is the reduct with some additional removable attributes. First stage of the algorithm is calculating the core using the discernibility matrix. Second stage is generating the superreduct by enriching the core with the most common attributes, i.e., attributes that are more frequent in the decision table. Described above algorithm has two disadvantages: i) generating the superreduct instead of reduct, ii) additional first stage may be unnecessary if the core is empty. But for the systems focused on the fast computation of the reduct the first disadvantage is not the key problem. The core calculation can be achieved with a combinational logic block, and thus add respectively little time to the whole process. Algorithm presented in this paper was implemented in Field Programmable Gate Array (FPGA) as a digital device consisting of blocks that process the data in a single step. Calculating the core is done by the comparators connected to the block called 'singleton detector', which detects if the input word contains only single 'one'. Calculating the number of occurrences of the attribute is performed in the combinational block made up of the cascade of the adders. The superreduct generation process is iterative and thus needs the sequential circuit for controlling the calculations. For the research purpose, the algorithm was also implemented in C language and run on a PC. The times of execution of the reduct calculation in a hardware and software were considered. Results show increase in the speed of data processing.Keywords: data reduction, digital systems design, field programmable gate array (FPGA), reduct, rough set
Procedia PDF Downloads 2191585 Risk of Disrupted Eating Attitudes in Disabled Athletes
Authors: Zehra Buyuktuncer, Aylin H. Büyükkaragöz, Tuğçe N. Balcı, Nevin Ergun
Abstract:
Background: Undergoing rigid dietary habits for enhancing athletic performance could lead to eating disorders. High prevalence of eating disorders among female athletes has been already reported. However, the risk of disordered eating among disabled athletes is not known. A better knowledge of the different eating behaviors and their prevalence in disabled athletes would be helpful to understand interactions between eating and health. This study aimed to examine the cognitive restraint, uncontrolled eating and emotional eating behaviors in a disabled athlete population. Method: A total of 70 disabled Turkish national athletes (33 female, 37 male) from 5 sport branches (soccer, weight lifting, shooting, table tennis and basketball) were involved in the study. The cognitive restraint, uncontrolled eating and emotional eating behaviors were assessed using the revised version of Three Factor Eating Questionnaire-R18 (TFEQ-R18). The questionnaires were conducted by dietitian during the preparation camps of athletes. Body weight, height and waist circumference (WC) were measured; and body composition was analyzed by bioelectrical impedance analysis method. Results: The TFEQ scales showed a cognitive dietary restraint score of 13.9±4.2, uncontrolled eating score of 17.7±5.8 and emotional eating score of 4.9±2.5. The mean score of total TFEQ-R18 was 36.5±8.62. Neither total TFEQ-R18 score nor subscale scores differed significantly by gender or sport branches (p>0.05, for each). The scores were also similar in BMI groups (n=63; p>0.05). Total TFEQ, uncontrolled eating and emotional eating scores were significantly higher among the athletes with congenital disabilities compared to the scores of the athletes with acquired disabilities (p<0.05, for each). Moreover, the cognitive dietary restraint score was significantly high in athletes who would like to lose weight (p=0.009). Conclusion: Disabled athletes might have a risk of disordered eating. The different eating behaviors among disabled athletes should be assessed using validated tools to develop personalized nutritional strategies for those athletes.Keywords: disabled athletes, eating behaviour, three-factor eating questionnaire-r18, body composition
Procedia PDF Downloads 3351584 Investigation on Dry Sliding Wear for Laser Cladding of Stellite 6 Produced on a P91 Steel Substrate
Authors: Alain Kusmoko, Druce Dunne, Huijun Li
Abstract:
Stellite 6 was deposited by laser cladding on a chromium bearing substrate (P91) with energy inputs of 1 kW (P91-1) and 1.8 kW (P91-1.8). The chemical compositions and microstructures of these coatings were characterized by atomic absorption spectroscopy, optical microscopy and scanning electron microscopy. The microhardness of the coatings was measured and the wear mechanism of the coatings was assessed using a pin-on-plate (reciprocating) wear testing machine. The results showed less cracking and pore development for Stellite 6 coatings applied to the P91 steel substrate with the lower heat input (P91-1). Further, the Stellite coating for P91-1 was significantly harder than that obtained for P91-1.8. The wear test results indicated that the weight loss for P91-1 was much lower than for P91-1.8. It is concluded that the lower hardness of the coating for P91-1.8, together with the softer underlying substrate structure, markedly reduced the wear resistance of the Stellite 6 coating.Keywords: friction and wear, laser cladding, P91 steel, Stellite 6 coating
Procedia PDF Downloads 4411583 Black-Box-Base Generic Perturbation Generation Method under Salient Graphs
Authors: Dingyang Hu, Dan Liu
Abstract:
DNN (Deep Neural Network) deep learning models are widely used in classification, prediction, and other task scenarios. To address the difficulties of generic adversarial perturbation generation for deep learning models under black-box conditions, a generic adversarial ingestion generation method based on a saliency map (CJsp) is proposed to obtain salient image regions by counting the factors that influence the input features of an image on the output results. This method can be understood as a saliency map attack algorithm to obtain false classification results by reducing the weights of salient feature points. Experiments also demonstrate that this method can obtain a high success rate of migration attacks and is a batch adversarial sample generation method.Keywords: adversarial sample, gradient, probability, black box
Procedia PDF Downloads 1041582 Evaluation of Batch Splitting in the Context of Load Scattering
Authors: S. Wesebaum, S. Willeke
Abstract:
Production companies are faced with an increasingly turbulent business environment, which demands very high production volumes- and delivery date flexibility. If a decoupling by storage stages is not possible (e.g. at a contract manufacturing company) or undesirable from a logistical point of view, load scattering effects the production processes. ‘Load’ characterizes timing and quantity incidence of production orders (e.g. in work content hours) to workstations in the production, which results in specific capacity requirements. Insufficient coordination between load (demand capacity) and capacity supply results in heavy load scattering, which can be described by deviations and uncertainties in the input behavior of a capacity unit. In order to respond to fluctuating loads, companies try to implement consistent and realizable input behavior using the capacity supply available. For example, a uniform and high level of equipment capacity utilization keeps production costs down. In contrast, strong load scattering at workstations leads to performance loss or disproportionately fluctuating WIP, whereby the logistics objectives are affected negatively. Options for reducing load scattering are e.g. shifting the start and end dates of orders, batch splitting and outsourcing of operations or shifting to other workstations. This leads to an adjustment of load to capacity supply, and thus to a reduction of load scattering. If the adaptation of load to capacity cannot be satisfied completely, possibly flexible capacity must be used to ensure that the performance of a workstation does not decrease for a given load. Where the use of flexible capacities normally raises costs, an adjustment of load to capacity supply reduces load scattering and, in consequence, costs. In the literature you mostly find qualitative statements for describing load scattering. Quantitative evaluation methods that describe load mathematically are rare. In this article the authors discuss existing approaches for calculating load scattering and their various disadvantages such as lack of opportunity for normalization. These approaches are the basis for the development of our mathematical quantification approach for describing load scattering that compensates the disadvantages of the current quantification approaches. After presenting our mathematical quantification approach, the method of batch splitting will be described. Batch splitting allows the adaptation of load to capacity to reduce load scattering. After describing the method, it will be explicitly analyzed in the context of the logistic curve theory by Nyhuis using the stretch factor α1 in order to evaluate the impact of the method of batch splitting on load scattering and on logistic curves. The conclusion of this article will be to show how the methods and approaches presented can help companies in a turbulent environment to quantify the occurring work load scattering accurately and apply an efficient method for adjusting work load to capacity supply. In this way, the achievements of the logistical objectives are increased without causing additional costs.Keywords: batch splitting, production logistics, production planning and control, quantification, load scattering
Procedia PDF Downloads 3991581 Analysis of the Relationship between the Unitary Impulse Response for the nth-Volterra Kernel of a Duffing Oscillator System
Authors: Guillermo Manuel Flores Figueroa, Juan Alejandro Vazquez Feijoo, Jose Navarro Antonio
Abstract:
A continuous nonlinear system response may be obtained by an infinite sum of the so-called Volterra operators. Each operator is obtained from multidimensional convolution of nth-order between the nth-order Volterra kernel and the system input. These operators can also be obtained from the Associated Linear Equations (ALEs) that are linear models of subsystems which inputs and outputs are of the same nth-order. Each ALEs produces a particular nth-Volterra operator. As linear models a unitary impulse response can be obtained from them. This work shows the relationship between this unitary impulse responses and the corresponding order Volterra kernel.Keywords: Volterra series, frequency response functions FRF, associated linear equations ALEs, unitary response function, Voterra kernel
Procedia PDF Downloads 6701580 Immediate Geometric Solution of Irregular Quadrilaterals: A Digital Tool Applied to Topography
Authors: Miguel Mariano Rivera Galvan
Abstract:
The purpose of this research was to create a digital tool by which users can obtain an immediate and accurate solution of the angular characteristics of an irregular quadrilateral. The development of this project arose because of the frequent absence of a polygon’s geometric information in land ownership accreditation documents. The researcher created a mathematical model using a linear approximation iterative method, employing various disciplines and techniques including trigonometry, geometry, algebra, and topography. This mathematical model uses as input data the surface of the quadrilateral, as well as the length of its sides, to obtain its interior angles and make possible its representation in a coordinate system. The results are as accurate and reliable as the user requires, offering the possibility of using this tool as a support to develop future engineering and architecture projects quickly and reliably.Keywords: digital tool, geometry, mathematical model, quadrilateral, solution
Procedia PDF Downloads 1471579 Investigating the Form of the Generalised Equations of Motion of the N-Bob Pendulum and Computing Their Solution Using MATLAB
Authors: Divij Gupta
Abstract:
Pendular systems have a range of both mathematical and engineering applications, ranging from modelling the behaviour of a continuous mass-density rope to utilisation as Tuned Mass Dampers (TMD). Thus, it is of interest to study the differential equations governing the motion of such systems. Here we attempt to generalise these equations of motion for the plane compound pendulum with a finite number of N point masses. A Lagrangian approach is taken, and we attempt to find the generalised form for the Euler-Lagrange equations of motion for the i-th bob of the N -bob pendulum. The co-ordinates are parameterized as angular quantities to reduce the number of degrees of freedom from 2N to N to simplify the form of the equations. We analyse the form of these equations up to N = 4 to determine the general form of the equation. We also develop a MATLAB program to compute a solution to the system for a given input value of N and a given set of initial conditions.Keywords: classical mechanics, differential equation, lagrangian analysis, pendulum
Procedia PDF Downloads 2081578 Adaptive Filtering in Subbands for Supervised Source Separation
Authors: Bruna Luisa Ramos Prado Vasques, Mariane Rembold Petraglia, Antonio Petraglia
Abstract:
This paper investigates MIMO (Multiple-Input Multiple-Output) adaptive filtering techniques for the application of supervised source separation in the context of convolutive mixtures. From the observation that there is correlation among the signals of the different mixtures, an improvement in the NSAF (Normalized Subband Adaptive Filter) algorithm is proposed in order to accelerate its convergence rate. Simulation results with mixtures of speech signals in reverberant environments show the superior performance of the proposed algorithm with respect to the performances of the NLMS (Normalized Least-Mean-Square) and conventional NSAF, considering both the convergence speed and SIR (Signal-to-Interference Ratio) after convergence.Keywords: adaptive filtering, multi-rate processing, normalized subband adaptive filter, source separation
Procedia PDF Downloads 435