Search results for: Four Phase Methodology
2195 Analysis of the Interference from Risk-Determining Factors of Cooperative and Conventional Construction Contracts
Authors: E. Harrer, M. Mauerhofer, T. Werginz
Abstract:
As a result of intensive competition, the building sector is suffering from a high degree of rivalry. Furthermore, there can be observed an unbalanced distribution of project risks. Clients are aimed to shift their own risks into the sphere of the constructors or planners. The consequence of this is that the number of conflicts between the involved parties is inordinately high or even increasing; an alternative approach to counter on that developments are cooperative project forms in the construction sector. This research compares conventional contract models and models with partnering agreements to examine the influence on project risks by an early integration of the involved parties. The goal is to show up deviations in different project stages from the design phase to the project transfer phase. These deviations are evaluated by a survey of experts from the three spheres: clients, contractors and planners. By rating the influence of the participants on specific risk factors it is possible to identify factors which are relevant for a smooth project execution.Keywords: Collaborative work, construction industry, contract-models, influence, partnering, project management, risk.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8692194 Implementation of Conceptual Real-Time Embedded Functional Design via Drive-by-Wire ECU Development
Authors: A. Ukaew, C. Chauypen
Abstract:
Design concepts of real-time embedded system can be realized initially by introducing novel design approaches. In this literature, model based design approach and in-the-loop testing were employed early in the conceptual and preliminary phase to formulate design requirements and perform quick real-time verification. The design and analysis methodology includes simulation analysis, model based testing, and in-the-loop testing. The design of conceptual driveby- wire, or DBW, algorithm for electronic control unit, or ECU, was presented to demonstrate the conceptual design process, analysis, and functionality evaluation. The concepts of DBW ECU function can be implemented in the vehicle system to improve electric vehicle, or EV, conversion drivability. However, within a new development process, conceptual ECU functions and parameters are needed to be evaluated. As a result, the testing system was employed to support conceptual DBW ECU functions evaluation. For the current setup, the system components were consisted of actual DBW ECU hardware, electric vehicle models, and control area network or CAN protocol. The vehicle models and CAN bus interface were both implemented as real-time applications where ECU and CAN protocol functionality were verified according to the design requirements. The proposed system could potentially benefit in performing rapid real-time analysis of design parameters for conceptual system or software algorithm development.Keywords: Drive-by-wire ECU, in-the-loop testing, modelbased design, real-time embedded system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21762193 Evaluation of Environmental, Technical, and Economic Indicators of a Fused Deposition Modeling Process
Authors: M. Yosofi, S. Ezeddini, A. Ollivier, V. Lavaste, C. Mayousse
Abstract:
Additive manufacturing processes have changed significantly in a wide range of industries and their application progressed from rapid prototyping to production of end-use products. However, their environmental impact is still a rather open question. In order to support the growth of this technology in the industrial sector, environmental aspects should be considered and predictive models may help monitor and reduce the environmental footprint of the processes. This work presents predictive models based on a previously developed methodology for the environmental impact evaluation combined with a technical and economical assessment. Here we applied the methodology to the Fused Deposition Modeling process. First, we present the predictive models relative to different types of machines. Then, we present a decision-making tool designed to identify the optimum manufacturing strategy regarding technical, economic, and environmental criteria.
Keywords: Additive manufacturing, decision-makings, environmental impact, predictive models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10602192 Visual Study on Flow Patterns and Heat Transfer during Convective Boiling Inside Horizontal Smooth and Microfin Tubes
Authors: V.D. Hatamipour, M.A. Akhavan-Behabadi
Abstract:
Evaporator is an important and widely used heat exchanger in air conditioning and refrigeration industries. Different methods have been used by investigators to increase the heat transfer rates in evaporators. One of the passive techniques to enhance heat transfer coefficient is the application of microfin tubes. The mechanism of heat transfer augmentation in microfin tubes is dependent on the flow regime of two-phase flow. Therefore many investigations of the flow patterns for in-tube evaporation have been reported in literatures. The gravitational force, surface tension and the vapor-liquid interfacial shear stress are known as three dominant factors controlling the vapor and liquid distribution inside the tube. A review of the existing literature reveals that the previous investigations were concerned with the two-phase flow pattern for flow boiling in horizontal tubes [12], [9]. Therefore, the objective of the present investigation is to obtain information about the two-phase flow patterns for evaporation of R-134a inside horizontal smooth and microfin tubes. Also Investigation of heat transfer during flow boiling of R-134a inside horizontal microfin and smooth tube have been carried out experimentally The heat transfer coefficients for annular flow in the smooth tube is shown to agree well with Gungor and Winterton-s correlation [4]. All the flow patterns occurred in the test can be divided into three dominant regimes, i.e., stratified-wavy flow, wavy-annular flow and annular flow. Experimental data are plotted in two kinds of flow maps, i.e., Weber number for the vapor versus weber number for the liquid flow map and mass flux versus vapor quality flow map. The transition from wavy-annular flow to annular or stratified-wavy flow is identified in the flow maps.Keywords: Flow boiling, Flow pattern, Heat transfer, Horizontal, Smooth tube, Microfin tube.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23302191 Haematology and Serum Biochemical Profile of Laying Chickens Reared on Deep Litter System with or without Access to Grass or Legume Pasture under Humid Tropical Climate
Authors: E. Oke, A. O. Ladokun, J. O. Daramola, O. M. Onagbesan
Abstract:
There has been a growing interest on the effects of access to pasture on poultry health status. However, there is a paucity of data on the relative benefits of grass and legume pastures. An experiment was conducted to determine the effects of rearing systems {deep litter system (DL), deep litter with access to legumes (LP) or grass (GP) pastures} haematology and serum chemistry of ISA Brown layers. The study involved the use of two hundred and forty 12 weeks old pullets. The birds were reared until 60 weeks of age. Eighty birds were assigned to each treatment; each treatment had four replicates of 20 birds each. Blood samples (2.5 ml) were collected from the wing vein of two birds per replicate and serum chemistry and haematological parameters were determined. The results showed that there were no significant differences between treatments in all the parameters considered at 18 weeks of age. At 24 weeks old, the percentage of heterophyl (HET) in DL and LP were similar but higher than that of GP. The ratio of H:L was higher (P<0.05) in DL than those of LP and GP while LP and GP were comparable. At week 38 of age, the percentage of PCV in the birds in LP and GP were similar but the birds in DL had significantly lower level than that of GP. In the early production phase, serum total protein of the birds in LP was similar to that of GP but higher (P<0.05) than that of DL. At the peak production phase (week 38), the total protein in GP and DL were similar but significantly lower than that of LP. The albumin level in LP was greater (P<0.05) than GP but similar to that of DL. In the late production phase, the total protein in LP was significantly higher than that of DL but similar to that of GP. It was concluded that rearing chickens in either grass or legume pasture did not have deleterious effects on the health of laying chickens but improved some parameters including blood protein and HET/lymphocyte.Keywords: Rearing systems, Stylosanthes, Cynodon serum chemistry, haematology, hen.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20652190 Optimization for Subcritical Water Extraction of Phenolic Compounds from Rambutan Peels
Authors: Nuttawan Yoswathana, M. N. Eshtiaghi
Abstract:
Rambutan is a tropical fruit which peel possesses antioxidant properties. This work was conducted to optimize extraction conditions of phenolic compounds from rambutan peel. Response surface methodology (RSM) was adopted to optimize subcritical water extraction (SWE) on temperature, extraction time and percent solvent mixture. The results demonstrated that the optimum conditions for SWE were as follows: temperature 160°C, extraction time 20min. and concentration of 50% ethanol. Comparison of the phenolic compounds from the rambutan peels in maceration 6h, soxhlet 4h, and SWE 20min., it indicated that total phenolic content (using Folin-Ciocalteu-s phenol reagent) was 26.42, 70.29, and 172.47mg of tannic acid equivalent (TAE) per g dry rambutan peel, respectively. The comparative study concluded that SWE was a promising technique for phenolic compounds extraction from rambutan peel, due to much more two times of conventional techniques and shorter extraction times.
Keywords: Subcritical water extraction, Rambutan peel, phenolic compounds, response surface methodology
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36532189 Performance of Subcarrier- OCDMA System with Complementary Subtraction Detection Technique
Authors: R. K. Z. Sahbudin, M. K. Abdullah, M. Mokhtar, S. B. A. Anas, S. Hitam
Abstract:
A subcarrier - spectral amplitude coding optical code division multiple access system using the Khazani-Syed code with Complementary subtraction detection technique is proposed. The proposed system has been analyzed by taking into account the effects of phase-induced intensity noise, shot noise, thermal noise and intermodulation distortion noise. The performance of the system has been compared with the spectral amplitude coding optical code division multiple access system using the Hadamard code and the Modified Quadratic Congruence code. The analysis shows that the proposed system can eliminate the multiple access interference using the Complementary subtraction detection technique, and hence improve the overall system performance.Keywords: Complementary subtraction, Khazani-Syed code, multiple access interference, phase-induced intensity noise
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17432188 Negative Selection as a Means of Discovering Unknown Temporal Patterns
Authors: Wanli Ma, Dat Tran, Dharmendra Sharma
Abstract:
The temporal nature of negative selection is an under exploited area. In a negative selection system, newly generated antibodies go through a maturing phase, and the survivors of the phase then wait to be activated by the incoming antigens after certain number of matches. These without having enough matches will age and die, while these with enough matches (i.e., being activated) will become active detectors. A currently active detector may also age and die if it cannot find any match in a pre-defined (lengthy) period of time. Therefore, what matters in a negative selection system is the dynamics of the involved parties in the current time window, not the whole time duration, which may be up to eternity. This property has the potential to define the uniqueness of negative selection in comparison with the other approaches. On the other hand, a negative selection system is only trained with “normal" data samples. It has to learn and discover unknown “abnormal" data patterns on the fly by itself. Consequently, it is more appreciate to utilize negation selection as a system for pattern discovery and recognition rather than just pattern recognition. In this paper, we study the potential of using negative selection in discovering unknown temporal patterns.
Keywords: Artificial Immune Systems, ComputationalIntelligence, Negative Selection, Pattern Discovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16652187 Applying Resilience Engineering to improve Safety Management in a Construction Site: Design and Validation of a Questionnaire
Authors: M. C. Pardo-Ferreira, J. C. Rubio-Romero, M. Martínez-Rojas
Abstract:
Resilience Engineering is a new paradigm of safety management that proposes to change the way of managing the safety to focus on the things that go well instead of the things that go wrong. Many complex and high-risk sectors such as air traffic control, health care, nuclear power plants, railways or emergencies, have applied this new vision of safety and have obtained very positive results. In the construction sector, safety management continues to be a problem as indicated by the statistics of occupational injuries worldwide. Therefore, it is important to improve safety management in this sector. For this reason, it is proposed to apply Resilience Engineering to the construction sector. The Construction Phase Health and Safety Plan emerges as a key element for the planning of safety management. One of the key tools of Resilience Engineering is the Resilience Assessment Grid that allows measuring the four essential abilities (respond, monitor, learn and anticipate) for resilient performance. The purpose of this paper is to develop a questionnaire based on the Resilience Assessment Grid, specifically on the ability to learn, to assess whether a Construction Phase Health and Safety Plans helps companies in a construction site to implement this ability. The research process was divided into four stages: (i) initial design of a questionnaire, (ii) validation of the content of the questionnaire, (iii) redesign of the questionnaire and (iii) application of the Delphi method. The questionnaire obtained could be used as a tool to help construction companies to evolve from Safety-I to Safety-II. In this way, companies could begin to develop the ability to learn, which will serve as a basis for the development of the other abilities necessary for resilient performance. The following steps in this research are intended to develop other questions that allow evaluating the rest of abilities for resilient performance such as monitoring, learning and anticipating.
Keywords: Resilience engineering, construction sector, resilience assessment grid, construction phase health and safety plan.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10022186 Mean Shift-based Preprocessing Methodology for Improved 3D Buildings Reconstruction
Authors: Nikolaos Vassilas, Theocharis Tsenoglou, Djamchid Ghazanfarpour
Abstract:
In this work, we explore the capability of the mean shift algorithm as a powerful preprocessing tool for improving the quality of spatial data, acquired from airborne scanners, from densely built urban areas. On one hand, high resolution image data corrupted by noise caused by lossy compression techniques are appropriately smoothed while at the same time preserving the optical edges and, on the other, low resolution LiDAR data in the form of normalized Digital Surface Map (nDSM) is upsampled through the joint mean shift algorithm. Experiments on both the edge-preserving smoothing and upsampling capabilities using synthetic RGB-z data show that the mean shift algorithm is superior to bilateral filtering as well as to other classical smoothing and upsampling algorithms. Application of the proposed methodology for 3D reconstruction of buildings of a pilot region of Athens, Greece results in a significant visual improvement of the 3D building block model.Keywords: 3D buildings reconstruction, data fusion, data upsampling, mean shift.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20052185 Development of a Nano-Alumina-Zirconia Composite Catalyst as an Active Thin Film in Biodiesel Production
Authors: N. Marzban, J. K. Heydarzadeh M. Pourmohammadbagher, M. H. Hatami, A. Samia
Abstract:
A nano-alumina-zirconia composite catalyst was synthesized by a simple aqueous sol-gel method using AlCl3.6H2O and ZrCl4 as precursors. Thermal decomposition of the precursor and subsequent formation of γ-Al2O3 and t-Zr were investigated by thermal analysis. XRD analysis showed that γ-Al2O3 and t-ZrO2 phases were formed at 700 °C. FT-IR analysis also indicated that the phase transition to γ-Al2O3 occurred in corroboration with X-ray studies. TEM analysis of the calcined powder revealed that spherical particles were in the range of 8-12 nm. The nano-alumina-zirconia composite particles were mesoporous and uniformly distributed in their crystalline phase. In order to measure the catalytic activity, esterification reaction was carried out. Biodiesel, as a renewable fuel, was formed in a continuous packed column reactor. Free fatty acid (FFA) was esterified with ethanol in a heterogeneous catalytic reactor. It was found that the synthesized γ-Al2O3/ZrO2 composite had the potential to be used as a heterogeneous base catalyst for biodiesel production processes.Keywords: Nano-alumina-zirconia, composite catalyst, thin film, biodiesel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17482184 Human Walking Vertical Force and Vertical Vibration of Pedestrian Bridge Induced by Its Higher Components
Authors: M. Yoneda
Abstract:
The purpose of this study is to identify human walking vertical force by using FFT power spectrum density from the experimental acceleration data of the human body. An experiment on human walking is carried out on a stationary floor especially paying attention to higher components of dynamic vertical walking force. Based on measured acceleration data of the human lumbar part, not only in-phase component with frequency of 2fw, 3fw, but also in-opposite-phase component with frequency of 0.5 fw, 1.5 fw, 2.5 fw where fw is the walking rate is observed. The vertical vibration of pedestrian bridge induced by higher components of human walking vertical force is also discussed in this paper. A full scale measurement for the existing pedestrian bridge with center span length of 33 m is carried out focusing on the resonance phenomenon due to higher components of human walking vertical force. Dynamic response characteristics excited by these vertical higher components of human walking are revealed from the dynamic design viewpoint of pedestrian bridge.
Keywords: Simplified method, Human walking vertical force, Higher component, Pedestrian bridge vibration.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18142183 Crude Distillation Process Simulation Using Unisim Design Simulator
Authors: C. Patrascioiu, M. Jamali
Abstract:
The paper deals with the simulation of the crude distillation process using the Unisim Design simulator. The necessity of simulating this process is argued both by considerations related to the design of the crude distillation column, but also by considerations related to the design of advanced control systems. In order to use the Unisim Design simulator to simulate the crude distillation process, the identification of the simulators used in Romania and an analysis of the PRO/II, HYSYS, and Aspen HYSYS simulators were carried out. Analysis of the simulators for the crude distillation process has allowed the authors to elaborate the conclusions of the success of the crude modelling. A first aspect developed by the authors is the implementation of specific problems of petroleum liquid-vapors equilibrium using Unisim Design simulator. The second major element of the article is the development of the methodology and the elaboration of the simulation program for the crude distillation process, using Unisim Design resources. The obtained results validate the proposed methodology and will allow dynamic simulation of the process.
Keywords: Crude oil, distillation, simulation, Unisim Design, simulators.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27022182 Effect of Visual Speech in Sign Speech Synthesis
Authors: Zdenek Krnoul
Abstract:
This article investigates a contribution of synthesized visual speech. Synthesis of visual speech expressed by a computer consists in an animation in particular movements of lips. Visual speech is also necessary part of the non-manual component of a sign language. Appropriate methodology is proposed to determine the quality and the accuracy of synthesized visual speech. Proposed methodology is inspected on Czech speech. Hence, this article presents a procedure of recording of speech data in order to set a synthesis system as well as to evaluate synthesized speech. Furthermore, one option of the evaluation process is elaborated in the form of a perceptual test. This test procedure is verified on the measured data with two settings of the synthesis system. The results of the perceptual test are presented as a statistically significant increase of intelligibility evoked by real and synthesized visual speech. Now, the aim is to show one part of evaluation process which leads to more comprehensive evaluation of the sign speech synthesis system.
Keywords: Perception test, Sign speech synthesis, Talking head, Visual speech.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14772181 Synthesizing CuFe2O4 Spinel Powders by a Combustion-Like Process for Solid Oxide Fuel Cell Interconnect Coatings
Authors: S. N. Hosseini, M. H. Enayati, F. Karimzadeh, N. M. Sammes
Abstract:
The synthesis of CuFe2O4 spinel powders by an optimized combustion-like process followed by calcination is described herein. The samples were characterized using X-ray diffraction (XRD), differential thermal analysis (TG/DTA), scanning electron microscopy (SEM), dilatometry and 4-probe DC methods. Different glycine to nitrate (G/N) ratios of 1 (fuel-deficient), 1.48 (stoichiometric) and 2 (fuel-rich) were employed. Calcining the asprepared powders at 800 and 1000°C for 5 hours showed that the G/N ratio of 2 results in the formation of the desired copper spinel single phase at both calcination temperatures. For G/N=1, formation of CuFe2O4 takes place in three steps. First, iron and copper nitrates decompose to iron oxide and pure copper. Then, copper transforms to copper oxide and finally, copper and iron oxides react with each other to form a copper ferrite spinel phase. The electrical conductivity and the coefficient of thermal expansion of the sintered pelletized samples were 2 S.cm-1 (800°C) and 11×10-6 °C-1 (25-800°C), respectively.Keywords: SOFC interconnect coatings, Copper ferrite, Spinels, Electrical conductivity, Glycine–nitrate process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24842180 Pattern Recognition Using Feature Based Die-Map Clusteringin the Semiconductor Manufacturing Process
Authors: Seung Hwan Park, Cheng-Sool Park, Jun Seok Kim, Youngji Yoo, Daewoong An, Jun-Geol Baek
Abstract:
Depending on the big data analysis becomes important, yield prediction using data from the semiconductor process is essential. In general, yield prediction and analysis of the causes of the failure are closely related. The purpose of this study is to analyze pattern affects the final test results using a die map based clustering. Many researches have been conducted using die data from the semiconductor test process. However, analysis has limitation as the test data is less directly related to the final test results. Therefore, this study proposes a framework for analysis through clustering using more detailed data than existing die data. This study consists of three phases. In the first phase, die map is created through fail bit data in each sub-area of die. In the second phase, clustering using map data is performed. And the third stage is to find patterns that affect final test result. Finally, the proposed three steps are applied to actual industrial data and experimental results showed the potential field application.
Keywords: Die-Map Clustering, Feature Extraction, Pattern Recognition, Semiconductor Manufacturing Process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31512179 Estimating the Absorption of Volatile Organic Compounds in Four Biodiesels Using the UNIFAC Procedure
Authors: Corina M Mateescu, Edison Muzenda, Mohamed Belaid, Saka Abdulkareem, Ayo S Afolabi
Abstract:
This work considered the thermodynamic feasibility of scrubbing volatile organic compounds into biodiesel in view of designing a gas treatment process with this absorbent. A detailed vapour – liquid equilibrium investigation was performed using the original UNIFAC group contribution method. The four biodiesels studied in this work are methyl oleate, methyl palmitate, methyl linolenate and ethyl stearate. The original UNIFAC procedure was used to estimate the infinite dilution activity coefficients of 13 selected volatile organic compounds in the biodiesels. The calculations were done at the VOC mole fraction of 9.213x10-8. Ethyl stearate gave the most favourable phase equilibrium. A close agreement was found between the infinite dilution activity coefficient of toluene found in this work and those reported in literature. Thermodynamic models can efficiently be used to calculate vast amount of phase equilibrium behaviour using limited number of experimental data.Keywords: Biodiesel, Equilibrium, Gas treatment, Infinitedilution, Thermodynamic
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18822178 Efficient System for Speech Recognition using General Regression Neural Network
Authors: Abderrahmane Amrouche, Jean Michel Rouvaen
Abstract:
In this paper we present an efficient system for independent speaker speech recognition based on neural network approach. The proposed architecture comprises two phases: a preprocessing phase which consists in segmental normalization and features extraction and a classification phase which uses neural networks based on nonparametric density estimation namely the general regression neural network (GRNN). The relative performances of the proposed model are compared to the similar recognition systems based on the Multilayer Perceptron (MLP), the Recurrent Neural Network (RNN) and the well known Discrete Hidden Markov Model (HMM-VQ) that we have achieved also. Experimental results obtained with Arabic digits have shown that the use of nonparametric density estimation with an appropriate smoothing factor (spread) improves the generalization power of the neural network. The word error rate (WER) is reduced significantly over the baseline HMM method. GRNN computation is a successful alternative to the other neural network and DHMM.Keywords: Speech Recognition, General Regression NeuralNetwork, Hidden Markov Model, Recurrent Neural Network, ArabicDigits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21852177 Inverse Problem Methodology for the Measurement of the Electromagnetic Parameters Using MLP Neural Network
Authors: T. Hacib, M. R. Mekideche, N. Ferkha
Abstract:
This paper presents an approach which is based on the use of supervised feed forward neural network, namely multilayer perceptron (MLP) neural network and finite element method (FEM) to solve the inverse problem of parameters identification. The approach is used to identify unknown parameters of ferromagnetic materials. The methodology used in this study consists in the simulation of a large number of parameters in a material under test, using the finite element method (FEM). Both variations in relative magnetic permeability and electrical conductivity of the material under test are considered. Then, the obtained results are used to generate a set of vectors for the training of MLP neural network. Finally, the obtained neural network is used to evaluate a group of new materials, simulated by the FEM, but not belonging to the original dataset. Noisy data, added to the probe measurements is used to enhance the robustness of the method. The reached results demonstrate the efficiency of the proposed approach, and encourage future works on this subject.Keywords: Inverse problem, MLP neural network, parametersidentification, FEM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17642176 Understanding Physical Activity Behavior of Type 2 Diabetics Using the Theory of Planned Behavior and Structural Equation Modeling
Authors: D. O. Omondi, M. K. Walingo, G. M. Mbagaya, L. O. A. Othuon
Abstract:
Understanding patient factors related to physical activity behavior is important in the management of Type 2 Diabetes. This study applied the Theory of Planned Behavior model to understand physical activity behavior among sampled Type 2 diabetics in Kenya. The study was conducted within the diabetic clinic at Kisii Level 5 Hospital and adopted sequential mixed methods design beginning with qualitative phase and ending with quantitative phase. Qualitative data was analyzed using grounded theory analysis method. Structural equation modeling using maximum likelihood was used to analyze quantitative data. The common fit indices revealed that the theory of planned behavior fitted the data acceptably well among the Type 2 diabetes and within physical activity behavior {¤ç2 = 213, df = 84, n=230, p = .061, ¤ç2/df = 2.53; TLI = .97; CFI =.96; RMSEA (90CI) = .073(.029, .08)}. This theory proved to be useful in understanding physical activity behavior among Type 2 diabetics.
Keywords: Physical activity, Theory of Planned Behavior, Type2 diabetes, Kenya.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19842175 An Improved Prediction Model of Ozone Concentration Time Series Based On Chaotic Approach
Authors: N. Z. A. Hamid, M. S. M. Noorani
Abstract:
This study is focused on the development of prediction models of the Ozone concentration time series. Prediction model is built based on chaotic approach. Firstly, the chaotic nature of the time series is detected by means of phase space plot and the Cao method. Then, the prediction model is built and the local linear approximation method is used for the forecasting purposes. Traditional prediction of autoregressive linear model is also built. Moreover, an improvement in local linear approximation method is also performed. Prediction models are applied to the hourly Ozone time series observed at the benchmark station in Malaysia. Comparison of all models through the calculation of mean absolute error, root mean squared error and correlation coefficient shows that the one with improved prediction method is the best. Thus, chaotic approach is a good approach to be used to develop a prediction model for the Ozone concentration time series.
Keywords: Chaotic approach, phase space, Cao method, local linear approximation method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17832174 Harmonic Pollution Control of the Electrical Network by Three-Phase Shunt Active Filter: Comparative Study of Controls, by Hysteresis and by Duty Cycle Modulation
Authors: T. Patrice Nna Nna, S. Ndjakomo Essiane, S. Pérabi Ngoffé, F. Amigue Fissou
Abstract:
This paper deals with the harmonic decontamination of current in an electrical grid by an active shunt filter in order to improve power quality. The contribution of this paper is mainly based on the proposal of a control strategy for an active filter based on Duty Cycle Modulation (DCM). First, three-monophase method is applied for the identification of disturbing currents. A Simulink model of this method is given for one phase of the grid. Secondly, two orders were designed: the first one is the Hysteresis Control and the second one is the DCM Control. Finally, a comparative study of the two controls was performed. The results obtained show a significant improvement in the rate of harmonic distortion for both controls. The harmonic distortion for the Hysteresis control is limited by the non-controllability of the switching frequencies of the inverter's switches and reduces the harmonic distortion rate (THD) to 3.12% as opposed to the DCM control which limits the THD to 2.82% which makes it better.Keywords: Harmonic pollution, shunt active filter, hysteresis, Duty Cycle Modulation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6332173 Biosecurity Control Systems in Two Phases for Poultry Farms
Authors: M. Peña Aguilar Juan, E. Nava Galván Claudia, Pastrana Palma Alberto
Abstract:
In this work was developed and implemented a thermal fogging disinfection system to counteract pathogens from poultry feces in agribusiness farms, to reduce mortality rates and increase biosafety in them. The control system consists of two phases for the conditioning of the farm during the sanitary break. In the first phase, viral and bacterial inactivation was performed by treating the stool dry cleaning, along with the development of a specialized product that foster the generation of temperatures above 55 °C in less than 24 hr, for virus inactivation. In the second phase, a process for disinfection by fogging was implemented, along with the development of a specialized disinfectant that guarantee no risk for the operators’ health or birds. As a result of this process, it was possible to minimize the level of mortality of chickens on farms from 12% to 5.49%, representing a reduction of 6.51% in the death rate, through the formula applied to the treatment of poultry litter based on oxidising agents used as antiseptics, hydrogen peroxide solutions, glacial acetic acid and EDTA in order to act on bacteria, viruses, micro bacteria and spores.
Keywords: Innovation, triple-helix, innovation, poultry farms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14532172 Probabilistic Damage Tolerance Methodology for Solid Fan Blades and Discs
Authors: Andrej Golowin, Viktor Denk, Axel Riepe
Abstract:
Solid fan blades and discs in aero engines are subjected to high combined low and high cycle fatigue loads especially around the contact areas between blade and disc. Therefore, special coatings (e.g. dry film lubricant) and surface treatments (e.g. shot peening or laser shock peening) are applied to increase the strength with respect to combined cyclic fatigue and fretting fatigue, but also to improve damage tolerance capability. The traditional deterministic damage tolerance assessment based on fracture mechanics analysis, which treats service damage as an initial crack, often gives overly conservative results especially in the presence of vibratory stresses. A probabilistic damage tolerance methodology using crack initiation data has been developed for fan discs exposed to relatively high vibratory stresses in cross- and tail-wind conditions at certain resonance speeds for limited time periods. This Monte-Carlo based method uses a damage databank from similar designs, measured vibration levels at typical aircraft operations and wind conditions and experimental crack initiation data derived from testing of artificially damaged specimens with representative surface treatment under combined fatigue conditions. The proposed methodology leads to a more realistic prediction of the minimum damage tolerance life for the most critical locations applicable to modern fan disc designs.Keywords: Damage tolerance, Monte-Carlo method, fan blade and disc, laser shock peening.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15772171 Semantic Support for Hypothesis-Based Research from Smart Environment Monitoring and Analysis Technologies
Authors: T. S. Myers, J. Trevathan
Abstract:
Improvements in the data fusion and data analysis phase of research are imperative due to the exponential growth of sensed data. Currently, there are developments in the Semantic Sensor Web community to explore efficient methods for reuse, correlation and integration of web-based data sets and live data streams. This paper describes the integration of remotely sensed data with web-available static data for use in observational hypothesis testing and the analysis phase of research. The Semantic Reef system combines semantic technologies (e.g., well-defined ontologies and logic systems) with scientific workflows to enable hypothesis-based research. A framework is presented for how the data fusion concepts from the Semantic Reef architecture map to the Smart Environment Monitoring and Analysis Technologies (SEMAT) intelligent sensor network initiative. The data collected via SEMAT and the inferred knowledge from the Semantic Reef system are ingested to the Tropical Data Hub for data discovery, reuse, curation and publication.
Keywords: Information architecture, Semantic technologies Sensor networks, Ontologies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17152170 Working Children and Adolescents and the Vicious Circle of Poverty from the Perspective of Gunnar Myrdal’s Theory of Circular Cumulative Causation: Analysis and Implementation of a Probit Model to Brazil
Authors: J. Leige Lopes, L. Aparecida Bastos, R. Monteiro da Silva
Abstract:
The objective of this paper is to study the work of children and adolescents and the vicious circle of poverty from the perspective of Guinar Myrdal’s Theory of Circular Cumulative Causation. The objective is to show that if a person starts working in the juvenile phase of life they will be classified as poor or extremely poor when they are adult, which can to be observed in the case of Brazil, more specifically in the north and northeast. To do this, the methodology used was statistical and econometric analysis by applying a probit model. The main results show that: if people reside in the northeastern region of Brazil, and if they have a low educational level and if they start their professional life before the age 18, they will increase the likelihood that they will be poor or extremely poor. There is a consensus in the literature that one of the causes of the intergenerational transmission of poverty is related to child labor, this because when one starts their professional life while still in the toddler or adolescence stages of life, they end up sacrificing their studies. Because of their low level of education, children or adolescents are forced to perform low-paid functions and abandon school, becoming in the future, people who will be classified as poor or extremely poor. As a result of poverty, parents may be forced to send their children out to work when they are young, so that in the future they will also become poor adults, a process that is characterized as the "vicious circle of poverty."Keywords: Children, adolescents, Gunnar Myrdal, poverty, vicious circle.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16892169 Exploring the Landscape of Information Visualization through a Mark Lombardi Lens
Authors: Alon Friedman, Antonio Sánchez Chinchón
Abstract:
This bibliometric study takes an artistic and storytelling approach to explore the term ”Information visualization.” Analyzing over 1008 titles collected from databases that specialize in data visualization research, we examine the titles of these publications to report on the characteristics and development trends in the field. Employing a qualitative methodology, we delve into the titles of these publications, extracting leading terms and exploring the co-occurrence of these terms to gain deeper insights. By systematically analyzing the leading terms and their relationships within the titles, we shed light on the prevailing themes that shape the landscape of ”Information visualization” by employing the artist Mark Lombardi’s techniques to visualize our findings. By doing so, this study provides valuable insights into bibliometrics visualization while also opening new avenues for leveraging art and storytelling to enhance data representation.
Keywords: Bibliometrics analysis, Mark Lombardi design, information visualization, qualitative methodology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1032168 A Two-Phase Mechanism for Agent's Action Selection in Soccer Simulation
Authors: Vahid Salmani, Mahmoud Naghibzadeh, Farid Seifi, Amirhossein Taherinia
Abstract:
Soccer simulation is an effort to motivate researchers and practitioners to do artificial and robotic intelligence research; and at the same time put into practice and test the results. Many researchers and practitioners throughout the world are continuously working to polish their ideas and improve their implemented systems. At the same time, new groups are forming and they bring bright new thoughts to the field. The research includes designing and executing robotic soccer simulation algorithms. In our research, a soccer simulation player is considered to be an intelligent agent that is capable of receiving information from the environment, analyze it and to choose the best action from a set of possible ones, for its next move. We concentrate on developing a two-phase method for the soccer player agent to choose its best next move. The method is then implemented into our software system called Nexus simulation team of Ferdowsi University. This system is based on TsinghuAeolus[1] team that was the champion of the world RoboCup soccer simulation contest in 2001 and 2002.
Keywords: RoboCup, Soccer simulation, multi-agent environment, intelligent soccer agent, ball controller agent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15452167 Software Engineering Inspired Cost Estimation for Process Modelling
Authors: Felix Baumann, Aleksandar Milutinovic, Dieter Roller
Abstract:
Up to this point business process management projects in general and business process modelling projects in particular could not rely on a practical and scientifically validated method to estimate cost and effort. Especially the model development phase is not covered by a cost estimation method or model. Further phases of business process modelling starting with implementation are covered by initial solutions which are discussed in the literature. This article proposes a method of filling this gap by deriving a cost estimation method from available methods in similar domains namely software development or software engineering. Software development is regarded as closely similar to process modelling as we show. After the proposition of this method different ideas for further analysis and validation of the method are proposed. We derive this method from COCOMO II and Function Point which are established methods of effort estimation in the domain of software development. For this we lay out similarities of the software development process and the process of process modelling which is a phase of the Business Process Management life-cycle.Keywords: Cost Estimation, Effort Estimation, Process Modelling, Business Process Management, COCOMO.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22942166 An Investigation into the Role of Market Beta in Asset Pricing: Evidence from the Romanian Stock Market
Authors: Ioan Popa, Radu Lupu, Cristiana Tudor
Abstract:
In this paper, we apply the FM methodology to the cross-section of Romanian-listed common stocks and investigate the explanatory power of market beta on the cross-section of commons stock returns from Bucharest Stock Exchange. Various assumptions are empirically tested, such us linearity, market efficiency, the “no systematic effect of non-beta risk" hypothesis or the positive expected risk-return trade-off hypothesis. We find that the Romanian stock market shows the same properties as the other emerging markets in terms of efficiency and significance of the linear riskreturn models. Our analysis included weekly returns from January 2002 until May 2010 and the portfolio formation, estimation and testing was performed in a rolling manner using 51 observations (one year) for each stage of the analysis.Keywords: Bucharest Stock Exchange, Fama-Macbeth methodology, systematic risk, non-linear risk-return dependence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1905