Search results for: time scale
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 22005

Search results for: time scale

21975 The Challenges of Scaling Agile to Large-Scale Distributed Development: An Overview of the Agile Factory Model

Authors: Bernard Doherty, Andrew Jelfs, Aveek Dasgupta, Patrick Holden

Abstract:

Many companies have moved to agile and hybrid agile methodologies where portions of the Software Design Life-cycle (SDLC) and Software Test Life-cycle (STLC) can be time boxed in order to enhance delivery speed, quality and to increase flexibility to changes in software requirements. Despite widespread proliferation of agile practices, implementation often fails due to lack of adequate project management support, decreased motivation or fear of increased interaction. Consequently, few organizations effectively adopt agile processes with tailoring often required to integrate agile methodology in large scale environments. This paper provides an overview of the challenges in implementing an innovative large-scale tailored realization of the agile methodology termed the Agile Factory Model (AFM), with the aim of comparing and contrasting issues of specific importance to organizations undertaking large scale agile development. The conclusions demonstrate that agile practices can be effectively translated to a globally distributed development environment.

Keywords: agile, agile factory model, globally distributed development, large-scale agile

Procedia PDF Downloads 266
21974 Grating Scale Thermal Expansion Error Compensation for Large Machine Tools Based on Multiple Temperature Detection

Authors: Wenlong Feng, Zhenchun Du, Jianguo Yang

Abstract:

To decrease the grating scale thermal expansion error, a novel method which based on multiple temperature detections is proposed. Several temperature sensors are installed on the grating scale and the temperatures of these sensors are recorded. The temperatures of every point on the grating scale are calculated by interpolating between adjacent sensors. According to the thermal expansion principle, the grating scale thermal expansion error model can be established by doing the integral for the variations of position and temperature. A novel compensation method is proposed in this paper. By applying the established error model, the grating scale thermal expansion error is decreased by 90% compared with no compensation. The residual positioning error of the grating scale is less than 15um/10m and the accuracy of the machine tool is significant improved.

Keywords: thermal expansion error of grating scale, error compensation, machine tools, integral method

Procedia PDF Downloads 334
21973 Imputation of Incomplete Large-Scale Monitoring Count Data via Penalized Estimation

Authors: Mohamed Dakki, Genevieve Robin, Marie Suet, Abdeljebbar Qninba, Mohamed A. El Agbani, Asmâa Ouassou, Rhimou El Hamoumi, Hichem Azafzaf, Sami Rebah, Claudia Feltrup-Azafzaf, Nafouel Hamouda, Wed a.L. Ibrahim, Hosni H. Asran, Amr A. Elhady, Haitham Ibrahim, Khaled Etayeb, Essam Bouras, Almokhtar Saied, Ashrof Glidan, Bakar M. Habib, Mohamed S. Sayoud, Nadjiba Bendjedda, Laura Dami, Clemence Deschamps, Elie Gaget, Jean-Yves Mondain-Monval, Pierre Defos Du Rau

Abstract:

In biodiversity monitoring, large datasets are becoming more and more widely available and are increasingly used globally to estimate species trends and con- servation status. These large-scale datasets challenge existing statistical analysis methods, many of which are not adapted to their size, incompleteness and heterogeneity. The development of scalable methods to impute missing data in incomplete large-scale monitoring datasets is crucial to balance sampling in time or space and thus better inform conservation policies. We developed a new method based on penalized Poisson models to impute and analyse incomplete monitoring data in a large-scale framework. The method al- lows parameterization of (a) space and time factors, (b) the main effects of predic- tor covariates, as well as (c) space–time interactions. It also benefits from robust statistical and computational capability in large-scale settings. The method was tested extensively on both simulated and real-life waterbird data, with the findings revealing that it outperforms six existing methods in terms of missing data imputation errors. Applying the method to 16 waterbird species, we estimated their long-term trends for the first time at the entire North African scale, a region where monitoring data suffer from many gaps in space and time series. This new approach opens promising perspectives to increase the accuracy of species-abundance trend estimations. We made it freely available in the r package ‘lori’ (https://CRAN.R-project.org/package=lori) and recommend its use for large- scale count data, particularly in citizen science monitoring programmes.

Keywords: biodiversity monitoring, high-dimensional statistics, incomplete count data, missing data imputation, waterbird trends in North-Africa

Procedia PDF Downloads 117
21972 Developing New Media Credibility Scale: A Multidimensional Perspective

Authors: Hanaa Farouk Saleh

Abstract:

The main purposes of this study are to develop a scale that reflects emerging theoretical understandings of new media credibility, based on the evolution of credibility studies in western researches, identification of the determinants of credibility in the media and its components by comparing traditional and new media credibility scales and building accumulative scale to test new media credibility. This approach was built on western researches using conceptualizations of media credibility, which focuses on four principal components: Source (journalist), message (article), medium (newspaper, radio, TV, web, etc.), and organization (owner of the medium), and adding user and cultural context as key components to assess new media credibility in particular. This study’s value lies in its contribution to the conceptualization and development of new media credibility through the creation of a theoretical measurement tool. Future studies should explore this scale to test new media credibility, which represents a promising new approach in the efforts to define and measure credibility of all media types.

Keywords: credibility scale, media credibility components, new media credibility scale, scale development

Procedia PDF Downloads 273
21971 Path Integrals and Effective Field Theory of Large Scale Structure

Authors: Revant Nayar

Abstract:

In this work, we recast the equations describing large scale structure, and by extension all nonlinear fluids, in the path integral formalism. We first calculate the well known two and three point functions using Schwinger Keldysh formalism used commonly to perturbatively solve path integrals in non- equilibrium systems. Then we include EFT corrections due to pressure, viscosity, and noise as effects on the time-dependent propagator. We are able to express results for arbitrary two and three point correlation functions in LSS in terms of differential operators acting on a triple K master intergral. We also, for the first time, get analytical results for more general initial conditions deviating from the usual power law P∝kⁿ by introducing a mass scale in the initial conditions. This robust field theoretic formalism empowers us with tools from strongly coupled QFT to study the strongly non-linear regime of LSS and turbulent fluid dynamics such as OPE and holographic duals. These could be used to capture fully the strongly non-linear dynamics of fluids and move towards solving the open problem of classical turbulence.

Keywords: quantum field theory, cosmology, effective field theory, renormallisation

Procedia PDF Downloads 105
21970 Impacts of Financial Development and Operational Scale on Bank Efficiencies in Taiwan

Authors: Ying-Hsiu Chen, Pao-Peng Hsu

Abstract:

This paper adopts a two-stage data envelopment analysis to explore the impacts of financial development and bank operational scale on bank efficiencies. The sample comprises of unbalanced panel data of 32 Taiwanese enlisted in domestic commercial banks over the period 1998 to 2013. Empirical results show that technical efficiency is positively related to financial development, whereas the effect of financial development on scale efficiency is insignificant. The effect of operational scale exerts a significantly positive effect on bank efficiencies, but the gain of efficiency is decreased gradually when operational scale increases. Furthermore, increase in capital adequacy ratio and market power of banks leads to a growth of bank efficiencies.

Keywords: financial development, operational scale, efficiency, DEA

Procedia PDF Downloads 492
21969 Prediction of Fire Growth of the Office by Real-Scale Fire Experiment

Authors: Kweon Oh-Sang, Kim Heung-Youl

Abstract:

Estimating the engineering properties of fires is important to be prepared for the complex and various fire risks of large-scale structures such as super-tall buildings, large stadiums, and multi-purpose structures. In this study, a mock-up of a compartment which was 2.4(L) x 3.6 (W) x 2.4 (H) meter in dimensions was fabricated at the 10MW LSC (Large Scale Calorimeter) and combustible office supplies were placed in the compartment for a real-scale fire test. Maximum heat release rate was 4.1 MW and total energy release obtained through the application of t2 fire growth rate was 6705.9 MJ.

Keywords: fire growth, fire experiment, t2 curve, large scale calorimeter

Procedia PDF Downloads 306
21968 Support for Planning of Mobile Personnel Tasks by Solving Time-Dependent Routing Problems

Authors: Wlodzimierz Ogryczak, Tomasz Sliwinski, Jaroslaw Hurkala, Mariusz Kaleta, Bartosz Kozlowski, Piotr Palka

Abstract:

Implementation concepts of a decision support system for planning and management of mobile personnel tasks (sales representatives and others) are discussed. Large-scale periodic time-dependent vehicle routing and scheduling problems with complex constraints are solved for this purpose. Complex nonuniform constraints with respect to frequency, time windows, working time, etc. are taken into account with additional fast adaptive procedures for operational rescheduling of plans in the presence of various disturbances. Five individual solution quality indicators with respect to a single personnel person are considered. This paper deals with modeling issues corresponding to the problem and general solution concepts. The research was supported by the European Union through the European Regional Development Fund under the Operational Programme ‘Innovative Economy’ for the years 2007-2013; Priority 1 Research and development of modern technologies under the project POIG.01.03.01-14-076/12: 'Decision Support System for Large-Scale Periodic Vehicle Routing and Scheduling Problems with Complex Constraints.'

Keywords: mobile personnel management, multiple criteria, time dependent, time windows, vehicle routing and scheduling

Procedia PDF Downloads 296
21967 Multiple Relaxation Times in the Gibbs Ensemble Monte Carlo Simulation of Phase Separation

Authors: Bina Kumari, Subir K. Sarkar, Pradipta Bandyopadhyay

Abstract:

The autocorrelation function of the density fluctuation is studied in each of the two phases in a Gibbs Ensemble Monte Carlo (GEMC) simulation of the problem of phase separation for a square well potential with various values of its range. We find that the normalized autocorrelation function is described very well as a linear combination of an exponential function with a time scale τ₂ and a stretched exponential function with a time scale τ₁ and an exponent α. Dependence of (α, τ₁, τ₂) on the parameters of the GEMC algorithm and the range of the square well potential is investigated and interpreted. We also analyse the issue of how to choose the parameters of the GEMC simulation optimally.

Keywords: autocorrelation function, density fluctuation, GEMC, simulation

Procedia PDF Downloads 156
21966 Hardy Type Inequalities of Two-Dimensional on Time Scales via Steklov Operator

Authors: Wedad Albalawi

Abstract:

The mathematical inequalities have been the core of mathematical study and used in almost all branches of mathematics as well in various areas of science and engineering. The inequalities by Hardy, Littlewood and Polya were the first significant composition of several science. This work presents fundamental ideas, results and techniques and it has had much influence on research in various branches of analysis. Since 1934, various inequalities have been produced and studied in the literature. Furthermore, some inequalities have been formulated by some operators; in 1989, weighted Hardy inequalities have been obtained for integration operators. Then, they obtained weighted estimates for Steklov operators that were used in the solution of the Cauchy problem for the wave equation. They were improved upon in 2011 to include the boundedness of integral operators from the weighted Sobolev space to the weighted Lebesgue space. Some inequalities have been demonstrated and improved using the Hardy–Steklov operator. Recently, a lot of integral inequalities have been improved by differential operators. Hardy inequality has been one of the tools that is used to consider integrity solutions of differential equations. Then dynamic inequalities of Hardy and Coposon have been extended and improved by various integral operators. These inequalities would be interesting to apply in different fields of mathematics (functional spaces, partial differential equations, mathematical modeling). Some inequalities have been appeared involving Copson and Hardy inequalities on time scales to obtain new special version of them. A time scale is defined as a closed subset contains real numbers. Then the inequalities of time scales version have received a lot of attention and has had a major field in both pure and applied mathematics. There are many applications of dynamic equations on time scales to quantum mechanics, electrical engineering, neural networks, heat transfer, combinatorics, and population dynamics. This study focuses on double integrals to obtain new time-scale inequalities of Copson driven by Steklov operator. They will be applied in the solution of the Cauchy problem for the wave equation. The proof can be done by introducing restriction on the operator in several cases. In addition, the obtained inequalities done by using some concepts in time scale version such as time scales calculus, theorem of Fubini and the inequality of H¨older.

Keywords: time scales, inequality of Hardy, inequality of Coposon, Steklov operator

Procedia PDF Downloads 43
21965 Space Tourism Pricing Model Revolution from Time Independent Model to Time-Space Model

Authors: Kang Lin Peng

Abstract:

Space tourism emerged in 2001 and became famous in 2021, following the development of space technology. The space market is twisted because of the excess demand. Space tourism is currently rare and extremely expensive, with biased luxury product pricing, which is the seller’s market that consumers can not bargain with. Spaceship companies such as Virgin Galactic, Blue Origin, and Space X have been charged space tourism prices from 200 thousand to 55 million depending on various heights in space. There should be a reasonable price based on a fair basis. This study aims to derive a spacetime pricing model, which is different from the general pricing model on the earth’s surface. We apply general relativity theory to deduct the mathematical formula for the space tourism pricing model, which covers the traditional time-independent model. In the future, the price of space travel will be different from current flight travel when space travel is measured in lightyear units. The pricing of general commodities mainly considers the general equilibrium of supply and demand. The pricing model considers risks and returns with the dependent time variable as acceptable when commodities are on the earth’s surface, called flat spacetime. Current economic theories based on the independent time scale in the flat spacetime do not consider the curvature of spacetime. Current flight services flying the height of 6, 12, and 19 kilometers are charging with a pricing model that measures time coordinate independently. However, the emergence of space tourism is flying heights above 100 to 550 kilometers that have enlarged the spacetime curvature, which means tourists will escape from a zero curvature on the earth’s surface to the large curvature of space. Different spacetime spans should be considered in the pricing model of space travel to echo general relativity theory. Intuitively, this spacetime commodity needs to consider changing the spacetime curvature from the earth to space. We can assume the value of each spacetime curvature unit corresponding to the gradient change of each Ricci or energy-momentum tensor. Then we know how much to spend by integrating the spacetime from the earth to space. The concept is adding a price p component corresponding to the general relativity theory. The space travel pricing model degenerates into a time-independent model, which becomes a model of traditional commodity pricing. The contribution is that the deriving of the space tourism pricing model will be a breakthrough in philosophical and practical issues for space travel. The results of the space tourism pricing model extend the traditional time-independent flat spacetime mode. The pricing model embedded spacetime as the general relativity theory can better reflect the rationality and accuracy of space travel on the universal scale. The universal scale from independent-time scale to spacetime scale will bring a brand-new pricing concept for space traveling commodities. Fair and efficient spacetime economics will also bring to humans’ travel when we can travel in lightyear units in the future.

Keywords: space tourism, spacetime pricing model, general relativity theory, spacetime curvature

Procedia PDF Downloads 87
21964 Preoperative Anxiety Evaluation: Comparing the Visual Facial Anxiety Scale/Yumul Faces Anxiety Scale, Numerical Verbal Rating Scale, Categorization Scale, and the State-Trait Anxiety Inventory

Authors: Roya Yumul, Chse, Ofelia Loani Elvir Lazo, David Chernobylsky, Omar Durra

Abstract:

Background: Preoperative anxiety has been shown to be caused by the fear associated with surgical and anesthetic complications; however, the current gold standard for assessing patient anxiety, the STAI, is problematic to use in the preoperative setting given the duration and concentration required to complete the 40-item extensive questionnaire. Our primary aim in the study is to investigate the correlation of the Visual Facial Anxiety Scale (VFAS) and Numerical Verbal Rating Scale (NVRS) to State-Trait Anxiety Inventory (STAI) to determine the optimal anxiety scale to use in the perioperative setting. Methods: A clinical study of patients undergoing various surgeries was conducted utilizing each of the preoperative anxiety scales. Inclusion criteria included patients undergoing elective surgeries, while exclusion criteria included patients with anesthesia contraindications, inability to comprehend instructions, impaired judgement, substance abuse history, and those pregnant or lactating. 293 patients were analyzed in terms of demographics, anxiety scale survey results, and anesthesia data via Spearman Coefficients, Chi-Squared Analysis, and Fischer’s exact test utilized for comparison analysis. Results: Statistical analysis showed that VFAS had a higher correlation to STAI than NVRS (rs=0.66, p<0.0001 vs. rs=0.64, p<0.0001). The combined VFAS-Categorization Scores showed the highest correlation with the gold standard (rs=0.72, p<0.0001). Subgroup analysis showed similar results. STAI evaluation time (247.7 ± 54.81 sec) far exceeds VFAS (7.29 ± 1.61 sec), NVRS (7.23 ± 1.60 sec), and Categorization scales (7.29 ± 1.99 sec). Patients preferred VFAS (54.4%), Categorization (11.6%), and NVRS (8.8%). Anesthesiologists preferred VFAS (63.9%), NVRS (22.1%), and Categorization Scales (14.0%). Of note, the top five causes of preoperative anxiety were determined to be waiting (56.5%), pain (42.5%), family concerns (40.5%), no information about surgery (40.1%), or anesthesia (31.6%). Conclusions: Combined VFAS-Categorization Score (VCS) demonstrates the highest correlation to the gold standard, STAI. Both VFAS and Categorization tests also take significantly less time than STAI, which is critical in the preoperative setting. Among both patients and anesthesiologists, VFAS was the most preferred scale. This forms the basis of the Yumul FACES Anxiety Scale, designed for quick quantization and assessment in the preoperative setting while maintaining a high correlation to the golden standard. Additional studies using the formulated Yumul FACES Anxiety Scale are merited.

Keywords: numerical verbal anxiety scale, preoperative anxiety, state-trait anxiety inventory, visual facial anxiety scale

Procedia PDF Downloads 106
21963 Unsteadiness Effects on Variable Thrust Nozzle Performance

Authors: A. M. Tahsini, S. Tadayon Mousavi

Abstract:

The purpose of this paper is to elucidate the flow unsteady behavior for moving plug in convergent-divergent variable thrust nozzle. Compressible axisymmetric Navier-Stokes equations are used to study this physical phenomenon. Different velocities are set for plug to investigate the effect of plug movement on flow unsteadiness. Variation of mass flow rate and thrust are compared under two conditions: First, the plug is placed at different positions and flow is simulated to reach the steady state (quasi steady simulation) and second, the plug is moved with assigned velocity and flow simulation is coupled with plug movement (unsteady simulation). If plug speed is high enough and its movement time scale is at the same order of the flow time scale, variation of the mass flow rate and thrust level versus plug position demonstrate a vital discrepancy under the quasi steady and unsteady conditions. This phenomenon should be considered especially from response time viewpoints in thrusters design.

Keywords: nozzle, numerical study, unsteady, variable thrust

Procedia PDF Downloads 321
21962 Investigation of the Relationship between Digital Game Playing, Internet Addiction and Perceived Stress Levels in University Students

Authors: Sevim Ugur, Cemile Kutmec Yilmaz, Omer Us, Sevdenur Koksaldi

Abstract:

Aim: This study aims to investigate the effect of digital game playing and Internet addiction on perceived stress levels in university students. Method: The descriptive study was conducted through face-to-face interview method with a total of 364 university students studying at Aksaray University between November 15 and December 30, 2017. The research data were collected using personal information form, a questionnaire to determine the characteristics of playing digital game, the Internet addiction scale and the perceived stress scale. In the evaluation of the data, Mann-Whitney U test was used for two-group comparison of the sample with non-normal distribution, Kruskal-Wallis H-test was used in the comparison of more than two groups, and the Spearman correlation test was used to determine the relationship between Internet addiction and the perceived stress level. Results: It was determined that the mean age of the students participated in the study was 20.13 ± 1.7 years, 67.6% was female, 35.7% was sophomore, and 62.1% had an income 500 TL or less. It was found that 83.5% of the students use the Internet every day and 70.6% uses the Internet for 5 hours or less per day. Of the students, 12.4% prefers digital games instead of spending time outdoors, 8% plays a game as the first activity in leisure time, 12.4% plays all day, 15.7% feels anger when he/she is prevented from playing, 14.8% prefers playing games to get away from his/her problems, 23.4% had his/her school achievement affected negatively because of game playing, and 8% argues with family members due to the time spent for gaming. Students who play games on the computer for a long time were found to feel back pain (30.8%), headache (28.6%), insomnia (26.9%), dryness and pain in the eyes (26.6%), pain in the wrist (21.2%), feeling excessive tension and anger (16.2%), humpback (12.9), vision loss (9.6%) and pain in the wrist and fingers (7.4%). In our study, students' Internet addiction scale mean score was found to be 45.47 ± 16.1 and mean perceived stress scale score was 28.56 ± 2.7. A significant and negative correlation (p=0.037) was found between the total score of the Internet addiction scale and the total score of the perceived stress scale (r=-0.110). Conclusion: It was found in the study that Internet addiction and perceived stress of the students were at a moderate level and that there was a negative correlation between Internet addiction and perceived stress levels. Internet addiction was found to increase with the increasing perceived stress levels of students, and students were found to have health problems such as back pain, dryness in the eyes, pain, insomnia, headache, and humpback. Therefore, it is recommended to inform students about different coping methods other than spending time on the Internet to cope with the stress they perceive.

Keywords: digital game, internet addiction, student, stress level

Procedia PDF Downloads 253
21961 Analysis of Efficiency Production of Grass Black Jelly (Mesona palustris) in Double Scale

Authors: Irvan Adhin Cholilie, Susinggih Wijana, Yusron Sugiarto

Abstract:

The aim of this research is to compare the results of black grass jelly produced using laboratory scale and double scale. In this research, the production from the laboratory scale is using ingredients of 1 kg black grass jelly added with 5 liters of water, while the double scale is using 5 kg black grass jelly and 75 liters of water. The results of organoleptic tests performed by 30 panelists (general) to the sample gels of grass black powder produced from both of laboratory and double scale are not different significantly in color, odor, flavor, and texture. Proximate test results conducted in both of grass black jelly powder produced in laboratory scale and double scale also have no significant differences in all parameters. Grass black jelly powder from double scale contains water, carbohydrate, crude fiber, and yield in the amount of 12,25 %; 43,7 %; 5,89 %; and 16,28 % respectively. The results of the energy efficiency analysis by boiling, draining, evaporation, drying, and milling processes are 85,11 %; 76,97 %; 99,64 %; 99,99% and 99,39% respectively. The utility needs including water needs for each batch amounted 0.1 m3 and cost Rp 220,5 per batch, the electricity needs for each batch is 20.01 kWh and cost Rp 18569.28 per batch, and LPG needs for each batch is 30 kg costed Rp 234,000.00 so that the total cost spent for the process is Rp 252,789.78 .

Keywords: black grass jelly, powder, mass balance, energy balance, cost

Procedia PDF Downloads 355
21960 Two Efficient Heuristic Algorithms for the Integrated Production Planning and Warehouse Layout Problem

Authors: Mohammad Pourmohammadi Fallah, Maziar Salahi

Abstract:

In the literature, a mixed-integer linear programming model for the integrated production planning and warehouse layout problem is proposed. To solve the model, the authors proposed a Lagrangian relax-and-fix heuristic that takes a significant amount of time to stop with gaps above 5$\%$ for large-scale instances. Here, we present two heuristic algorithms to solve the problem. In the first one, we use a greedy approach by allocating warehouse locations with less reservation costs and also less transportation costs from the production area to locations and from locations to the output point to items with higher demands. Then a smaller model is solved. In the second heuristic, first, we sort items in descending order according to the fraction of the sum of the demands for that item in the time horizon plus the maximum demand for that item in the time horizon and the sum of all its demands in the time horizon. Then we categorize the sorted items into groups of 3, 4, or 5 and solve a small-scale optimization problem for each group, hoping to improve the solution of the first heuristic. Our preliminary numerical results show the effectiveness of the proposed heuristics.

Keywords: capacitated lot-sizing, warehouse layout, mixed-integer linear programming, heuristics algorithm

Procedia PDF Downloads 155
21959 Comparative Study of Isothermal and Cyclic Oxidation on Titanium Alloys

Authors: Poonam Yadav, Dong Bok Lee

Abstract:

Isothermal oxidation at 800°C for 50h and Cyclic oxidation at 600°C and 800°C for 40h of Pure Ti and Ti64 were performed in a muffle furnace. In Cyclic oxidation, massive scale spallation occurred, and the oxide scale cracks and peels off were observed at high temperature, it represents oxide scale that formed during cyclic oxidation was spalled out owing to stresses due to thermal shock generated during repetitive oxidation and subsequent cooling. The thickness of scale is larger in cyclic oxidation than the isothermal case. This is due to inward diffusion of oxygen through oxide scales and/or pores and cracks in cyclic oxidation.

Keywords: cyclic, diffusion, isothermal, cyclic

Procedia PDF Downloads 886
21958 Study of Aqueous Solutions: A Dielectric Spectroscopy Approach

Authors: Kumbharkhane Ashok

Abstract:

The time domain dielectric relaxation spectroscopy (TDRS) probes the interaction of a macroscopic sample with a time-dependent electrical field. The resulting complex permittivity spectrum, characterizes amplitude (voltage) and time scale of the charge-density fluctuations within the sample. These fluctuations may arise from the reorientation of the permanent dipole moments of individual molecules or from the rotation of dipolar moieties in flexible molecules, like polymers. The time scale of these fluctuations depends on the sample and its relative relaxation mechanism. Relaxation times range from some picoseconds in low viscosity liquids to hours in glasses, Therefore the DRS technique covers an extensive dynamical process, its corresponding frequency range from 10-4 Hz to 1012 Hz. This inherent ability to monitor the cooperative motion of molecular ensemble distinguishes dielectric relaxation from methods like NMR or Raman spectroscopy which yield information on the motions of individual molecules. An experimental set up for Time Domain Reflectometry (TDR) technique from 10 MHz to 30 GHz has been developed for the aqueous solutions. This technique has been very simple and covers a wide band of frequencies in the single measurement. Dielectric Relaxation Spectroscopy is especially sensitive to intermolecular interactions. The complex permittivity spectra of aqueous solutions have been fitted using Cole-Davidson (CD) model to determine static dielectric constants and relaxation times for entire concentrations. The heterogeneous molecular interactions in aqueous solutions have been discussed through Kirkwood correlation factor and excess properties.

Keywords: liquid, aqueous solutions, time domain reflectometry

Procedia PDF Downloads 417
21957 Adjustment and Scale-Up Strategy of Pilot Liquid Fermentation Process of Azotobacter sp.

Authors: G. Quiroga-Cubides, A. Díaz, M. Gómez

Abstract:

The genus Azotobacter has been widely used as bio-fertilizer due to its significant effects on the stimulation and promotion of plant growth in various agricultural species of commercial interest. In order to obtain significantly viable cellular concentration, a scale-up strategy for a liquid fermentation process (SmF) with two strains of A. chroococcum (named Ac1 and Ac10) was validated and adjusted at laboratory and pilot scale. A batch fermentation process under previously defined conditions was carried out on a biorreactor Infors®, model Minifors of 3.5 L, which served as a baseline for this research. For the purpose of increasing process efficiency, the effect of the reduction of stirring speed was evaluated in combination with a fed-batch-type fermentation laboratory scale. To reproduce the efficiency parameters obtained, a scale-up strategy with geometric and fluid dynamic behavior similarities was evaluated. According to the analysis of variance, this scale-up strategy did not have significant effect on cellular concentration and in laboratory and pilot fermentations (Tukey, p > 0.05). Regarding air consumption, fermentation process at pilot scale showed a reduction of 23% versus the baseline. The percentage of reduction related to energy consumption reduction under laboratory and pilot scale conditions was 96.9% compared with baseline.

Keywords: Azotobacter chroococcum, scale-up, liquid fermentation, fed-batch process

Procedia PDF Downloads 411
21956 System Survivability in Networks in the Context of Defense/Attack Strategies: The Large Scale

Authors: Asma Ben Yaghlane, Mohamed Naceur Azaiez, Mehdi Mrad

Abstract:

We investigate the large scale of networks in the context of network survivability under attack. We use appropriate techniques to evaluate and the attacker-based- and the defender-based-network survivability. The attacker is unaware of the operated links by the defender. Each attacked link has some pre-specified probability to be disconnected. The defender choice is so that to maximize the chance of successfully sending the flow to the destination node. The attacker however will select the cut-set with the highest chance to be disabled in order to partition the network. Moreover, we extend the problem to the case of selecting the best p paths to operate by the defender and the best k cut-sets to target by the attacker, for arbitrary integers p,k > 1. We investigate some variations of the problem and suggest polynomial-time solutions.

Keywords: defense/attack strategies, large scale, networks, partitioning a network

Procedia PDF Downloads 244
21955 Scaling-Down an Agricultural Waste Biogas Plant Fermenter

Authors: Matheus Pessoa, Matthias Kraume

Abstract:

Scale-Down rules in process engineering help us to improve and develop Industrial scale parameters into lab scale. Several scale-down rules available in the literature like Impeller Power Number, Agitation device Power Input, Substrate Tip Speed, Reynolds Number and Cavern Development were investigated in order to stipulate the rotational speed to operate an 11 L working volume lab-scale bioreactor within industrial process parameters. Herein, xanthan gum was used as a fluid with a representative viscosity of a hypothetical biogas plant, with H/D = 1 and central agitation, fermentation broth using sewage sludge and sugar beet pulp as substrate. The results showed that the cavern development strategy was the best method for establishing a rotational speed for the bioreactor operation, while the other rules presented values out of reality for this article proposes.

Keywords: anaerobic digestion, cavern development, scale down rules, xanthan gum

Procedia PDF Downloads 452
21954 A Comparative Performance of Polyaspartic Acid and Sodium Polyacrylate on Silicate Scale Inhibition

Authors: Ismail Bin Mohd Saaid, Abubakar Abubakar Umar

Abstract:

Despite the successes recorded by Alkaline/Surfactant/Polymer (ASP) flooding as an effective chemical EOR technique, the combination CEOR is not unassociated with stern glitches, one of which is the scaling of downhole equipment. One of the major issues inside the oil industry is how to control scale formation, regardless of whether it is in the wellhead equipment, down-hole pipelines or even the actual field formation. The best approach to handle the challenge associated with oilfield scale formation is the application of scale inhibitors to avert the scale formation. Chemical inhibitors have been employed in doing such. But due to environmental regulations, the industry have focused on using green scale inhibitors to mitigate the formation of scales. This paper compares the scale inhibition performance of Polyaspartic acid and sodium polyacrylic acid, both commercial green scale inhibitors, in mitigating silicate scales formed during Alkaline/Surfactant/polymer flooding under static conditions. Both PASP and TH5000 are non-threshold inhibitors, therefore their efficiency was only seeing in delaying the deposition of the silicate scales.

Keywords: alkaline/surfactant/polymer flooding (ASP), polyaspartic acid (PASP), sodium polyacrylate (SPA)

Procedia PDF Downloads 314
21953 Seismic Behavior of Concrete Filled Steel Tube Reinforced Concrete Column

Authors: Raghabendra Yadav, Baochun Chen, Huihui Yuan, Zhibin Lian

Abstract:

Pseudo-dynamic test (PDT) method is an advanced seismic test method that combines loading technology with computer technology. Large-scale models or full scale seismic tests can be carried out by using this method. CFST-RC columns are used in civil engineering structures because of their better seismic performance. A CFST-RC column is composed of four CFST limbs which are connected with RC web in longitudinal direction and with steel tube in transverse direction. For this study, a CFST-RC pier is tested under Four different earthquake time histories having scaled PGA of 0.05g. From the experiment acceleration, velocity, displacement and load time histories are observed. The dynamic magnification factors for acceleration due to Elcentro, Chi-Chi, Imperial Valley and Kobe ground motions are observed as 15, 12, 17 and 14 respectively. The natural frequency of the pier is found to be 1.40 Hz. The result shows that this type of pier has excellent static and earthquake resistant properties.

Keywords: bridge pier, CFST-RC pier, pseudo dynamic test, seismic performance, time history

Procedia PDF Downloads 152
21952 Experimental and Numerical Studies on Hydrogen Behavior in a Small-Scale Container with Passive Autocatalytic Recombiner

Authors: Kazuyuki Takase, Yoshihisa Hiraki, Gaku Takase, Isamu Kudo

Abstract:

One of the most important issue is to ensure the safety of long-term waste storage containers in which fuel debris and radioactive materials are accumulated. In this case, hydrogen generated by water decomposition by radiation is accumulated in the container for a long period of time, so it is necessary to reduce the concentration of hydrogen in the container. In addition, a condition that any power supplies from the outside of the container are unnecessary is requested. Then, radioactive waste storage containers with the passive autocatalytic recombiner (PAR) would be effective. The radioactive waste storage container with PAR was used for moving the fuel debris of the Three Mile Island Unit 2 to the storage location. However, the effect of PAR is not described in detail. Moreover, the reduction of hydrogen concentration during the long-term storage period was performed by the venting system, which was installed on the top of the container. Therefore, development of a long-term storage container with PAR was started with the aim of safely storing fuel debris picked up at the Fukushima Daiichi Nuclear Power Plant for a long period of time. A fundamental experiment for reducing the concentration of hydrogen which generates in a nuclear waste long-term storage container was carried out using a small-scale container with PAR. Moreover, the circulation flow behavior of hydrogen in the small-scale container resulting from the natural convection by the decay heat was clarified. In addition, preliminary numerical analyses were performed to predict the experimental results regarding the circulation flow behavior and the reduction of hydrogen concentration in the small-scale container. From the results of the present study, the validity of the container with PAR was experimentally confirmed on the reduction of hydrogen concentration. In addition, it was predicted numerically that the circulation flow behavior of hydrogen in the small-scale container is blocked by steam which generates by chemical reaction of hydrogen and oxygen.

Keywords: hydrogen behavior, reduction of concentration, long-term storage container, small-scale, PAR, experiment, analysis

Procedia PDF Downloads 140
21951 Comparison of the Yumul Faces Anxiety Scale to the Categorization Scale, the Numerical Verbal Rating Scale, and the State-Trait Anxiety Inventory for Preoperative Anxiety Evaluation

Authors: Ofelia Loani Elvir Lazo, Roya Yumul, David Chernobylsky, Omar Durra

Abstract:

Background: It is crucial to detect the patient’s existing anxiety to assist patients in a perioperative setting which is to be caused by the fear associated with surgical and anesthetic complications. However, the current gold standard for assessing patient anxiety, the STAI, is problematic to use in the preoperative setting, given the duration and concentration required to complete the 40-item questionnaire. Our primary aim in the study is to investigate the correlation of the Yumul Visual Facial Anxiety Scale (VFAS) and Numerical Verbal Rating Scale (NVRS) to State-Trait Anxiety Inventory (STAI) to determine the optimal anxiety scale to use in the perioperative setting. Methods: A clinical study of patients undergoing various surgeries was conducted utilizing each of the preoperative anxiety scales. Inclusion criteria included patients undergoing elective surgeries, while exclusion criteria included patients with anesthesia contraindications, inability to comprehend instructions, impaired judgement, substance abuse history, and those pregnant or lactating. 293 patients were analyzed in terms of demographics, anxiety scale survey results, and anesthesia data via Spearman Coefficients, Chi-Squared Analysis, and Fischer’s exact test utilized for comparative analysis. Results: Statistical analysis showed that VFAS had a higher correlation to STAI than NVRS (rs=0.66, p<0.0001 vs. rs=0.64, p<0.0001). The combined VFAS-Categorization Scores showed the highest correlation with the gold standard (rs=0.72, p<0.0001). Subgroup analysis showed similar results. STAI evaluation time (247.7 ± 54.81 sec) far exceeds VFAS (7.29 ± 1.61 sec), NVRS (7.23 ± 1.60 sec), and Categorization scales (7.29 ± 1.99 sec). Patients preferred VFAS (54.4%), Categorization (11.6%), and NVRS (8.8%). Anesthesiologists preferred VFAS (63.9%), NVRS (22.1%), and Categorization Scales (14.0%). Of note, the top five causes of preoperative anxiety were determined to be waiting (56.5%), pain (42.5%), family concerns (40.5%), no information about surgery (40.1%), or anesthesia (31.6%). Conclusıons: Both VFAS and Categorization tests also take significantly less time than STAI, which is critical in the preoperative setting. Combined VFAS-Categorization Score (VCS) demonstrates the highest correlation to the gold standard, STAI. Among both patients and anesthesiologists, VFAS was the most preferred scale. This forms the basis of the Yumul Faces Anxiety Scale, designed for quick quantization and assessment in the preoperative setting while maintaining a high correlation to the golden standard. Additional studies using the formulated Yumul Faces Anxiety Scale are merited.

Keywords: numerical verbal anxiety scale, preoperative anxiety, state-trait anxiety inventory, visual facial anxiety scale

Procedia PDF Downloads 90
21950 Effects of Employees’ Training Program on the Performance of Small Scale Enterprises in Oyo State

Authors: Itiola Kehinde Adeniran

Abstract:

The study examined the effect of employees’ training on the performance of small scale enterprises in Oyo State. A structured questionnaire was used to collect data from 150 respondents through purposive sampling method. Linear regression was used with the aid of statistical package for social science (SPSS) version 20 to analyze the data collected in order to examine the effect of independent variable, employees’ training on dependent variable, performance (profit) of small scale enterprises. The result revealed that employees’ training has a significant effect on the performance of small scale enterprises. It was concluded that predictor variable namely (training) is 55.5% variance of enterprises performance (profitability). Therefore, the paper recommended that all small scale enterprises in Nigeria should embrace manpower training and development in order to improve employees’ performance leading to organizational profitability.

Keywords: training, employee performance, small scale enterprise, organizational profitability

Procedia PDF Downloads 337
21949 Validating the Theme Park Service Quality Scale: A Case Study of Zhuhai Chimelong Ocean Kingdom

Authors: Kat Jingjing Luo

Abstract:

The development of theme parks in China has been through a rapid growth in the past decades. Increasing competition within service quality has forced theme park managers concerned the relationship between service quality and visitors’ satisfaction. Even though those existing service quality measurements such as SERVQUAL and THEMEQUAL have been applied in related researches, none of them is exclusive for Chinese theme park service quality. This study aims to investigate the service quality of the most popular theme park in China currently and develop a unique, reliable and valid scale. The reliability and validity analysis results from a survey of over 200 tourists in Chimelong ocean kingdom in Zhuhai city, south of China, indicate that the dimension of waiting time is a discover factor in the measurement of Chinese theme park service quality excluding in the THEMEQUAL instrument (i.e., tangibles, reliability, responsiveness and access, assurance, empathy and courtesy). The newly developed scale gives a better understand service quality in Chinese theme park industry, and the managerial implications in regard to the research, how to improve theme park service quality are discussed.

Keywords: theme park, scale development, China, service quality

Procedia PDF Downloads 243
21948 Spatial Scale of Clustering of Residential Burglary and Its Dependence on Temporal Scale

Authors: Mohammed A. Alazawi, Shiguo Jiang, Steven F. Messner

Abstract:

Research has long focused on two main spatial aspects of crime: spatial patterns and spatial processes. When analyzing these patterns and processes, a key issue has been to determine the proper spatial scale. In addition, it is important to consider the possibility that these patterns and processes might differ appreciably for different temporal scales and might vary across geographic units of analysis. We examine the spatial-temporal dependence of residential burglary. This dependence is tested at varying geographical scales and temporal aggregations. The analyses are based on recorded incidents of crime in Columbus, Ohio during the 1994-2002 period. We implement point pattern analysis on the crime points using Ripley’s K function. The results indicate that spatial point patterns of residential burglary reveal spatial scales of clustering relatively larger than the average size of census tracts of the study area. Also, spatial scale is independent of temporal scale. The results of our analyses concerning the geographic scale of spatial patterns and processes can inform the development of effective policies for crime control.

Keywords: inhomogeneous K function, residential burglary, spatial point pattern, spatial scale, temporal scale

Procedia PDF Downloads 309
21947 Produced Water Treatment Using Novel Solid Scale Inhibitors Based on Silver Tungstate Loaded Kit-6: Static and Modeling Evaluation

Authors: R. Hosny, Mahmoud F. Mubarak, Heba M. Salem, Asmaa A. Abdelrahman

Abstract:

Oilfield scaling is a major problem in the oil and gas industry. Scale issues cost the industry millions of dollars in damage and lost production every year. One of the main causes of global production decline is scale. In this study, solid scale inhibitors based on silver tungstate loaded KIT-6 were synthesized and evaluated in both static and scale inhibition modeling. The silver tungstate loaded KIT-6 catalysts were synthesized via a simple impregnated method using 3D mesoporous KIT-6 as support. The synthesized materials were characterized using wide and low XRD, N2 adsorption–desorption analysis, TGA analysis, and FTIR, SEM, and XPS analysis. The scale inhibition efficiency of the synthesized materials was evaluated using a static scale inhibition test. The results of this study demonstrate the potential application of silver tungstate-loaded KIT-6 solid scale inhibitors for the oil and gas industry. The results of this study will contribute to the development of new and innovative solid scale inhibitors based on silver tungstate-loaded KIT-6. The inhibition efficiency of the scale inhibitor increases, and calcite scale inhibitor decreases with increasing pH (2 to8), it proposes that the scale inhibitor was more effective under alkaline conditions. An inhibition efficiency of 99% on calcium carbonate can be achieved at the optimal dosage of 7.5 ppm at 55oC, indicating that the scale inhibitor exhibits a relatively good inhibition performance on calcium carbonate. The use of these materials can potentially lead to more efficient and cost-effective solutions for scaling inhibition in various industrial processes.

Keywords: produced water treatment, solid scale inhibitors, calcite, silver tungestate, 3 D mesoporous KIT-6, oilfield scales, adsorption

Procedia PDF Downloads 119
21946 Coupled Space and Time Homogenization of Viscoelastic-Viscoplastic Composites

Authors: Sarra Haouala, Issam Doghri

Abstract:

In this work, a multiscale computational strategy is proposed for the analysis of structures, which are described at a refined level both in space and in time. The proposal is applied to two-phase viscoelastic-viscoplastic (VE-VP) reinforced thermoplastics subjected to large numbers of cycles. The main aim is to predict the effective long time response while reducing the computational cost considerably. The proposed computational framework is a combination of the mean-field space homogenization based on the generalized incrementally affine formulation for VE-VP composites, and the asymptotic time homogenization approach for coupled isotropic VE-VP homogeneous solids under large numbers of cycles. The time homogenization method is based on the definition of micro and macro-chronological time scales, and on asymptotic expansions of the unknown variables. First, the original anisotropic VE-VP initial-boundary value problem of the composite material is decomposed into coupled micro-chronological (fast time scale) and macro-chronological (slow time-scale) problems. The former is purely VE, and solved once for each macro time step, whereas the latter problem is nonlinear and solved iteratively using fully implicit time integration. Second, mean-field space homogenization is used for both micro and macro-chronological problems to determine the micro and macro-chronological effective behavior of the composite material. The response of the matrix material is VE-VP with J2 flow theory assuming small strains. The formulation exploits the return-mapping algorithm for the J2 model, with its two steps: viscoelastic predictor and plastic corrections. The proposal is implemented for an extended Mori-Tanaka scheme, and verified against finite element simulations of representative volume elements, for a number of polymer composite materials subjected to large numbers of cycles.

Keywords: asymptotic expansions, cyclic loadings, inclusion-reinforced thermoplastics, mean-field homogenization, time homogenization

Procedia PDF Downloads 338