Search results for: general merge calculation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6149

Search results for: general merge calculation

6119 Numerical Methods versus Bjerksund and Stensland Approximations for American Options Pricing

Authors: Marasovic Branka, Aljinovic Zdravka, Poklepovic Tea

Abstract:

Numerical methods like binomial and trinomial trees and finite difference methods can be used to price a wide range of options contracts for which there are no known analytical solutions. American options are the most famous of that kind of options. Besides numerical methods, American options can be valued with the approximation formulas, like Bjerksund-Stensland formulas from 1993 and 2002. When the value of American option is approximated by Bjerksund-Stensland formulas, the computer time spent to carry out that calculation is very short. The computer time spent using numerical methods can vary from less than one second to several minutes or even hours. However to be able to conduct a comparative analysis of numerical methods and Bjerksund-Stensland formulas, we will limit computer calculation time of numerical method to less than one second. Therefore, we ask the question: Which method will be most accurate at nearly the same computer calculation time?

Keywords: Bjerksund and Stensland approximations, computational analysis, finance, options pricing, numerical methods

Procedia PDF Downloads 419
6118 Modeling Nanomechanical Behavior of ZnO Nanowires as a Function of Nano-Diameter

Authors: L. Achou, A. Doghmane

Abstract:

Elastic performances, as an essential property of nanowires (NWs), play a significant role in the design and fabrication of modern nanodevices. In this paper, our interest is focused on ZnO NWs to investigate wire diameter (Dwire ≤ 400 nm) effects on elastic properties. The plotted data reveal that a strong size dependence of the elastic constants exists when the wire diameter is smaller than ~ 100 nm. For larger diameters (Dwire > 100 nm), these ones approach their corresponding bulk values. To enrich this study, we make use of the scanning acoustic microscopy simulation technique. The calculation methodology consists of several steps: determination of longitudinal and transverse wave velocities, calculation of refection coefficients, calculation of acoustic signatures and Rayleigh velocity determination. Quantitatively, it was found that changes in ZnO diameters over the ranges 1 nm ≤ Dwire ≤ 100 nm lead to similar exponential variations, for all elastic parameters, of the from: A = a + b exp(-Dwire/c) where a, b, and c are characteristic constants of a given parameter. The developed relation can be used to predict elastic properties of such NW by just knowing its diameter and vice versa.

Keywords: elastic properties, nanowires, semiconductors, theoretical model, ZnO

Procedia PDF Downloads 137
6117 Placement Characteristics of Major Stream Vehicular Traffic at Median Openings

Authors: Tathagatha Khan, Smruti Sourava Mohapatra

Abstract:

Median openings are provided in raised median of multilane roads to facilitate U-turn movement. The U-turn movement is a highly complex and risky maneuver because U-turning vehicle (minor stream) makes 180° turns at median openings and merge with the approaching through traffic (major stream). A U-turning vehicle requires a suitable gap in the major stream to merge, and during this process, the possibility of merging conflict develops. Therefore, these median openings are potential hot spot of conflict and posses concern pertaining to safety. The traffic at the median openings could be managed efficiently with enhanced safety when the capacity of a traffic facility has been estimated correctly. The capacity of U-turns at median openings is estimated by Harder’s formula, which requires three basic parameters namely critical gap, follow up time and conflict flow rate. The estimation of conflicting flow rate under mixed traffic condition is very much complicated due to absence of lane discipline and discourteous behavior of the drivers. The understanding of placement of major stream vehicles at median opening is very much important for the estimation of conflicting traffic faced by U-turning movement. The placement data of major stream vehicles at different section in 4-lane and 6-lane divided multilane roads were collected. All the test sections were free from the effect of intersection, bus stop, parked vehicles, curvature, pedestrian movements or any other side friction. For the purpose of analysis, all the vehicles were divided into 6 categories such as motorized 2W, autorickshaw (3-W), small car, big car, light commercial vehicle, and heavy vehicle. For the collection of placement data of major stream vehicles, the entire road width was divided into sections of 25 cm each and these were numbered seriatim from the pavement edge (curbside) to the end of the road. The placement major stream vehicle crossing the reference line was recorded by video graphic technique on various weekdays. The collected data for individual category of vehicles at all the test sections were converted into a frequency table with a class interval of 25 cm each and the placement frequency curve. Separate distribution fittings were tried for 4- lane and 6-lane divided roads. The variation of major stream traffic volume on the placement characteristics of major stream vehicles has also been explored. The findings of this study will be helpful to determine the conflict volume at the median openings. So, the present work holds significance in traffic planning, operation and design to alleviate the bottleneck, prospect of collision and delay at median opening in general and at median opening in developing countries in particular.

Keywords: median opening, U-turn, conflicting traffic, placement, mixed traffic

Procedia PDF Downloads 112
6116 Acoustic Finite Element Analysis of a Slit Model with Consideration of Air Viscosity

Authors: M. Sasajima, M. Watanabe, T. Yamaguchi Y. Kurosawa, Y. Koike

Abstract:

In very narrow pathways, the speed of sound propagation and the phase of sound waves change due to the air viscosity. We have developed a new Finite Element Method (FEM) that includes the effects of air viscosity for modeling a narrow sound pathway. This method is developed as an extension of the existing FEM for porous sound-absorbing materials. The numerical calculation results for several three-dimensional slit models using the proposed FEM are validated against existing calculation methods.

Keywords: simulation, FEM, air viscosity, slit

Procedia PDF Downloads 340
6115 Evaluation of Carbon Dioxide Pressure through Radial Velocity Difference in Arterial Blood Modeled by Drift Flux Model

Authors: Aicha Rima Cheniti, Hatem Besbes, Joseph Haggege, Christophe Sintes

Abstract:

In this paper, we are interested to determine the carbon dioxide pressure in the arterial blood through radial velocity difference. The blood was modeled as a two phase mixture (an aqueous carbon dioxide solution with carbon dioxide gas) by Drift flux model and the Young-Laplace equation. The distributions of mixture velocities determined from the considered model permitted the calculation of the radial velocity distributions with different values of mean mixture pressure and the calculation of the mean carbon dioxide pressure knowing the mean mixture pressure. The radial velocity distributions are used to deduce a calculation method of the mean mixture pressure through the radial velocity difference between two positions which is measured by ultrasound. The mean carbon dioxide pressure is then deduced from the mean mixture pressure.

Keywords: mean carbon dioxide pressure, mean mixture pressure, mixture velocity, radial velocity difference

Procedia PDF Downloads 391
6114 Modeling and Calculation of Physical Parameters of the Pollution of Water by Oil and Materials in Suspensions

Authors: Ainas Belkacem, Fourar Ali

Abstract:

The present study focuses on the mathematical modeling and calculation of physical parameters of water pollution by oil and sand in regime fully dispersed in water. In this study, the sand particles and oil are suspended in the case of fully developed turbulence. The study consists to understand, model and predict the viscosity, the structure and dynamics of these types of mixtures. The work carried out is Numerical and validated by experience.

Keywords: multi phase flow, pollution, suspensions, turbulence

Procedia PDF Downloads 210
6113 Determination Power and Sample Size Zero-Inflated Negative Binomial Dependent Death Rate of Age Model (ZINBD): Regression Analysis Mortality Acquired Immune Deficiency De ciency Syndrome (AIDS)

Authors: Mohd Asrul Affendi Bin Abdullah

Abstract:

Sample size calculation is especially important for zero inflated models because a large sample size is required to detect a significant effect with this model. This paper verify how to present percentage of power approximation for categorical and then extended to zero inflated models. Wald test was chosen to determine power sample size of AIDS death rate because it is frequently used due to its approachability and its natural for several major recent contribution in sample size calculation for this test. Power calculation can be conducted when covariates are used in the modeling ‘excessing zero’ data and assist categorical covariate. Analysis of AIDS death rate study is used for this paper. Aims of this study to determine the power of sample size (N = 945) categorical death rate based on parameter estimate in the simulation of the study.

Keywords: power sample size, Wald test, standardize rate, ZINBDR

Procedia PDF Downloads 408
6112 Characterization of the in 0.53 Ga 0.47 as n+nn+ Photodetectors

Authors: Fatima Zohra Mahi, Luca Varani

Abstract:

We present an analytical model for the calculation of the sensitivity, the spectral current noise and the detectivity for an optically illuminated In0.53Ga0.47As n+nn+ diode. The photocurrent due to the excess carrier is obtained by solving the continuity equation. Moreover, the current noise level is evaluated at room temperature and under a constant voltage applied between the diode terminals. The analytical calculation of the current noise in the n+nn+ structure is developed. The responsivity and the detectivity are discussed as functions of the doping concentrations and the emitter layer thickness in one-dimensional homogeneous n+nn+ structure.

Keywords: detectivity, photodetectors, continuity equation, current noise

Procedia PDF Downloads 610
6111 Calculation and Comparison of a Turbofan Engine Performance Parameters with Various Definitions

Authors: O. Onal, O. Turan

Abstract:

In this paper, some performance parameters of a selected turbofan engine (JT9D) are analyzed. The engine is a high bypass turbofan engine which powers a wide-body aircraft and it produces 206 kN thrust force (thrust/weight ratio is 5.4). The objective parameters for the engine include calculation of power, specific fuel consumption, specific thrust, engine propulsive, thermal and overall efficiencies according to the various definitions given in the literature. Furthermore, in the case study, wasted energy from the exhaust is calculated at the maximum power setting (i.e. take off phase) for the engine.

Keywords: turbofan, power, efficiency, trust

Procedia PDF Downloads 270
6110 Investigating Jacket-Type Offshore Structures Failure Probability by Applying the Reliability Analyses Methods

Authors: Majid Samiee Zonoozian

Abstract:

For such important constructions as jacket type platforms, scrupulous attention in analysis, design and calculation processes is needed. The reliability assessment method has been established into an extensively used method to behavior safety calculation of jacket platforms. In the present study, a methodology for the reliability calculation of an offshore jacket platform in contradiction of the extreme wave loading state is available. Therefore, sensitivity analyses are applied to acquire the nonlinear response of jacket-type platforms against extreme waves. The jacket structure is modeled by applying a nonlinear finite-element model with regards to the tubular members' behave. The probability of a member’s failure under extreme wave loading is figured by a finite-element reliability code. The FORM and SORM approaches are applied for the calculation of safety directories and reliability indexes have been detected. A case study for a fixed jacket-type structure positioned in the Persian Gulf is studied by means of the planned method. Furthermore, to define the failure standards, equations suggested by the 21st version of the API RP 2A-WSD for The jacket-type structures’ tubular members designing by applying the mixed axial bending and axial pressure. Consequently, the effect of wave Loades in the reliability index was considered.

Keywords: Jacket-Type structure, reliability, failure probability, tubular members

Procedia PDF Downloads 147
6109 Statistical Characteristics of Code Formula for Design of Concrete Structures

Authors: Inyeol Paik, Ah-Ryang Kim

Abstract:

In this research, a statistical analysis is carried out to examine the statistical properties of the formula given in the design code for concrete structures. The design formulas of the Korea highway bridge design code - the limit state design method (KHBDC) which is the current national bridge design code and the design code for concrete structures by Korea Concrete Institute (KCI) are applied for the analysis. The safety levels provided by the strength formulas of the design codes are defined based on the probabilistic and statistical theory.KHBDC is a reliability-based design code. The load and resistance factors of this code were calibrated to attain the target reliability index. It is essential to define the statistical properties for the design formulas in this calibration process. In general, the statistical characteristics of a member strength are due to the following three factors. The first is due to the difference between the material strength of the actual construction and that used in the design calculation. The second is the difference between the actual dimensions of the constructed sections and those used in design calculation. The third is the difference between the strength of the actual member and the formula simplified for the design calculation. In this paper, the statistical study is focused on the third difference. The formulas for calculating the shear strength of concrete members are presented in different ways in KHBDC and KCI. In this study, the statistical properties of design formulas were obtained through comparison with the database which comprises the experimental results from the reference publications. The test specimen was either reinforced with the shear stirrup or not. For an applied database, the bias factor was about 1.12 and the coefficient of variation was about 0.18. By applying the statistical properties of the design formula to the reliability analysis, it is shown that the resistance factors of the current design codes satisfy the target reliability indexes of both codes. Also, the minimum resistance factors of the KHBDC which is written in the material resistance factor format and KCE which is in the member resistance format are obtained and the results are presented. A further research is underway to calibrate the resistance factors of the high strength and high-performance concrete design guide.

Keywords: concrete design code, reliability analysis, resistance factor, shear strength, statistical property

Procedia PDF Downloads 289
6108 Characteristic Sentence Stems in Academic English Texts: Definition, Identification, and Extraction

Authors: Jingjie Li, Wenjie Hu

Abstract:

Phraseological units in academic English texts have been a central focus in recent corpus linguistic research. A wide variety of phraseological units have been explored, including collocations, chunks, lexical bundles, patterns, semantic sequences, etc. This paper describes a special category of clause-level phraseological units, namely, Characteristic Sentence Stems (CSSs), with a view to describing their defining criteria and extraction method. CSSs are contiguous lexico-grammatical sequences which contain a subject-predicate structure and which are frame expressions characteristic of academic writing. The extraction of CSSs consists of six steps: Part-of-speech tagging, n-gram segmentation, structure identification, significance of occurrence calculation, text range calculation, and overlapping sequence reduction. Significance of occurrence calculation is the crux of this study. It includes the computing of both the internal association and the boundary independence of a CSS and tests the occurring significance of the CSS from both inside and outside perspectives. A new normalization algorithm is also introduced into the calculation of LocalMaxs for reducing overlapping sequences. It is argued that many sentence stems are so recurrent in academic texts that the most typical of them have become the habitual ways of making meaning in academic writing. Therefore, studies of CSSs could have potential implications and reference value for academic discourse analysis, English for Academic Purposes (EAP) teaching and writing.

Keywords: characteristic sentence stem, extraction method, phraseological unit, the statistical measure

Procedia PDF Downloads 137
6107 Human Resource Development Climate (HRDC) in Nigerian Banks: General and Gender Perceptions

Authors: Akinyemi Benjamin

Abstract:

This study investigates the prevailing HRDC Nigerian commercial banks as perceived by employees in general. The perceptional differences on the state of HRDC by gender category are also examined. Using Abraham and Rao’s HRDC 38-item questionnaire, data from 310 respondents, with 303 valid responses, were entered into excel sheet and analysed to determine frequencies, mean scores, standard deviation and percentages for four variables: HRDC, general climate, HRD mechanism, and OCTAPAC culture. Results of analysis indicate that generally, employees perceive the overall HRDC and its three dimensions of general climate, HRD mechanism and OCTAPAC culture to be at an average or moderate level. The perceptions of both male and female subjects also indicate an average HRDC level although participants report slightly higher scores than their male subjects but these scores are still at an average level on all the dimensions of HRDC measured. The implications of this result for organizations in general and the banking industry in particular are discussed.

Keywords: HRDC, HRD mechanism, general climate, OCTAPAC culture, gender

Procedia PDF Downloads 384
6106 Conduction Transfer Functions for the Calculation of Heat Demands in Heavyweight Facade Systems

Authors: Mergim Gasia, Bojan Milovanovica, Sanjin Gumbarevic

Abstract:

Better energy performance of the building envelope is one of the most important aspects of energy savings if the goals set by the European Union are to be achieved in the future. Dynamic heat transfer simulations are being used for the calculation of building energy consumption because they give more realistic energy demands compared to the stationary calculations that do not take the building’s thermal mass into account. Software used for these dynamic simulation use methods that are based on the analytical models since numerical models are insufficient for longer periods. The analytical models used in this research fall in the category of the conduction transfer functions (CTFs). Two methods for calculating the CTFs covered by this research are the Laplace method and the State-Space method. The literature review showed that the main disadvantage of these methods is that they are inadequate for heavyweight façade elements and shorter time periods used for the calculation. The algorithms for both the Laplace and State-Space methods are implemented in Mathematica, and the results are compared to the results from EnergyPlus and TRNSYS since these software use similar algorithms for the calculation of the building’s energy demand. This research aims to check the efficiency of the Laplace and the State-Space method for calculating the building’s energy demand for heavyweight building elements and shorter sampling time, and it also gives the means for the improvement of the algorithms used by these methods. As the reference point for the boundary heat flux density, the finite difference method (FDM) is used. Even though the dynamic heat transfer simulations are superior to the calculation based on the stationary boundary conditions, they have their limitations and will give unsatisfactory results if not properly used.

Keywords: Laplace method, state-space method, conduction transfer functions, finite difference method

Procedia PDF Downloads 98
6105 Comparison of Different Intraocular Lens Power Calculation Formulas in People With Very High Myopia

Authors: Xia Chen, Yulan Wang

Abstract:

purpose: To compare the accuracy of Haigis, SRK/T, T2, Holladay 1, Hoffer Q, Barrett Universal II, Emmetropia Verifying Optical (EVO) and Kane for intraocular lens power calculation in patients with axial length (AL) ≥ 28 mm. Methods: In this retrospective single-center study, 50 eyes of 41 patients with AL ≥ 28 mm that underwent uneventful cataract surgery were enrolled. The actual postoperative refractive results were compared to the predicted refraction calculated with different formulas (Haigis, SRK/T, T2, Holladay 1, Hoffer Q, Barrett Universal II, EVO and Kane). The mean absolute prediction errors (MAE) 1 month postoperatively were compared. Results: The MAE of different formulas were as follows: Haigis (0.509), SRK/T (0.705), T2 (0.999), Holladay 1 (0.714), Hoffer Q (0.583), Barrett Universal II (0.552), EVO (0.463) and Kane (0.441). No significant difference was found among the different formulas (P = .122). The Kane and EVO formulas achieved the lowest level of mean prediction error (PE) and median absolute error (MedAE) (p < 0.05). Conclusion: The Kane and EVO formulas had a better success rate than others in predicting IOL power in high myopic eyes with AL longer than 28 mm in this study.

Keywords: cataract, power calculation formulas, intraocular lens, long axial length

Procedia PDF Downloads 37
6104 Exposure to Bullying and General Psychopathology: A Prospective, Longitudinal Study

Authors: Jolien Rijlaarsdam, Charlotte A. M. Cecil, J. Marieke Buil, Pol A. C. Van Lier, Edward D. Barker

Abstract:

Although there is mounting evidence that the experience of being bullied associates with both internalizing and externalizing symptoms, it is not known yet whether the identified associations are specific to these symptoms or shared between them. The primary focus of this study is to assess the prospective associations of bullying exposure with both general and specific (i.e., internalizing, externalizing) factors of psychopathology. This study included data from 6,210 children participating in the Avon Longitudinal Study of Parents and Children (ALSPAC). Child bullying was measured by self-report at ages 8 and 10 years. Child psychopathology symptoms were assessed by parent-interview, using the Development and Well-being Assessment (DAWBA) at ages 7 and 13 years. Bullying exposure is significantly associated with the general psychopathology factor in early adolescence. In particular, chronically victimized youth exposed to multiple forms of bullying (i.e., both overt and relational) showed the highest levels of general psychopathology. Bullying exposure is also associated with both internalizing and externalizing factors from the correlated-factors model. However, the effect estimates for these factors decreased considerably in size and dropped to insignificant for the internalizing factor after extracting the shared variance that belongs to the general factor of psychopathology. In an integrative longitudinal model, higher levels of general psychopathology at age seven are associated with bullying exposure at age eight, which, in turn, is associated with general psychopathology at age 13 through its two-year continuity. Findings suggest that exposure to bullying is a risk factor for a more general vulnerability to psychopathology through mutually influencing relationships.

Keywords: bullying exposure, externalizing, general psychopathology, internalizing, longitudinal

Procedia PDF Downloads 116
6103 Optimization of Heat Insulation Structure and Heat Flux Calculation Method of Slug Calorimeter

Authors: Zhu Xinxin, Wang Hui, Yang Kai

Abstract:

Heat flux is one of the most important test parameters in the ground thermal protection test. Slug calorimeter is selected as the main sensor measuring heat flux in arc wind tunnel test due to the convenience and low cost. However, because of excessive lateral heat transfer and the disadvantage of the calculation method, the heat flux measurement error of the slug calorimeter is large. In order to enhance measurement accuracy, the heat insulation structure and heat flux calculation method of slug calorimeter were improved. The heat transfer model of the slug calorimeter was built according to the energy conservation principle. Based on the heat transfer model, the insulating sleeve of the hollow structure was designed, which helped to greatly decrease lateral heat transfer. And the slug with insulating sleeve of hollow structure was encapsulated using a package shell. The improved insulation structure reduced heat loss and ensured that the heat transfer characteristics were almost the same when calibrated and tested. The heat flux calibration test was carried out in arc lamp system for heat flux sensor calibration, and the results show that test accuracy and precision of slug calorimeter are improved greatly. In the meantime, the simulation model of the slug calorimeter was built. The heat flux values in different temperature rise time periods were calculated by the simulation model. The results show that extracting the data of the temperature rise rate as soon as possible can result in a smaller heat flux calculation error. Then the different thermal contact resistance affecting calculation error was analyzed by the simulation model. The contact resistance between the slug and the insulating sleeve was identified as the main influencing factor. The direct comparison calibration correction method was proposed based on only heat flux calibration. The numerical calculation correction method was proposed based on the heat flux calibration and simulation model of slug calorimeter after the simulation model was solved by solving the contact resistance between the slug and the insulating sleeve. The simulation and test results show that two methods can greatly reduce the heat flux measurement error. Finally, the improved slug calorimeter was tested in the arc wind tunnel. And test results show that the repeatability accuracy of improved slug calorimeter is less than 3%. The deviation of measurement value from different slug calorimeters is less than 3% in the same fluid field. The deviation of measurement value between slug calorimeter and Gordon Gage is less than 4% in the same fluid field.

Keywords: correction method, heat flux calculation, heat insulation structure, heat transfer model, slug calorimeter

Procedia PDF Downloads 90
6102 [Keynote Talk]: Applying p-Balanced Energy Technique to Solve Liouville-Type Problems in Calculus

Authors: Lina Wu, Ye Li, Jia Liu

Abstract:

We are interested in solving Liouville-type problems to explore constancy properties for maps or differential forms on Riemannian manifolds. Geometric structures on manifolds, the existence of constancy properties for maps or differential forms, and energy growth for maps or differential forms are intertwined. In this article, we concentrate on discovery of solutions to Liouville-type problems where manifolds are Euclidean spaces (i.e. flat Riemannian manifolds) and maps become real-valued functions. Liouville-type results of vanishing properties for functions are obtained. The original work in our research findings is to extend the q-energy for a function from finite in Lq space to infinite in non-Lq space by applying p-balanced technique where q = p = 2. Calculation skills such as Hölder's Inequality and Tests for Series have been used to evaluate limits and integrations for function energy. Calculation ideas and computational techniques for solving Liouville-type problems shown in this article, which are utilized in Euclidean spaces, can be universalized as a successful algorithm, which works for both maps and differential forms on Riemannian manifolds. This innovative algorithm has a far-reaching impact on research work of solving Liouville-type problems in the general settings involved with infinite energy. The p-balanced technique in this algorithm provides a clue to success on the road of q-energy extension from finite to infinite.

Keywords: differential forms, holder inequality, Liouville-type problems, p-balanced growth, p-harmonic maps, q-energy growth, tests for series

Procedia PDF Downloads 208
6101 Friction Calculation and Simulation of Column Electric Power Steering System

Authors: Seyed Hamid Mirmohammad Sadeghi, Raffaella Sesana, Daniela Maffiodo

Abstract:

This study presents a procedure for friction calculation of column electric power steering (C-EPS) system which affects handling and comfort in driving. The friction losses estimation is obtained from experimental tests and mathematical calculation. Parts in C-EPS mainly involved in friction losses are bearings and worm gear. In the theoretical approach, the gear geometry and Hertz law were employed to measure the normal load and the sliding velocity and contact areas from the worm gears driving conditions. The viscous friction generated in the worm gear was obtained with a theoretical approach and the result was applied to model the friction in the steering system. Finally, by viscous friction coefficient and Coulomb friction coefficient, values of friction in worm gear were calculated. According to the Bearing Company and the characteristics of each bearing, the friction torques due to load and due to speed were calculated. A MATLAB Simulink model for calculating the friction in bearings and worm gear in C-EPS were done and the total friction value was estimated.

Keywords: friction, worm gear, column electric power steering system, simulink, bearing, EPS

Procedia PDF Downloads 326
6100 The Relation between Subtitling and General Translation from a Didactic Perspective

Authors: Sonia Gonzalez Cruz

Abstract:

Subtitling activities allow for acquiring and developing certain translation skills, and they also have a great impact on the students' motivation. Active subtitling is a relatively recent activity that has generated a lot of interest particularly in the field of second-language acquisition, but it is also present within both the didactics of general translation and language teaching for translators. It is interesting to analyze the level of inclusion of these new resources into the existent curricula and observe to what extent these different teaching methods are being used in the translation classroom. Although subtitling has already become an independent discipline of study and it is considered to be a type of translation on its own, it is necessary to do further research on the different didactic varieties that this type of audiovisual translation offers. Therefore, this project is framed within the field of the didactics of translation, and it focuses on the relationship between the didactics of general translation and active subtitling as a didactic tool. Its main objective is to analyze the inclusion of interlinguistic active subtitling in general translation curricula at different universities. As it has been observed so far, the analyzed curricula do not make any type of reference to the use of this didactic tool in general translation classrooms. However, they do register the inclusion of other audiovisual activities such as dubbing, script translation or video watching, among others. By means of online questionnaires and interviews, the main goal is to confirm the results obtained after the observation of the curricula and find out to what extent subtitling has actually been included into general translation classrooms.

Keywords: subtitling, general translation, didactics, translation competence

Procedia PDF Downloads 142
6099 A Heuristic Approach for the General Flowshop Scheduling Problem to Minimize the Makespan

Authors: Mohsen Ziaee

Abstract:

Almost all existing researches on the flowshop scheduling problems focus on the permutation schedules and there is insufficient study dedicated to the general flowshop scheduling problems in the literature, since the modeling and solving of the general flowshop scheduling problems are more difficult than the permutation ones, especially for the large-size problem instances. This paper considers the general flowshop scheduling problem with the objective function of the makespan (F//Cmax). We first find the optimal solution of the problem by solving a mixed integer linear programming model. An efficient heuristic method is then presented to solve the problem. An ant colony optimization algorithm is also proposed for the problem. In order to evaluate the performance of the methods, computational experiments are designed and performed. Numerical results show that the heuristic algorithm can result in reasonable solutions with low computational effort and even achieve optimal solutions in some cases.

Keywords: scheduling, general flow shop scheduling problem, makespan, heuristic

Procedia PDF Downloads 178
6098 Accurate Calculation of the Penetration Depth of a Bullet Using ANSYS

Authors: Eunsu Jang, Kang Park

Abstract:

In developing an armored ground combat vehicle (AGCV), it is a very important step to analyze the vulnerability (or the survivability) of the AGCV against enemy’s attack. In the vulnerability analysis, the penetration equations are usually used to get the penetration depth and check whether a bullet can penetrate the armor of the AGCV, which causes the damage of internal components or crews. The penetration equations are derived from penetration experiments which require long time and great efforts. However, they usually hold only for the specific material of the target and the specific type of the bullet used in experiments. Thus, penetration simulation using ANSYS can be another option to calculate penetration depth. However, it is very important to model the targets and select the input parameters in order to get an accurate penetration depth. This paper performed a sensitivity analysis of input parameters of ANSYS on the accuracy of the calculated penetration depth. Two conflicting objectives need to be achieved in adopting ANSYS in penetration analysis: maximizing the accuracy of calculation and minimizing the calculation time. To maximize the calculation accuracy, the sensitivity analysis of the input parameters for ANSYS was performed and calculated the RMS error with the experimental data. The input parameters include mesh size, boundary condition, material properties, target diameter are tested and selected to minimize the error between the calculated result from simulation and the experiment data from the papers on the penetration equation. To minimize the calculation time, the parameter values obtained from accuracy analysis are adjusted to get optimized overall performance. As result of analysis, the followings were found: 1) As the mesh size gradually decreases from 0.9 mm to 0.5 mm, both the penetration depth and calculation time increase. 2) As diameters of the target decrease from 250mm to 60 mm, both the penetration depth and calculation time decrease. 3) As the yield stress which is one of the material property of the target decreases, the penetration depth increases. 4) The boundary condition with the fixed side surface of the target gives more penetration depth than that with the fixed side and rear surfaces. By using above finding, the input parameters can be tuned to minimize the error between simulation and experiments. By using simulation tool, ANSYS, with delicately tuned input parameters, penetration analysis can be done on computer without actual experiments. The data of penetration experiments are usually hard to get because of security reasons and only published papers provide them in the limited target material. The next step of this research is to generalize this approach to anticipate the penetration depth by interpolating the known penetration experiments. This result may not be accurate enough to be used to replace the penetration experiments, but those simulations can be used in the early stage of the design process of AGCV in modelling and simulation stage.

Keywords: ANSYS, input parameters, penetration depth, sensitivity analysis

Procedia PDF Downloads 361
6097 Treatment of the Modern Management Mechanism of the Debris Flow Processes Expected in the Mletiskhevi

Authors: G. Chakhaia, S. Gogilava, L. Tsulukidze, Z. Laoshvili, I. Khubulava, S. Bosikashvili, T. Gugushvili

Abstract:

The work reviewed and evaluated various genesis debris flow phenomena recently formatted in the Mletiskhevi, accordingly it revealed necessity of treatment modern debris flow against measures. Based on this, it is proposed the debris flow against truncated semi cone shape construction, which elements are contained in the car’s secondary tires. its constituent elements (sections), due to the possibilities of amortization and geometric shapes is effective and sustainable towards debris flow hitting force. The construction is economical, because after crossing the debris flows in the river bed, the riverbed is not cleanable, also the elements of the building are resource saving. For assessment of influence of cohesive debris flow at the construction and evaluation of the construction effectiveness have been implemented calculation in the specific assumptions with approved methodology. According to the calculation, it was established that after passing debris flow in the debris flow construction (in 3 row case) its hitting force reduces 3 times, that causes reduce of debris flow speed and kinetic energy, as well as sedimentation on a certain section of water drain in the lower part of the construction. Based on the analysis and report on the debris flow against construction, it can be said that construction is effective, inexpensive, technically relatively easy-to-reach measure, that’s why its implementation is prospective.

Keywords: construction, debris flow, sections, theoretical calculation

Procedia PDF Downloads 156
6096 Location Quotient Analysis: Case Study

Authors: Seyed Habib A. Rahmati, Mohamad Hasan Sadeghpour, Parsa Fallah Sheikhlari

Abstract:

Location quotient (LQ) is a comparison technique that represents emphasized economic structure of single zone versus the standard area to identify specialty for every zone. In another words, the exact calculation of this metric can show the main core competencies and critical capabilities of an area to the decision makers. This research focus on the exact calculation of the LQ for an Iranian Province called Qazvin and within a case study introduces LQ of the capable industries of Qazvin. Finally, through different graphs and tables, it creates an opportunity to compare the recognized capabilities.

Keywords: location quotient, case study, province analysis, core competency

Procedia PDF Downloads 618
6095 Selectivity Mechanism of Cobalt Precipitation by an Imidazole Linker From an Old Battery Solution

Authors: Anna-Caroline Lavergne-Bril, Jean-François Colin, David Peralta, Pascale Maldivi

Abstract:

Cobalt is a critical material, widely used in Li-ion batteries. Due to the planned electrification of European vehicles, cobalt needs are expending – and resources are limited. To meet the needs in cobalt to come, it is necessary to develop new efficient ways to recycle cobalt. One of the biggest sources comes from old electrical vehicles batteries (batteries sold in 2019: 500 000 tons of waste to be). A closed loop process of cobalt recycling has been developed and this presentation aims to present the selectivity mechanism of cobalt over manganese and nickel in solution. Cobalt precipitation as a ZIF material (Zeolitic Imidazolate framework) from a starting solution composed of equimolar nickel, manganese and cobalt is studied. A 2-MeIm (2-methylimidazole) linker is introduced in a multimetallic Ni, Mn, Co solution and the resulting ZIF-67 is 100% pure Co among its metallic centers. Selectivity of Co over Ni is experimentally studied and DFT modelisation calculation are conducted to understand the geometry of ligand-metal-solvent complexes in solution. Selectivity of Co over Mn is experimentally studied, and DFT modelisation calcucation are conducted to understand the link between pKa of the ligand and precipitration of Mn impurities within the final material. Those calculation open the way to other ligand being used in the same process, with more efficiency. Experimental material are synthetized from bimetallic (Ni²⁺/Co²⁺, Mn²⁺/Co²⁺, Mn²⁺/Ni²⁺) solutions. Their crystallographic structure is analysed by XRD diffraction (Brüker AXS D8 diffractometer, Cu anticathode). Morphology is studied by scanning electron microscopy, using a LEO 1530 FE-SEM microscope. The chemical analysis is performed by using ICP-OES (Agilent Technologies 700 series ICP-OES). Modelisation calculation are DFT calculation (density functional theory), using B3LYP, conducted with Orca 4.2.

Keywords: MOFs, ZIFs, recycling, closed-loop, cobalt, li-ion batteries

Procedia PDF Downloads 99
6094 Calculation of Lungs Physiological Lung Motion in External Lung Irradiation

Authors: Yousif Mohamed Y. Abdallah, Khalid H. Eltom

Abstract:

This is an experimental study deals with measurement of the periodic physiological organ motion during lung external irradiation in order to reduce the exposure of healthy tissue during radiation treatments. The results showed for left lung displacement reading (4.52+1.99 mm) and right lung is (8.21+3.77 mm) which the radiotherapy physician should take suitable countermeasures in case of significant errors. The motion ranged between 2.13 mm and 12.2 mm (low and high). In conclusion, the calculation of tumour mobility can improve the accuracy of target areas definition in patients undergo Sterostatic RT for stage I, II and III lung cancer (NSCLC). Definition of the target volume based on a high resolution CT scan with a margin of 3-5 mm is appropriate.

Keywords: physiological motion, lung, external irradiation, radiation medicine

Procedia PDF Downloads 384
6093 Heterogeneous-Resolution and Multi-Source Terrain Builder for CesiumJS WebGL Virtual Globe

Authors: Umberto Di Staso, Marco Soave, Alessio Giori, Federico Prandi, Raffaele De Amicis

Abstract:

The increasing availability of information about earth surface elevation (Digital Elevation Models DEM) generated from different sources (remote sensing, Aerial Images, Lidar) poses the question about how to integrate and make available to the most than possible audience this huge amount of data. In order to exploit the potential of 3D elevation representation the quality of data management plays a fundamental role. Due to the high acquisition costs and the huge amount of generated data, highresolution terrain surveys tend to be small or medium sized and available on limited portion of earth. Here comes the need to merge large-scale height maps that typically are made available for free at worldwide level, with very specific high resolute datasets. One the other hand, the third dimension increases the user experience and the data representation quality, unlocking new possibilities in data analysis for civil protection, real estate, urban planning, environment monitoring, etc. The open-source 3D virtual globes, which are trending topics in Geovisual Analytics, aim at improving the visualization of geographical data provided by standard web services or with proprietary formats. Typically, 3D Virtual globes like do not offer an open-source tool that allows the generation of a terrain elevation data structure starting from heterogeneous-resolution terrain datasets. This paper describes a technological solution aimed to set up a so-called “Terrain Builder”. This tool is able to merge heterogeneous-resolution datasets, and to provide a multi-resolution worldwide terrain services fully compatible with CesiumJS and therefore accessible via web using traditional browser without any additional plug-in.

Keywords: Terrain Builder, WebGL, Virtual Globe, CesiumJS, Tiled Map Service, TMS, Height-Map, Regular Grid, Geovisual Analytics, DTM

Procedia PDF Downloads 397
6092 Comparative Analysis of the Computer Methods' Usage for Calculation of Hydrocarbon Reserves in the Baltic Sea

Authors: Pavel Shcherban, Vlad Golovanov

Abstract:

Nowadays, the depletion of hydrocarbon deposits on the land of the Kaliningrad region leads to active geological exploration and development of oil and natural gas reserves in the southeastern part of the Baltic Sea. LLC 'Lukoil-Kaliningradmorneft' implements a comprehensive program for the development of the region's shelf in 2014-2023. Due to heterogeneity of reservoir rocks in various open fields, as well as with ambiguous conclusions on the contours of deposits, additional geological prospecting and refinement of the recoverable oil reserves are carried out. The key element is use of an effective technique of computer stock modeling at the first stage of processing of the received data. The following step uses information for the cluster analysis, which makes it possible to optimize the field development approaches. The article analyzes the effectiveness of various methods for reserves' calculation and computer modelling methods of the offshore hydrocarbon fields. Cluster analysis allows to measure influence of the obtained data on the development of a technical and economic model for mining deposits. The relationship between the accuracy of the calculation of recoverable reserves and the need of modernization of existing mining infrastructure, as well as the optimization of the scheme of opening and development of oil deposits, is observed.

Keywords: cluster analysis, computer modelling of deposits, correction of the feasibility study, offshore hydrocarbon fields

Procedia PDF Downloads 139
6091 Comparing the Contribution of General Vocabulary Knowledge and Academic Vocabulary Knowledge to Learners' Academic Achievement

Authors: Reem Alsager, James Milton

Abstract:

Coxhead’s (2000) Academic Word List (AWL) believed to be essential for students pursuing higher education and helps differentiate English for Academic Purposes (EAP) from General English as a course of study, and it is thought to be important for comprehending English academic texts. It has been described that AWL is an infrequent, discrete set of vocabulary items unreachable from general language. On the other hand, it has been known for a period of time that general vocabulary knowledge is a good predictor of academic achievement. This study, however, is an attempt to measure and compare the contribution of academic knowledge and general vocabulary knowledge to learners’ GPA and examine what knowledge is a better predictor of academic achievement and investigate whether AWL as a specialised list of infrequent words relates to the frequency effect. The participants were comprised of 44 international postgraduate students in Swansea University, all from the School of Management, following the taught MSc (Master of Science). The study employed the Academic Vocabulary Size Test (AVST) and the XK_Lex vocabulary size test. The findings indicate that AWL is a list based on word frequency rather than a discrete and unique word list and that the AWL performs the same function as general vocabulary, with tests of each found to measure largely the same quality of knowledge. The findings also suggest that the contribution that AWL knowledge provides for academic success is not sufficient and that general vocabulary knowledge is better in predicting academic achievement. Furthermore, the contribution that academic knowledge added above the contribution of general vocabulary knowledge when combined is really small and noteworthy. This study’s results are in line with the argument and suggest that it is the development of general vocabulary size is an essential quality for academic success and acquiring the words of the AWL will form part of this process. The AWL by itself does not provide sufficient coverage, and is probably not specialised enough, for knowledge of this list to influence this general process. It can be concluded that AWL as an academic word list epitomizes only a fraction of words that are actually needed for academic success in English and that knowledge of academic vocabulary combined with general vocabulary knowledge above the most frequent 3000 words is what matters most to ultimate academic success.

Keywords: academic achievement, academic vocabulary, general vocabulary, vocabulary size

Procedia PDF Downloads 196
6090 The Success and Failure of the Solicitor General When the U.S. Government Appears as a Direct Party before the U.S. Supreme Court

Authors: Joseph Ignagni, Rebecca Deen

Abstract:

This paper analyzes the extent to which the U.S. Supreme Court votes to support the position of the United States in cases where the government is a party to the litigation. This study considers the relationship between the Solicitor General’s Office and the U.S. Supreme Court. The Solicitor General has the unique position of being the representative of the Executive Branch and the U.S. government before the Supreme Court. While a great deal of research has looked at the Solicitor General’s success as a “friend of the court,” far less has considered this relationship when the U.S. is a direct party in the litigation. This paper investigates the success rate of the Solicitor General’s Office in these cases. We find that there is considerable variation in the U.S. government’s success rate before the Court depending on the issue, Supreme Court leadership, the ideological direction of the Court and whether the U.S. approached the Court as a petitioner or respondent. We conduct our analysis on the Court’s decisions from 1953-2009. This study adds to our understanding of checks and balances, separation of powers, and inter-institutional relationships between the branches of the federal government of the United States.

Keywords: U.S. president, solicitor general, U.S. Supreme Court, separation of power, checks and balances

Procedia PDF Downloads 327