Search results for: slope calculation
1435 Interaction Evaluation of Silver Ion and Silver Nanoparticles with Dithizone Complexes Using DFT Calculations and NMR Analysis
Authors: W. Nootcharin, S. Sujittra, K. Mayuso, K. Kornphimol, M. Rawiwan
Abstract:
Silver has distinct antibacterial properties and has been used as a component of commercial products with many applications. An increasing number of commercial products cause risks of silver effects for human and environment such as the symptoms of Argyria and the release of silver to the environment. Therefore, the detection of silver in the aquatic environment is important. The colorimetric chemosensor is designed by the basic of ligand interactions with a metal ion, leading to the change of signals for the naked-eyes which are very useful method to this application. Dithizone ligand is considered as one of the effective chelating reagents for metal ions due to its high selectivity and sensitivity of a photochromic reaction for silver as well as the linear backbone of dithizone affords the rotation of various isomeric forms. The present study is focused on the conformation and interaction of silver ion and silver nanoparticles (AgNPs) with dithizone using density functional theory (DFT). The interaction parameters were determined in term of binding energy of complexes and the geometry optimization, frequency of the structures and calculation of binding energies using density functional approaches B3LYP and the 6-31G(d,p) basis set. Moreover, the interaction of silver–dithizone complexes was supported by UV–Vis spectroscopy, FT-IR spectrum that was simulated by using B3LYP/6-31G(d,p) and 1H NMR spectra calculation using B3LYP/6-311+G(2d,p) method compared with the experimental data. The results showed the ion exchange interaction between hydrogen of dithizone and silver atom, with minimized binding energies of silver–dithizone interaction. However, the result of AgNPs in the form of complexes with dithizone. Moreover, the AgNPs-dithizone complexes were confirmed by using transmission electron microscope (TEM). Therefore, the results can be the useful information for determination of complex interaction using the analysis of computer simulations.Keywords: silver nanoparticles, dithizone, DFT, NMR
Procedia PDF Downloads 2071434 Improving the Global Competitiveness of SMEs by Logistics Transportation Management: Case Study Chicken Meat Supply Chain
Authors: P. Vanichkobchinda
Abstract:
The Logistics Transportation techniques, Open Vehicle Routing (OVR) is an approach toward transportation cost reduction, especially for long distance pickup and delivery nodes. The outstanding characteristic of OVR is that the route starting node and ending node are not necessary the same as in typical vehicle routing problems. This advantage enables the routing to flow continuously and the vehicle does not always return to its home base. This research aims to develop a heuristic for the open vehicle routing problem with pickup and delivery under time window and loading capacity constraints to minimize the total distance. The proposed heuristic is developed based on the Insertion method, which is a simple method and suitable for the rapid calculation that allows insertion of the new additional transportation requirements along the original paths. According to the heuristic analysis, cost comparisons between the proposed heuristic and companies are using method, nearest neighbor method show that the insertion heuristic. Moreover, the proposed heuristic gave superior solutions in all types of test problems. In conclusion, the proposed heuristic can effectively and efficiently solve the open vehicle routing. The research indicates that the improvement of new transport's calculation and the open vehicle routing with "Insertion Heuristic" represent a better outcome with 34.3 percent in average. in cost savings. Moreover, the proposed heuristic gave superior solutions in all types of test problems. In conclusion, the proposed heuristic can effectively and efficiently solve the open vehicle routing.Keywords: business competitiveness, cost reduction, SMEs, logistics transportation, VRP
Procedia PDF Downloads 6851433 DEMs: A Multivariate Comparison Approach
Authors: Juan Francisco Reinoso Gordo, Francisco Javier Ariza-López, José Rodríguez Avi, Domingo Barrera Rosillo
Abstract:
The evaluation of the quality of a data product is based on the comparison of the product with a reference of greater accuracy. In the case of MDE data products, quality assessment usually focuses on positional accuracy and few studies consider other terrain characteristics, such as slope and orientation. The proposal that is made consists of evaluating the similarity of two DEMs (a product and a reference), through the joint analysis of the distribution functions of the variables of interest, for example, elevations, slopes and orientations. This is a multivariable approach that focuses on distribution functions, not on single parameters such as mean values or dispersions (e.g. root mean squared error or variance). This is considered to be a more holistic approach. The use of the Kolmogorov-Smirnov test is proposed due to its non-parametric nature, since the distributions of the variables of interest cannot always be adequately modeled by parametric models (e.g. the Normal distribution model). In addition, its application to the multivariate case is carried out jointly by means of a single test on the convolution of the distribution functions of the variables considered, which avoids the use of corrections such as Bonferroni when several statistics hypothesis tests are carried out together. In this work, two DEM products have been considered, DEM02 with a resolution of 2x2 meters and DEM05 with a resolution of 5x5 meters, both generated by the National Geographic Institute of Spain. DEM02 is considered as the reference and DEM05 as the product to be evaluated. In addition, the slope and aspect derived models have been calculated by GIS operations on the two DEM datasets. Through sample simulation processes, the adequate behavior of the Kolmogorov-Smirnov statistical test has been verified when the null hypothesis is true, which allows calibrating the value of the statistic for the desired significance value (e.g. 5%). Once the process has been calibrated, the same process can be applied to compare the similarity of different DEM data sets (e.g. the DEM05 versus the DEM02). In summary, an innovative alternative for the comparison of DEM data sets based on a multinomial non-parametric perspective has been proposed by means of a single Kolmogorov-Smirnov test. This new approach could be extended to other DEM features of interest (e.g. curvature, etc.) and to more than three variablesKeywords: data quality, DEM, kolmogorov-smirnov test, multivariate DEM comparison
Procedia PDF Downloads 1151432 Sequence Stratigraphy and Petrophysical Analysis of Sawan Gas Field, Central Indus Basin, Pakistan
Authors: Saeed Ur Rehman Chaudhry
Abstract:
The objectives of the study are to reconstruct sequence stratigraphic framework and petrophysical analysis of the reservoir marked by using sequence stratigraphy of Sawan Gas Field. The study area lies in Central Indus Basin, District Khairpur, Sindh province, Pakistan. The study area lies tectonically in an extensional regime. Lower Goru Formation and Sembar Formation act as a reservoir and source respectively. To achieve objectives, data set of seismic lines, consisting of seismic lines PSM96-114, PSM96-115, PSM96-133, PSM98-201, PSM98-202 and well logs of Sawan-01, Sawan-02 and Gajwaro-01 has been used. First of all interpretation of seismic lines has been carried out. Interpretation of seismic lines shows extensional regime in the area and cut entire Cretaceous section. Total of seven reflectors has been marked on each seismic line. Lower Goru Formation is thinning towards west. Seismic lines also show eastward tilt of stratigraphy due to uplift at the western side. Sequence stratigraphic reconstruction has been done by integrating seismic and wireline log data. Total of seven sequence boundaries has been interpreted between the top of Chiltan Limestone to Top of Lower Goru Formation. It has been observed on seismic lines that Sembar Formation initially generated shelf margin profile and then ramp margin on which Lower Goru deposition took place. Shelf edge deltas and slope fans have been observed on seismic lines, and signatures of slope fans are also observed on wireline logs as well. Total of six sequences has been interpreted. Stratigraphic and sequence stratigraphic correlation has been carried out by using Sawan 01, Sawan 02 and Gajwaro 01 and a Low Stand Systems tract (LST) within Lower Goru C sands has been marked as a zone of interest. The petrophysical interpretation includes shale volume, effective porosity, permeability, saturation of water and hydrocarbon. On the basis of good effective porosity and hydrocarbon saturation petrophysical analysis confirms that the LST in Sawan-01 and Sawan-02 has good hydrocarbon potential.Keywords: petrophysical analysis, reservoir potential, Sawan Gas Field, sequence stratigraphy
Procedia PDF Downloads 2621431 Quantification of Dispersion Effects in Arterial Spin Labelling Perfusion MRI
Authors: Rutej R. Mehta, Michael A. Chappell
Abstract:
Introduction: Arterial spin labelling (ASL) is an increasingly popular perfusion MRI technique, in which arterial blood water is magnetically labelled in the neck before flowing into the brain, providing a non-invasive measure of cerebral blood flow (CBF). The accuracy of ASL CBF measurements, however, is hampered by dispersion effects; the distortion of the ASL labelled bolus during its transit through the vasculature. In spite of this, the current recommended implementation of ASL – the white paper (Alsop et al., MRM, 73.1 (2015): 102-116) – does not account for dispersion, which leads to the introduction of errors in CBF. Given that the transport time from the labelling region to the tissue – the arterial transit time (ATT) – depends on the region of the brain and the condition of the patient, it is likely that these errors will also vary with the ATT. In this study, various dispersion models are assessed in comparison with the white paper (WP) formula for CBF quantification, enabling the errors introduced by the WP to be quantified. Additionally, this study examines the relationship between the errors associated with the WP and the ATT – and how this is influenced by dispersion. Methods: Data were simulated using the standard model for pseudo-continuous ASL, along with various dispersion models, and then quantified using the formula in the WP. The ATT was varied from 0.5s-1.3s, and the errors associated with noise artefacts were computed in order to define the concept of significant error. The instantaneous slope of the error was also computed as an indicator of the sensitivity of the error with fluctuations in ATT. Finally, a regression analysis was performed to obtain the mean error against ATT. Results: An error of 20.9% was found to be comparable to that introduced by typical measurement noise. The WP formula was shown to introduce errors exceeding 20.9% for ATTs beyond 1.25s even when dispersion effects were ignored. Using a Gaussian dispersion model, a mean error of 16% was introduced by using the WP, and a dispersion threshold of σ=0.6 was determined, beyond which the error was found to increase considerably with ATT. The mean error ranged from 44.5% to 73.5% when other physiologically plausible dispersion models were implemented, and the instantaneous slope varied from 35 to 75 as dispersion levels were varied. Conclusion: It has been shown that the WP quantification formula holds only within an ATT window of 0.5 to 1.25s, and that this window gets narrower as dispersion occurs. Provided that the dispersion levels fall below the threshold evaluated in this study, however, the WP can measure CBF with reasonable accuracy if dispersion is correctly modelled by the Gaussian model. However, substantial errors were observed with other common models for dispersion with dispersion levels similar to those that have been observed in literature.Keywords: arterial spin labelling, dispersion, MRI, perfusion
Procedia PDF Downloads 3691430 Study on the Process of Detumbling Space Target by Laser
Authors: Zhang Pinliang, Chen Chuan, Song Guangming, Wu Qiang, Gong Zizheng, Li Ming
Abstract:
The active removal of space debris and asteroid defense are important issues in human space activities. Both of them need a detumbling process, for almost all space debris and asteroid are in a rotating state, and it`s hard and dangerous to capture or remove a target with a relatively high tumbling rate. So it`s necessary to find a method to reduce the angular rate first. The laser ablation method is an efficient way to tackle this detumbling problem, for it`s a contactless technique and can work at a safe distance. In existing research, a laser rotational control strategy based on the estimation of the instantaneous angular velocity of the target has been presented. But their calculation of control torque produced by a laser, which is very important in detumbling operation, is not accurate enough, for the method they used is only suitable for the plane or regularly shaped target, and they did not consider the influence of irregular shape and the size of the spot. In this paper, based on the triangulation reconstruction of the target surface, we propose a new method to calculate the impulse of the irregularly shaped target under both the covered irradiation and spot irradiation of the laser and verify its accuracy by theoretical formula calculation and impulse measurement experiment. Then we use it to study the process of detumbling cylinder and asteroid by laser. The result shows that the new method is universally practical and has high precision; it will take more than 13.9 hours to stop the rotation of Bennu with 1E+05kJ laser pulse energy; the speed of the detumbling process depends on the distance between the spot and the centroid of the target, which can be found an optimal value in every particular case.Keywords: detumbling, laser ablation drive, space target, space debris remove
Procedia PDF Downloads 841429 Kinetic Aspect Investigation of Chitosan / Nanohydroxyapatite / Na ₂CO₃ Gel System
Authors: P. S. D. Perera, S. U. Adikary
Abstract:
The gelation behavior of Chitosan/nanohydroxyapatite sol in the presence of a crosslinking agent Na ₂CO₃ was investigated experimentally. In this case, the gelation time(tgel) was determined by the rheological measurements of the final mixture. The tgel has been determined from dynamic viscosity slope experiments. We found that chitosan/nHA sol with 1% nano-hydroxyapatite and 1.6% Na2CO3 required coagulant performance. Hence Na ₂CO₃ and nanohydroxyapatite concentrations remain constant over the experiment. The order of reaction was first order with respect to chitosan and rate constant of the gel system was 9.0 x 10-4 s-1, respectively, depending on the temperature of the system. The gelation temperature was carried out at 37 ⁰C.Keywords: kinetics, gelation, sol-gel system, chitosan/ nHA/ Na ₂CO₃ composite
Procedia PDF Downloads 1661428 Modeling of Sediment Yield and Streamflow of Watershed Basin in the Philippines Using the Soil Water Assessment Tool Model for Watershed Sustainability
Authors: Warda L. Panondi, Norihiro Izumi
Abstract:
Sedimentation is a significant threat to the sustainability of reservoirs and their watershed. In the Philippines, the Pulangi watershed experienced a high sediment loss mainly due to land conversions and plantations that showed critical erosion rates beyond the tolerable limit of -10 ton/ha/yr in all of its sub-basin. From this event, the prediction of runoff volume and sediment yield is essential to examine using the country's soil conservation techniques realistically. In this research, the Pulangi watershed was modeled using the soil water assessment tool (SWAT) to predict its watershed basin's annual runoff and sediment yield. For the calibration and validation of the model, the SWAT-CUP was utilized. The model was calibrated with monthly discharge data for 1990-1993 and validated for 1994-1997. Simultaneously, the sediment yield was calibrated in 2014 and validated in 2015 because of limited observed datasets. Uncertainty analysis and calculation of efficiency indexes were accomplished through the SUFI-2 algorithm. According to the coefficient of determination (R2), Nash Sutcliffe efficiency (NSE), King-Gupta efficiency (KGE), and PBIAS, the calculation of streamflow indicates a good performance for both calibration and validation periods while the sediment yield resulted in a satisfactory performance for both calibration and validation. Therefore, this study was able to identify the most critical sub-basin and severe needs of soil conservation. Furthermore, this study will provide baseline information to prevent floods and landslides and serve as a useful reference for land-use policies and watershed management and sustainability in the Pulangi watershed.Keywords: Pulangi watershed, sediment yield, streamflow, SWAT model
Procedia PDF Downloads 2091427 [Keynote Talk]: Applying p-Balanced Energy Technique to Solve Liouville-Type Problems in Calculus
Authors: Lina Wu, Ye Li, Jia Liu
Abstract:
We are interested in solving Liouville-type problems to explore constancy properties for maps or differential forms on Riemannian manifolds. Geometric structures on manifolds, the existence of constancy properties for maps or differential forms, and energy growth for maps or differential forms are intertwined. In this article, we concentrate on discovery of solutions to Liouville-type problems where manifolds are Euclidean spaces (i.e. flat Riemannian manifolds) and maps become real-valued functions. Liouville-type results of vanishing properties for functions are obtained. The original work in our research findings is to extend the q-energy for a function from finite in Lq space to infinite in non-Lq space by applying p-balanced technique where q = p = 2. Calculation skills such as Hölder's Inequality and Tests for Series have been used to evaluate limits and integrations for function energy. Calculation ideas and computational techniques for solving Liouville-type problems shown in this article, which are utilized in Euclidean spaces, can be universalized as a successful algorithm, which works for both maps and differential forms on Riemannian manifolds. This innovative algorithm has a far-reaching impact on research work of solving Liouville-type problems in the general settings involved with infinite energy. The p-balanced technique in this algorithm provides a clue to success on the road of q-energy extension from finite to infinite.Keywords: differential forms, holder inequality, Liouville-type problems, p-balanced growth, p-harmonic maps, q-energy growth, tests for series
Procedia PDF Downloads 2351426 Research the Causes of Defects and Injuries of Reinforced Concrete and Stone Construction
Authors: Akaki Qatamidze
Abstract:
Implementation of the project will be a step forward in terms of reliability in Georgia and the improvement of the construction and the development of construction. Completion of the project is expected to result in a complete knowledge, which is expressed in concrete and stone structures of assessing the technical condition of the processing. This method is based on a detailed examination of the structure, in order to establish the injuries and the elimination of the possibility of changing the structural scheme of the new requirements and architectural preservationists. Reinforced concrete and stone structures research project carried out in a systematic analysis of the important approach is to optimize the process of research and development of new knowledge in the neighboring areas. In addition, the problem of physical and mathematical models of rational consent, the main pillar of the physical (in-situ) data and mathematical calculation models and physical experiments are used only for the calculation model specification and verification. Reinforced concrete and stone construction defects and failures the causes of the proposed research to enhance the effectiveness of their maximum automation capabilities and expenditure of resources to reduce the recommended system analysis of the methodological concept-based approach, as modern science and technology major particularity of one, it will allow all family structures to be identified for the same work stages and procedures, which makes it possible to exclude subjectivity and addresses the problem of the optimal direction. It discussed the methodology of the project and to establish a major step forward in the construction trades and practical assistance to engineers, supervisors, and technical experts in the construction of the settlement of the problem.Keywords: building, reinforced concrete, expertise, stone structures
Procedia PDF Downloads 3361425 Automation of Embodied Energy Calculations for Buildings through Building Information Modelling
Authors: Ahmad Odeh
Abstract:
Researchers are currently more concerned about the calculations of energy at the operational stage, mainly due to its larger environmental impact, but the fact remains, embodied energies represent a substantial contributor unaccounted for in the overall energy computation method. The calculation of materials’ embodied energy during the construction stage is complicated. This is due to the various factors involved. The equipment used, fuel needed, and electricity required for each type of materials varies with location and thus the embodied energy will differ for each project. Moreover, the method used in manufacturing, transporting and putting in place will have significant influence on the materials’ embodied energy. This anomaly has made it difficult to calculate or even bench mark the usage of such energies. This paper presents a model aimed at calculating embodied energies based on such variabilities. It presents a systematic approach that uses an efficient method of calculation to provide a new insight for the selection of construction materials. The model is developed in a BIM environment. The quantification of materials’ energy is determined over the three main stages of their lifecycle: manufacturing, transporting and placing. The model uses three major databases each of which contains set of the construction materials that are most commonly used in building projects. The first dataset holds information about the energy required to manufacture any type of materials, the second includes information about the energy required for transporting the materials while the third stores information about the energy required by machinery to place the materials in their intended locations. Through geospatial data analysis, the model automatically calculates the distances between the suppliers and construction sites and then uses dataset information for energy computations. The computational sum of all the energies is automatically calculated and then the model provides designers with a list of usable equipment along with the associated embodied energies.Keywords: BIM, lifecycle energy assessment, building automation, energy conservation
Procedia PDF Downloads 1891424 Wave Transmitting Boundary in Dynamic Analysis for an Elastoplastic Medium Using the Material Point Method
Authors: Chinh Phuong Do
Abstract:
Dynamic analysis of slope under seismic condition requires the elimination of spurious reflection at the bounded domain. This paper studies the performances of wave transmitting boundaries, including the standard viscous boundary and the viscoelastic boundary to the material point method (MPM) framework. First, analytical derivations of these non-reflecting conditions particularly to the implicit MPM are presented. Then, a number of benchmark and geotechnical examples will be shown. Overall, the results agree well with analytical solutions, indicating the ability to accurately simulate the radiation at the bounded domain.Keywords: dynamic analysis, implicit, MPM, non-reflecting boundary
Procedia PDF Downloads 2021423 Correlation Between Diastolic Function and Lower GLS in Hypertensive Patients
Authors: A. Kherraf, S. Ouarrak, L. Azzouzi, R. Habbal
Abstract:
Introduction: Preserved LVEF heart failure is an important cause of mortality and morbidity in hypertensive patients. A strong correlation between impaired diastolic function and longitudinal systolic dysfunction. could have several explanations, first, the diastole is an energy dependent process, especially during its first phase, it also includes active systolic components during the phase of iso volumetric relaxation, in addition, the impairment of the intrinsic myocytic function is part of hypertensive pathology as evidenced by recent studies. METHODS AND MATERIALS: This work consists of performing in a series of 333 hypertensive patients (aged 25 to 75 years) a complete echocardiographic study, including LVEF by Simpson biplane method, the calculation of the indexed left ventricular mass, the analysis of the diastolic function, and finally, the study of the longitudinal deformation of the LV by the technique of speckletracking (calculation of the GLS). Patients with secondary hypertension, leaky or stenosing valve disease, arrhythmia, and a history of coronary insufficiency were excluded from this study. RESULTS: Of the 333 hypertensive patients, 225 patients (67.5%) had impaired diastolic function, of which 60 patients (18%) had high filling pressures. 49.39% had echocardigraphic HVG, Almost all of these patients (60 patients) had low GLS. There is a statistically very significant relationship between lower GLS and increased left ventricular filling pressures in hypertensive patients. These results suggest that increased filling pressures are closely associated with atrioventricular interaction in patients with hypertension, with a strong correlation with impairment of longitudinal systolic function and diastolic function CONCLUSION: Overall, a linear relationship is established between increased left ventricular mass, diastolic dysfunction, and longitudinal LV systolic dysfunctionKeywords: hypertension, diastolic function, left ventricle, heart failure
Procedia PDF Downloads 1261422 Analysis of Force Convection in Bandung Triga Reactor Core Plate Types Fueled Using Coolod-N2
Authors: K. A. Sudjatmi, Endiah Puji Hastuti, Surip Widodo, Reinaldy Nazar
Abstract:
Any pretensions to stop the production of TRIGA fuel elements by TRIGA reactor fuel elements manufacturer should be anticipated by the operating agency of TRIGA reactor to replace the cylinder type fuel element with plate type fuel element, that available on the market. This away was performed the calculation on U3Si2Al fuel with uranium enrichment of 19.75% and a load level of 2.96 gU/cm3. Maximum power that can be operated on free convection cooling mode at the BANDUNG TRIGA reactor fuel plate was 600 kW. This study has been conducted thermalhydraulic characteristic calculation model of the reactor core power 2MW. BANDUNG TRIGA reactor core fueled plate type is composed of 16 fuel elements, 4 control elements and one irradiation facility which is located right in the middle of the core. The reactor core is cooled using a pump which is already available with flow rate 900 gpm. Analysis on forced convection cooling mode with flow from the top down from 10%, 20%, 30% and so on up to a 100% rate of coolant flow. performed using the COOLOD-N2 code. The calculations result showed that the 2 MW power with inlet coolant temperature at 37 °C and cooling rate percentage of 50%, then the coolant temperature, maximum cladding and meat respectively 64.96 oC, 124.81 oC, and 125.08 oC, DNBR (departure from nucleate boiling ratio)=1.23 and OFIR (onset of flow instability ratio)=1:00. The results are expected to be used as a reference for determining the power and cooling rate level of the BANDUNG TRIGA reactor core plate types fueled.Keywords: TRIGA, COOLOD-N2, plate type fuel element, force convection, thermal hydraulic characteristic
Procedia PDF Downloads 3001421 Training of Sensors for Early Warning System of Rainfall Induced Landslides
Authors: M. Naresh, Pratik Chaturvedi, Srishti Yadav, Varun Dutt, K. V. Uday
Abstract:
Changes in the Earth’s climate are likely to increase natural hazards such as drought, floods, earthquakes, landslides, etc. The present study focusing on to early warning systems (EWS) of landslides, major issues in Himalayan region without prominence to deforestation, encroachments and un-engineered cutting of slopes and reforming for infrastructural purposes. EWS can be depicted by conducting a series of flume tests using micro-electro mechanical systems sensors data after reaching threshold values under controlled laboratory conditions. Based on the threshold value database, an alert will be sent via SMS.Keywords: slope-instability, flume test, sensors, early warning system
Procedia PDF Downloads 2651420 Sediment Wave and Cyclic Steps as Mechanism for Sediment Transport in Submarine Canyons Thalweg
Authors: Taiwo Olusoji Lawrence, Peace Mawo Aaron
Abstract:
Seismic analysis of bedforms has proven to be one of the best ways to study deepwater sedimentary features. Canyons are known to be sediment transportation conduit. Sediment wave are large-scale depositional bedforms in various parts of the world's oceans formed predominantly by suspended load transport. These undulating objects usually have tens of meters to a few kilometers in wavelength and a height of several meters. Cyclic steps have long long-wave upstream-migrating bedforms confined by internal hydraulic jumps. They usually occur in regions with high gradients and slope breaks. Cyclic steps and migrating sediment waves are the most common bedform on the seafloor. Cyclic steps and related sediment wave bedforms are significant to the morpho-dynamic evolution of deep-water depositional systems architectural elements, especially those located along tectonically active margins with high gradients and slope breaks that can promote internal hydraulic jumps in turbidity currents. This report examined sedimentary activities and sediment transportation in submarine canyons and provided distinctive insight into factors that created a complex seabed canyon system in the Ceara Fortaleza basin Brazilian Equatorial Margin (BEM). The growing importance of cyclic steps made it imperative to understand the parameters leading to their formation, migration, and architecture as well as their controls on sediment transport in canyon thalweg. We extracted the parameters of the observed bedforms and evaluated the aspect ratio and asymmetricity. We developed a relationship between the hydraulic jump magnitude, depth of the hydraulic fall and the length of the cyclic step therein. It was understood that an increase in the height of the cyclic step increases the magnitude of the hydraulic jump and thereby increases the rate of deposition on the preceding stoss side. An increase in the length of the cyclic steps reduces the magnitude of the hydraulic jump and reduces the rate of deposition at the stoss side. Therefore, flat stoss side was noticed at most preceding cyclic step and sediment wave.Keywords: Ceara Fortaleza, submarine canyons, cyclic steps, sediment wave
Procedia PDF Downloads 1141419 Extreme Value Theory Applied in Reliability Analysis: Case Study of Diesel Generator Fans
Authors: Jelena Vucicevic
Abstract:
Reliability analysis represents a very important task in different areas of work. In any industry, this is crucial for maintenance, efficiency, safety and monetary costs. There are ways to calculate reliability, unreliability, failure density and failure rate. In this paper, the results for the reliability of diesel generator fans were calculated through Extreme Value Theory. The Extreme Value Theory is not widely used in the engineering field. Its usage is well known in other areas such as hydrology, meteorology, finance. The significance of this theory is in the fact that unlike the other statistical methods it is focused on rare and extreme values, and not on average. It should be noted that this theory is not designed exclusively for extreme events, but for extreme values in any event. Therefore, this is a great opportunity to apply the theory and test if it could be applied in this situation. The significance of the work is the calculation of time to failure or reliability in a new way, using statistic. Another advantage of this calculation is that there is no need for technical details and it can be implemented in any part for which we need to know the time to fail in order to have appropriate maintenance, but also to maximize usage and minimize costs. In this case, calculations have been made on diesel generator fans but the same principle can be applied to any other part. The data for this paper came from a field engineering study of the time to failure of diesel generator fans. The ultimate goal was to decide whether or not to replace the working fans with a higher quality fan to prevent future failures. The results achieved in this method will show the approximation of time for which the fans will work as they should, and the percentage of probability of fans working more than certain estimated time. Extreme Value Theory can be applied not only for rare and extreme events, but for any event that has values which we can consider as extreme.Keywords: extreme value theory, lifetime, reliability analysis, statistic, time to failure
Procedia PDF Downloads 3271418 Optimization Method of the Number of Berth at Bus Rapid Transit Stations Based on Passenger Flow Demand
Authors: Wei Kunkun, Cao Wanyang, Xu Yujie, Qiao Yuzhi, Liu Yingning
Abstract:
The reasonable design of bus parking spaces can improve the traffic capacity of the station and reduce traffic congestion. In order to reasonably determine the number of berths at BRT (Bus Rapid Transit) stops, it is based on the actual bus rapid transit station observation data, scheduling data, and passenger flow data. Optimize the number of station berths from the perspective of optimizing the balance of supply and demand at the site. Combined with the classical capacity calculation model, this paper first analyzes the important factors affecting the traffic capacity of BRT stops by using SPSS PRO and MATLAB programming software, namely the distribution of BRT stops and the distribution of BRT stop time. Secondly, the method of calculating the number of the classic human capital management (HCM) model is optimized based on the actual passenger demand of the station, and the method applicable to the actual number of station berths is proposed. Taking Gangding Station of Zhongshan Avenue Bus Rapid Transit Corridor in Guangzhou as an example, based on the calculation method proposed in this paper, the number of berths of sub-station 1, sub-station 2 and sub-station 3 is 2, which reduces the road space of the station by 33.3% compared with the previous berth 3 of each sub-station, and returns to social vehicles. Therefore, under the condition of ensuring the passenger flow demand of BRT stations, the road space of the station is reduced, and the road is returned to social vehicles, the traffic capacity of social vehicles is improved, and the traffic capacity and efficiency of the BRT corridor system are improved as a whole.Keywords: urban transportation, bus rapid transit station, HCM model, capacity, number of berths
Procedia PDF Downloads 951417 Performance Analysis of BPJLT with Different Gate and Spacer Materials
Authors: Porag Jyoti Ligira, Gargi Khanna
Abstract:
The paper presents a simulation study of the electrical characteristic of Bulk Planar Junctionless Transistor (BPJLT) using spacer. The BPJLT is a transistor without any PN junctions in the vertical direction. It is a gate controlled variable resistor. The characteristics of BPJLT are analyzed by varying the oxide material under the gate. It can be shown from the simulation that an ideal subthreshold slope of ~60 mV/decade can be achieved by using highk dielectric. The effects of variation of spacer length and material on the electrical characteristic of BPJLT are also investigated in the paper. The ION / IOFF ratio improvement is of the order of 107 and the OFF current reduction of 10-4 is obtained by using gate dielectric of HfO2 instead of SiO2.Keywords: spacer, BPJLT, high-k, double gate
Procedia PDF Downloads 4291416 Numerical Calculation and Analysis of Fine Echo Characteristics of Underwater Hemispherical Cylindrical Shell
Authors: Hongjian Jia
Abstract:
A finite-length cylindrical shell with a spherical cap is a typical engineering approximation model of actual underwater targets. The research on the omni-directional acoustic scattering characteristics of this target model can provide a favorable basis for the detection and identification of actual underwater targets. The elastic resonance characteristics of the target are the results of the comprehensive effect of the target length, shell-thickness ratio and materials. Under the conditions of different materials and geometric dimensions, the coincidence resonance characteristics of the target have obvious differences. Aiming at this problem, this paper obtains the omni-directional acoustic scattering field of the underwater hemispherical cylindrical shell by numerical calculation and studies the influence of target geometric parameters (length, shell-thickness ratio) and material parameters on the coincidence resonance characteristics of the target in turn. The study found that the formant interval is not a stable value and changes with the incident angle. Among them, the formant interval is less affected by the target length and shell-thickness ratio and is significantly affected by the material properties, which is an effective feature for classifying and identifying targets of different materials. The quadratic polynomial is utilized to fully fit the change relationship between the formant interval and the angle. The results show that the three fitting coefficients of the stainless steel and aluminum targets are significantly different, which can be used as an effective feature parameter to characterize the target materials.Keywords: hemispherical cylindrical shell;, fine echo characteristics;, geometric and material parameters;, formant interval
Procedia PDF Downloads 1091415 Performance Evaluation of the CareSTART S1 Analyzer for Quantitative Point-Of-Care Measurement of Glucose-6-Phosphate Dehydrogenase Activity
Authors: Haiyoung Jung, Mi Joung Leem, Sun Hwa Lee
Abstract:
Background & Objective: Glucose-6-phosphate dehydrogenase (G6PD) deficiency is a genetic abnormality that results in an inadequate amount of G6PD, leading to increased susceptibility of red blood cells to reactive oxygen species and hemolysis. The present study aimed to evaluate the careSTARTTM S1 analyzer for measuring G6PD activity to hemoglobin (Hb) ratio. Methods: Precision for G6PD activity and hemoglobin measurement was evaluated using control materials with two levels on five repeated runs per day for five days. The analytic performance of the careSTARTTM S1 analyzer was compared with spectrophotometry in 40 patient samples. Reference ranges suggested by the manufacturer were validated in 20 healthy males and females each. Results: The careSTARTTM S1 analyzer demonstrated precision of 6.0% for low-level (14~45 U/dL) and 2.7% for high-level (60~90 U/dL) control in G6PD activity, and 1.4% in hemoglobin (7.9~16.3 u/g Hb). A comparison study of G6PD to Hb ratio between the careSTARTTM S1 analyzer and spectrophotometry showed an average difference of 29.1% with a positive bias of the careSTARTTM S1 analyzer. All normal samples from the healthy population were validated for the suggested reference range for males (≥2.19 U/g Hb) and females (≥5.83 U/g Hb). Conclusion: The careSTARTTM S1 analyzer demonstrated good analytical performance and can replace the current spectrophotometric measurement of G6PD enzyme activity. In the aspect of the management of clinical laboratories, it can be a reasonable option as a point-of-care analyzer with minimal handling of samples and reagents, in addition to the automatic calculation of the ratio of measured G6PD activity and Hb concentration, to minimize any clerical errors involved with manual calculation.Keywords: POCT, G6PD, performance evaluation, careSTART
Procedia PDF Downloads 631414 Isotopic Evidence (He, Ne, Ar) for Deep Fluid in the Caucasus Continental Collision Zone
Authors: Larisa Liamina, Vasily Lavrushin, Salvatore Inguaggiato
Abstract:
This study presents and summarizes the results of researching the isotopic signature of helium in the deep fluid eastern part of the Southern slope of the Greater Caucasus and the Lesser Caucasus (Azerbaijan and Armenia) for the period from 2010 to 2016. The results of isotope ratios of 3He/4He in 59 samples of the gas phase of geothermal fluids and mud volcanoes are presented. New data have been obtained not only on the isotopic ratios of helium, but also neon and argon. The R/Ra ratio was analyzed along the Ankara-Sevan ophiolite structure. The patterns of lateral variations of the 3He/4He ratio of different geological structural elements of the studied region are revealed.Keywords: isotopes helium, deep fluids, tectonic structures, Caucasus
Procedia PDF Downloads 451413 Study of Evapotranspiration for Pune District
Authors: Ranjeet Sable, Mahotsavi Patil, Aadesh Nimbalkar, Prajakta Palaskar, Ritu Sagar
Abstract:
The exact amount of water used by various crops in different climatic conditions is necessary to step for design, planning, and management of irrigation schemes, water resources, scheduling of irrigation systems. Evaporation and transpiration are combinable called as evapotranspiration. Water loss from trees during photosynthesis is called as transpiration and when water gets converted into gaseous state is called evaporation. For calculation of correct evapotranspiration, we have to choose the method in such way that is should be suitable and require minimum climatic data also it should be applicable for wide range of climatic conditions. In hydrology, there are multiple correlations and regression is generally used to develop relationships between three or more hydrological variables by knowing the dependence between them. This research work includes the study of various methods for calculation of evapotranspiration and selects reasonable and suitable one Pune region (Maharashtra state). As field methods are very costly, time-consuming and not give appropriate results if the suitable climate is not maintained. Observation recorded at Pune metrological stations are used to calculate evapotranspiration with the help of Radiation Method (RAD), Modified Penman Method (MPM), Thornthwaite Method (THW), Blaney-Criddle (BCL), Christiansen Equation (CNM), Hargreaves Method (HGM), from which Hargreaves and Thornthwaite are temperature based methods. Performance of all these methods are compared with Modified Penman method and method which showing less variation with standard Modified Penman method (MPM) is selected as the suitable one. Evapotranspiration values are estimated on a monthly basis. Comparative analysis in this research used for selection for raw data-dependent methods in case of missing data.Keywords: Blaney-Criddle, Christiansen equation evapotranspiration, Hargreaves method, precipitations, Penman method, water use efficiency
Procedia PDF Downloads 2711412 Study on the Geometric Similarity in Computational Fluid Dynamics Calculation and the Requirement of Surface Mesh Quality
Authors: Qian Yi Ooi
Abstract:
At present, airfoil parameters are still designed and optimized according to the scale of conventional aircraft, and there are still some slight deviations in terms of scale differences. However, insufficient parameters or poor surface mesh quality is likely to occur if these small deviations are embedded in a future civil aircraft with a size that is quite different from conventional aircraft, such as a blended-wing-body (BWB) aircraft with future potential, resulting in large deviations in geometric similarity in computational fluid dynamics (CFD) simulations. To avoid this situation, the study on the CFD calculation on the geometric similarity of airfoil parameters and the quality of the surface mesh is conducted to obtain the ability of different parameterization methods applied on different airfoil scales. The research objects are three airfoil scales, including the wing root and wingtip of conventional civil aircraft and the wing root of the giant hybrid wing, used by three parameterization methods to compare the calculation differences between different sizes of airfoils. In this study, the constants including NACA 0012, a Reynolds number of 10 million, an angle of attack of zero, a C-grid for meshing, and the k-epsilon (k-ε) turbulence model are used. The experimental variables include three airfoil parameterization methods: point cloud method, B-spline curve method, and class function/shape function transformation (CST) method. The airfoil dimensions are set to 3.98 meters, 17.67 meters, and 48 meters, respectively. In addition, this study also uses different numbers of edge meshing and the same bias factor in the CFD simulation. Studies have shown that with the change of airfoil scales, different parameterization methods, the number of control points, and the meshing number of divisions should be used to improve the accuracy of the aerodynamic performance of the wing. When the airfoil ratio increases, the most basic point cloud parameterization method will require more and larger data to support the accuracy of the airfoil’s aerodynamic performance, which will face the severe test of insufficient computer capacity. On the other hand, when using the B-spline curve method, average number of control points and meshing number of divisions should be set appropriately to obtain higher accuracy; however, the quantitative balance cannot be directly defined, but the decisions should be made repeatedly by adding and subtracting. Lastly, when using the CST method, it is found that limited control points are enough to accurately parameterize the larger-sized wing; a higher degree of accuracy and stability can be obtained by using a lower-performance computer.Keywords: airfoil, computational fluid dynamics, geometric similarity, surface mesh quality
Procedia PDF Downloads 2221411 A Compressor Map Optimizing Tool for Prediction of Compressor Off-Design Performance
Authors: Zhongzhi Hu, Jie Shen, Jiqiang Wang
Abstract:
A high precision aeroengine model is needed when developing the engine control system. Compared with other main components, the axial compressor is the most challenging component to simulate. In this paper, a compressor map optimizing tool based on the introduction of a modifiable β function is developed for FWorks (FADEC Works). Three parameters (d density, f fitting coefficient, k₀ slope of the line β=0) are introduced to the β function to make it modifiable. The comparison of the traditional β function and the modifiable β function is carried out for a certain type of compressor. The interpolation errors show that both methods meet the modeling requirements, while the modifiable β function can predict compressor performance more accurately for some areas of the compressor map where the users are interested in.Keywords: beta function, compressor map, interpolation error, map optimization tool
Procedia PDF Downloads 2671410 Modeling and Simulating Productivity Loss Due to Project Changes
Authors: Robert Pellerin, Michel Gamache, Remi Trudeau, Nathalie Perrier
Abstract:
The context of large engineering projects is particularly favorable to the appearance of engineering changes and contractual modifications. These elements are potential causes for claims. In this paper, we investigate one of the critical components of the claim management process: the calculation of the impacts of changes in terms of losses of productivity due to the need to accelerate some project activities. When project changes are initiated, delays can arise. Indeed, project activities are often executed in fast-tracking in an attempt to respect the completion date. But the acceleration of project execution and the resulting rework can entail important costs as well as induce productivity losses. In the past, numerous methods have been proposed to quantify the duration of delays, the gains achieved by project acceleration, and the loss of productivity. The calculation related to those changes can be divided into two categories: direct cost and indirect cost. The direct cost is easily quantifiable as opposed to indirect costs which are rarely taken into account during the calculation of the cost of an engineering change or contract modification despite several research projects have been made on this subject. However, proposed models have not been accepted by companies yet, nor they have been accepted in court. Those models require extensive data and are often seen as too specific to be used for all projects. These techniques are also ignoring the resource constraints and the interdependencies between the causes of delays and the delays themselves. To resolve this issue, this research proposes a simulation model that mimics how major engineering changes or contract modifications are handled in large construction projects. The model replicates the use of overtime in a reactive scheduling mode in order to simulate the loss of productivity present when a project change occurs. Multiple tests were conducted to compare the results of the proposed simulation model with statistical analysis conducted by other researchers. Different scenarios were also conducted in order to determine the impact the number of activities, the time of occurrence of the change, the availability of resources, and the type of project changes on productivity loss. Our results demonstrate that the number of activities in the project is a critical variable influencing the productivity of a project. When changes occur, the presence of a large number of activities leads to a much lower productivity loss than a small number of activities. The speed of reducing productivity for 30-job projects is about 25 percent faster than the reduction speed for 120-job projects. The moment of occurrence of a change also shows a significant impact on productivity. Indeed, the sooner the change occurs, the lower the productivity of the labor force. The availability of resources also impacts the productivity of a project when a change is implemented. There is a higher loss of productivity when the amount of resources is restricted.Keywords: engineering changes, indirect costs overtime, productivity, scheduling, simulation
Procedia PDF Downloads 2381409 TiO₂ Nanotube Array Based Selective Vapor Sensors for Breath Analysis
Authors: Arnab Hazra
Abstract:
Breath analysis is a quick, noninvasive and inexpensive technique for disease diagnosis can be used on people of all ages without any risk. Only a limited number of volatile organic compounds (VOCs) can be associated with the occurrence of specific diseases. These VOCs can be considered as disease markers or breath markers. Selective detection with specific concentration of breath marker in exhaled human breath is required to detect a particular disease. For example, acetone (C₃H₆O), ethanol (C₂H₅OH), ethane (C₂H₆) etc. are the breath markers and abnormal concentrations of these VOCs in exhaled human breath indicates the diseases like diabetes mellitus, renal failure, breast cancer respectively. Nanomaterial-based vapor sensors are inexpensive, small and potential candidate for the detection of breath markers. In practical measurement, selectivity is the most crucial issue where trace detection of breath marker is needed to identify accurately in the presence of several interfering vapors and gases. Current article concerns a novel technique for selective and lower ppb level detection of breath markers at very low temperature based on TiO₂ nanotube array based vapor sensor devices. Highly ordered and oriented TiO₂ nanotube array was synthesized by electrochemical anodization of high purity tatinium (Ti) foil. 0.5 wt% NH₄F, ethylene glycol and 10 vol% H₂O was used as the electrolyte and anodization was carried out for 90 min with 40 V DC potential. Au/TiO₂ Nanotube/Ti, sandwich type sensor device was fabricated for the selective detection of VOCs in low concentration range. Initially, sensor was characterized where resistive and capacitive change of the sensor was recorded within the valid concentration range for individual breath markers (or organic vapors). Sensor resistance was decreased and sensor capacitance was increased with the increase of vapor concentration. Now, the ratio of resistive slope (mR) and capacitive slope (mC) provided a concentration independent constant term (M) for a particular vapor. For the detection of unknown vapor, ratio of resistive change and capacitive change at any concentration was same to the previously calculated constant term (M). After successful identification of the target vapor, concentration was calculated from the straight line behavior of resistance as a function of concentration. Current technique is suitable for the detection of particular vapor from a mixture of other interfering vapors.Keywords: breath marker, vapor sensors, selective detection, TiO₂ nanotube array
Procedia PDF Downloads 1551408 Analytical Study of Cobalt(II) and Nickel(II) Extraction with Salicylidene O-, M-, and P-Toluidine in Chloroform
Authors: Sana Almi, Djamel Barkat
Abstract:
The solvent extraction of cobalt (II) and nickel (II) from aqueous sulfate solutions were investigated with the analytical methods of slope analysis using salicylidene aniline and the three isomeric o-, m- and p-salicylidene toluidine diluted with chloroform at 25°C. By a statistical analysis of the extraction data, it was concluded that the extracted species are CoL2 with CoL2(HL) and NiL2 (HL denotes HSA, HSOT, HSMT, and HSPT). The extraction efficiency of Co(II) was higher than Ni(II). This tendency is confirmed from numerical extraction constants for each metal cations. The best extraction was according to the following order: HSMT > HSPT > HSOT > HSA for Co2+ and Ni2+.Keywords: solvent extraction, nickel(II), cobalt(II), salicylidene aniline, o-, m-, and p-salicylidene toluidine
Procedia PDF Downloads 4841407 Full-Scale Test of a Causeway Embankment Supported by Raft-Aggregate Column Foundation on Soft Clay Deposit
Authors: Tri Harianto, Lawalenna Samang, St. Hijraini Nur, Arwin
Abstract:
Recently, a port development is constructed in Makassar city, South Sulawesi Province, Indonesia. Makassar city is located in lowland area that dominated by soft marine clay deposit. A two kilometers causeway construction was built which is situated on the soft clay layer. In order to investigate the behavior of causeway embankment, a full-scale test was conducted of high embankment built on a soft clay deposit. The embankment with 3,5 m high was supported by two types of reinforcement such as raft and raft-aggregate column foundation. Since the ground was undergoing consolidation due to the preload, the raft and raft-aggregate column foundations were monitored in order to analyze the vertical ground movement by inducing the settlement of the foundation. In this study, two types of foundation (raft and raft-aggregate column) were tested to observe the effectiveness of raft-aggregate column compare to raft foundation in reducing the settlement. The settlement monitored during the construction stage by using the settlement plates, which is located in the center and toe of the embankment. Measurements were taken every day for each embankment construction stage (4 months). In addition, an analytical calculation was conducted in this study to compare the full-scale test result. The result shows that the raft-aggregate column foundation significantly reduces the settlement by 30% compared to the raft foundation. A raft-aggregate column foundation also reduced the time period of each loading stage. The Good agreement of analytical calculation compared to the full-scale test result also found in this study.Keywords: full-scale, preloading, raft-aggregate column, soft clay
Procedia PDF Downloads 2981406 Seismic Vulnerability Analysis of Arch Dam Based on Response Surface Method
Authors: Serges Mendomo Meye, Li Guowei, Shen Zhenzhong
Abstract:
Earthquake is one of the main loads threatening dam safety. Once the dam is damaged, it will bring huge losses of life and property to the country and people. Therefore, it is very important to research the seismic safety of the dam. Due to the complex foundation conditions, high fortification intensity, and high scientific and technological content, it is necessary to adopt reasonable methods to evaluate the seismic safety performance of concrete arch dams built and under construction in strong earthquake areas. Structural seismic vulnerability analysis can predict the probability of structural failure at all levels under different intensity earthquakes, which can provide a scientific basis for reasonable seismic safety evaluation and decision-making. In this paper, the response surface method (RSM) is applied to the seismic vulnerability analysis of arch dams, which improves the efficiency of vulnerability analysis. Based on the central composite test design method, the material-seismic intensity samples are established. The response surface model (RSM) with arch crown displacement as performance index is obtained by finite element (FE) calculation of the samples, and then the accuracy of the response surface model (RSM) is verified. To obtain the seismic vulnerability curves, the seismic intensity measure ??(?1) is chosen to be 0.1~1.2g, with an interval of 0.1g and a total of 12 intensity levels. For each seismic intensity level, the arch crown displacement corresponding to 100 sets of different material samples can be calculated by algebraic operation of the response surface model (RSM), which avoids 1200 times of nonlinear dynamic calculation of arch dam; thus, the efficiency of vulnerability analysis is improved greatly.Keywords: high concrete arch dam, performance index, response surface method, seismic vulnerability analysis, vector-valued intensity measure
Procedia PDF Downloads 240