Search results for: initial cluster center
427 Study of Compost Maturity during Humification Process using UV-Spectroscopy
Authors: N. Sanmanee, K. Panishkan, K. Obsuwan, S. Dharmvanij
Abstract:
The increments of aromatic structures are widely used to monitor the degree of humification. Compost derived from mix manures mixed with agricultural wastes was studied. The compost collected at day 0, 7, 14, 21, 28, 35, 49, 77, 91, 105, and 119 was divided into 3 stages, initial stage at day 0, thermophilic stage during day 1-48, and mature stage during day 49-119. The change of highest absorptions at wavelength range between 210-235 nm during day 0- 49 implied that small molecules such as nitrates and carboxylic occurred faster than the aromatic molecules that were found at wavelength around 280 nm. The ratio of electron-transfer band at wavelength 253 nm by the benzonoid band at wavelength 230 nm (E253/E230) also gradually increased during the fermenting period indicating the presence of O-containing functional groups. This was in agreement with the shift change from aliphatic to aromatic structures as shown by the relationship with C/N and H/C ratios (r = - 0.631 and -0.717, p< 0.05) since both were decreasing. Although the amounts of humic acid (HA) were not different much during the humification process, the UV spectral deconvolution showed better qualitative characteristics to help in determining the compost quality. From this study, the compost should be used at day 49 and should not be kept longer than 3 months otherwise the quality of HA would decline regardless of the amounts of HA that might be rising. This implied that other processes, such as mineralization had an influence on the humification process changing HA-s structure and its qualities.
Keywords: Compost maturity, UV spectroscopy, humification, humic acid
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1659426 Principal Type of Water Responsible for Damage of Concrete Repeated Freeze-Thaw Cycles
Authors: L. Dahmani
Abstract:
The first and basic cause of the failure of concrete is repeated freezing (thawing) of moisture contained in the pores, microcracks, and cavities of the concrete. On transition to ice, water existing in the free state in cracks increases in volume, expanding the recess in which freezing occurs. A reduction in strength below the initial value is to be expected and further cycle of freezing and thawing have a further marked effect. By using some experimental parameters like nuclear magnetic resonance variation (NMR), enthalpy-temperature (or heat capacity) variation, we can resolve between the various water states and their effect on concrete properties during cooling through the freezing transition temperature range. The main objective of this paper is to describe the principal type of water responsible for the reduction in strength and structural damage (frost damage) of concrete following repeated freeze –thaw cycles. Some experimental work was carried out at the institute of cryogenics to determine what happens to water in concrete during the freezing transition.
Keywords: Concrete, frost proof, strength, water diffusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1793425 Effect of Exercise on Sexual Behavior and Semen Quality of Sahiwal Bulls
Authors: Abdelrasoul, Khalid Ahmed Elrabie
Abstract:
The study was conducted on Sahiwal cattle bulls maintained at the Artificial Breeding Complex, NDRI, Karnal, Hayana, India, to determine the effect of exercise on the sexual behavior and semen quality. Fourteen Sahiwal bulls were classified into two groups of seven each. Group-1, bulls were exercised by walking in a bull exerciser once a week one hour before semen collection, whereas bulls in group-2 were exercised daily. Sexual behavior and semen quality traits studied were: Reaction time (RT), Dismounting time (DMT), Total time taken in mounts (TTTM), Flehmen response (FR), Erection Score (ES), Protrusion Score (PS), Intensity of thrust (ITS), Temperament Score (TS), Libido Score (LS), Semen volume, Physical appearance, Mass activity, Initial progressive motility, Non-eosinophilic spermatozoa count (NESC) and post thaw motility percent. Data were analyzed by least squares technique. Group-2 showed significantly (p < 0.01) higher value in RT (sec), DMT (sec), TTTM (sec), ES, PS, ITS, LS, semen volume, semen color density and mass activity.
Keywords: Exercise, Sahiwal bulls, semen quality, sexual behavior.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1246424 Low Power and Less Area Architecture for Integer Motion Estimation
Authors: C Hisham, K Komal, Amit K Mishra
Abstract:
Full search block matching algorithm is widely used for hardware implementation of motion estimators in video compression algorithms. In this paper we are proposing a new architecture, which consists of a 2D parallel processing unit and a 1D unit both working in parallel. The proposed architecture reduces both data access power and computational power which are the main causes of power consumption in integer motion estimation. It also completes the operations with nearly the same number of clock cycles as compared to a 2D systolic array architecture. In this work sum of absolute difference (SAD)-the most repeated operation in block matching, is calculated in two steps. The first step is to calculate the SAD for alternate rows by a 2D parallel unit. If the SAD calculated by the parallel unit is less than the stored minimum SAD, the SAD of the remaining rows is calculated by the 1D unit. Early termination, which stops avoidable computations has been achieved with the help of alternate rows method proposed in this paper and by finding a low initial SAD value based on motion vector prediction. Data reuse has been applied to the reference blocks in the same search area which significantly reduced the memory access.
Keywords: Sum of absolute difference, high speed DSP.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491423 X-Ray Intensity Measurement Using Frequency Output Sensor for Computed Tomography
Authors: R. M. Siddiqui, D. Z. Moghaddam, T. R. Turlapati, S. H. Khan, I. Ul Ahad
Abstract:
Quality of 2D and 3D cross-sectional images produce by Computed Tomography primarily depend upon the degree of precision of primary and secondary X-Ray intensity detection. Traditional method of primary intensity detection is apt to errors. Recently the X-Ray intensity measurement system along with smart X-Ray sensors is developed by our group which is able to detect primary X-Ray intensity unerringly. In this study a new smart X-Ray sensor is developed using Light-to-Frequency converter TSL230 from Texas Instruments which has numerous advantages in terms of noiseless data acquisition and transmission. TSL230 construction is based on a silicon photodiode which converts incoming X-Ray radiation into the proportional current signal. A current to frequency converter is attached to this photodiode on a single monolithic CMOS integrated circuit which provides proportional frequency count to incoming current signal in the form of the pulse train. The frequency count is delivered to the center of PICDEM FS USB board with PIC18F4550 microcontroller mounted on it. With highly compact electronic hardware, this Demo Board efficiently read the smart sensor output data. The frequency output approaches overcome nonlinear behavior of sensors with analog output thus un-attenuated X-Ray intensities could be measured precisely and better normalization could be acquired in order to attain high resolution.Keywords: Computed tomography, detector technology, X-Ray intensity measurement
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2608422 Using A Hybrid Algorithm to Improve the Quality of Services in Multicast Routing Problem
Authors: Mohammad Reza Karami Nejad
Abstract:
A hybrid learning automata-genetic algorithm (HLGA) is proposed to solve QoS routing optimization problem of next generation networks. The algorithm complements the advantages of the learning Automato Algorithm(LA) and Genetic Algorithm(GA). It firstly uses the good global search capability of LA to generate initial population needed by GA, then it uses GA to improve the Quality of Service(QoS) and acquiring the optimization tree through new algorithms for crossover and mutation operators which are an NP-Complete problem. In the proposed algorithm, the connectivity matrix of edges is used for genotype representation. Some novel heuristics are also proposed for mutation, crossover, and creation of random individuals. We evaluate the performance and efficiency of the proposed HLGA-based algorithm in comparison with other existing heuristic and GA-based algorithms by the result of simulation. Simulation results demonstrate that this paper proposed algorithm not only has the fast calculating speed and high accuracy but also can improve the efficiency in Next Generation Networks QoS routing. The proposed algorithm has overcome all of the previous algorithms in the literature.
Keywords: Routing, Quality of Service, Multicaset, Learning Automata, Genetic, Next Generation Networks.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1737421 Investigating Polynomial Interpolation Functions for Zooming Low Resolution Digital Medical Images
Authors: Maninder Pal
Abstract:
Medical digital images usually have low resolution because of nature of their acquisition. Therefore, this paper focuses on zooming these images to obtain better level of information, required for the purpose of medical diagnosis. For this purpose, a strategy for selecting pixels in zooming operation is proposed. It is based on the principle of analog clock and utilizes a combination of point and neighborhood image processing. In this approach, the hour hand of clock covers the portion of image to be processed. For alignment, the center of clock points at middle pixel of the selected portion of image. The minute hand is longer in length, and is used to gain information about pixels of the surrounding area. This area is called neighborhood pixels region. This information is used to zoom the selected portion of the image. The proposed algorithm is implemented and its performance is evaluated for many medical images obtained from various sources such as X-ray, Computerized Tomography (CT) scan and Magnetic Resonance Imaging (MRI). However, for illustration and simplicity, the results obtained from a CT scanned image of head is presented. The performance of algorithm is evaluated in comparison to various traditional algorithms in terms of Peak signal-to-noise ratio (PSNR), maximum error, SSIM index, mutual information and processing time. From the results, the proposed algorithm is found to give better performance than traditional algorithms.
Keywords: Zooming, interpolation, medical images, resolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575420 Distributed Estimation Using an Improved Incremental Distributed LMS Algorithm
Authors: Amir Rastegarnia, Mohammad Ali Tinati, Azam Khalili
Abstract:
In this paper we consider the problem of distributed adaptive estimation in wireless sensor networks for two different observation noise conditions. In the first case, we assume that there are some sensors with high observation noise variance (noisy sensors) in the network. In the second case, different variance for observation noise is assumed among the sensors which is more close to real scenario. In both cases, an initial estimate of each sensor-s observation noise is obtained. For the first case, we show that when there are such sensors in the network, the performance of conventional distributed adaptive estimation algorithms such as incremental distributed least mean square (IDLMS) algorithm drastically decreases. In addition, detecting and ignoring these sensors leads to a better performance in a sense of estimation. In the next step, we propose a simple algorithm to detect theses noisy sensors and modify the IDLMS algorithm to deal with noisy sensors. For the second case, we propose a new algorithm in which the step-size parameter is adjusted for each sensor according to its observation noise variance. As the simulation results show, the proposed methods outperforms the IDLMS algorithm in the same condition.
Keywords: Distributes estimation, sensor networks, adaptive filter, IDLMS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1445419 Bearing Capacity of Sheet Hanger Connection to the Trapezoidal Metal Sheet
Authors: Kateřina Jurdová
Abstract:
Hanging to the trapezoidal sheet by decking hanger is a very widespread solution used in civil engineering to lead the distribution of energy, sanitary, air distribution system etc. under the roof or floor structure. The trapezoidal decking hanger is usually a part of the whole installation system for specific distribution medium. The leading companies offer installation systems for each specific distribution e.g. pipe rings, sprinkler systems, installation channels etc. Every specific part is connected to the base connector which is decking hanger. The own connection has three main components: decking hanger, threaded bar with nuts and web of trapezoidal sheet. The aim of this contribution is determinate the failure mechanism of each component in connection. Load bearing capacity of most components in connection could be calculated by formulas in European codes. This contribution is focused on problematic of bearing resistance of threaded bar in web of trapezoidal sheet. This issue is studied by experimental research and numerical modelling. This contribution presented the initial results of experiment which is compared with numerical model of specimen.
Keywords: Decking hanger, concentrated load, connection, load bearing capacity, trapezoidal metal sheet.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2649418 Automatic Reusability Appraisal of Software Components using Neuro-fuzzy Approach
Authors: Parvinder S. Sandhu, Hardeep Singh
Abstract:
Automatic reusability appraisal could be helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable components from existing legacy systems; that can save cost of developing the software from scratch. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. In this paper, we have mentioned two-tier approach by studying the structural attributes as well as usability or relevancy of the component to a particular domain. Latent semantic analysis is used for the feature vector representation of various software domains. It exploits the fact that FeatureVector codes can be seen as documents containing terms -the idenifiers present in the components- and so text modeling methods that capture co-occurrence information in low-dimensional spaces can be used. Further, we devised Neuro- Fuzzy hybrid Inference System, which takes structural metric values as input and calculates the reusability of the software component. Decision tree algorithm is used to decide initial set of fuzzy rules for the Neuro-fuzzy system. The results obtained are convincing enough to propose the system for economical identification and retrieval of reusable software components.Keywords: Clustering, ID3, LSA, Neuro-fuzzy System, SVD
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1662417 Study of the Antimicrobial Activity of Aminoreductone against Pathogenic Bacteria in Comparison with Other Antibiotics
Authors: Vu Thu Trang, Lam Xuan Thanh, Samira Sarter, Tomoko Shimamura, Hiroaki Takeuchi
Abstract:
Antimicrobial activities of aminoreductone (AR), a product formed in the initial stage of Maillard reaction, were screened against pathogenic bacteria. A significant growth inhibition of AR against all 7 isolates (Staphylococcus aureus ATCC® 25923™, Salmonella typhimurium ATCC® 14028™, Bacillus cereus ATCC® 13061™, Bacillus subtilis ATCC® 11774™, Escherichia coli ATCC® 25922™, Enterococcus faecalis ATCC® 29212™, Listeria innocua ATCC® 33090™) were observed by the standard disc diffusion methods. The inhibition zone for each isolate by AR (2.5 mg) ranged from 15±0mm to 28.3±0.4mm in diameter. The minimum inhibitory concentration (MIC) of AR ranging from 20mM to 26mM was proven in the 7 isolates tested. AR also showed the similar effect of growth inhibition in comparison with antibiotics frequently used for the treatment of infections bacteria, such as amikacin, ciprofloxacin, meropennem and levofloxacin. The results indicated that foods containing AR are valuable sources of bioactive compounds towards pathogenic bacteria.
Keywords: Pathogenic bacteria, aminoreductone, Maillard reaction, antimicrobial activity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304416 Binary Decision Diagrams: An Improved Variable Ordering using Graph Representation of Boolean Functions
Authors: P.W. C. Prasad, A. Assi, A. Harb, V.C. Prasad
Abstract:
This paper presents an improved variable ordering method to obtain the minimum number of nodes in Reduced Ordered Binary Decision Diagrams (ROBDD). The proposed method uses the graph topology to find the best variable ordering. Therefore the input Boolean function is converted to a unidirectional graph. Three levels of graph parameters are used to increase the probability of having a good variable ordering. The initial level uses the total number of nodes (NN) in all the paths, the total number of paths (NP) and the maximum number of nodes among all paths (MNNAP). The second and third levels use two extra parameters: The shortest path among two variables (SP) and the sum of shortest path from one variable to all the other variables (SSP). A permutation of the graph parameters is performed at each level for each variable order and the number of nodes is recorded. Experimental results are promising; the proposed method is found to be more effective in finding the variable ordering for the majority of benchmark circuits.
Keywords: Binary decision diagrams, graph representation, Boolean functions representation, variable ordering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2116415 Mechanical Characterization and Impact Study on the Environment of Raw Sediments and Sediments Dehydrated by Addition of Polymer
Authors: A. Kasmi, N. E. Abriak, M. Benzerzour, I. Shahrour
Abstract:
Large volumes of river sediments are dredged each year in Europe in order to maintain harbour activities and prevent floods. The management of this sediment has become increasingly complex. Several European projects were implemented to find environmentally sound solutions for these materials. The main objective of this study is to show the ability of river sediment to be used in road. Since sediments contain a high amount of water, then a dehydrating treatment by addition of the flocculation aid has been used. Firstly, a lot of physical characteristics are measured and discussed for a better identification of the raw sediment and this dehydrated sediment by addition the flocculation aid. The identified parameters are, for example, the initial water content, the density, the organic matter content, the grain size distribution, the liquid limit and plastic limit and geotechnical parameters. The environmental impacts of the used material were evaluated. The results obtained show that there is a slight change on the physical-chemical and geotechnical characteristics of sediment after dehydration by the addition of polymer. However, these sediments cannot be used in road construction.
Keywords: River sediment, dehydration, flocculation aid, characteristics, environmental impacts, road construction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1297414 Oxidation of Selected Pharmaceuticals in Water Matrices by Bromine and Chlorine
Authors: Juan L. Acero, F. Javier Benitez, Francisco J. Real, Gloria Roldan, Francisco Casas
Abstract:
The bromination of five selected pharmaceuticals (metoprolol, naproxen, amoxicillin, hydrochlorotiazide and phenacetin) in ultrapure water and in three water matrices (a groundwater, a surface water from a public reservoir and a secondary effluent from a WWTP) was investigated. The apparent rate constants for the bromination reaction were determined as a function of the pH, and the sequence obtained for the reaction rate was amoxicillin > naproxen >> hydrochlorotiazide ≈ phenacetin ≈ metoprolol. The proposal of a kinetic mechanism, which specifies the dissociation of bromine and each pharmaceutical according to their pKa values and the pH allowed the determination of the intrinsic rate constants for every elementary reaction. The influence of the main operating conditions (pH, initial bromine dose, and the water matrix) on the degradation of pharmaceuticals was established. In addition, the presence of bromide in chlorination experiments was investigated. The presence of bromide in wastewaters and drinking waters in the range of 10 to several hundred μg L-1 accelerated slightly the oxidation of the selected pharmaceuticals during chorine disinfection.Keywords: Pharmaceuticals, bromine, chlorine, apparent andintrinsic rate constants, water matrices, degradation rates
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2330413 Computing Continuous Skyline Queries without Discriminating between Static and Dynamic Attributes
Authors: Ibrahim Gomaa, Hoda M. O. Mokhtar
Abstract:
Although most of the existing skyline queries algorithms focused basically on querying static points through static databases; with the expanding number of sensors, wireless communications and mobile applications, the demand for continuous skyline queries has increased. Unlike traditional skyline queries which only consider static attributes, continuous skyline queries include dynamic attributes, as well as the static ones. However, as skyline queries computation is based on checking the domination of skyline points over all dimensions, considering both the static and dynamic attributes without separation is required. In this paper, we present an efficient algorithm for computing continuous skyline queries without discriminating between static and dynamic attributes. Our algorithm in brief proceeds as follows: First, it excludes the points which will not be in the initial skyline result; this pruning phase reduces the required number of comparisons. Second, the association between the spatial positions of data points is examined; this phase gives an idea of where changes in the result might occur and consequently enables us to efficiently update the skyline result (continuous update) rather than computing the skyline from scratch. Finally, experimental evaluation is provided which demonstrates the accuracy, performance and efficiency of our algorithm over other existing approaches.
Keywords: Continuous query processing, dynamic database, moving object, skyline queries.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1242412 Modeling Drying and Pyrolysis of Moist Wood Particles at Slow Heating Rates
Authors: Avdhesh K. Sharma
Abstract:
Formulation for drying and pyrolysis process in packed beds at slow heating rates is presented. Drying of biomass particles bed is described by mass diffusion equation and local moisture-vapour-equilibrium relations. In gasifiers, volatilization rate during pyrolysis of biomass is modeled by using apparent kinetic rate expression, while product compositions at slow heating rates is modeled using empirical fitted mass ratios (i.e., CO/CO2, ME/CO2, H2O/CO2) in terms of pyrolysis temperature. The drying module is validated fairly with available chemical kinetics scheme and found that the testing zone in gasifier bed constituted of relatively smaller particles having high airflow with high isothermal temperature expedite the drying process. Further, volatile releases more quickly within the shorter zone height at high temperatures (isothermal). Both, moisture loss and volatile release profiles are found to be sensitive to temperature, although the influence of initial moisture content on volatile release profile is not so sensitive.
Keywords: Modeling downdraft gasifier, drying, pyrolysis, moist woody biomass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 832411 Numerical Investigation of the Performance of a Vorsyl Separator Using a Euler-Lagrange Approach
Authors: Guozhen Li, Philip Hall, Nick Miles, Tao Wu, Jie Dong
Abstract:
This paper presents a Euler-Lagrange model of the water-particles multiphase flows in a Vorsyl separator where particles with different densities are separated. A series of particles with their densities ranging from 760 kg/m3 to 1380 kg/m3 were fed into the Vorsyl separator with water by means of tangential inlet. The simulation showed that the feed materials acquired centrifugal force which allows most portion of the particles with a density less than water to move to the center of the separator, enter the vortex finder and leave the separator through the bottom outlet. While the particles heavier than water move to the wall, reach the throat area and leave the separator through the side outlet. The particles were thus separated and particles collected at the bottom outlet are pure and clean. The influence of particle density on separation efficiency was investigated which demonstrated a positive correlation of the separation efficiency with increasing density difference between medium liquid and the particle. In addition, the influence of the split ratio on the performance was studied which showed that the separation efficiency of the Vorsyl separator can be improved by the increase of split ratio. The simulation also suggested that the Vorsyl separator may not function when the feeding velocity is smaller than a certain critical feeding in velocity. In addition, an increasing feeding velocity gives rise to increased pressure drop, however does not necessarily increase the separation efficiency.Keywords: Vorsyl separator, separation efficiency, CFD, split ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1198410 Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method
Authors: Prashant P. Bedekar, Sudhir R. Bhide, Vijay S. Kale
Abstract:
Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.Keywords: Constrained optimization, LPP, Overcurrent relaycoordination, Two-phase simplex method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3007409 Enhancing Protein Incorporation in Calcium Phosphate Coating on Titanium by Rapid Biomimetic Co-Precipitation Technique
Authors: J. Suwanprateeb, F. Thammarakcharoen
Abstract:
Calcium phosphate coating (CaP) has been employed for protein delivery, but the typical direct protein adsorption on the coating led to low incorporation content and fast release of the protein from the coating. By using bovine serum albumin (BSA) as a model protein, rapid biomimetic co-precipitation between calcium phosphate and BSA was employed to control the distribution of BSA within calcium phosphate coating during biomimetic formation on titanium surface for only 6 h at 50oC in an accelerated calcium phosphate solution. As a result, the amount of BSA incorporation and release duration could be increased by using a rapid biomimetic coprecipitation technique. Up to 43 fold increases in the BSA incorporation content and the increase from 6 h to more than 360 h in release duration compared to typical direct adsorption technique were observed depending on the initial BSA concentration used during coprecipitation (1, 10 and 100 μg.ml-1). From x-ray diffraction and Fourier transform infrared spectroscopy studies, the coating composition was not altered with the incorporation of BSA by this rapid biomimetic co-precipitation and mainly comprised octacalcium phosphate and hydroxyapatite. However, the microstructure of calcium phosphate crystals changed from straight, plate-like units to curved, plate-like units with increasing BSA content.
Keywords: Biomimetic, Calcium Phosphate Coating, Protein, Titanium.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2378408 Interaction Effect of Feed Rate and Cutting Speed in CNC-Turning on Chip Micro-Hardness of 304- Austenitic Stainless Steel
Authors: G. H. Senussi
Abstract:
The present work is concerned with the effect of turning process parameters (cutting speed, feed rate, and depth of cut) and distance from the center of work piece as input variables on the chip micro-hardness as response or output. Three experiments were conducted; they were used to investigate the chip micro-hardness behavior at diameter of work piece for 30[mm], 40[mm], and 50[mm]. Response surface methodology (R.S.M) is used to determine and present the cause and effect of the relationship between true mean response and input control variables influencing the response as a two or three dimensional hyper surface. R.S.M has been used for designing a three factor with five level central composite rotatable factors design in order to construct statistical models capable of accurate prediction of responses. The results obtained showed that the application of R.S.M can predict the effect of machining parameters on chip micro-hardness. The five level factorial designs can be employed easily for developing statistical models to predict chip micro-hardness by controllable machining parameters. Results obtained showed that the combined effect of cutting speed at it?s lower level, feed rate and depth of cut at their higher values, and larger work piece diameter can result increasing chi micro-hardness.Keywords: Machining Parameters, Chip Micro-Hardness, CNCMachining, 304-Austenic Stainless Steel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3284407 Sediment Patterns from Fluid-Bed Interactions: A Direct Numerical Simulations Study on Fluvial Turbulent Flows
Authors: Nadim Zgheib, Sivaramakrishnan Balachandar
Abstract:
We present results on the initial formation of ripples from an initially flattened erodible bed. We use direct numerical simulations (DNS) of turbulent open channel flow over a fixed sinusoidal bed coupled with hydrodynamic stability analysis. We use the direct forcing immersed boundary method to account for the presence of the sediment bed. The resolved flow provides the bed shear stress and consequently the sediment transport rate, which is needed in the stability analysis of the Exner equation. The approach is different from traditional linear stability analysis in the sense that the phase lag between the bed topology, and the sediment flux is obtained from the DNS. We ran 11 simulations at a fixed shear Reynolds number of 180, but for different sediment bed wavelengths. The analysis allows us to sweep a large range of physical and modelling parameters to predict their effects on linear growth. The Froude number appears to be the critical controlling parameter in the early linear development of ripples, in contrast with the dominant role of particle Reynolds number during the equilibrium stage.Keywords: Direct numerical simulation, immersed boundary method, sediment-bed interactions, turbulent multiphase flow, linear stability analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 673406 Economic Analysis, Growth and Yield of Grafting Tomato Varieties for Solanum torvum as a Rootstock
Authors: Evy Latifah, Eko Widaryanto, M. Dawam Maghfoer, Arifin
Abstract:
Tomato (Lycopersicon esculentum Mill.) is potential vegetables to develop, because it has high economic value and has the potential to be exported. There is a decrease in tomato productivity due to unfavorable growth conditions such as bacterial wilt, fusarium wilt, high humidity, high temperature and inappropriate production technology. Grafting technology is one alternative technology. In addition to being able to control the disease in the soil, grafting is also able to increase the growth and yield of production. Besides, it is also necessary to know the economic benefits if using grafting technology. A promising eggplant rootstock for tomato grafting is Solanum torvum. S. torvum is selected as a rootstock with high compatibility. The purpose of this research is to know the effect of grafting several varieties of tomatoes with Solanum torvum as a rootstock. The experiment was conducted in Agricultural Extension Center Pare. Experimental Garden of Pare Kediri sub-district from July to early December 2016. The materials used were tomato Cervo varieties, Karina, Timoty, and Solanum torvum. Economic analysis, growth, and yield including plant height, number of leaves, percentage of disease and tomato production were used as performance measures. The study showed that grafting tomato Timoty scion with Solanum torvum as rootstock had higher production. Financially, grafting tomato Timoty and Cervo scion had higher profit about. 28,6% and 16,3% compared to Timoty and Cervo variety treatment without grafting.
Keywords: Grafting technology, economic analysis, growth, yield of tomato, Solanum torvum.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1446405 The Implementation of the Javanese Lettered-Manuscript Image Preprocessing Stage Model on the Batak Lettered-Manuscript Image
Authors: Anastasia Rita Widiarti, Agus Harjoko, Marsono, Sri Hartati
Abstract:
This paper presents the results of a study to test whether the Javanese character manuscript image preprocessing model that have been more widely applied, can also be applied to segment of the Batak characters manuscripts. The treatment process begins by converting the input image into a binary image. After the binary image is cleaned of noise, then the segmentation lines using projection profile is conducted. If unclear histogram projection is found, then the smoothing process before production indexes line segments is conducted. For each line image which has been produced, then the segmentation scripts in the line is applied, with regard of the connectivity between pixels which making up the letters that there is no characters are truncated. From the results of manuscript preprocessing system prototype testing, it is obtained the information about the system truth percentage value on pieces of Pustaka Batak Podani Ma AjiMamisinon manuscript ranged from 65% to 87.68% with a confidence level of 95%. The value indicates the truth percentage shown the initial processing model in Javanese characters manuscript image can be applied also to the image of the Batak characters manuscript.Keywords: Connected component, preprocessing manuscript image, projection profiles.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 923404 Implication of Taliban’s Recent Relationship with Neighboring Countries and Its Impact on the Current Peace Process
Authors: Lutfurahman Aftab
Abstract:
The Taliban’s relationships with the neighboring countries are a complex political issue that local people interpret one way, and politicians have different perceptions; therefore, it is a current issue that needs to be analyzed broadly and impartially. In this article, we investigate the Taliban’s current relationships with the neighboring countries, as well as look at the effects these relationships have on the current peace negotiations in Doha, which began on September 12, 2020. The issue of Taliban and the current peace process has turned to be the center-of-attention for most of the neighboring countries, and every country has opened new pages in their foreign policies because after the Taliban-US peace agreement, the neighboring countries are meticulously and closely observing the situation and they believe that the Taliban is on the verge to tighten their grips on the future political power of Afghanistan. Every neighboring country of Afghanistan has political, economic, and social interests in this land-locked country. The Taliban’s current role within the peace talks and anticipated future position within the Afghan government will have great political, economic, and social implications on countries in the region as they assess their foreign policies. As these countries move to form closer ties with the Taliban, the government of Afghanistan is worried that this may hinder the peace process. Afghanistan has long blamed Pakistan for sheltering the Taliban and providing safe havens for the terrorist groups, including Al Qaeda, and the recent visits of Taliban’s delegations to Islamabad, Pakistan, have raised concern among government officials in Afghanistan who believe that the Taliban is not independent in their decisions, and for every step they take, are consulting with Pakistan’s political leadership.
Keywords: peace process, USA, Afghanistan, Taliban
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 447403 A Constructivist Approach and Tool for Autonomous Agent Bottom-up Sequential Learning
Authors: Jianyong Xue, Olivier L. Georgeon, Salima Hassas
Abstract:
During the initial phase of cognitive development, infants exhibit amazing abilities to generate novel behaviors in unfamiliar situations, and explore actively to learn the best while lacking extrinsic rewards from the environment. These abilities set them apart from even the most advanced autonomous robots. This work seeks to contribute to understand and replicate some of these abilities. We propose the Bottom-up hiErarchical sequential Learning algorithm with Constructivist pAradigm (BEL-CA) to design agents capable of learning autonomously and continuously through interactions. The algorithm implements no assumption about the semantics of input and output data. It does not rely upon a model of the world given a priori in the form of a set of states and transitions as well. Besides, we propose a toolkit to analyze the learning process at run time called GAIT (Generating and Analyzing Interaction Traces). We use GAIT to report and explain the detailed learning process and the structured behaviors that the agent has learned on each decision making. We report an experiment in which the agent learned to successfully interact with its environment and to avoid unfavorable interactions using regularities discovered through interaction.Keywords: Cognitive development, constructivist learning, hierarchical sequential learning, self-adaptation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 533402 Accuracy of Autonomy Navigation of Unmanned Aircraft Systems through Imagery
Authors: Sidney A. Lima, Hermann J. H. Kux, Elcio H. Shiguemori
Abstract:
The Unmanned Aircraft Systems (UAS) usually navigate through the Global Navigation Satellite System (GNSS) associated with an Inertial Navigation System (INS). However, GNSS can have its accuracy degraded at any time or even turn off the signal of GNSS. In addition, there is the possibility of malicious interferences, known as jamming. Therefore, the image navigation system can solve the autonomy problem, because if the GNSS is disabled or degraded, the image navigation system would continue to provide coordinate information for the INS, allowing the autonomy of the system. This work aims to evaluate the accuracy of the positioning though photogrammetry concepts. The methodology uses orthophotos and Digital Surface Models (DSM) as a reference to represent the object space and photograph obtained during the flight to represent the image space. For the calculation of the coordinates of the perspective center and camera attitudes, it is necessary to know the coordinates of homologous points in the object space (orthophoto coordinates and DSM altitude) and image space (column and line of the photograph). So if it is possible to automatically identify in real time the homologous points the coordinates and attitudes can be calculated whit their respective accuracies. With the methodology applied in this work, it is possible to verify maximum errors in the order of 0.5 m in the positioning and 0.6º in the attitude of the camera, so the navigation through the image can reach values equal to or higher than the GNSS receivers without differential correction. Therefore, navigating through the image is a good alternative to enable autonomous navigation.
Keywords: Autonomy, navigation, security, photogrammetry, remote sensing, spatial resection, UAS.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1321401 Using Field Indices of Rill and Gully in order to Erosion Estimating and Sediment Analysis (Case Study: Menderjan Watershed in Isfahan Province, Iran)
Authors: Masoud Nasri, Sadat Feiznia, Mohammad Jafari, Hasan Ahmadi
Abstract:
Today, incorrect use of lands and land use changes, excessive grazing, no suitable using of agricultural farms, plowing on steep slopes, road construct, building construct, mine excavation etc have been caused increasing of soil erosion and sediment yield. For erosion and sediment estimation one can use statistical and empirical methods. This needs to identify land unit map and the map of effective factors. However, these empirical methods are usually time consuming and do not give accurate estimation of erosion. In this study, we applied GIS techniques to estimate erosion and sediment of Menderjan watershed at upstream Zayandehrud river in center of Iran. Erosion faces at each land unit were defined on the basis of land use, geology and land unit map using GIS. The UTM coordinates of each erosion type that showed more erosion amounts such as rills and gullies were inserted in GIS using GPS data. The frequency of erosion indicators at each land unit, land use and their sediment yield of these indices were calculated. Also using tendency analysis of sediment yield changes in watershed outlet (Menderjan hydrometric gauge station), was calculated related parameters and estimation errors. The results of this study according to implemented watershed management projects can be used for more rapid and more accurate estimation of erosion than traditional methods. These results can also be used for regional erosion assessment and can be used for remote sensing image processing.Keywords: Erosion and sedimentation, Gully, Rill, GIS, GPS, Menderjan Watershed
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1908400 Experimental and Numerical Simulation of Fire in a Scaled Underground Station
Authors: Nuri Yucel, Muhammed Ilter Berberoglu, Salih Karaaslan, Nureddin Dinler
Abstract:
The objective of this study is to investigate fire behaviors, experimentally and numerically, in a scaled version of an underground station. The effect of ventilation velocity on the fire is examined. Fire experiments are simulated by burning 10 ml isopropyl alcohol fuel in a fire pool with dimensions 5cm x 10cm x 4 mm at the center of 1/100 scaled underground station model. A commercial CFD program FLUENT was used in numerical simulations. For air flow simulations, k-ω SST turbulence model and for combustion simulation, non-premixed combustion model are used. This study showed that, the ventilation velocity is increased from 1 m/s to 3 m/s the maximum temperature in the station is found to be less for ventilation velocity of 1 m/s. The reason for these experimental result lies on the relative dominance of oxygen supply effect on cooling effect. Without piston effect, maximum temperature occurs above the fuel pool. However, when the ventilation velocity increased the flame was tilted in the direction of ventilation and the location of maximum temperature moves along the flow direction. The velocities measured experimentally in the station at different locations are well matched by the CFD simulation results. The prediction of general flow pattern is satisfactory with the smoke visualization tests. The backlayering in velocity is well predicted by CFD simulation. However, all over the station, the CFD simulations predicted higher temperatures compared to experimental measurements.Keywords: Fire, underground station, flame propagation, CFDsimulation, k-ω SST turbulence model, non-premixed combustionmodel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2643399 A Follow up Study on the Elderly Survivors - Mental Health Two Years after the Wenchuan Earthquake
Authors: Ting Wang, Huiqin Yang, Buxin Han
Abstract:
Background: This investigated the mental health of the elderly survivors six months, ten months and two years after the “5.12 Wenchuan" earthquake. Methods: Two hundred and thirty-two physically healthy older survivors from earthquake-affected Mianyang County were interviewed. The measures included the Revised Impact of Event Scale (IES-R, Chinese version, for PTSD) and a Chinese Mental Health Inventory for the Elderly (MHIE). A repeated measures ANOVA test was used for statistical analysis. Results: The follow-up group had a statistically significant lower IES-R score and lower MHIE score than the initial group ten months after the earthquake. Two years later, the score of IES-R in follow-up group were still lower than that of non-follow-up group, but no differences were significant on the score of MHIE between groups. Furthermore, a negative relationship was found between scores of IES-R and MHIE. Conclusion: The earthquake has had a persistent negative impact on older survivors- mental health within the two-year period and that although the PTSD level declined significantly with time, it did not disappear completely.
Keywords: Elderly survivors, follow-up, mental health, post-Wenchuan earthquake.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2025398 An Inverse Heat Transfer Algorithm for Predicting the Thermal Properties of Tumors during Cryosurgery
Authors: Mohamed Hafid, Marcel Lacroix
Abstract:
This study aimed at developing an inverse heat transfer approach for predicting the time-varying freezing front and the temperature distribution of tumors during cryosurgery. Using a temperature probe pressed against the layer of tumor, the inverse approach is able to predict simultaneously the metabolic heat generation and the blood perfusion rate of the tumor. Once these parameters are predicted, the temperature-field and time-varying freezing fronts are determined with the direct model. The direct model rests on one-dimensional Pennes bioheat equation. The phase change problem is handled with the enthalpy method. The Levenberg-Marquardt Method (LMM) combined to the Broyden Method (BM) is used to solve the inverse model. The effect (a) of the thermal properties of the diseased tissues; (b) of the initial guesses for the unknown thermal properties; (c) of the data capture frequency; and (d) of the noise on the recorded temperatures is examined. It is shown that the proposed inverse approach remains accurate for all the cases investigated.
Keywords: Cryosurgery, inverse heat transfer, Levenberg-Marquardt method, thermal properties, Pennes model, enthalpy method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1499