Search results for: exponential time differencing method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32195

Search results for: exponential time differencing method

31175 Effective Editable Emoticon Description Schema for Mobile Applications

Authors: Jiwon Lee, Si-hwan Jang, Sanghyun Joo

Abstract:

The popularity of emoticons are on the rise since the mobile messengers are generalized. At the same time, few problems of emoticons are also occurred due to innate characteristics of emoticons. Too many emoticons make difficult people to select one which is well-suited for user's intention. On the contrary to this, sometimes user cannot find the emoticon which expresses user's exact intention. Poor information delivery of emoticon is another problem due to a major part of current emoticons are focused on emotion delivery. In this situation, we propose a new concept of emoticons, editable emoticons, to solve above drawbacks of emoticons. User can edit the components inside the proposed editable emoticon and send it to express his exact intention. By doing so, the number of editable emoticons can be maintained reasonable, and it can express user's exact intention. Further, editable emoticons can be used as information deliverer according to user's intention and editing skills. In this paper, we propose the concept of editable emoticons and schema based editable emoticon description method. The proposed description method is 200 times superior to the compared screen capturing method in the view of transmission bandwidth. Further, the description method is designed to have compatibility since it follows MPEG-UD international standard. The proposed editable emoticons can be exploited not only mobile applications, but also various fields such as education and medical field.

Keywords: description schema, editable emoticon, emoticon transmission, mobile applications

Procedia PDF Downloads 285
31174 The Trajectory of the Ball in Football Game

Authors: Mahdi Motahari, Mojtaba Farzaneh, Ebrahim Sepidbar

Abstract:

Tracking of moving and flying targets is one of the most important issues in image processing topic. Estimating of trajectory of desired object in short-term and long-term scale is more important than tracking of moving and flying targets. In this paper, a new way of identifying and estimating of future trajectory of a moving ball in long-term scale is estimated by using synthesis and interaction of image processing algorithms including noise removal and image segmentation, Kalman filter algorithm in order to estimating of trajectory of ball in football game in short-term scale and intelligent adaptive neuro-fuzzy algorithm based on time series of traverse distance. The proposed system attain more than 96% identify accuracy by using aforesaid methods and relaying on aforesaid algorithms and data base video in format of synthesis and interaction. Although the present method has high precision, it is time consuming. By comparing this method with other methods we realize the accuracy and efficiency of that.

Keywords: tracking, signal processing, moving targets and flying, artificial intelligent systems, estimating of trajectory, Kalman filter

Procedia PDF Downloads 450
31173 Prospect and Challenges of Public Bicycle Sharing System in Indian Cities

Authors: Anil Kumar

Abstract:

Public Bicycle System (PBS), generally known as Public Bicycle Share System or Bike-Share, is a service provided to the everyday commuters in which several cycles are available on the shared system. The concept of PBS is new to the people of India and requires more study in the fields of essential requirements, major infrastructural requirements, social acceptability, and various challenges. In various Indian cities, MRTS, BRTS, Monorail, and other modes of transport have been adopted for the main haul of transport. These modes take more time, space and are also expensive to implement. At the same time, the PBS system is more economical and takes less time to implement. The main benefit of the PBS system is that it is more environmentally friendly. PBS is being implemented in many Indian cities for public use, but various challenges are associated with this. The study aims to determine what are the basic infrastructural requirements for PBS in India, as well as to determine to what extent a Public Bike Sharing System can provide a quality and efficient service to passengers as a primary method of transportation.

Keywords: public bicycle sharing system, sustainable transport, infrastructure, smart city

Procedia PDF Downloads 182
31172 Characterization of Bacteria by a Nondestructive Sample Preparation Method in a TEM System

Authors: J. Shiue, I. H. Chen, S. W. Y. Chiu, Y. L. Wang

Abstract:

In this work, we present a nondestructive method to characterize bacteria in a TEM system. Unlike the conventional TEM specimen preparation method, which needs to thin the specimen in a destructive way, or spread the samples on a tiny millimeter sized carbon grid, our method is easy to operate without the need of sample pretreatment. With a specially designed transparent chip that allows the electron beam to pass through, and a custom made chip holder to fit into a standard TEM sample holder, the bacteria specimen can be easily prepared on the chip without any pretreatment, and then be observed under TEM. The centimeter-sized chip is covered with Au nanoparticles in the surface as the markers which allow the bacteria to be observed easily on the chip. We demonstrate the success of our method by using E. coli as an example, and show that high-resolution TEM images of E. coli can be obtained with the method presented. Some E. coli morphology characteristics imaged using this method are also presented.

Keywords: bacteria, chip, nanoparticles, TEM

Procedia PDF Downloads 303
31171 Spectroscopic Study of Tb³⁺ Doped Calcium Aluminozincate Phosphor for Display and Solid-State Lighting Applications

Authors: Sumandeep Kaur, Allam Srinivasa Rao, Mula Jayasimhadri

Abstract:

In recent years, rare earth (RE) ions doped inorganic luminescent materials are seeking great attention due to their excellent physical and chemical properties. These materials offer high thermal and chemical stability and exhibit good luminescence properties due to the presence of RE ions. The luminescent properties of these materials are attributed to their intra-configurational f-f transitions in RE ions. A series of Tb³⁺ doped calcium aluminozincate has been synthesized via sol-gel method. The structural and morphological studies have been carried out by recording X-ray diffraction patterns and SEM image. The luminescent spectra have been recorded for a comprehensive study of their luminescence properties. The XRD profile reveals the single-phase orthorhombic crystal structure with an average crystallite size of 65 nm as calculated by using DebyeScherrer equation. The SEM image exhibits completely random, irregular morphology of micron size particles of the prepared samples. The optimization of luminescence has been carried out by varying the dopant Tb³⁺ concentration within the range from 0.5 to 2.0 mol%. The as-synthesized phosphors exhibit intense emission at 544 nm pumped at 478 nm excitation wavelength. The optimized Tb³⁺ concentration has been found to be 1.0 mol% in the present host lattice. The decay curves show bi-exponential fitting for the as-synthesized phosphor. The colorimetric studies show green emission with CIE coordinates (0.334, 0.647) lying in green region for the optimized Tb³⁺ concentration. This report reveals the potential utility of Tb³⁺ doped calcium aluminozincate phosphors for display and solid-state lighting devices.

Keywords: concentration quenching, phosphor, photoluminescence, XRD

Procedia PDF Downloads 139
31170 Micropillar-Assisted Electric Field Enhancement for High-Efficiency Inactivation of Bacteria

Authors: Sanam Pudasaini, A. T. K. Perera, Ahmed Syed Shaheer Uddin, Sum Huan Ng, Chun Yang

Abstract:

Development of high-efficiency and environment friendly bacterial inactivation methods is of great importance for preventing waterborne diseases which are one of the leading causes of death in the world. Traditional bacterial inactivation methods (e.g., ultraviolet radiation and chlorination) have several limitations such as longer treatment time, formation of toxic byproducts, bacterial regrowth, etc. Recently, an electroporation-based inactivation method was introduced as a substitute. Here, an electroporation-based continuous flow microfluidic device equipped with an array of micropillars is developed, and the device achieved high bacterial inactivation performance ( > 99.9%) within a short exposure time ( < 1 s). More than 99.9% reduction of Escherichia coli bacteria was obtained for the flow rate of 1 mL/hr, and no regrowth of bacteria was observed. Images from scanning electron microscope confirmed the formation of electroporation-induced nano-pore within the cell membrane. Through numerical simulation, it has been shown that sufficiently large electric field strength (3 kV/cm), required for bacterial electroporation, were generated using PDMS micropillars for an applied voltage of 300 V. Further, in this method of inactivation, there is no involvement of chemicals and the formation of harmful by-products is also minimum.

Keywords: electroporation, high-efficiency, inactivation, microfluidics, micropillar

Procedia PDF Downloads 168
31169 Investigation the Photocatalytic Properties of Fe3O4-TiO2 Nanocomposites Prepared by Sonochemical Method

Authors: Zh. Saffari, A. Naeimi, M. S. Ekrami-Kakhki, F. Hamidi

Abstract:

Fe3O4 is one of the important magnetic oxides with spinel structure; it has exhibited unique electric and magnetic properties based on the electron transfer between Fe2+ and Fe3+ in the octahedral sites. Fe3O4 has received considerable attention in various areas such as cancer therapy, drug targeting, enzyme immobilization catalysis, magnetic cell separation, magnetic refrigeration systems and super-paramagnetic materials Fe3O4–TiO2 nanostructures were synthesized by simple, effective and new co-precipitation method assisted by ultrasonic reaction at room temperatures with organic surfactant. The effect of various parameters such as temperature, time, and power on the size and morphology of the product was investigated. Alternating gradient force magnetometer shows that Fe3O4 nanoparticles exhibit super-paramagnetic behaviour at room temperature. For preparation of nanocomposite, 1 g of TiO2 nanostructures were dispersed in 100 mL of ethanol. 0.25 g of Fe(NO3)2 and 2 mL of octanoic acid was added to the solution as a surfactant. Then, NaOH solution (1.5 M) was slowly added into the solution until the pH of the mixture was 7–8. After complete precipitation, the solution placed under the ultrasonic irradiation for 30 min. The product was centrifuged, washed with distilled water and dried in an oven at 100 °C for 3 h. The resulting red powder was calcinated at 800 °C for 3 h to remove any organic residue. The photocatalytic behaviour of Fe3O4–TiO2 nanoparticles was evaluated using the degradation of a Methyl Violet (MV) aqueous solution under ultraviolet light irradiation. As time increased, more and more MV was adsorbed on the nanoparticles catalyst, until the absorption peak vanish. The MV concentration decreased rapidly with increasing UV-irradiation time

Keywords: magnetic, methyl violet, nanocomposite, photocatalytic

Procedia PDF Downloads 245
31168 Teaching Academic Vocabulary: A Recent and Old Approach

Authors: Sara Fine-Meltzer

Abstract:

An obvious, but ill-addressed hindrance to reading comprehension in academic English is poor vocabulary. Unfortunately, dealing with the problem is usually delayed until university entrance. It is the contention of this paper that the chore should be confronted much earlier and by using a very old-fashioned method. This presentation is accompanied by vocabulary lists for advanced level university students with explanations concerning the content and justification for the 500-word lists: how they change over time in accordance with evolving styles of academic writing. There are also sample quizzes and methods to ensure that the words are “absorbed” over time. There is a discussion of other vocabulary acquisition methods and conclusions drawn from the drawbacks of such methods. The paper concludes with the rationale for beginning the study of “academic” vocabulary earlier than is generally acceptable.

Keywords: academic vocabulary, old-fashioned methods, quizzes, vocabulary lists

Procedia PDF Downloads 110
31167 Scenario Based Reaction Time Analysis for Seafarers

Authors: Umut Tac, Leyla Tavacioglu, Pelin Bolat

Abstract:

Human factor has been one of the elements that cause vulnerabilities which can be resulted with accidents in maritime transportation. When the roots of human factor based accidents are analyzed, gaps in performing cognitive abilities (reaction time, attention, memory…) are faced as the main reasons for the vulnerabilities in complex environment of maritime systems. Thus cognitive processes in maritime systems have arisen important subject that should be investigated comprehensively. At this point, neurocognitive tests such as reaction time analysis tests have been used as coherent tools that enable us to make valid assessments for cognitive status. In this respect, the aim of this study is to evaluate the reaction time (response time or latency) of seafarers due to their occupational experience and age. For this study, reaction time for different maneuverers has been taken while the participants were performing a sea voyage through a simulator which was run up with a certain scenario. After collecting the data for reaction time, a statistical analyze has been done to understand the relation between occupational experience and cognitive abilities.

Keywords: cognitive abilities, human factor, neurocognitive test battery, reaction time

Procedia PDF Downloads 291
31166 Application of the Material Point Method as a New Fast Simulation Technique for Textile Composites Forming and Material Handling

Authors: Amir Nazemi, Milad Ramezankhani, Marian Kӧrber, Abbas S. Milani

Abstract:

The excellent strength to weight ratio of woven fabric composites, along with their high formability, is one of the primary design parameters defining their increased use in modern manufacturing processes, including those in aerospace and automotive. However, for emerging automated preform processes under the smart manufacturing paradigm, complex geometries of finished components continue to bring several challenges to the designers to cope with manufacturing defects on site. Wrinklinge. g. is a common defectoccurring during the forming process and handling of semi-finished textile composites. One of the main reasons for this defect is the weak bending stiffness of fibers in unconsolidated state, causing excessive relative motion between them. Further challenges are represented by the automated handling of large-area fiber blanks with specialized gripper systems. For fabric composites forming simulations, the finite element (FE)method is a longstanding tool usedfor prediction and mitigation of manufacturing defects. Such simulations are predominately meant, not only to predict the onset, growth, and shape of wrinkles but also to determine the best processing condition that can yield optimized positioning of the fibers upon forming (or robot handling in the automated processes case). However, the need for use of small-time steps via explicit FE codes, facing numerical instabilities, as well as large computational time, are among notable drawbacks of the current FEtools, hindering their extensive use as fast and yet efficient digital twins in industry. This paper presents a novel woven fabric simulation technique through the application of the material point method (MPM), which enables the use of much larger time steps, facing less numerical instabilities, hence the ability to run significantly faster and efficient simulationsfor fabric materials handling and forming processes. Therefore, this method has the ability to enhance the development of automated fiber handling and preform processes by calculating the physical interactions with the MPM fiber models and rigid tool components. This enables the designers to virtually develop, test, and optimize their processes based on either algorithmicor Machine Learning applications. As a preliminary case study, forming of a hemispherical plain weave is shown, and the results are compared to theFE simulations, as well as experiments.

Keywords: material point method, woven fabric composites, forming, material handling

Procedia PDF Downloads 173
31165 Analysis of Fault Tolerance on Grid Computing in Real Time Approach

Authors: Parampal Kaur, Deepak Aggarwal

Abstract:

In the computational Grid, fault tolerance is an imperative issue to be considered during job scheduling. Due to the widespread use of resources, systems are highly prone to errors and failures. Hence, fault tolerance plays a key role in the grid to avoid the problem of unreliability. Scheduling the task to the appropriate resource is a vital requirement in computational Grid. The fittest resource scheduling algorithm searches for the appropriate resource based on the job requirements, in contrary to the general scheduling algorithms where jobs are scheduled to the resources with best performance factor. The proposed method is to improve the fault tolerance of the fittest resource scheduling algorithm by scheduling the job in coordination with job replication when the resource has low reliability. Based on the reliability index of the resource, the resource is identified as critical. The tasks are scheduled based on the criticality of the resources. Results show that the execution time of the tasks is comparatively reduced with the proposed algorithm using real-time approach rather than a simulator.

Keywords: computational grid, fault tolerance, task replication, job scheduling

Procedia PDF Downloads 426
31164 Displacement Based Design of a Dual Structural System

Authors: Romel Cordova Shedan

Abstract:

The traditional seismic design is the methodology of Forced Based Design (FBD). The Displacement Based Design (DBD) is a seismic design that considers structural damage to achieve a failure mechanism of the structure before the collapse. It is easier to quantify damage of a structure with displacements rather than forces. Therefore, a structure to achieve an inelastic displacement design with good ductility, it is necessary to be damaged. The first part of this investigation is about differences between the methodologies of DBD and FBD with some DBD advantages. In the second part, there is a study case about a dual building 5-story, which is regular in plan and elevation. The building is located in a seismic zone, which acceleration in firm soil is 45% of the acceleration of gravity. Then it is applied both methodologies into the study case to compare its displacements, shear forces and overturning moments. In the third part, the Dynamic Time History Analysis (DTHA) is done, to compare displacements with DBD and FBD methodologies. Three accelerograms were used and the magnitude of the acceleration scaled to be spectrum compatible with design spectrum. Then, using ASCE 41-13 guidelines, the hinge plastics were assigned to structure. Finally, both methodologies results about study case are compared. It is important to take into account that the seismic performance level of the building for DBD is greater than FBD method. This is due to drifts of DBD are in the order of 2.0% and 2.5% comparing with FBD drifts of 0.7%. Therefore, displacements of DBD is greater than the FBD method. Shear forces of DBD result greater than FBD methodology. These strengths of DBD method ensures that structure achieves design inelastic displacements, because those strengths were obtained due to a displacement spectrum reduction factor which depends on damping and ductility of the dual system. Also, the displacements for the study case for DBD results to be greater than FBD and DTHA. In that way, it proves that the seismic performance level of the building for DBD is greater than FBD method. Due to drifts of DBD which are in the order of 2.0% and 2.5% compared with little FBD drifts of 0.7%.

Keywords: displacement-based design, displacement spectrum reduction factor, dynamic time history analysis, forced based design

Procedia PDF Downloads 222
31163 Design and Development of Motorized Placer for Balloon Uterine Stents in Gynecology

Authors: Metehan Mutlu, Meltem Elitas

Abstract:

This study aims to provide an automated method for placing the balloon uterine stents after hysteroscopy adhesiolysis. Currently, there are no automatized tools to place the balloon uterine stent; therefore, surgeons into the endometrial cavity manually fit it. However, it is very hard to pass the balloon stent through the cervical canal, which is roughly 10mm after the surgery. Our method aims to provide an effective and practical way of placing the stent, by automating the procedure through our designed device. Furthermore, our device does the required tasks fast compared to traditional methods, reduces the narcosis time, and decreases the bacterial contamination risks.

Keywords: balloon uterine stent, endometrial cavity, hysteroscopy, motorized-tool

Procedia PDF Downloads 268
31162 Compensation of Cable Attenuation in Step Current Generators to Enable the Convolution Method for Calibration of Current Transducers

Authors: P. Treyer, M. Kujda, H. Urs

Abstract:

The purpose of this paper is to digitally compensate for the apparent discharge time constant of the coaxial cable so that the current step response is flat and can be used to calibrate current transducers using the convolution method. For proper use of convolution, the step response record length is required to be at least the same as the waveform duration to be evaluated. The current step generator based on the cable discharge is compared to the Blumlein generator. Moreover, the influence of each component of the system on the performance of the step is described, which allows building the appropriate measurement set-up. In the end, the calibration of current viewing resistors dedicated to high current impulse is computed.

Keywords: Blumlein generator, cable attenuation, convolution, current step generator

Procedia PDF Downloads 141
31161 Logistics Model for Improving Quality in Railway Transport

Authors: Eva Nedeliakova, Juraj Camaj, Jaroslav Masek

Abstract:

This contribution is focused on the methodology for identifying levels of quality and improving quality through new logistics model in railway transport. It is oriented on the application of dynamic quality models, which represent an innovative method of evaluation quality services. Through this conception, time factor, expected, and perceived quality in each moment of the transportation process within logistics chain can be taken into account. Various models describe the improvement of the quality which emphases the time factor throughout the whole transportation logistics chain. Quality of services in railway transport can be determined by the existing level of service quality, by detecting the causes of dissatisfaction employees but also customers, to uncover strengths and weaknesses. This new logistics model is able to recognize critical processes in logistic chain. It includes service quality rating that must respect its specific properties, which are unrepeatability, impalpability, their use right at the time they are provided and particularly changeability, which is significant factor in the conditions of rail transport as well. These peculiarities influence the quality of service regarding the constantly increasing requirements and that result in new ways of finding progressive attitudes towards the service quality rating.

Keywords: logistics model, quality, railway transport

Procedia PDF Downloads 552
31160 Decomposition of the Discount Function Into Impatience and Uncertainty Aversion. How Neurofinance Can Help to Understand Behavioral Anomalies

Authors: Roberta Martino, Viviana Ventre

Abstract:

Intertemporal choices are choices under conditions of uncertainty in which the consequences are distributed over time. The Discounted Utility Model is the essential reference for describing the individual in the context of intertemporal choice. The model is based on the idea that the individual selects the alternative with the highest utility, which is calculated by multiplying the cardinal utility of the outcome, as if the reception were instantaneous, by the discount function that determines a decrease in the utility value according to how the actual reception of the outcome is far away from the moment the choice is made. Initially, the discount function was assumed to have an exponential trend, whose decrease over time is constant, in line with a profile of a rational investor described by classical economics. Instead, empirical evidence called for the formulation of alternative, hyperbolic models that better represented the actual actions of the investor. Attitudes that do not comply with the principles of classical rationality are termed anomalous, i.e., difficult to rationalize and describe through normative models. The development of behavioral finance, which describes investor behavior through cognitive psychology, has shown that deviations from rationality are due to the limited rationality condition of human beings. What this means is that when a choice is made in a very difficult and information-rich environment, the brain does a compromise job between the cognitive effort required and the selection of an alternative. Moreover, the evaluation and selection phase of the alternative, the collection and processing of information, are dynamics conditioned by systematic distortions of the decision-making process that are the behavioral biases involving the individual's emotional and cognitive system. In this paper we present an original decomposition of the discount function to investigate the psychological principles of hyperbolic discounting. It is possible to decompose the curve into two components: the first component is responsible for the smaller decrease in the outcome as time increases and is related to the individual's impatience; the second component relates to the change in the direction of the tangent vector to the curve and indicates how much the individual perceives the indeterminacy of the future indicating his or her aversion to uncertainty. This decomposition allows interesting conclusions to be drawn with respect to the concept of impatience and the emotional drives involved in decision-making. The contribution that neuroscience can make to decision theory and inter-temporal choice theory is vast as it would allow the description of the decision-making process as the relationship between the individual's emotional and cognitive factors. Neurofinance is a discipline that uses a multidisciplinary approach to investigate how the brain influences decision-making. Indeed, considering that the decision-making process is linked to the activity of the prefrontal cortex and amygdala, neurofinance can help determine the extent to which abnormal attitudes respect the principles of rationality.

Keywords: impatience, intertemporal choice, neurofinance, rationality, uncertainty

Procedia PDF Downloads 116
31159 Modelling Structural Breaks in Stock Price Time Series Using Stochastic Differential Equations

Authors: Daniil Karzanov

Abstract:

This paper studies the effect of quarterly earnings reports on the stock price. The profitability of the stock is modeled by geometric Brownian diffusion and the Constant Elasticity of Variance model. We fit several variations of stochastic differential equations to the pre-and after-report period using the Maximum Likelihood Estimation and Grid Search of parameters method. By examining the change in the model parameters after reports’ publication, the study reveals that the reports have enough evidence to be a structural breakpoint, meaning that all the forecast models exploited are not applicable for forecasting and should be refitted shortly.

Keywords: stock market, earnings reports, financial time series, structural breaks, stochastic differential equations

Procedia PDF Downloads 193
31158 Electro Magnetic Tractor (E. M. Tractor)

Authors: Sijo Varghese

Abstract:

A space craft (E. M. Tractor) which is intended to deflect or tug the asteroids which possesses threat towards the planets is the whole idea behind this paper. In this case "Electro Magnetic Induction" is used where it is known that when two separate circuits are connected to the electro magnet and on application of electric current through the one circuit in to the coil induces magnetic fields which repels the other circuit.( Faraday's law of Electromagnetic Induction). Basically a Spacecraft is used to attach a large sheet of aluminum on to the surface of the asteroid, the Spacecraft acts as an electro magnet and the induced magnetic field would eventually repel the aluminum intern repelling the asteroid. This method would take less time as compared to use of gravity( which requires a larger spacecraft and process will take a long time).

Keywords: asteroids, electro magnetic induction, gravity, electro magnetic tractor

Procedia PDF Downloads 480
31157 Optimal Load Control Strategy in the Presence of Stochastically Dependent Renewable Energy Sources

Authors: Mahmoud M. Othman, Almoataz Y. Abdelaziz, Yasser G. Hegazy

Abstract:

This paper presents a load control strategy based on modification of the Big Bang Big Crunch optimization method. The proposed strategy aims to determine the optimal load to be controlled and the corresponding time of control in order to minimize the energy purchased from substation. The presented strategy helps the distribution network operator to rely on the renewable energy sources in supplying the system demand. The renewable energy sources used in the presented study are modeled using the diagonal band Copula method and sequential Monte Carlo method in order to accurately consider the multivariate stochastic dependence between wind power, photovoltaic power and the system demand. The proposed algorithms are implemented in MATLAB environment and tested on the IEEE 37-node feeder. Several case studies are done and the subsequent discussions show the effectiveness of the proposed algorithm.

Keywords: big bang big crunch, distributed generation, load control, optimization, planning

Procedia PDF Downloads 332
31156 Simultaneous Saccharification and Fermentation for D-Lactic Acid Production from Dried Distillers Grains with Solubles

Authors: Nurul Aqilah Mohd Zaini, Afroditi Chatzifragkou, Dimitris Charalampopoulos

Abstract:

D-Lactic acid production is gaining increasing attention due to the thermostable properties of its polymer, Polylactic Acid (PLA). In this study, D-lactic acid was produced in microbial cultures using Lactobacillus coryniformis subsp. torquens as D-lactic acid producer and hydrolysates of Dried Distillers Grains with Solubles (DDGS) as fermentation substrate. Prior to fermentation, DDGS was first alkaline pretreated with 5% (w/v) NaOH, for 15 minutes (121oC/ ~16 psi). This led to the generation of DDGS solid residues, rich in carbohydrates and especially cellulose (~52%). The carbohydrate-rich solids were then subjected to enzymatic hydrolysis with Accellerase® 1500. For Separate Hydrolysis and Fermentation (SHF), enzymatic hydrolysis was carried out at 50oC for 24 hours, followed by fermentation of D-lactic acid at 37oC in controlled pH 6. The obtained hydrolysate contained 24 g/l glucose, 5.4 g/l xylose and 0.6 g/l arabinose. In the case of Simultaneous Saccharification and Fermentation (SSF), hydrolysis and fermentation were conducted in a single step process at 37oC in pH 5. The enzymatic hydrolysis of DGGS pretreated solids took place mostly during lag phase of L. coryniformis fermentation, with only a small amount of glucose consumed during the first 6 h. When exponential phase was started, glucose generation reduced as the microorganism started to consume glucose for D-lactic acid production. Higher concentrations of D-lactic acid were produced when SSF approach was applied, with 28 g/l D-lactic acid after 24 h of fermentation (84.5% yield). In contrast, 21.2 g/l D-lactic acid were produced when SHF was used. The optical pu rity of D-lactic acid produced from both experiments was 99.9%. Besides, approximately 2 g/l acetic acid was also generated due to lactic acid degradation after glucose depletion in SHF. SSF was proved an efficient towards DDGS ulilisation and D-lactic acid production, by reducing the overall processing time, yielding sufficient D-lactic acid concentrations without the generation of fermentation by-products.

Keywords: DDGS, alkaline pretreatment, SSF, D-lactic acid

Procedia PDF Downloads 325
31155 Evaluation of Cytotoxic Effect of Two Diterpenes from Plectranthus barbatus

Authors: Nawal Al Musayeib, Musarat Amina, Perwez Alam

Abstract:

Plectranthus barbatus Andrews (Lamiaceae) is the most common species of genus Plectranthus. It is used for treating various ailments. In this study, two rare diterpenes 11,14-dihydroxy-8,11,13-abietatrien-7-one (1) and 12-hydroxyabieta-8(14),9(11),12-trien-7-one (2) were isolated for the first time from P. barbatus. Their chemical structures were verified utilizing various spectroscopic experiments. The effect of diterpenes against undifferentiated/anaplastic thyroid cancer cell line (FRO) was evaluated and they were quantitatively analysed using HPTLC method. The two diterpenes were found to be cytotoxic, however compound 1 showed significant cytotoxic effects where 95% reduction in the cell viability was observed in different time intervals. The quantity of compound 1 and compound 2 in PBCE were found to be 2.04 and15.97 μg/mg, respectively of dried weight of the extract.

Keywords: abietatrien, cancer, diterpenes, Plectranthus barbatus

Procedia PDF Downloads 241
31154 Self-Attention Mechanism for Target Hiding Based on Satellite Images

Authors: Hao Yuan, Yongjian Shen, Xiangjun He, Yuheng Li, Zhouzhou Zhang, Pengyu Zhang, Minkang Cai

Abstract:

Remote sensing data can provide support for decision-making in disaster assessment or disaster relief. The traditional processing methods of sensitive targets in remote sensing mapping are mainly based on manual retrieval and image editing tools, which are inefficient. Methods based on deep learning for sensitive target hiding are faster and more flexible. But these methods have disadvantages in training time and cost of calculation. This paper proposed a target hiding model Self Attention (SA) Deepfill, which used self-attention modules to replace part of gated convolution layers in image inpainting. By this operation, the calculation amount of the model becomes smaller, and the performance is improved. And this paper adds free-form masks to the model’s training to enhance the model’s universal. The experiment on an open remote sensing dataset proved the efficiency of our method. Moreover, through experimental comparison, the proposed method can train for a longer time without over-fitting. Finally, compared with the existing methods, the proposed model has lower computational weight and better performance.

Keywords: remote sensing mapping, image inpainting, self-attention mechanism, target hiding

Procedia PDF Downloads 116
31153 Desulfurization of Crude Oil Using Bacteria

Authors: Namratha Pai, K. Vasantharaj, K. Haribabu

Abstract:

Our Team is developing an innovative cost effective biological technique to desulfurize crude oil. ’Sulphur’ is found to be present in crude oil samples from .05% - 13.95% and its elimination by industrial methods is expensive currently. Materials required :- Alicyclobacillus acidoterrestrius, potato dextrose agar, oxygen, Pyragallol and inert gas(nitrogen). Method adapted and proposed:- 1) Growth of bacteria studied, energy needs. 2) Compatibility with crude-oil. 3) Reaction rate of bacteria studied and optimized. 4) Reaction development by computer simulation. 5) Simulated work tested by building the reactor. The method being developed requires the use of bacteria Alicyclobacillus acidoterrestrius - an acidothermophilic heterotrophic, soil dwelling aerobic, Sulfur bacteria. The bacteria are fed to crude oil in a unique manner. Its coated onto potato dextrose agar beads, cultured for 24 hours (growth time coincides with time when it begins reacting) and fed into the reactor. The beads are to be replenished with O2 by passing them through a jacket around the reactor which has O2 supply. The O2 can’t be supplied directly as crude oil is inflammable, hence the process. Beads are made to move around based on the concept of fluidized bed reactor. By controlling the velocity of inert gas pumped , the beads are made to settle down when exhausted of O2. It is recycled through the jacket where O2 is re-fed and beads which were inside the ring substitute the exhausted ones. Crude-oil is maintained between 1 atm-270 M Pa pressure and 45°C treated with tartaric acid (Ph reason for bacteria growth) for optimum output. Bacteria being of oxidising type react with Sulphur in crude-oil and liberate out SO4^2- and no gas. SO4^2- is absorbed into H2O. NaOH is fed once reaction is complete and beads separated. Crude-oil is thus separated of SO4^2-, thereby Sulphur, tartaric acid and other acids which are separated out. Bio-corrosion is taken care of by internal wall painting (phenolepoxy paints). Earlier methods used included use of Pseudomonas and Rhodococcus species. They were found to be inefficient, time and energy consuming and reduce the fuel value as they fed on skeleton.

Keywords: alicyclobacillus acidoterrestrius, potato dextrose agar, fluidized bed reactor principle, reaction time for bacteria, compatibility with crude oil

Procedia PDF Downloads 308
31152 Copula Autoregressive Methodology for Simulation of Solar Irradiance and Air Temperature Time Series for Solar Energy Forecasting

Authors: Andres F. Ramirez, Carlos F. Valencia

Abstract:

The increasing interest in renewable energies strategies application and the path for diminishing the use of carbon related energy sources have encouraged the development of novel strategies for integration of solar energy into the electricity network. A correct inclusion of the fluctuating energy output of a photovoltaic (PV) energy system into an electric grid requires improvements in the forecasting and simulation methodologies for solar energy potential, and the understanding not only of the mean value of the series but the associated underlying stochastic process. We present a methodology for synthetic generation of solar irradiance (shortwave flux) and air temperature bivariate time series based on copula functions to represent the cross-dependence and temporal structure of the data. We explore the advantages of using this nonlinear time series method over traditional approaches that use a transformation of the data to normal distributions as an intermediate step. The use of copulas gives flexibility to represent the serial variability of the real data on the simulation and allows having more control on the desired properties of the data. We use discrete zero mass density distributions to assess the nature of solar irradiance, alongside vector generalized linear models for the bivariate time series time dependent distributions. We found that the copula autoregressive methodology used, including the zero mass characteristics of the solar irradiance time series, generates a significant improvement over state of the art strategies. These results will help to better understand the fluctuating nature of solar energy forecasting, the underlying stochastic process, and quantify the potential of a photovoltaic (PV) energy generating system integration into a country electricity network. Experimental analysis and real data application substantiate the usage and convenience of the proposed methodology to forecast solar irradiance time series and solar energy across northern hemisphere, southern hemisphere, and equatorial zones.

Keywords: copula autoregressive, solar irradiance forecasting, solar energy forecasting, time series generation

Procedia PDF Downloads 310
31151 Ranking of Inventory Policies Using Distance Based Approach Method

Authors: Gupta Amit, Kumar Ramesh, P. C. Tewari

Abstract:

Globalization is putting enormous pressure on the business organizations specially manufacturing one to rethink the supply chain in innovative manners. Inventory consumes major portion of total sale revenue. Effective and efficient inventory management plays a vital role for the successful functioning of any organization. Selection of inventory policy is one of the important purchasing activities. This paper focuses on selection and ranking of alternative inventory policies. A deterministic quantitative model-based on Distance Based Approach (DBA) method has been developed for evaluation and ranking of inventory policies. We have employed this concept first time for this type of the selection problem. Four inventory policies Economic Order Quantity (EOQ), Just in Time (JIT), Vendor Managed Inventory (VMI) and monthly policy are considered. Improper selection could affect a company’s competitiveness in terms of the productivity of its facilities and quality of its products. The ranking of inventory policies is a multi-criteria problem. There is a need to first identify the selection criteria and then processes the information with reference to relative importance of attributes for comparison. Criteria values for each inventory policy can be obtained either analytically or by using a simulation technique or they are linguistic subjective judgments defined by fuzzy sets, like, for example, the values of criteria. A methodology is developed and applied to rank the inventory policies.

Keywords: inventory policy, ranking, DBA, selection criteria

Procedia PDF Downloads 373
31150 Comparative Analysis of Various Waste Oils for Biodiesel Production

Authors: Olusegun Ayodeji Olagunju, Christine Tyreesa Pillay

Abstract:

Biodiesel from waste sources is regarded as an economical and most viable fuel alternative to depleting fossil fuels. In this work, biodiesel was produced from three different sources of waste cooking oil; from cafeterias, which is vegetable-based using the transesterification method. The free fatty acids (% FFA) of the feedstocks were conducted successfully through the titration method. The results for sources 1, 2, and 3 were 0.86 %, 0.54 % and 0.20 %, respectively. The three variables considered in this process were temperature, reaction time, and catalyst concentration within the following range: 50 oC – 70 oC, 30 min – 90 min, and 0.5 % – 1.5 % catalyst. Produced biodiesel was characterized using ASTM standard methods for biodiesel property testing to determine the fuel properties, including kinematic viscosity, specific gravity, flash point, pour point, cloud point, and acid number. The results obtained indicate that the biodiesel yield from source 3 was greater than the other sources. All produced biodiesel fuel properties are within the standard biodiesel fuel specifications ASTM D6751. The optimum yield of biodiesel was obtained at 98.76%, 96.4%, and 94.53% from source 3, source 2, and source 1, respectively at optimum operating variables of 65 oC temperature, 90 minutes reaction time, and 0.5 wt% potassium hydroxide.

Keywords: waste cooking oil, biodiesel, free fatty acid content, potassium hydroxide catalyst, optimization analysis

Procedia PDF Downloads 65
31149 Hierarchical Piecewise Linear Representation of Time Series Data

Authors: Vineetha Bettaiah, Heggere S. Ranganath

Abstract:

This paper presents a Hierarchical Piecewise Linear Approximation (HPLA) for the representation of time series data in which the time series is treated as a curve in the time-amplitude image space. The curve is partitioned into segments by choosing perceptually important points as break points. Each segment between adjacent break points is recursively partitioned into two segments at the best point or midpoint until the error between the approximating line and the original curve becomes less than a pre-specified threshold. The HPLA representation achieves dimensionality reduction while preserving prominent local features and general shape of time series. The representation permits course-fine processing at different levels of details, allows flexible definition of similarity based on mathematical measures or general time series shape, and supports time series data mining operations including query by content, clustering and classification based on whole or subsequence similarity.

Keywords: data mining, dimensionality reduction, piecewise linear representation, time series representation

Procedia PDF Downloads 267
31148 Evaluation of Lemongrass (Cymbopogon citratus) as Mosquito Repellent Extracted by Supercritical Carbon Dioxide Assisted Process

Authors: Chia-Yu Lin, Chun-Ying Lee, Chih-Jer Lin

Abstract:

Lemongrass (Cymbopogon citratus), grown in tropical and subtropical regions over the world, has many potential uses in pharmaceutical, cosmetics, food and flavor, and agriculture industries. In this study, because of its affinity to human body and friendliness to the environment, lemongrass extract was prepared from different processes to evaluate its effectiveness as mosquito repellent. Moreover, the supercritical fluid extraction method has been widely used as an effective and environmental friendly process in the preparation of a variety of compounds. Thus, both the extracts from lemongrass by the conventional hydrodistillation method and the supercritical CO₂ assisted method were compared. The effects of pressure, temperature and time duration on the supercritical CO₂ extraction were also investigated. The compositions of different extracts were examined using mass spectrometer. As for the experiment of mosquito repellence, the extract was placed inside a mosquito trap along with syrup. The mosquito counts in each trap with extracts prepared from different processes were employed in the quantitative evaluation. It was found that the extract from the supercritical CO₂ assisted process contained higher citronellol content than the conventional hydrodistillation method. The extract with higher citronellol content also demonstrated more effective as a mosquito repellent.

Keywords: lemongrass (Cymbopogon citratus), hydrodistillation, supercritical fluid extraction, mosquito repellent

Procedia PDF Downloads 170
31147 Estimating Gait Parameter from Digital RGB Camera Using Real Time AlphaPose Learning Architecture

Authors: Murad Almadani, Khalil Abu-Hantash, Xinyu Wang, Herbert Jelinek, Kinda Khalaf

Abstract:

Gait analysis is used by healthcare professionals as a tool to gain a better understanding of the movement impairment and track progress. In most circumstances, monitoring patients in their real-life environments with low-cost equipment such as cameras and wearable sensors is more important. Inertial sensors, on the other hand, cannot provide enough information on angular dynamics. This research offers a method for tracking 2D joint coordinates using cutting-edge vision algorithms and a single RGB camera. We provide an end-to-end comprehensive deep learning pipeline for marker-less gait parameter estimation, which, to our knowledge, has never been done before. To make our pipeline function in real-time for real-world applications, we leverage the AlphaPose human posture prediction model and a deep learning transformer. We tested our approach on the well-known GPJATK dataset, which produces promising results.

Keywords: gait analysis, human pose estimation, deep learning, real time gait estimation, AlphaPose, transformer

Procedia PDF Downloads 109
31146 Heroin Withdrawal, Prison and Multiple Temporalities

Authors: Ian Walmsley

Abstract:

The aim of this paper is to explore the influence of time and temporality on the experience of coming off heroin in prison. The presentation draws on qualitative data collected during a small-scale pilot study of the role of self-care in the process of coming off drugs in prison. Time and temporality emerged as a key theme in the interview transcripts. Drug dependent prisoners experience of time in prison has not been recognized in the research literature. Instead, the literature on prison time typically views prisoners as a homogenous group or tends to focus on the influence of aging and gender on prison time. Furthermore, there is a tendency in the literature on prison drug treatment and recovery to conceptualize drug dependent prisoners as passive recipients of prison healthcare, rather than active agents. In building on these gaps, this paper argues that drug dependent prisoners experience multiple temporalities which involve an interaction between the body-times of the drug dependent prisoner and the economy of time in prison. One consequence of this interaction is the feeling that they are doing, at this point in their prison sentence, double prison time. The second part of the argument is that time and temporality were a means through which they governed their withdrawing bodies. In addition, this paper will comment on the challenges of prison research in England.

Keywords: heroin withdrawal, time and temporality, prison, body

Procedia PDF Downloads 266