Search results for: weight improved particle swarm optimization (WIPSO)
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 12327

Search results for: weight improved particle swarm optimization (WIPSO)

9657 Optimization of 3D Printing Parameters Using Machine Learning to Enhance Mechanical Properties in Fused Deposition Modeling (FDM) Technology

Authors: Darwin Junnior Sabino Diego, Brando Burgos Guerrero, Diego Arroyo Villanueva

Abstract:

Additive manufacturing, commonly known as 3D printing, has revolutionized modern manufacturing by enabling the agile creation of complex objects. However, challenges persist in the consistency and quality of printed parts, particularly in their mechanical properties. This study focuses on addressing these challenges through the optimization of printing parameters in FDM technology, using Machine Learning techniques. Our aim is to improve the mechanical properties of printed objects by optimizing parameters such as speed, temperature, and orientation. We implement a methodology that combines experimental data collection with Machine Learning algorithms to identify relationships between printing parameters and mechanical properties. The results demonstrate the potential of this methodology to enhance the quality and consistency of 3D printed products, with significant applications across various industrial fields. This research not only advances understanding of additive manufacturing but also opens new avenues for practical implementation in industrial settings.

Keywords: 3D printing, additive manufacturing, machine learning, mechanical properties

Procedia PDF Downloads 54
9656 2D Numerical Modeling of Ultrasonic Measurements in Concrete: Wave Propagation in a Multiple-Scattering Medium

Authors: T. Yu, L. Audibert, J. F. Chaix, D. Komatitsch, V. Garnier, J. M. Henault

Abstract:

Linear Ultrasonic Techniques play a major role in Non-Destructive Evaluation (NDE) for civil engineering structures in concrete since they can meet operational requirements. Interpretation of ultrasonic measurements could be improved by a better understanding of ultrasonic wave propagation in a multiple scattering medium. This work aims to develop a 2D numerical model of ultrasonic wave propagation in a heterogeneous medium, like concrete, integrating the multiple scattering phenomena in SPECFEM software. The coherent field of multiple scattering is obtained by averaging numerical wave fields, and it is used to determine the effective phase velocity and attenuation corresponding to an equivalent homogeneous medium. First, this model is applied to one scattering element (a cylinder) in a homogenous medium in a linear-elastic system, and its validation is completed thanks to the comparison with analytical solution. Then, some cases of multiple scattering by a set of randomly located cylinders or polygons are simulated to perform parametric studies on the influence of frequency and scatterer size, concentration, and shape. Also, the effective properties are compared with the predictions of Waterman-Truell model to verify its validity. Finally, the mortar viscoelastic behavior is introduced in the simulation in order to considerer the dispersion and the attenuation due to porosity included in the cement paste. In the future, different steps will be developed: The comparisons with experimental results, the interpretation of NDE measurements, and the optimization of NDE parameters before an auscultation.

Keywords: attenuation, multiple-scattering medium, numerical modeling, phase velocity, ultrasonic measurements

Procedia PDF Downloads 277
9655 Multi Universe Existence Based-On Quantum Relativity using DJV Circuit Experiment Interpretation

Authors: Muhammad Arif Jalil, Somchat Sonasang, Preecha Yupapin

Abstract:

This study hypothesizes that the universe is at the center of the universe among the white and black holes, which are the entangled pairs. The coupling between them is in terms of spacetime forming the universe and things. The birth of things is based on exchange energy between the white and black sides. That is, the transition from the white side to the black side is called wave-matter, where it has a speed faster than light with positive gravity. The transition from the black to the white side has a speed faster than light with negative gravity called a wave-particle. In the part where the speed is equal to light, the particle rest mass is formed. Things can appear to take shape here. Thus, the gravity is zero because it is the center. The gravitational force belongs to the Earth itself because it is in a position that is twisted towards the white hole. Therefore, it is negative. The coupling of black-white holes occurs directly on both sides. The mass is formed at the saturation and will create universes and other things. Therefore, it can be hundreds of thousands of universes on both sides of the B and white holes before reaching the saturation point of multi-universes. This work will use the DJV circuit that the research team made as an entangled or two-level system circuit that has been experimentally demonstrated. Therefore, this principle has the possibility for interpretation. This work explains the emergence of multiple universes and can be applied as a practical guideline for searching for universes in the future. Moreover, the results indicate that the DJV circuit can create the elementary particles according to Feynman's diagram with rest mass conditions, which will be discussed for fission and fusion applications.

Keywords: multi-universes, feynman diagram, fission, fusion

Procedia PDF Downloads 64
9654 Multi-Objective Optimization (Pareto Sets) and Multi-Response Optimization (Desirability Function) of Microencapsulation of Emamectin

Authors: Victoria Molina, Wendy Franco, Sergio Benavides, José M. Troncoso, Ricardo Luna, Jose R. PéRez-Correa

Abstract:

Emamectin Benzoate (EB) is a crystal antiparasitic that belongs to the avermectin family. It is one of the most common treatments used in Chile to control Caligus rogercresseyi in Atlantic salmon. However, the sea lice acquired resistance to EB when it is exposed at sublethal EB doses. The low solubility rate of EB and its degradation at the acidic pH in the fish digestive tract are the causes of the slow absorption of EB in the intestine. To protect EB from degradation and enhance its absorption, specific microencapsulation technologies must be developed. Amorphous Solid Dispersion techniques such as Spray Drying (SD) and Ionic Gelation (IG) seem adequate for this purpose. Recently, Soluplus® (SOL) has been used to increase the solubility rate of several drugs with similar characteristics than EB. In addition, alginate (ALG) is a widely used polymer in IG for biomedical applications. Regardless of the encapsulation technique, the quality of the obtained microparticles is evaluated with the following responses, yield (Y%), encapsulation efficiency (EE%) and loading capacity (LC%). In addition, it is important to know the percentage of EB released from the microparticles in gastric (GD%) and intestinal (ID%) digestions. In this work, we microencapsulated EB with SOL (EB-SD) and with ALG (EB-IG) using SD and IG, respectively. Quality microencapsulation responses and in vitro gastric and intestinal digestions at pH 3.35 and 7.8, respectively, were obtained. A central composite design was used to find the optimum microencapsulation variables (amount of EB, amount of polymer and feed flow). In each formulation, the behavior of these variables was predicted with statistical models. Then, the response surface methodology was used to find the best combination of the factors that allowed a lower EB release in gastric conditions, while permitting a major release at intestinal digestion. Two approaches were used to determine this. The desirability approach (DA) and multi-objective optimization (MOO) with multi-criteria decision making (MCDM). Both microencapsulation techniques allowed to maintain the integrity of EB in acid pH, given the small amount of EB released in gastric medium, while EB-IG microparticles showed greater EB release at intestinal digestion. For EB-SD, optimal conditions obtained with MOO plus MCDM yielded a good compromise among the microencapsulation responses. In addition, using these conditions, it is possible to reduce microparticles costs due to the reduction of 60% of BE regard the optimal BE proposed by (DA). For EB-GI, the optimization techniques used (DA and MOO) yielded solutions with different advantages and limitations. Applying DA costs can be reduced 21%, while Y, GD and ID showed 9.5%, 84.8% and 2.6% lower values than the best condition. In turn, MOO yielded better microencapsulation responses, but at a higher cost. Overall, EB-SD with operating conditions selected by MOO seems the best option, since a good compromise between costs and encapsulation responses was obtained.

Keywords: microencapsulation, multiple decision-making criteria, multi-objective optimization, Soluplus®

Procedia PDF Downloads 132
9653 Hyper Parameter Optimization of Deep Convolutional Neural Networks for Pavement Distress Classification

Authors: Oumaima Khlifati, Khadija Baba

Abstract:

Pavement distress is the main factor responsible for the deterioration of road structure durability, damage vehicles, and driver comfort. Transportation agencies spend a high proportion of their funds on pavement monitoring and maintenance. The auscultation of pavement distress was based on the manual survey, which was extremely time consuming, labor intensive, and required domain expertise. Therefore, the automatic distress detection is needed to reduce the cost of manual inspection and avoid more serious damage by implementing the appropriate remediation actions at the right time. Inspired by recent deep learning applications, this paper proposes an algorithm for automatic road distress detection and classification using on the Deep Convolutional Neural Network (DCNN). In this study, the types of pavement distress are classified as transverse or longitudinal cracking, alligator, pothole, and intact pavement. The dataset used in this work is composed of public asphalt pavement images. In order to learn the structure of the different type of distress, the DCNN models are trained and tested as a multi-label classification task. In addition, to get the highest accuracy for our model, we adjust the structural optimization hyper parameters such as the number of convolutions and max pooling, filers, size of filters, loss functions, activation functions, and optimizer and fine-tuning hyper parameters that conclude batch size and learning rate. The optimization of the model is executed by checking all feasible combinations and selecting the best performing one. The model, after being optimized, performance metrics is calculated, which describe the training and validation accuracies, precision, recall, and F1 score.

Keywords: distress pavement, hyperparameters, automatic classification, deep learning

Procedia PDF Downloads 94
9652 Load Management Using Multiple Sequential Load Shaping Techniques

Authors: Amira M. Attia, Karim H. Youssef, Nabil H. Abbasi

Abstract:

Demand Side Management (DSM) is an essential characteristic of current and future smart grid systems. As one of DSM functions, load management aims to control customers’ total electric consumption and utility’s load factor by using various load shaping techniques. However, applying load shaping techniques such as load shifting, peak clipping, or strategic conservation individually does not provide the desired level of improvement for load factor increment and/or customer’s bill reduction. In this paper, two load shaping techniques will be simulated as constrained optimization problems. The purpose is to reflect the application of combined load shifting and strategic conservation model together at the same time, and the application of combined load shifting and peak clipping model as well. The problem will be formulated and solved by using disciplined convex programming (CVX) based MATLAB® R2013b. Simulation results will be evaluated and compared for studying the most impactful multi-techniques model in improving load curve.

Keywords: convex programing, demand side management, load shaping, multiple, building energy optimization

Procedia PDF Downloads 313
9651 Optimal Allocation of PHEV Parking Lots to Minimize Dstribution System Losses

Authors: Mohsen Mazidi, Ali Abbaspour, Mahmud Fotuhi-Firuzabad, Mohamamd Rastegar

Abstract:

To tackle the air pollution issues, Plug-in Hybrid Electric Vehicles (PHEVs) are proposed as an appropriate solution. Charging a large amount of PHEV batteries, if not controlled, would have negative impacts on the distribution system. The control process of charging of these vehicles can be centralized in parking lots that may provide a chance for better coordination than the individual charging in houses. In this paper, an optimization-based approach is proposed to determine the optimum PHEV parking capacities in candidate nodes of the distribution system. In so doing, a profile for charging and discharging of PHEVs is developed in order to flatten the network load profile. Then, this profile is used in solving an optimization problem to minimize the distribution system losses. The outputs of the proposed method are the proper place for PHEV parking lots and optimum capacity for each parking. The application of the proposed method on the IEEE-34 node test feeder verifies the effectiveness of the method.

Keywords: loss, plug-in hybrid electric vehicle (PHEV), PHEV parking lot, V2G

Procedia PDF Downloads 543
9650 Adequacy of Antenatal Care and Its Relationship with Low Birth Weight in Botucatu, São Paulo, Brazil: A Case-Control Study

Authors: Cátia Regina Branco da Fonseca, Maria Wany Louzada Strufaldi, Lídia Raquel de Carvalho, Rosana Fiorini Puccini

Abstract:

Background: Birth weight reflects gestational conditions and development during the fetal period. Low birth weight (LBW) may be associated with antenatal care (ANC) adequacy and quality. The purpose of this study was to analyze ANC adequacy and its relationship with LBW in the Unified Health System in Brazil. Methods: A case-control study was conducted in Botucatu, São Paulo, Brazil, 2004 to 2008. Data were collected from secondary sources (the Live Birth Certificate), and primary sources (the official medical records of pregnant women). The study population consisted of two groups, each with 860 newborns. The case group comprised newborns weighing less than 2,500 grams, while the control group comprised live newborns weighing greater than or equal to 2,500 grams. Adequacy of ANC was evaluated according to three measurements: 1. Adequacy of the number of ANC visits adjusted to gestational age; 2. Modified Kessner Index; and 3. Adequacy of ANC laboratory studies and exams summary measure according to parameters defined by the Ministry of Health in the Program for Prenatal and Birth Care Humanization. Results: Analyses revealed that LBW was associated with the number of ANC visits adjusted to gestational age (OR = 1.78, 95% CI 1.32-2.34) and the ANC laboratory studies and exams summary measure (OR = 4.13, 95% CI 1.36-12.51). According to the modified Kessner Index, 64.4% of antenatal visits in the LBW group were adequate, with no differences between groups. Conclusions: Our data corroborate the association between inadequate number of ANC visits, laboratory studies and exams, and increased risk of LBW newborns. No association was found between the modified Kessner Index as a measure of adequacy of ANC and LBW. This finding reveals the low indices of coverage for basic actions already well regulated in the Health System in Brazil. Despite the association found in the study, we cannot conclude that LBW would be prevented only by an adequate ANC, as LBW is associated with factors of complex and multifactorial etiology. The results could be used to plan monitoring measures and evaluate programs of health care assistance during pregnancy, at delivery and to newborns, focusing on reduced LBW rates.

Keywords: low birth weight, antenatal care, prenatal care, adequacy of health care, health evaluation, public health system

Procedia PDF Downloads 433
9649 Optimization of Ultrasound Assisted Extraction of Polysaccharides from Plant Waste Materials: Selected Model Material is Hazelnut Skin

Authors: T. Yılmaz, Ş. Tavman

Abstract:

In this study, optimization of ultrasound assisted extraction (UAE) of hemicellulose based polysaccharides from plant waste material has been studied. Selected material is hazelnut skin. Extraction variables for the operation are extraction time, amplitude and application temperature. Optimum conditions have been evaluated depending on responses such as amount of wet crude polysaccharide, total carbohydrate content and dried sample. Pretreated hazelnut skin powders were used for the experiments. 10 grams of samples were suspended in 100 ml water in a jacketed vessel with additional magnetic stirring. Mixture was sonicated by immersing ultrasonic probe processor. After the extraction procedures, ethanol soluble and insoluble sides were separated for further examinations. The obtained experimental data were analyzed by analysis of variance (ANOVA). Second order polynomial models were developed using multiple regression analysis. The individual and interactive effects of applied variables were evaluated by Box Behnken Design. The models developed from the experimental design were predictive and good fit with the experimental data with high correlation coefficient value (R2 more than 0.95). Extracted polysaccharides from hazelnut skin are assumed to be pectic polysaccharides according to the literature survey of Fourier Transform Spectrometry (FTIR) analysis results. No more change can be observed between spectrums of different sonication times. Application of UAE at optimized condition has an important effect on extraction of hemicellulose from plant material by satisfying partial hydrolysis to break the bounds with other components in plant cell wall material. This effect can be summarized by varied intensity of microjets and microstreaming at varied sonication conditions.

Keywords: hazelnut skin, optimization, polysaccharide, ultrasound assisted extraction

Procedia PDF Downloads 333
9648 An Efficient Fundamental Matrix Estimation for Moving Object Detection

Authors: Yeongyu Choi, Ju H. Park, S. M. Lee, Ho-Youl Jung

Abstract:

In this paper, an improved method for estimating fundamental matrix is proposed. The method is applied effectively to monocular camera based moving object detection. The method consists of corner points detection, moving object’s motion estimation and fundamental matrix calculation. The corner points are obtained by using Harris corner detector, motions of moving objects is calculated from pyramidal Lucas-Kanade optical flow algorithm. Through epipolar geometry analysis using RANSAC, the fundamental matrix is calculated. In this method, we have improved the performances of moving object detection by using two threshold values that determine inlier or outlier. Through the simulations, we compare the performances with varying the two threshold values.

Keywords: corner detection, optical flow, epipolar geometry, RANSAC

Procedia PDF Downloads 409
9647 A Comparison of Alternative Traffic Controls for Interchange Ramp Areas Using Synchro Software

Authors: Mohamed Mesbah, Bruce Janson

Abstract:

An interchange is the most important component of freeway and highway facilities. It is working as a connector between the highway’s elements. The main goal of designing interchanges is to provide an acceptable level of service and delay to make vehicles move smoothly when they are entering and exiting the interchange. There are many factors that can have a significant impact on the level of service; the main factors are traffic volumes, and type of interchange. This paper will discuss interchange with roundabouts under various values of traffic volumes to determine the level of service of the interchanges that will be studied in this paper and replace the system of interchange from roundabout to traffic signal to make a significant compression between these systems. A secondary goal is to propose improvements for scenarios where the level of service is deemed unacceptable. This will be achieved using Synchro traffic simulation software, which facilitates the simulation and optimization of interchanges to enhance operational efficiency and safety.

Keywords: interchange, roundabout, traffic signal, Synchro, delay, level of service, traffic volumes, vehicles, simulation, optimization, adjustment

Procedia PDF Downloads 27
9646 A Deterministic Approach for Solving the Hull and White Interest Rate Model with Jump Process

Authors: Hong-Ming Chen

Abstract:

This work considers the resolution of the Hull and White interest rate model with the jump process. A deterministic process is adopted to model the random behavior of interest rate variation as deterministic perturbations, which is depending on the time t. The Brownian motion and jumps uncertainty are denoted as the integral functions piecewise constant function w(t) and point function θ(t). It shows that the interest rate function and the yield function of the Hull and White interest rate model with jump process can be obtained by solving a nonlinear semi-infinite programming problem. A relaxed cutting plane algorithm is then proposed for solving the resulting optimization problem. The method is calibrated for the U.S. treasury securities at 3-month data and is used to analyze several effects on interest rate prices, including interest rate variability, and the negative correlation between stock returns and interest rates. The numerical results illustrate that our approach essentially generates the yield functions with minimal fitting errors and small oscillation.

Keywords: optimization, interest rate model, jump process, deterministic

Procedia PDF Downloads 161
9645 Authentication and Legal Admissibility of 'Computer Evidence from Electronic Voting Machines' in Electoral Litigation: A Qualitative Legal Analysis of Judicial Opinions of Appellate Courts in the USA

Authors: Felix O. Omosele

Abstract:

Several studies have established that electronic voting machines are prone to multi-faceted challenges. One of which is their capacity to lose votes after the ballots might have been cast. Therefore, the international consensus appears to favour the use of electronic voting machines that are accompanied with verifiable audit paper audit trail (VVPAT). At present, there is no known study that has evaluated the impacts (or otherwise) of this verification and auditing on the authentication, admissibility and evidential weight of electronically-obtained electoral data. This legal inquiry is important as elections are sometimes won or lost in courts and on the basis of such data. This gap will be filled by the present research work. Using the United States of America as a case study, this paper employed a qualitative legal analysis of several of its appellate courts’ judicial opinions. This analysis equally unearths the necessary statutory rules and regulations that are important to the research problem. The objective of the research is to highlight the roles played by VVPAT on electoral evidence- as seen from the eyes of the court. The preliminary outcome of this qualitative analysis shows that the admissibility and weight attached to ‘Computer Evidence from e-voting machines (CEEM)’ are often treated with general standards applied to other computer-stored evidence. These standards sometimes fail to embrace the peculiar challenges faced by CEEM, particularly with respect to their tabulation and transmission. This paper, therefore, argues that CEEM should be accorded unique consideration by courts. It proposes the development of a legal standard which recognises verification and auditing as ‘weight enhancers’ for electronically-obtained electoral data.

Keywords: admissibility of computer evidence, electronic voting, qualitative legal analysis, voting machines in the USA

Procedia PDF Downloads 199
9644 A Practical Survey on Zero-Shot Prompt Design for In-Context Learning

Authors: Yinheng Li

Abstract:

The remarkable advancements in large language models (LLMs) have brought about significant improvements in natural language processing tasks. This paper presents a comprehensive review of in-context learning techniques, focusing on different types of prompts, including discrete, continuous, few-shot, and zero-shot, and their impact on LLM performance. We explore various approaches to prompt design, such as manual design, optimization algorithms, and evaluation methods, to optimize LLM performance across diverse tasks. Our review covers key research studies in prompt engineering, discussing their methodologies and contributions to the field. We also delve into the challenges faced in evaluating prompt performance, given the absence of a single ”best” prompt and the importance of considering multiple metrics. In conclusion, the paper highlights the critical role of prompt design in harnessing the full potential of LLMs and provides insights into the combination of manual design, optimization techniques, and rigorous evaluation for more effective and efficient use of LLMs in various Natural Language Processing (NLP) tasks.

Keywords: in-context learning, prompt engineering, zero-shot learning, large language models

Procedia PDF Downloads 86
9643 Estimation of Fourier Coefficients of Flux Density for Surface Mounted Permanent Magnet (SMPM) Generators by Direct Search Optimization

Authors: Ramakrishna Rao Mamidi

Abstract:

It is essential for Surface Mounted Permanent Magnet (SMPM) generators to determine the performance prediction and analyze the magnet’s air gap flux density wave shape. The flux density wave shape is neither a pure sine wave or square wave nor a combination. This is due to the variation of air gap reluctance between the stator and permanent magnets. The stator slot openings and the number of slots make the wave shape highly complicated. To reduce the complexity of analysis, approximations are made to the wave shape using Fourier analysis. In contrast to the traditional integration method, the Fourier coefficients, an and bn, are obtained by direct search method optimization. The wave shape with optimized coefficients gives a wave shape close to the desired wave shape. Harmonics amplitudes are worked out and compared with initial values. It can be concluded that the direct search method can be used for estimating Fourier coefficients for irregular wave shapes.

Keywords: direct search, flux plot, fourier analysis, permanent magnets

Procedia PDF Downloads 217
9642 Damage Assessment of Reinforced Concrete Slabs Subjected to Blast Loading

Authors: W. Badla

Abstract:

A numerical investigation has been carried out to examine the behaviour of reinforced concrete slabs to uniform blast loading. The aim of this work is to determine the effects of various parameters on the results. Finite element simulations were performed in the non linear dynamic range using an elasto-plastic damage model. The main parameters considered are: the negative phase of blast loading, time duration, equivalent weight of TNT, distance of the explosive and slab dimensions. Numerical modelling has been performed using ABAQUS/Explicit. The results obtained in terms of displacements and propagation of damage show that the above parameters influence considerably the nonlinear dynamic behaviour of reinforced concrete slabs under uniform blast loading.

Keywords: blast loading, reinforced concrete slabs, elasto-plastic damage model, negative phase, time duration, equivalent weight of TNT, explosive distance, slab dimensions

Procedia PDF Downloads 534
9641 Optimization of Ultrasonic Assisted Extraction of Antioxidants and Phenolic Compounds from Coleus Using Response Surface Methodology

Authors: Reihaneh Ahmadzadeh Ghavidel

Abstract:

Free radicals such as reactive oxygen species (ROS) have detrimental effects on human health through several mechanisms. On the other hand, antioxidant molecules reduce free radical generation in biologic systems. Synthetic antioxidants, which are used in food industry, have also negative impact on human health. Therefore recognition of natural antioxidants such as anthocyanins can solve these problems simultaneously. Coleus (Solenostemon scutellarioides) with red leaves is a rich source of anthocyanins compounds. In this study we evaluated the effect of time (10, 20 and 30 min) and temperature (40, 50 and 60° C) on optimization of anthocyanin extraction using surface response method. In addition, the study was aimed to determine maximum extraction for anthocyanin from coleus plant using ultrasound method. The results indicated that the optimum conditions for extraction were 39.84 min at 69.25° C. At this point, total compounds were achieved 3.7451 mg 100 ml⁻¹. Furthermore, under optimum conditions, anthocyanin concentration, extraction efficiency, ferric reducing ability, total phenolic compounds and EC50 were registered 3.221931, 6.692765, 223.062, 3355.605 and 2.614045, respectively.

Keywords: anthocyanin, antioxidant, coleus, extraction, sonication

Procedia PDF Downloads 321
9640 Heart and Plasma LDH and CK in Response to Intensive Treadmill Running and Aqueous Extraction of Red Crataegus pentagyna in Male Rats

Authors: A. Abdi, A. Barari, A. Hojatollah Nikbakht, Khosro Ebrahim

Abstract:

Aim: The purpose of the current study was to investigate the effect of a high intensity treadmill running training (8 weeks) with or without aqueous extraction of Crataegus pentagyna on heart and plasma LDH and CK. Design: Thirty-two Wistar male rats (4-6 weeks old, 125-135 gr weight) were used. Animals were randomly assigned into training (n = 16) and control (n = 16) groups and further divided into saline-control (SC, n = 8), saline-training (ST, n = 8), red Crataegus pentagyna extraction -control (CPEC, n = 8), and red Crataegus pentagyna extraction -training (CPET, n = 8) groups. Training groups have performed a high-intensity running program 34 m/min on 0% grade, 60 min/day, 5 days/week) on a motor-driven treadmill for 8 weeks. Animals were fed orally with Crataegus extraction and saline solution (500mg/kg body weight/or 10ml/kg body weight) for last six weeks. Seventy- two hours after the last training session, rats were sacrificed; plasma and heart were excised and immediately frozen in liquid nitrogen. LDH and CK levels were measured by colorimetric method. Statistical analysis was performed using a one way analysis of variance and Tukey test. Significance was accepted at P = 0.05. Results: Result showed that consumption crataegus lowers LDH and CK in heart and plasma. Also the heart LDH and CK were lower in the CPET compared to the ST, while plasma LDH and CK in CPET was higher than the ST. The results of ANOVA showed that the due high-intensity exercise and consumption crataegus, there are significant differences between levels of hearth LDH (P < 0/001), plasma (P < 0/006) and hearth (P < 0/001) CK. Conclusion: It appears that high-intensity exercise led to increased tissue damage and inflammatory factors in plasma. In other hand, consumption aqueous extraction of Red Crataegus maybe inhibits these factors and prevents muscle and heart damage.

Keywords: LDH, CK, crataegus, intensity

Procedia PDF Downloads 439
9639 Direct Contact Ultrasound Assisted Drying of Mango Slices

Authors: E. K. Mendez, N. A. Salazar, C. E. Orrego

Abstract:

There is undoubted proof that increasing the intake of fruit lessens the risk of hypertension, coronary heart disease, stroke, and probable evidence that lowers the risk of cancer. Proper fruit drying is an excellent alternative to make their shelf-life longer, commercialization easier, and ready-to-eat healthy products or ingredients. The conventional way of drying is by hot air forced convection. However, this process step often requires a very long residence time; furthermore, it is highly energy consuming and detrimental to the product quality. Nowadays, power ultrasound (US) technic has been considered as an emerging and promising technology for industrial food processing. Most of published works dealing with drying food assisted by US have studied the effect of ultrasonic pre-treatment prior to air-drying on food and the airborne US conditions during dehydration. In this work a new approach was tested taking in to account drying time and two quality parameters of mango slices dehydrated by convection assisted by 20 KHz power US applied directly using a holed plate as product support and sound transmitting surface. During the drying of mango (Mangifera indica L.) slices (ca. 6.5 g, 0.006 m height and 0.040 m diameter), their weight was recorded every hour until final moisture content (10.0±1.0 % wet basis) was reached. After previous tests, optimization of three drying parameters - frequencies (2, 5 and 8 minutes each half-hour), air temperature (50-55-60⁰C) and power (45-70-95W)- was attempted by using a Box–Behnken design under the response surface methodology for the optimal drying time, color parameters and rehydration rate of dried samples. Assays involved 17 experiments, including a quintuplicate of the central point. Dried samples with and without US application were packed in individual high barrier plastic bags under vacuum, and then stored in the dark at 8⁰C until their analysis. All drying assays and sample analysis were performed in triplicate. US drying experimental data were fitted with nine models, among which the Verna model resulted in the best fit with R2 > 0.9999 and reduced χ2 ≤ 0.000001. Significant reductions in drying time were observed for the assays that used lower frequency and high US power. At 55⁰C, 95 watts and 2 min/30 min of sonication, 10% moisture content was reached in 211 min, as compared with 320 min for the same test without the use of US (blank). Rehydration rates (RR), defined as the ratio of rehydrated sample weight to that of dry sample and measured, was also larger than those of blanks and, in general, the higher the US power, the greater the RR. The direct contact and intermittent US treatment of mango slices used in this work improve drying rates and dried fruit rehydration ability. This technique can thus be used to reduce energy processing costs and the greenhouse gas emissions of fruit dehydration.

Keywords: ultrasonic assisted drying, fruit drying, mango slices, contact ultrasonic drying

Procedia PDF Downloads 348
9638 Effect of Organophilic Clay on the Stability and Rheological Behavior of Oil-Based Drilling Muds

Authors: Hammadi Larbi

Abstract:

The major problem with oil-based drilling muds (reverse emulsions) is their thermodynamic instability and their high tendency to coalescence over time, irreversibly leading to destabilization. Water/Oil reverse emulsion drilling Muds are highly recommended when significant depths are reached. This study aimed to contribute experimentally to the knowledge of the structure (stability) and rheological behavior of drilling mud systems based on water/crude oil inverse emulsions through the investigation of the effect of organophilic clay. The chemical composition of organophilic clay such as VG69 shows a strong presence of silicon oxide (SiO2), followed by aluminum oxide (Al2O3), so these two elements are considered to be the main constituents of organophilic clays. The study also shows that the SiO2/Al2O3 ratio is equal to 3.52, which can be explained by the high content of free silica contained in the organophile clay used. The particle size analysis of the organophilic clays showed that the size of the of the particles analysed is in the range of 30 to 80 μm, this result ensures the correct particle size quality of organophilic clays and allows these powders to be used in Drilling mud systems.The experimental data of steady-state flow measurements are analyzed in the classic way by the Herschel-Bulkley model. Microscopic observation shows that the addition of quantities of organophilic clay type VG69 less than or equal to 3 g leading to the stability of the water/oil inverse emulsions, on the other hand, for quantities greater than 3 g, the emulsions are destabilized. The results obtained also showed that adding 3 g of organophilic clay to the crude oil drilling mud improves their stability by 70%.

Keywords: drilling muds, inverse emulsions, rheological behavior, yield stress, stability, organophilic clay

Procedia PDF Downloads 17
9637 Determination of Slope of Hilly Terrain by Using Proposed Method of Resolution of Forces

Authors: Reshma Raskar-Phule, Makarand Landge, Saurabh Singh, Vijay Singh, Jash Saparia, Shivam Tripathi

Abstract:

For any construction project, slope calculations are necessary in order to evaluate constructability on the site, such as the slope of parking lots, sidewalks, and ramps, the slope of sanitary sewer lines, slope of roads and highways. When slopes and grades are to be determined, designers are concerned with establishing proper slopes and grades for their projects to assess cut and fill volume calculations and determine inverts of pipes. There are several established instruments commonly used to determine slopes, such as Dumpy level, Abney level or Hand Level, Inclinometer, Tacheometer, Henry method, etc., and surveyors are very familiar with the use of these instruments to calculate slopes. However, they have some other drawbacks which cannot be neglected while major surveying works. Firstly, it requires expert surveyors and skilled staff. The accessibility, visibility, and accommodation to remote hilly terrain with these instruments and surveying teams are difficult. Also, determination of gentle slopes in case of road and sewer drainage constructions in congested urban places with these instruments is not easy. This paper aims to develop a method that requires minimum field work, minimum instruments, no high-end technology or instruments or software, and low cost. It requires basic and handy surveying accessories like a plane table with a fixed weighing machine, standard weights, alidade, tripod, and ranging rods should be able to determine the terrain slope in congested areas as well as in remote hilly terrain. Also, being simple and easy to understand and perform the people of that local rural area can be easily trained for the proposed method. The idea for the proposed method is based on the principle of resolution of weight components. When any object of standard weight ‘W’ is placed on an inclined surface with a weighing machine below it, then its cosine component of weight is presently measured by that weighing machine. The slope can be determined from the relation between the true or actual weight and the apparent weight. A proper procedure is to be followed, which includes site location, centering and sighting work, fixing the whole set at the identified station, and finally taking the readings. A set of experiments for slope determination, mild and moderate slopes, are carried out by the proposed method and by the theodolite instrument in a controlled environment, on the college campus, and uncontrolled environment actual site. The slopes determined by the proposed method were compared with those determined by the established instruments. For example, it was observed that for the same distances for mild slope, the difference in the slope obtained by the proposed method and by the established method ranges from 4’ for a distance of 8m to 2o15’20” for a distance of 16m for an uncontrolled environment. Thus, for mild slopes, the proposed method is suitable for a distance of 8m to 10m. The correlation between the proposed method and the established method shows a good correlation of 0.91 to 0.99 for various combinations, mild and moderate slope, with the controlled and uncontrolled environment.

Keywords: surveying, plane table, weight component, slope determination, hilly terrain, construction

Procedia PDF Downloads 97
9636 Molecular Dynamic Simulation of Cold Spray Process

Authors: Aneesh Joshi, Sagil James

Abstract:

Cold Spray (CS) process is deposition of solid particles over a substrate above a certain critical impact velocity. Unlike thermal spray processes, CS process does not melt the particles thus retaining their original physical and chemical properties. These characteristics make CS process ideal for various engineering applications involving metals, polymers, ceramics and composites. The bonding mechanism involved in CS process is extremely complex considering the dynamic nature of the process. Though CS process offers great promise for several engineering applications, the realization of its full potential is limited by the lack of understanding of the complex mechanisms involved in this process and the effect of critical process parameters on the deposition efficiency. The goal of this research is to understand the complex nanoscale mechanisms involved in CS process. The study uses Molecular Dynamics (MD) simulation technique to understand the material deposition phenomenon during the CS process. Impact of a single crystalline copper nanoparticle on copper substrate is modelled under varying process conditions. The quantitative results of the impacts at different velocities, impact angle and size of the particles are evaluated using flattening ratio, von Mises stress distribution and local shear strain. The study finds that the flattening ratio and hence the quality of deposition was highest for an impact velocity of 700 m/s, particle size of 20 Å and an impact angle of 90°. The stress and strain analysis revealed regions of shear instabilities in the periphery of impact and also revealed plastic deformation of the particles after the impact. The results of this study can be used to augment our existing knowledge in the field of CS processes.

Keywords: cold spray process, molecular dynamics simulation, nanoparticles, particle impact

Procedia PDF Downloads 370
9635 The Effect of Different Level Crop Load and Humic Substance Applications on Yield and Yield Components of Alphonse Lavallee Grape Cultivar

Authors: A. Sarıkaya, A. Akın

Abstract:

This study was carried out to investigate effects of Control (C), 18 bud/vine, 23 bud/vine, 28 bud/vine, 18 bud/vine + TKI-Humas (soil), 23 bud/vine + TKI-Humas (soil), 28 bud/vine + TKI-Humas (soil) applications on yield and yield components of Alphonse Lavallee grape cultivar. The results were obtained as the highest cluster weight (302.31 g) with 18 bud/vine application; the highest berry weight (6.31 g) with 23 bud/vine + TKI-Humas (soil) and (6.79 g) with 28 bud/vine + TKI-Humas (soil) applications; the highest maturity index (36.95) with 18 bud/vine + TKI-Humas (soil) application; the highest L* color intensity (33.99) with 18 bud/vine + TKI-Humas (soil); the highest a* color intensity (1.53) with 23 bud/vine + TKI-Humas (soil) application. The effects of applications on grape fresh yield, grape juice yield and b* color intensity values were not found statistically significant.

Keywords: Alphonse Lavallee grape cultivar, crop load, TKI-Humas substances (soil), yield, quality

Procedia PDF Downloads 292
9634 Optimization of Stevia Concentration in Rasgulla (Sweet Syrup Cheese Ball) Based on Quality

Authors: Gurveer Kaur, T. K. Goswami

Abstract:

Rasgulla (a sweet syrup cheese ball), a sweet, spongy dessert represents traditional sweet dish of an Indian subcontinent prepared by chhana. 100 g of Rasgulla contains 186 calories, and so it is a driving force behind obesity and diabetes. To reduce Rasgulla’s energy value sucrose mainly should be minimized, so instead of sucrose, stevia (zero calories natural sweetener) is used to prepare Rasgulla. In this study three samples were prepared with sucrose to stevia ratio taking 100:0 (as control sample), (i) 50:50 (T1); (ii) 25:75 (T2), and (iii) 0:100 (T3) from 4% fat milk. It was found that as the sucrose concentration decreases the percentage of fat increase in the Rasgulla slightly. Sample T2 showed < 0.1% (±0.06) sucrose content. But there was no significant difference on protein and ash content of the samples. Whitening index was highest (78.0 ± 0.13) for T2 and lowest (65.7 ± 0.21) for the control sample since less sucrose in syrup reduces the browning of the sample (T2). Energy value per 100 g was calculated to be 50, 72, 98, and 184 calories for T3, T2, T1 and control samples, respectively. According to optimization study, the preferred (high quality) order of samples was as follows: T1 > T1 > control > T3. Low sugar content Rasgulla with acceptable quality can be prepared with 25:75 ratio of sucrose to stevia.

Keywords: composition, rasgulla, sensory, stevia

Procedia PDF Downloads 206
9633 Performance Analysis of Curved U-Slot Patch Antenna with Enhanced Bandwidth and Isolation for Mimo Systems

Authors: Umesh Kumar, Arun Kumar Shukla, B. V. V. Ravindra Babu

Abstract:

The paper presents a compact tri band Curved U-Slot patch antenna with improved bandwidth and isolation characteristics. The proposed antenna excited by coaxial feed resonates at tri band of 2.8 GHz, 4.1 GHz and 5.7 GHz for VSWR ≤ 1.5 with an improved bandwidth of 99.7% and also for getting high gain antenna of 11.31 dB. A 2×2 MIMO is developed using the proposed antenna giving an excellent isolation of 28 dB between the two antennas. The simulation results of return loss, Mutual Coupling, Gain, VSWR, Surface Current Distribution and Electrical Distribution are presented. By keeping the substrate thickness constant over various dielectric constants, simulations were carried out using MATLAB® and HFSS (High Frequency Structure Simulator) software.

Keywords: performance analysis, curved U-slot patch, antenna with enhanced bandwidth, isolation for mimo systems

Procedia PDF Downloads 587
9632 Production of Rhamnolipids from Different Resources and Estimating the Kinetic Parameters for Bioreactor Design

Authors: Olfat A. Mohamed

Abstract:

Rhamnolipids biosurfactants have distinct properties given them importance in many industrial applications, especially their great new future applications in cosmetic and pharmaceutical industries. These applications have encouraged the search for diverse and renewable resources to control the cost of production. The experimental results were then applied to find a suitable mathematical model for obtaining the design criteria of the batch bioreactor. This research aims to produce Rhamnolipids from different oily wastewater sources such as petroleum crude oil (PO) and vegetable oil (VO) by using Pseudomonas aeruginosa ATCC 9027. Different concentrations of the PO and the VO are added to the media broth separately are in arrangement (0.5 1, 1.5, 2, 2.5 % v/v) and (2, 4, 6, 8 and 10%v/v). The effect of the initial concentration of oil residues and the addition of glycerol and palmitic acid was investigated as an inducer in the production of rhamnolipid and the surface tension of the broth. It was found that 2% of the waste (PO) and 6% of the waste (VO) was the best initial substrate concentration for the production of rhamnolipids (2.71, 5.01 g rhamnolipid/l) as arrangement. Addition of glycerol (10-20% v glycerol/v PO) to the 2% PO fermentation broth led to increase the rhamnolipid production (about 1.8-2 times fold). However, the addition of palmitic acid (5 and 10 g/l) to fermentation broth contained 6% VO rarely enhanced the production rate. The experimental data for 2% initially (PO) was used to estimate the various kinetic parameters. The following results were obtained, maximum rate or velocity of reaction (Vmax) = 0.06417 g/l.hr), yield of cell weight per unit weight of substrate utilized (Yx/s = 0.324 g Cx/g Cs) maximum specific growth rate (μmax = 0.05791 hr⁻¹), yield of rhamnolipid weight per unit weight of substrate utilized (Yp/s)=0.2571gCp/g Cs), maintenance coefficient (Ms =0.002419), Michaelis-Menten constant, (Km=6.1237 gmol/l), endogenous decay coefficient (Kd=0.002375 hr⁻¹). Predictive parameters and advanced mathematical models were applied to evaluate the time of the batch bioreactor. The results were as follows: 123.37, 129 and 139.3 hours in respect of microbial biomass, substrate and product concentration, respectively compared with experimental batch time of 120 hours in all cases. The expected mathematical models are compatible with the laboratory results and can, therefore, be considered as tools for expressing the actual system.

Keywords: batch bioreactor design, glycerol, kinetic parameters, petroleum crude oil, Pseudomonas aeruginosa, rhamnolipids biosurfactants, vegetable oil

Procedia PDF Downloads 133
9631 Therapeutic Effect of 12 Weeks of Sensorimotor Exercise on Pain, Functionality and Quality of Life in Non-athlete Women With Patellofemoral Pain Syndrome

Authors: Kasbparast Mehdi, Hassani Zainab

Abstract:

Aim: The purpose of this research was to investigate the effectiveness of therapeutical sensorimotor exercise. The statistical population of women who were diagnosed with patellofemoral pain syndrome by a doctor and were between the ages of 35 and 45 and registered for the first time in a sports club in the 4th district of Tehran, 30 people by random sampling and according to The include and exclude criteria were selected and divided into 2 equal control and experimental and homogeneous groups (in terms of height, weight and BMI).In both control and experimental groups, the pain was measured using a Visual Analog Scale(VAS) functionality was measured using the step-down test and quality of life was measured using a World Health Organization Quality of Life Scale (WHOQOL-BREF) (pre-test). Then, only the experimental group performed sensorimotor exercises for 12 weeks and 3 sessions each week, a total of 24 sessions and each session for 1 hour, and during this period, the control group only continued their daily activities. After the end of the training period, the desired factors were evaluated again (post-test) in the same way as the pre-test was done for them (experimental group and control group), with the same quality. Findings: The statistical results showed that in the experimental group, the amount of pain, function and quality of life had a statistical improvement (P≤0.05). Conclusion: In general conclusion, it can be stated that using sensorimotor exercises not only improved functionality and quality of life but also reduced the amount of pain in people with patellofemoral pain syndrome.

Keywords: pain, PFPS, sensori motor training, functionality

Procedia PDF Downloads 77
9630 Key Frame Based Video Summarization via Dependency Optimization

Authors: Janya Sainui

Abstract:

As a rapid growth of digital videos and data communications, video summarization that provides a shorter version of the video for fast video browsing and retrieval is necessary. Key frame extraction is one of the mechanisms to generate video summary. In general, the extracted key frames should both represent the entire video content and contain minimum redundancy. However, most of the existing approaches heuristically select key frames; hence, the selected key frames may not be the most different frames and/or not cover the entire content of a video. In this paper, we propose a method of video summarization which provides the reasonable objective functions for selecting key frames. In particular, we apply a statistical dependency measure called quadratic mutual informaion as our objective functions for maximizing the coverage of the entire video content as well as minimizing the redundancy among selected key frames. The proposed key frame extraction algorithm finds key frames as an optimization problem. Through experiments, we demonstrate the success of the proposed video summarization approach that produces video summary with better coverage of the entire video content while less redundancy among key frames comparing to the state-of-the-art approaches.

Keywords: video summarization, key frame extraction, dependency measure, quadratic mutual information

Procedia PDF Downloads 270
9629 A Statistical Approach to Predict and Classify the Commercial Hatchability of Chickens Using Extrinsic Parameters of Breeders and Eggs

Authors: M. S. Wickramarachchi, L. S. Nawarathna, C. M. B. Dematawewa

Abstract:

Hatchery performance is critical for the profitability of poultry breeder operations. Some extrinsic parameters of eggs and breeders cause to increase or decrease the hatchability. This study aims to identify the affecting extrinsic parameters on the commercial hatchability of local chicken's eggs and determine the most efficient classification model with a hatchability rate greater than 90%. In this study, seven extrinsic parameters were considered: egg weight, moisture loss, breeders age, number of fertilised eggs, shell width, shell length, and shell thickness. Multiple linear regression was performed to determine the most influencing variable on hatchability. First, the correlation between each parameter and hatchability were checked. Then a multiple regression model was developed, and the accuracy of the fitted model was evaluated. Linear Discriminant Analysis (LDA), Classification and Regression Trees (CART), k-Nearest Neighbors (kNN), Support Vector Machines (SVM) with a linear kernel, and Random Forest (RF) algorithms were applied to classify the hatchability. This grouping process was conducted using binary classification techniques. Hatchability was negatively correlated with egg weight, breeders' age, shell width, shell length, and positive correlations were identified with moisture loss, number of fertilised eggs, and shell thickness. Multiple linear regression models were more accurate than single linear models regarding the highest coefficient of determination (R²) with 94% and minimum AIC and BIC values. According to the classification results, RF, CART, and kNN had performed the highest accuracy values 0.99, 0.975, and 0.972, respectively, for the commercial hatchery process. Therefore, the RF is the most appropriate machine learning algorithm for classifying the breeder outcomes, which are economically profitable or not, in a commercial hatchery.

Keywords: classification models, egg weight, fertilised eggs, multiple linear regression

Procedia PDF Downloads 89
9628 Improved Active Constellation Extension for the PAPR Reduction of FBMC-OQAM Signals

Authors: Mounira Laabidi, Rafik Zayani, Ridha Bouallegue, Daniel Roviras

Abstract:

The Filter Bank multicarrier with Offset Quadrature Amplitude Modulation (FBMC-OQAM) has been introduced to overcome the poor spectral characteristics and the waste in both bandwidth and energy caused by the use of the cyclic prefix. However, the FBMC-OQAM signals suffer from the high Peak to Average Power Ratio (PAPR) problem. Due to the overlapping structure of the FBMC-OQAM signals, directly applying the PAPR reduction schemes conceived for the OFDM one turns out to be ineffective. In this paper, we address the problem of PAPR reduction for FBMC-OQAM systems by suggesting a new scheme based on an improved version of Active Constellation Extension scheme (ACE) of OFDM. The proposed scheme, named Rolling Window ACE, takes into consideration the overlapping naturally emanating from the FBMC-OQAM signals.

Keywords: ACE, FBMC, OQAM, OFDM, PAPR, rolling-window

Procedia PDF Downloads 547