Search results for: Leakage reduction
1152 Effects of Entomopathogenic Nematodes on Suppressing Hairy Rose Beetle, Tropinota squalida Scop. (Coleoptera: Scarabaeidae) Population in Cauliflower Field in Egypt
Authors: A. S. Abdel-Razek, M. M. M. Abd-Elgawad
Abstract:
The potential of entomopathogenic nematodes in suppressing T. squalida population on cauliflower from transplanting to harvest was evaluated. Significant reductions in plant infestation percentage and population density (/m2) were recorded throughout the plantation seasons, 2011 and 2012 before and after spraying the plants. The percent reduction in numbers/m2 was the highest in March for the treatments with Heterorhabditis indica Behera and Heterorhabditis bacteriophora Giza during the plantation season 2011, while at the plantation season 2012, the reduction in population density was the highest in January for Heterorhabditis Indica Behera and in February for H . bacteriophora Giza treatments. In a comparison test with conventional insecticides Hostathion and Lannate, there were no significant differences in control measures resulting from treatments with H. indica Behera, H. bacteriophora Giza and Lannate. At the plantation season is 2012. Also, the treatments reduced the economic threshold of T. squalida on cauliflower in this experiment as compared with before and after spraying with both the two entomopathogenic nematodes at both seasons 2011 and 2012. This means an increase in the marketability of heads harvested as a consequence of monthly treatments.
Keywords: Cruciferous plants, chemical insecticides, microbial control, Scarabiead beetles, seasonal monitoring.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18621151 Robotics and Embedded Systems Applied to the Buried Pipeline Inspection
Authors: Robson C. Santos, Julio C. P. Ribeiro, Iorran M. de Castro, Luan C. F. Rodrigues, Sandro R. L. Silva, Diego M. Quesada
Abstract:
The work aims to develop a robot in the form of autonomous vehicle to detect, inspection and mapping of underground pipelines through the ATmega328 Arduino platform. Hardware prototyping is very similar to C / C ++ language that facilitates its use in robotics open source, resembles PLC used in large industrial processes. The robot will traverse the surface independently of direct human action, in order to automate the process of detecting buried pipes, guided by electromagnetic induction. The induction comes from coils that send the signal to the Arduino microcontroller contained in that will make the difference in intensity and the treatment of the information, and then this determines actions to electrical components such as relays and motors, allowing the prototype to move on the surface and getting the necessary information. This change of direction is performed by a stepper motor with a servo motor. The robot was developed by electrical and electronic assemblies that allowed test your application. The assembly is made up of metal detector coils, circuit boards and microprocessor, which interconnected circuits previously developed can determine, process control and mechanical actions for a robot (autonomous car) that will make the detection and mapping of buried pipelines plates. This type of prototype can prevent and identifies possible landslides and they can prevent the buried pipelines suffer an external pressure on the walls with the possibility of oil leakage and thus pollute the environment.Keywords: Robotic, metal detector, embedded system, pipeline.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21601150 Flow Regime Characterization in a Diseased Artery Model
Authors: Anis S. Shuib, Peter R. Hoskins, William J. Easson
Abstract:
Cardiovascular disease mostly in the form of atherosclerosis is responsible for 30% of all world deaths amounting to 17 million people per year. Atherosclerosis is due to the formation of plaque. The fatty plaque may be at risk of rupture, leading typically to stroke and heart attack. The plaque is usually associated with a high degree of lumen reduction, called a stenosis. The initiation and progression of the disease is strongly linked to the hemodynamic environment near the vessel wall. The aim of this study is to validate the flow of blood mimic through an arterial stenosis model with computational fluid dynamics (CFD) package. In experiment, an axisymmetric model constructed consists of contraction and expansion region that follow a mathematical form of cosine function. A 30% diameter reduction was used in this study. Particle image velocimetry (PIV) was used to characterize the flow. The fluid consists of rigid spherical particles suspended in waterglycerol- NaCl mixture. The particles with 20 μm diameter were selected to follow the flow of fluid. The flow at Re=155, 270 and 390 were investigated. The experimental result is compared with FLUENT simulated flow that account for viscous laminar flow model. The results suggest that laminar flow model was sufficient to predict flow velocity at the inlet but the velocity at stenosis throat at Re =390 was overestimated. Hence, a transition to turbulent regime might have been developed at throat region as the flow rate increases.
Keywords: Atherosclerosis, Particle-laden flow, Particle imagevelocimetry, Stenosis artery
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17231149 Optimization of Air Pollution Control Model for Mining
Authors: Zunaira Asif, Zhi Chen
Abstract:
The sustainable measures on air quality management are recognized as one of the most serious environmental concerns in the mining region. The mining operations emit various types of pollutants which have significant impacts on the environment. This study presents a stochastic control strategy by developing the air pollution control model to achieve a cost-effective solution. The optimization method is formulated to predict the cost of treatment using linear programming with an objective function and multi-constraints. The constraints mainly focus on two factors which are: production of metal should not exceed the available resources, and air quality should meet the standard criteria of the pollutant. The applicability of this model is explored through a case study of an open pit metal mine, Utah, USA. This method simultaneously uses meteorological data as a dispersion transfer function to support the practical local conditions. The probabilistic analysis and the uncertainties in the meteorological conditions are accomplished by Monte Carlo simulation. Reasonable results have been obtained to select the optimized treatment technology for PM2.5, PM10, NOx, and SO2. Additional comparison analysis shows that baghouse is the least cost option as compared to electrostatic precipitator and wet scrubbers for particulate matter, whereas non-selective catalytical reduction and dry-flue gas desulfurization are suitable for NOx and SO2 reduction respectively. Thus, this model can aid planners to reduce these pollutants at a marginal cost by suggesting control pollution devices, while accounting for dynamic meteorological conditions and mining activities.
Keywords: Air pollution, linear programming, mining, optimization, treatment technologies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16061148 Bioremediation Potential in Recalcitrant Areas of PCE in Alluvial Fan Deposits
Authors: J. Herrero, D. Puigserver, I. Nijenhuis, K. Kuntze, J. M. Carmona
Abstract:
In the transition zone between aquifers and basal aquitards, the perchloroethene (PCE)-pools are more recalcitrant than those elsewhere in the aquifer. Although biodegradation of chloroethenes occur in this zone, it is a slow process and a remediation strategy is needed. The aim of this study is to demonstrate that combined strategy of biostimulation and in situ chemical reduction (ISCR) is more efficient than the two separated strategies. Four different microcosm experiments with sediment and groundwater of a selected field site where an aged pool exists at the bottom of a transition zone were designed under i) natural conditions, ii) biostimulation with lactic acid, iii) ISCR with zero-value iron (ZVI) and under iv) a combined strategy with lactic acid and ZVI. Biotic and abiotic dehalogenation, terminal electron acceptor processes and evolution of microbial communities were determined for each experiment. The main results were: i) reductive dehalogenation of PCE-pools occurs under sulfate-reducing conditions; ii) biostimulation with lactic acid supports more pronounced reductive dehalogenation of PCE and trichloroethene (TCE), but results in an accumulation of 1,2-cis-dichloroethene (cDCE); iii) ISCR with ZVI produces a sustained dehalogenation of PCE and its metabolites iv) combined strategy of biostimulation and ISCR results in a fast dehalogenation of PCE and TCE and a sustained dehalogenation of cisDCE. These findings suggest that biostimulation and ISCR with ZVI are the most suitable strategies for a complete reductive dehalogenation of PCE-pools in the transition zone and further to enable the dissolution of dense non-aqueous phase liquids.
Keywords: Aged PCE-pool, anaerobic microcosm experiment, biostimulation, in situ chemical reduction, natural attenuation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7341147 Optimized Brain Computer Interface System for Unspoken Speech Recognition: Role of Wernicke Area
Authors: Nassib Abdallah, Pierre Chauvet, Abd El Salam Hajjar, Bassam Daya
Abstract:
In this paper, we propose an optimized brain computer interface (BCI) system for unspoken speech recognition, based on the fact that the constructions of unspoken words rely strongly on the Wernicke area, situated in the temporal lobe. Our BCI system has four modules: (i) the EEG Acquisition module based on a non-invasive headset with 14 electrodes; (ii) the Preprocessing module to remove noise and artifacts, using the Common Average Reference method; (iii) the Features Extraction module, using Wavelet Packet Transform (WPT); (iv) the Classification module based on a one-hidden layer artificial neural network. The present study consists of comparing the recognition accuracy of 5 Arabic words, when using all the headset electrodes or only the 4 electrodes situated near the Wernicke area, as well as the selection effect of the subbands produced by the WPT module. After applying the articial neural network on the produced database, we obtain, on the test dataset, an accuracy of 83.4% with all the electrodes and all the subbands of 8 levels of the WPT decomposition. However, by using only the 4 electrodes near Wernicke Area and the 6 middle subbands of the WPT, we obtain a high reduction of the dataset size, equal to approximately 19% of the total dataset, with 67.5% of accuracy rate. This reduction appears particularly important to improve the design of a low cost and simple to use BCI, trained for several words.Keywords: Brain-computer interface, speech recognition, electroencephalography EEG, Wernicke area, artificial neural network.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9181146 Conventional and PSO Based Approaches for Model Reduction of SISO Discrete Systems
Authors: S. K. Tomar, R. Prasad, S. Panda, C. Ardil
Abstract:
Reduction of Single Input Single Output (SISO) discrete systems into lower order model, using a conventional and an evolutionary technique is presented in this paper. In the conventional technique, the mixed advantages of Modified Cauer Form (MCF) and differentiation are used. In this method the original discrete system is, first, converted into equivalent continuous system by applying bilinear transformation. The denominator of the equivalent continuous system and its reciprocal are differentiated successively, the reduced denominator of the desired order is obtained by combining the differentiated polynomials. The numerator is obtained by matching the quotients of MCF. The reduced continuous system is converted back into discrete system using inverse bilinear transformation. In the evolutionary technique method, Particle Swarm Optimization (PSO) is employed to reduce the higher order model. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example.
Keywords: Discrete System, Single Input Single Output (SISO), Bilinear Transformation, Reduced Order Model, Modified CauerForm, Polynomial Differentiation, Particle Swarm Optimization, Integral Squared Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19431145 Hybrid Advanced Oxidative Pretreatment of Complex Industrial Effluent for Biodegradability Enhancement
Authors: K. Paradkar, S. N. Mudliar, A. Sharma, A. B. Pandit, R. A. Pandey
Abstract:
The study explores the hybrid combination of Hydrodynamic Cavitation (HC) and Subcritical Wet Air Oxidation-based pretreatment of complex industrial effluent to enhance the biodegradability selectively (without major COD destruction) to facilitate subsequent enhanced downstream processing via anaerobic or aerobic biological treatment. Advanced oxidation based techniques can be less efficient as standalone options and a hybrid approach by combining Hydrodynamic Cavitation (HC), and Wet Air Oxidation (WAO) can lead to a synergistic effect since both the options are based on common free radical mechanism. The HC can be used for initial turbulence and generation of hotspots which can begin the free radical attack and this agitating mixture then can be subjected to less intense WAO since initial heat (to raise the activation energy) can be taken care by HC alone. Lab-scale venturi-based hydrodynamic cavitation and wet air oxidation reactor with biomethanated distillery wastewater (BMDWW) as a model effluent was examined for establishing the proof-of-concept. The results indicated that for a desirable biodegradability index (BOD: COD - BI) enhancement (up to 0.4), the Cavitation (standalone) pretreatment condition was: 5 bar and 88 min reaction time with a COD reduction of 36 % and BI enhancement of up to 0.27 (initial BI - 0.17). The optimum WAO condition (standalone) was: 150oC, 6 bar and 30 minutes with 31% COD reduction and 0.33 BI. The hybrid pretreatment (combined Cavitation + WAO) worked out to be 23.18 min HC (at 5 bar) followed by 30 min WAO at 150oC, 6 bar, at which around 50% COD was retained yielding a BI of 0.55. FTIR & NMR analysis of pretreated effluent indicated dissociation and/or reorientation of complex organic compounds in untreated effluent to simpler organic compounds post-pretreatment.
Keywords: BI, hybrid, hydrodynamic cavitation, wet air oxidation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17571144 Numerical Investigation of Delamination in Carbon-Epoxy Composite using Arcan Specimen
Authors: M. Nikbakht, N. Choupani
Abstract:
In this paper delamination phenomenon in Carbon-Epoxy laminated composite material is investigated numerically. Arcan apparatus and specimen is modeled in ABAQUS finite element software for different loading conditions and crack geometries. The influence of variation of crack geometry on interlaminar fracture stress intensity factor and energy release rate for various mixed mode ratios and pure mode I and II was studied. Also, correction factors for this specimen for different crack length ratios were calculated. The finite element results indicate that for loading angles close to pure mode-II loading, a high ratio of mode-II to mode-I fracture is dominant and there is an opposite trend for loading angles close to pure mode-I loading. It confirms that by varying the loading angle of Arcan specimen pure mode-I, pure mode-II and a wide range of mixed-mode loading conditions can be created and tested. Also, numerical results confirm that the increase of the mode- II loading contribution leads to an increase of fracture resistance in the CF/PEI composite (i.e., a reduction in the total strain energy release rate) and the increase of the crack length leads to a reduction of interlaminar fracture resistance in the CF/PEI composite (i.e., an increase in the total interlaminar strain energy release rate).Keywords: Fracture Mechanics, Mixed Mode, Arcan Specimen, Finite Element.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19091143 Feature Reduction of Nearest Neighbor Classifiers using Genetic Algorithm
Authors: M. Analoui, M. Fadavi Amiri
Abstract:
The design of a pattern classifier includes an attempt to select, among a set of possible features, a minimum subset of weakly correlated features that better discriminate the pattern classes. This is usually a difficult task in practice, normally requiring the application of heuristic knowledge about the specific problem domain. The selection and quality of the features representing each pattern have a considerable bearing on the success of subsequent pattern classification. Feature extraction is the process of deriving new features from the original features in order to reduce the cost of feature measurement, increase classifier efficiency, and allow higher classification accuracy. Many current feature extraction techniques involve linear transformations of the original pattern vectors to new vectors of lower dimensionality. While this is useful for data visualization and increasing classification efficiency, it does not necessarily reduce the number of features that must be measured since each new feature may be a linear combination of all of the features in the original pattern vector. In this paper a new approach is presented to feature extraction in which feature selection, feature extraction, and classifier training are performed simultaneously using a genetic algorithm. In this approach each feature value is first normalized by a linear equation, then scaled by the associated weight prior to training, testing, and classification. A knn classifier is used to evaluate each set of feature weights. The genetic algorithm optimizes a vector of feature weights, which are used to scale the individual features in the original pattern vectors in either a linear or a nonlinear fashion. By this approach, the number of features used in classifying can be finely reduced.Keywords: Feature reduction, genetic algorithm, pattern classification, nearest neighbor rule classifiers (k-NNR).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17651142 Reduction of False Positives in Head-Shoulder Detection Based on Multi-Part Color Segmentation
Authors: Lae-Jeong Park
Abstract:
The paper presents a method that utilizes figure-ground color segmentation to extract effective global feature in terms of false positive reduction in the head-shoulder detection. Conventional detectors that rely on local features such as HOG due to real-time operation suffer from false positives. Color cue in an input image provides salient information on a global characteristic which is necessary to alleviate the false positives of the local feature based detectors. An effective approach that uses figure-ground color segmentation has been presented in an effort to reduce the false positives in object detection. In this paper, an extended version of the approach is presented that adopts separate multipart foregrounds instead of a single prior foreground and performs the figure-ground color segmentation with each of the foregrounds. The multipart foregrounds include the parts of the head-shoulder shape and additional auxiliary foregrounds being optimized by a search algorithm. A classifier is constructed with the feature that consists of a set of the multiple resulting segmentations. Experimental results show that the presented method can discriminate more false positive than the single prior shape-based classifier as well as detectors with the local features. The improvement is possible because the presented approach can reduce the false positives that have the same colors in the head and shoulder foregrounds.
Keywords: Pedestrian detection, color segmentation, false positives, feature extraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11441141 Incentive Policies to Promote Green Infrastructure in Urban Jordan
Authors: Zayed Freah Zeadat
Abstract:
The wellbeing of urban dwellers is strongly associated with the quality and quantity of green infrastructure. Nevertheless, urban green infrastructure is still lagging in many Arab cities, and Jordan is no exception. The capital city of Jordan, Amman, is becoming more urban dense with limited green spaces. The unplanned urban growth in Amman has caused several environmental problems such as urban heat islands, air pollution and lack of green spaces. This study aims to investigate the most suitable drivers to leverage the implementation of urban green infrastructure in Jordan through qualitative and quantitative analysis. The qualitative research includes an extensive literature review to discuss the most common drivers used internationally to promote urban green infrastructure implementation in the literature. The quantitative study employs a questionnaire survey to rank the suitability of each driver. Consultants, contractors and policymakers were invited to fill the research questionnaire according to their judgments and opinions. Relative Importance Index has been used to calculate the weighted average of all drivers and the Kruskal-Wallis test to check the degree of agreement among groups. This study finds that research participants agreed that indirect financial incentives (i.e., tax reductions, reduction in stormwater utility fee, reduction of interest rate, density bonus etc.) are the most effective incentive policy whilst granting sustainability certificate policy is the least effective driver to ensure widespread of UGI is elements in Jordan.
Keywords: sustainable development, urban green infrastructure, relative importance index, urban Jordan
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5641140 A Set Theory Based Factoring Technique and Its Use for Low Power Logic Design
Authors: Padmanabhan Balasubramanian, Ryuta Arisaka
Abstract:
Factoring Boolean functions is one of the basic operations in algorithmic logic synthesis. A novel algebraic factorization heuristic for single-output combinatorial logic functions is presented in this paper and is developed based on the set theory paradigm. The impact of factoring is analyzed mainly from a low power design perspective for standard cell based digital designs in this paper. The physical implementation of a number of MCNC/IWLS combinational benchmark functions and sub-functions are compared before and after factoring, based on a simple technology mapping procedure utilizing only standard gate primitives (readily available as standard cells in a technology library) and not cells corresponding to optimized complex logic. The power results were obtained at the gate-level by means of an industry-standard power analysis tool from Synopsys, targeting a 130nm (0.13μm) UMC CMOS library, for the typical case. The wire-loads were inserted automatically and the simulations were performed with maximum input activity. The gate-level simulations demonstrate the advantage of the proposed factoring technique in comparison with other existing methods from a low power perspective, for arbitrary examples. Though the benchmarks experimentation reports mixed results, the mean savings in total power and dynamic power for the factored solution over a non-factored solution were 6.11% and 5.85% respectively. In terms of leakage power, the average savings for the factored forms was significant to the tune of 23.48%. The factored solution is expected to better its non-factored counterpart in terms of the power-delay product as it is well-known that factoring, in general, yields a delay-efficient multi-level solution.
Keywords: Factorization, Set theory, Logic function, Standardcell based design, Low power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17911139 Treatment of Paper and Pulp Mill Effluent by Coagulation
Authors: Pradeep Kumar, Tjoon Tow Teng, Shri Chand, Kailas L. Wasewar
Abstract:
The pulp and paper mill effluent is one of the high polluting effluent amongst the effluents obtained from polluting industries. All the available methods for treatment of pulp and paper mill effluent have certain drawbacks. The coagulation is one of the cheapest process for treatment of various organic effluents. Thus, the removal of chemical oxygen demand (COD) and colour of paper mill effluent is studied using coagulation process. The batch coagulation process was performed using various coagulants like: aluminium chloride, poly aluminium chloride and copper sulphate. The initial pH of the effluent (Coagulation pH) has tremendous effect on COD and colour removal. Poly aluminium chloride (PAC) as coagulant reduced COD to 84 % and 92 % of colour was removed at an optimum pH 5 and coagulant dose of 8 ml l-1. With aluminium chloride at an optimum pH = 4 and coagulant dose of 5 g l-1, 74 % COD and 86 % colour removal were observed. The results using copper sulphate as coagulant (a less commercial coagulant) were encouraging. At an optimum pH 6 and mass loading of 5 g l-1, 76 % COD reduction and 78 % colour reduction were obtained. It was also observed that after addition of coagulant, the pH of the effluent decreases. The decrease in pH was highest for AlCl3, which was followed by PAC and CuSO4. Significant amount of COD reductions was obtained by coagulation process. Since the coagulation process is the first stage for treatment of effluent and some of the coagulant cations usually remain in the treated effluents. Thus, cation like copper may be one of the good catalyst for second stage of treatment process like wet oxidation. The copper has been found to be good oxidation catalyst then iron and aluminum.Keywords: Aluminium based coagulants, Coagulation, Copper, PAC, Pulp and paper mill effluent, Wastewater treatment
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 69961138 Effect of Phosphate Solubilization Microorganisms (PSM) and Plant Growth Promoting Rhizobacteria (PGPR) on Yield and Yield Components of Corn (Zea mays L.)
Authors: Mohammad Yazdani, Mohammad Ali Bahmanyar, Hemmatollah Pirdashti, Mohammad Ali Esmaili
Abstract:
In order to study the effect of phosphate solubilization microorganisms (PSM) and plant growth promoting rhizobacteria (PGPR) on yield and yield components of corn Zea mays (L. cv. SC604) an experiment was conducted at research farm of Sari Agricultural Sciences and Natural Resources University, Iran during 2007. Experiment laid out as split plot based on randomized complete block design with three replications. Three levels of manures (consisted of 20 Mg.ha-1 farmyard manure, 15 Mg.ha-1 green manure and check or without any manures) as main plots and eight levels of biofertilizers (consisted of 1-NPK or conventional fertilizer application; 2-NPK+PSM+PGPR; 3 NP50%K+PSM+PGPR; 4- N50%PK+PSM +PGPR; 5-N50%P50%K+PSM+ PGPR; 6-PK+PGPR; 7- NK+PSM and 8-PSM+PGPR) as sub plots were treatments. Results showed that farmyard manure application increased row number, ear weight, grain number per ear, grain yield, biological yield and harvest index compared to check. Furthermore, using of PSM and PGPR in addition to conventional fertilizer applications (NPK) could improve ear weight, row number and grain number per row and ultimately increased grain yield in green manure and check plots. According to results in all fertilizer treatments application of PSM and PGPR together could reduce P application by 50% without any significant reduction of grain yield. However, this treatment could not compensate 50% reduction of N application.Keywords: Biofertilizers, corn, PSM, PGPR, grain yield.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 66851137 Data Privacy and Safety with Large Language Models
Authors: Ashly Joseph, Jithu Paulose
Abstract:
Large language models (LLMs) have revolutionized natural language processing capabilities, enabling applications such as chatbots, dialogue agents, image, and video generators. Nevertheless, their trainings on extensive datasets comprising personal information poses notable privacy and safety hazards. This study examines methods for addressing these challenges, specifically focusing on approaches to enhance the security of LLM outputs, safeguard user privacy, and adhere to data protection rules. We explore several methods including post-processing detection algorithms, content filtering, reinforcement learning from human and AI inputs, and the difficulties in maintaining a balance between model safety and performance. The study also emphasizes the dangers of unintentional data leakage, privacy issues related to user prompts, and the possibility of data breaches. We highlight the significance of corporate data governance rules and optimal methods for engaging with chatbots. In addition, we analyze the development of data protection frameworks, evaluate the adherence of LLMs to General Data Protection Regulation (GDPR), and examine privacy legislation in academic and business policies. We demonstrate the difficulties and remedies involved in preserving data privacy and security in the age of sophisticated artificial intelligence by employing case studies and real-life instances. This article seeks to educate stakeholders on practical strategies for improving the security and privacy of LLMs, while also assuring their responsible and ethical implementation.
Keywords: Data privacy, large language models, artificial intelligence, machine learning, cybersecurity, general data protection regulation, data safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1041136 Formant Tracking Linear Prediction Model using HMMs for Noisy Speech Processing
Authors: Zaineb Ben Messaoud, Dorra Gargouri, Saida Zribi, Ahmed Ben Hamida
Abstract:
This paper presents a formant-tracking linear prediction (FTLP) model for speech processing in noise. The main focus of this work is the detection of formant trajectory based on Hidden Markov Models (HMM), for improved formant estimation in noise. The approach proposed in this paper provides a systematic framework for modelling and utilization of a time- sequence of peaks which satisfies continuity constraints on parameter; the within peaks are modelled by the LP parameters. The formant tracking LP model estimation is composed of three stages: (1) a pre-cleaning multi-band spectral subtraction stage to reduce the effect of residue noise on formants (2) estimation stage where an initial estimate of the LP model of speech for each frame is obtained (3) a formant classification using probability models of formants and Viterbi-decoders. The evaluation results for the estimation of the formant tracking LP model tested in Gaussian white noise background, demonstrate that the proposed combination of the initial noise reduction stage with formant tracking and LPC variable order analysis, results in a significant reduction in errors and distortions. The performance was evaluated with noisy natual vowels extracted from international french and English vocabulary speech signals at SNR value of 10dB. In each case, the estimated formants are compared to reference formants.Keywords: Formants Estimation, HMM, Multi Band Spectral Subtraction, Variable order LPC coding, White Gauusien Noise.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19621135 Effect of Jatropha curcas Leaf Extract on Castor Oil Induced Diarrhea in Albino Rats
Authors: Fatima U. Maigari, Musa Halilu, M. Maryam Umar, Rabiu Zainab
Abstract:
Plants as therapeutic agents are used as drug in many parts of the world. Medicinal plants are mostly used in developing countries due to culture acceptability, belief or due to lack of easy access to primary health care services. Jatropha curcas is a plant from the Euphorbiaceae family which is widely used in Northern Nigeria as an anti-diarrheal agent. This study was conducted to determine the anti-diarrheal effect of the leaf extract on castor oil induced diarrhea in albino rats. The leaves of J. curcas were collected from Balanga Local government in Gombe State, north-eastern Nigeria; due to its bioavailability. The leaves were air-dried at room temperature and ground to powder. Phytochemical screening was done and different concentrations of the extract was prepared and administered to the different categories of experimental animals. From the results, aqueous leaf extract of Jatropha curcas at doses of 200mg/Kg and 400mg/Kg was found to reduce the mean stool score as compared to control rats, however, maximum reduction was achieved with the standard drug of Loperamide (5mg/Kg). Treatment of diarrhea with 200mg/Kg of the extract did not produce any significant decrease in stool fluid content but was found to be significant in those rats that were treated with 400mg/Kg of the extract at 2hours (0.05±0.02) and 4hours (0.01±0.01). A significant reduction of diarrhea in the experimental animals signifies it to possess some anti-diarrheal activity.Keywords: Anti-diarrhea, Diarrhea, Jatropha curcas, Loperamide.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18581134 Performance of BLDC Motor under Kalman Filter Sensorless Drive
Authors: Yuri Boiko, Ci Lin, Iluju Kiringa, Tet Yeap
Abstract:
The performance of a permanent magnet brushless direct current (BLDC) motor controlled by the Kalman filter based position-sensorless drive is studied in terms of its dependence from the system’s parameters variations. The effects of the system’s parameters changes on the dynamic behavior of state variables are verified. Simulated is the closed loop control scheme with Kalman filter in the feedback line. Distinguished are two separate data sampling modes in analyzing feedback output from the BLDC motor: (1) equal angular separation and (2) equal time intervals. In case (1), the data are collected via equal intervals of rotor’s angular position i, i.e. keeping = const. In case (2), the data collection time points ti are separated by equal sampling time intervals t = const. Demonstrated are the effects of the parameters changes on the sensorless control flow, in particular, reduction of the instability torque ripples, switching spikes, and torque load balancing. It is specifically shown that an efficient suppression of commutation induced instability torque ripples is an achievable selection of the sampling rate in the Kalman filter settings above a certain critical value. The computational cost of such suppression is shown to be higher for the motors with lower induction values of the windings.
Keywords: BLDC motor, Kalman filter, sensorless drive, state variables, instability torque ripples reduction, sampling rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7291133 Vibration Transmission across Junctions of Walls and Floors in an Apartment Building: An Experimental Investigation
Authors: Hugo Sampaio Libero, Max de Castro Magalhaes
Abstract:
The perception of sound radiated from a building floor is greatly influenced by the rooms in which it is immersed and by the position of both listener and source. The main question that remains unanswered is related to the influence of the source position on the sound power radiated by a complex wall-floor system in buildings. This research is concerned with the investigation of vibration transmission across walls and floors in buildings. It is primarily based on the determination of vibration reduction index via experimental tests. Knowledge of this parameter may help in predicting noise and vibration propagation in building components. First, the physical mechanisms involving vibration transmission across structural junctions is described. An experimental set-up is performed to aid this investigation. The experimental tests have showed that the vibration generation in the walls and floors are directed related to their size and boundary conditions. It is also shown that the vibration source position can affect the overall vibration spectrum significantly. Second, the characteristics of the noise spectra inside the rooms due to an impact source (tapping machine) are also presented. Conclusions are drawn for the general trend of vibration and noise spectrum of the structural components and rooms respectively. In summary, the aim of this paper is to investigate the vibro-acoustical behavior of building floors and walls under floor impact excitation. The impact excitation was at distinct positions on the slab. The analysis has highlighted the main physical characteristics of the vibration transmission mechanism.
Keywords: Vibration transmission, Vibration Reduction Index, Impact excitation, building acoustics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3331132 Cost Valuation Method for Development Concurrent Phase Appropriate Requirement Valuation Using the Example of Load Carrier Development in the Lithium-Ion-Battery Production
Authors: Achim Kampker, Christoph Deutskens, Heiner Hans Heimes, Mathias Ordung, Felix Optehostert
Abstract:
In the past years electric mobility became part of a public discussion. The trend to fully electrified vehicles instead of vehicles fueled with fossil energy has notably gained momentum. Today nearly every big car manufacturer produces and sells fully electrified vehicles, but electrified vehicles are still not as competitive as conventional powered vehicles. As the traction battery states the largest cost driver, lowering its price is a crucial objective. In addition to improvements in product and production processes a nonnegligible, but widely underestimated cost driver of production can be found in logistics, since the production technology is not continuous yet and neither are the logistics systems. This paper presents an approach to evaluate cost factors on different designs of load carrier systems. Due to numerous interdependencies, the combination of costs factors for a particular scenario is not transparent. This is effecting actions for cost reduction negatively, but still cost reduction is one of the major goals for simultaneous engineering processes. Therefore a concurrent and phase appropriate cost valuation method is necessary to serve cost transparency. In this paper the four phases of this cost valuation method are defined and explained, which based upon a new approach integrating the logistics development process in to the integrated product and process development.Keywords: Research and development, technology and Innovation, lithium-ion-battery production, load carrier development process, cost valuation method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22771131 Simulation of Lean Principles Impact in a Multi-Product Supply Chain
Authors: M. Rossini, A. Portioli Studacher
Abstract:
The market competition is moving from the single firm to the whole supply chain because of increasing competition and growing need for operational efficiencies and customer orientation. Supply chain management allows companies to look beyond their organizational boundaries to develop and leverage resources and capabilities of their supply chain partners. This creates competitive advantages in the marketplace and because of this SCM has acquired strategic importance. Lean Approach is a management strategy that focuses on reducing every type of waste present in an organization. This approach is becoming more and more popular among supply chain managers. The supply chain application of lean approach is not frequent. In particular, it is not well studied which are the impacts of lean approach principles in a supply chain context. In literature there are only few studies aimed at understanding the qualitative impact of the lean approach in supply chains. Therefore, the goal of this research work is to study the impacts of lean principles implementation along a supply chain. To achieve this, a simulation model of a threeechelon multi-product supply chain has been built. Kanban system (and several priority policies) and setup time reduction degrees are implemented in the lean-configured supply chain to apply pull and lot-sizing decrease principles respectively. To evaluate the benefits of lean approach, lean supply chain is compared with an EOQ-configured supply chain. The simulation results show that Kanban system and setup-time reduction improve inventory stock level. They also show that logistics efforts are affected to lean implementation degree. The paper concludes describing performances of lean supply chain in different contexts.Keywords: Inventory policy, Kanban, lean supply chain, simulation study, supply chain management, planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21291130 The Coverage of the Object-Oriented Framework Application Class-Based Test Cases
Authors: Jehad Al Dallal, Paul Sorenson
Abstract:
An application framework provides a reusable design and implementation for a family of software systems. Frameworks are introduced to reduce the cost of a product line (i.e., family of products that share the common features). Software testing is a time consuming and costly ongoing activity during the application software development process. Generating reusable test cases for the framework applications at the framework development stage, and providing and using the test cases to test part of the framework application whenever the framework is used reduces the application development time and cost considerably. Framework Interface Classes (FICs) are classes introduced by the framework hooks to be implemented at the application development stage. They can have reusable test cases generated at the framework development stage and provided with the framework to test the implementations of the FICs at the application development stage. In this paper, we conduct a case study using thirteen applications developed using three frameworks; one domain oriented and two application oriented. The results show that, in general, the percentage of the number of FICs in the applications developed using domain frameworks is, on average, greater than the percentage of the number of FICs in the applications developed using application frameworks. Consequently, the reduction of the application unit testing time using the reusable test cases generated for domain frameworks is, in general, greater than the reduction of the application unit testing time using the reusable test cases generated for application frameworks.Keywords: FICs, object-oriented framework, object-orientedframework application, software testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14471129 Geosynthetic Reinforced Unpaved Road: Literature Study and Design Example
Authors: D. Jayalakshmi, S. Bhosale
Abstract:
This paper, in its first part, presents the state-of-the-art literature of design approaches for geosynthetic reinforced unpaved roads. The literature starting since 1970 and the critical appraisal of flexible pavement design by Giroud and Han (2004) and Jonathan Fannin (2006) is presented. The design example is illustrated for Indian conditions. The example emphasizes the results computed by Giroud and Han's (2004) design method with the Indian road congress guidelines by IRC SP 72 -2015. The input data considered are related to the subgrade soil condition of Maharashtra State in India. The unified soil classification of the subgrade soil is inorganic clay with high plasticity (CH), which is expansive with a California bearing ratio (CBR) of 2% to 3%. The example exhibits the unreinforced case and geotextile as reinforcement by varying the rut depth from 25 mm to 100 mm. The present result reveals the base thickness for the unreinforced case from the IRC design catalogs is in good agreement with Giroud and Han (2004) approach for a range of 75 mm to 100 mm rut depth. Since Giroud and Han (2004) method is applicable for both reinforced and unreinforced cases, for the same data with appropriate Nc factor, for the same rut depth, the base thickness for the reinforced case has arrived for the Indian condition. From this trial, for the CBR of 2%, the base thickness reduction due to geotextile inclusion is 35%. For the CBR range of 2% to 5% with different stiffness in geosynthetics, the reduction in base course thickness will be evaluated, and the validation will be executed by the full-scale accelerated pavement testing set up at the College of Engineering Pune (COE), India.
Keywords: Base thickness, design approach, equation, full scale accelerated pavement set up, Indian condition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6551128 Energy Consumption, Emission Absorption and Carbon Emission Reduction on Semarang State University Campus
Authors: Dewi Liesnoor Setyowati, Puji Hardati, Tri Marhaeni Puji Astuti, Muhammad Amin
Abstract:
Universitas Negeri Semarang (UNNES) is a university with a vision of conservation. The impact of the UNNES conservation is the existence of a positive response from the community for the effort of greening the campus and the planting of conservation value in the academic community. But in reality, energy consumption in UNNES campus tends to increase. The objectives of the study were to analyze the energy consumption in the campus area, to analyze the absorption of emissions by trees and the awareness of UNNES citizens in reducing emissions. Research focuses on energy consumption, carbon emissions, and awareness of citizens in reducing emissions. Research subjects in this study are UNNES citizens (lecturers, students and employees). The research area covers 6 faculties and one administrative center building. Data collection is done by observation, interview and documentation. The research used a quantitative descriptive method to analyze the data. The number of trees in UNNES is 10,264. Total emission on campus UNNES is 7.862.281.56 kg/year, the tree absorption is 6,289,250.38 kg/year. In UNNES campus area there are still 1,575,031.18 kg/year of emissions, not yet absorbed by trees. There are only two areas of the faculty whose trees are capable of absorbing emissions. The awareness of UNNES citizens in reducing energy consumption is seen in change the habit of: using energy-saving equipment (65%); reduce energy consumption per unit (68%); do energy literacy for UNNES citizens (74%). UNNES leaders always provide motivation to the citizens of UNNES, to reduce and change patterns of energy consumption.
Keywords: Energy consumption, carbon emission absorption, emission reduction, energy literation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8441127 Machine Learning Facing Behavioral Noise Problem in an Imbalanced Data Using One Side Behavioral Noise Reduction: Application to a Fraud Detection
Authors: Salma El Hajjami, Jamal Malki, Alain Bouju, Mohammed Berrada
Abstract:
With the expansion of machine learning and data mining in the context of Big Data analytics, the common problem that affects data is class imbalance. It refers to an imbalanced distribution of instances belonging to each class. This problem is present in many real world applications such as fraud detection, network intrusion detection, medical diagnostics, etc. In these cases, data instances labeled negatively are significantly more numerous than the instances labeled positively. When this difference is too large, the learning system may face difficulty when tackling this problem, since it is initially designed to work in relatively balanced class distribution scenarios. Another important problem, which usually accompanies these imbalanced data, is the overlapping instances between the two classes. It is commonly referred to as noise or overlapping data. In this article, we propose an approach called: One Side Behavioral Noise Reduction (OSBNR). This approach presents a way to deal with the problem of class imbalance in the presence of a high noise level. OSBNR is based on two steps. Firstly, a cluster analysis is applied to groups similar instances from the minority class into several behavior clusters. Secondly, we select and eliminate the instances of the majority class, considered as behavioral noise, which overlap with behavior clusters of the minority class. The results of experiments carried out on a representative public dataset confirm that the proposed approach is efficient for the treatment of class imbalances in the presence of noise.Keywords: Machine learning, Imbalanced data, Data mining, Big data.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11371126 Phyllantus niruri Protects against Fe2+ and SNP Induced Oxidative Damage in Mitochondrial Enriched Fractions of Rats Brain
Authors: Olusola Olalekan Elekofehinti, Isaac Gbadura Adanlawo, Joao Batista Teixeira Rocha
Abstract:
The potential neuroprotective effect of Phyllantus nuriri against Fe2+ and sodium nitroprusside (SNP) induced oxidative stress in mitochondria of rats brain was evaluated. Cellular viability was assessed by MTT reduction, reactive oxygen species (ROS) generation was measured using the probe 2,7-dichlorofluoresce indiacetate (DCFH-DA). Glutathione content was measured using dithionitrobenzoic acid (DTNB). Fe2+ (10μM) and SNP (5μM) significantly decreased mitochondrial activity, assessed by MTT reduction assay, in a dose-dependent manner, this occurred in parallel with increased glutathione oxidation, ROS production and lipid peroxidation end-products (thiobarbituric acid reactive substances, TBARS). The co-incubation with methanolic extract of Phyllantus nuriri (10-200 μg/ml) reduced the disruption of mitochondrial activity, gluthathione oxidation, ROS production as well as the increase in TBARS levels caused by both Fe2+ and SNP in a dose dependent manner. HPLC analysis of the extract revealed the presence of gallic acid (20.540.01), caffeic acid (7.930.02), rutin (25.310.05), quercetin (31.280.03) and kaemferol (14.360.01). This result suggests that these phytochemicals account for the protective actions of P. niruri against Fe2+ and SNP -induced oxidative stress. Our results show that P. nuriri consist important bioactive molecules in the search for an improved therapy against the deleterious effects of Fe2+, an intrinsic producer of reactive oxygen species (ROS), that leads to neuronal oxidative stress and neurodegeneration.Keywords: Phyllantus niruri, mitochondria, antioxidant, oxidative stress, synaptosome.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17451125 Effects of High-Protein, Low-Energy Diet on Body Composition in Overweight and Obese Adults: A Clinical Trial
Authors: Makan Cheraghpour, Seyed Ahmad Hosseini, Damoon Ashtary-Larky, Saeed Shirali, Matin Ghanavati, Meysam Alipour
Abstract:
Background: In addition to reducing body weight, the low-calorie diets can reduce the lean body mass. It is hypothesized that in addition to reducing the body weight, the low-calorie diets can maintain the lean body mass. So, the current study aimed at evaluating the effects of high-protein diet with calorie restriction on body composition in overweight and obese individuals. Methods: 36 obese and overweight subjects were divided randomly into two groups. The first group received a normal-protein, low-energy diet (RDA), and the second group received a high-protein, low-energy diet (2×RDA). The anthropometric indices including height, weight, body mass index, body fat mass, fat free mass, and body fat percentage were evaluated before and after the study. Results: A significant reduction was observed in anthropometric indices in both groups (high-protein, low-energy diets and normal-protein, low-energy diets). In addition, more reduction in fat free mass was observed in the normal-protein, low-energy diet group compared to the high -protein, low-energy diet group. In other the anthropometric indices, significant differences were not observed between the two groups. Conclusion: Independently of the type of diet, low-calorie diet can improve the anthropometric indices, but during a weight loss, high-protein diet can help the fat free mass to be maintained.
Keywords: Diet, high-protein, body mass index, body fat percentage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12741124 Very Large Scale Integration Architecture of Finite Impulse Response Filter Implementation Using Retiming Technique
Authors: S. Jalaja, A. M. Vijaya Prakash
Abstract:
Recursive combination of an algorithm based on Karatsuba multiplication is exploited to design a generalized transpose and parallel Finite Impulse Response (FIR) Filter. Mid-range Karatsuba multiplication and Carry Save adder based on Karatsuba multiplication reduce time complexity for higher order multiplication implemented up to n-bit. As a result, we design modified N-tap Transpose and Parallel Symmetric FIR Filter Structure using Karatsuba algorithm. The mathematical formulation of the FFA Filter is derived. The proposed architecture involves significantly less area delay product (APD) then the existing block implementation. By adopting retiming technique, hardware cost is reduced further. The filter architecture is designed by using 90 nm technology library and is implemented by using cadence EDA Tool. The synthesized result shows better performance for different word length and block size. The design achieves switching activity reduction and low power consumption by applying with and without retiming for different combination of the circuit. The proposed structure achieves more than a half of the power reduction by adopting with and without retiming techniques compared to the earlier design structure. As a proof of the concept for block size 16 and filter length 64 for CKA method, it achieves a 51% as well as 70% less power by applying retiming technique, and for CSA method it achieves a 57% as well as 77% less power by applying retiming technique compared to the previously proposed design.Keywords: Carry save adder Karatsuba multiplication, mid-range Karatsuba multiplication, modified FFA, transposed filter, retiming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9101123 Reducing Defects through Organizational Learning within a Housing Association Environment
Authors: T. Hopkin, S. Lu, P. Rogers, M. Sexton
Abstract:
Housing Associations (HAs) contribute circa 20% of the UK’s housing supply. HAs are however under increasing pressure as a result of funding cuts and rent reductions. Due to the increased pressure, a number of processes are currently being reviewed by HAs, especially how they manage and learn from defects. Learning from defects is considered a useful approach to achieving defect reduction within the UK housebuilding industry. This paper contributes to our understanding of how HAs learn from defects by undertaking an initial round table discussion with key HA stakeholders as part of an ongoing collaborative research project with the National House Building Council (NHBC) to better understand how house builders and HAs learn from defects to reduce their prevalence. The initial discussion shows that defect information runs through a number of groups, both internal and external of a HA during both the defects management process and organizational learning (OL) process. Furthermore, HAs are reliant on capturing and recording defect data as the foundation for the OL process. During the OL process defect data analysis is the primary enabler to recognizing a need for a change to organizational routines. When a need for change has been recognized, new options are typically pursued to design out defects via updates to a HAs Employer’s Requirements. Proposed solutions are selected by a review board and committed to organizational routine. After implementing a change, both structured and unstructured feedback is sought to establish the change’s success. The findings from the HA discussion demonstrates that OL can achieve defect reduction within the house building sector in the UK. The paper concludes by outlining a potential ‘learning from defects model’ for the housebuilding industry as well as describing future work.
Keywords: Defects, new homes, housing associations, organizational learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1897