Search results for: biocontrol methods
4027 Direct Displacement-Based Design Procedure for Performance-Based Seismic Design of Structures
Authors: Haleh Hamidpour
Abstract:
Since the seismic damageability of structures is controlled by the inelastic deformation capacities of structural elements, seismic design of structure based on force analogy methods is not appropriate. In recent year, the basic approach of design codes have been changed from force-based approach to displacement-based. In this regard, a Direct Displacement-Based Design (DDBD) and a Performance-Based Plastic Design (PBPD) method are proposed. In this study, the efficiency of these two methods on seismic performance of structures is evaluated through a sample 12-story reinforced concrete moment frame. The building is designed separately based on the DDBD and the PBPD methods. Once again the structure is designed by the traditional force analogy method according to the FEMA P695 regulation. Different design method results in different structural elements. Seismic performance of these three structures is evaluated through nonlinear static and nonlinear dynamic analysis. The results show that the displacement-based design methods accommodate the intended performance objectives better than the traditional force analogy method.Keywords: direct performance-based design, ductility demands, inelastic seismic performance, yield mechanism
Procedia PDF Downloads 3334026 Corrosion of Concrete Reinforcing Steel Bars Tested and Compared Between Various Protection Methods
Authors: P. van Tonder, U. Bagdadi, B. M. D. Lario, Z. Masina, T. R. Motshwari
Abstract:
This paper analyses how concrete reinforcing steel bars corrode and how it can be minimised through the use of various protection methods against corrosion, such as metal-based paint, alloying, cathodic protection and electroplating. Samples of carbon steel bars were protected, using these four methods. Tests performed on the samples included durability, electrical resistivity and bond strength. Durability results indicated relatively low corrosion rates for alloying, cathodic protection, electroplating and metal-based paint. The resistivity results indicate all samples experienced a downward trend, despite erratic fluctuations in the data, indicating an inverse relationship between electrical resistivity and corrosion rate. The results indicated lowered bond strengths when the reinforced concrete was cured in seawater compared to being cured in normal water. It also showed that higher design compressive strengths lead to higher bond strengths which can be used to compensate for the loss of bond strength due to corrosion in a real-world application. In terms of implications, all protection methods have the potential to be effective at resisting corrosion in real-world applications, especially the alloying, cathodic protection and electroplating methods. The metal-based paint underperformed by comparison, most likely due to the nature of paint in general which can fade and chip away, revealing the steel samples and exposing them to corrosion. For alloying, stainless steel is the suggested material of choice, where Y-bars are highly recommended as smooth bars have a much-lowered bond strength. Cathodic protection performed the best of all in protecting the sample from corrosion, however, its real-world application would require significant evaluation into the feasibility of such a method.Keywords: protection methods, corrosion, concrete, reinforcing steel bars
Procedia PDF Downloads 1734025 On the Solution of Fractional-Order Dynamical Systems Endowed with Block Hybrid Methods
Authors: Kizito Ugochukwu Nwajeri
Abstract:
This paper presents a distinct approach to solving fractional dynamical systems using hybrid block methods (HBMs). Fractional calculus extends the concept of derivatives and integrals to non-integer orders and finds increasing application in fields such as physics, engineering, and finance. However, traditional numerical techniques often struggle to accurately capture the complex behaviors exhibited by these systems. To address this challenge, we develop HBMs that integrate single-step and multi-step methods, enabling the simultaneous computation of multiple solution points while maintaining high accuracy. Our approach employs polynomial interpolation and collocation techniques to derive a system of equations that effectively models the dynamics of fractional systems. We also directly incorporate boundary and initial conditions into the formulation, enhancing the stability and convergence properties of the numerical solution. An adaptive step-size mechanism is introduced to optimize performance based on the local behavior of the solution. Extensive numerical simulations are conducted to evaluate the proposed methods, demonstrating significant improvements in accuracy and efficiency compared to traditional numerical approaches. The results indicate that our hybrid block methods are robust and versatile, making them suitable for a wide range of applications involving fractional dynamical systems. This work contributes to the existing literature by providing an effective numerical framework for analyzing complex behaviors in fractional systems, thereby opening new avenues for research and practical implementation across various disciplines.Keywords: fractional calculus, numerical simulation, stability and convergence, Adaptive step-size mechanism, collocation methods
Procedia PDF Downloads 434024 Framework for Integrating Big Data and Thick Data: Understanding Customers Better
Authors: Nikita Valluri, Vatcharaporn Esichaikul
Abstract:
With the popularity of data-driven decision making on the rise, this study focuses on providing an alternative outlook towards the process of decision-making. Combining quantitative and qualitative methods rooted in the social sciences, an integrated framework is presented with a focus on delivering a much more robust and efficient approach towards the concept of data-driven decision-making with respect to not only Big data but also 'Thick data', a new form of qualitative data. In support of this, an example from the retail sector has been illustrated where the framework is put into action to yield insights and leverage business intelligence. An interpretive approach to analyze findings from both kinds of quantitative and qualitative data has been used to glean insights. Using traditional Point-of-sale data as well as an understanding of customer psychographics and preferences, techniques of data mining along with qualitative methods (such as grounded theory, ethnomethodology, etc.) are applied. This study’s final goal is to establish the framework as a basis for providing a holistic solution encompassing both the Big and Thick aspects of any business need. The proposed framework is a modified enhancement in lieu of traditional data-driven decision-making approach, which is mainly dependent on quantitative data for decision-making.Keywords: big data, customer behavior, customer experience, data mining, qualitative methods, quantitative methods, thick data
Procedia PDF Downloads 1624023 Generation of Quasi-Measurement Data for On-Line Process Data Analysis
Authors: Hyun-Woo Cho
Abstract:
For ensuring the safety of a manufacturing process one should quickly identify an assignable cause of a fault in an on-line basis. To this end, many statistical techniques including linear and nonlinear methods have been frequently utilized. However, such methods possessed a major problem of small sample size, which is mostly attributed to the characteristics of empirical models used for reference models. This work presents a new method to overcome the insufficiency of measurement data in the monitoring and diagnosis tasks. Some quasi-measurement data are generated from existing data based on the two indices of similarity and importance. The performance of the method is demonstrated using a real data set. The results turn out that the presented methods are able to handle the insufficiency problem successfully. In addition, it is shown to be quite efficient in terms of computational speed and memory usage, and thus on-line implementation of the method is straightforward for monitoring and diagnosis purposes.Keywords: data analysis, diagnosis, monitoring, process data, quality control
Procedia PDF Downloads 4814022 Implant Guided Surgery and Immediate Loading
Authors: Omid Tavakol, Mahnaz Gholami
Abstract:
Introduction : In this oral presentation the main goal is discussing immediate loading in dental implants , from treatment planning and surgical guide designing to delivery , follow up and occlusal consideration . Methods and materials : first of all systematic reviews about immediate loading will be considered . besides , a comparison will be made between immediate loading and conventional loading in terms of success rate and complications . After that different methods , prosthetic options and materials best used in immediate loading will be explained. Particularly multi unit abutments and their mechanism of function will be explained .Digital impressions and designing the temporaries is the next topic we are to explicate .Next issue is the differences between single unit , multiple unit and full arch implantation in immediate loading .Following we are going to describe methods for tissue engineering and papilla formation after extraction . Last slides are about a full mouth rehabilitation via immediate loading technique from surgical designing to follow up .At the end we would talk about potential complications , how to prevent from occurrence and what to do if we face up with .Keywords: guided surgery, digital implantology, immediate loading, digital dentistry
Procedia PDF Downloads 444021 Predicting Consolidation Coefficient of Busan Clay by Time-Displacement-Velocity Methods
Authors: Thang Minh Le, Hadi Khabbaz
Abstract:
The coefficient of consolidation is a parameter governing the rate at which saturated soil particularly clay undergoes consolidation when subjected to an increase in pressure. The rate and amount of compression in soil varies with the rate that pore water is lost; and hence depends on soil permeability. Over many years, various methods have been proposed to determine the coefficient of consolidation, cv, which is an indication of the rate of foundation settlement on soft ground. However, defining this parameter is often problematic and heavily relies on graphical techniques, which are subject to some uncertainties. This paper initially presents an overview of many well-established methods to determine the vertical coefficient of consolidation from the incremental loading consolidation tests. An array of consolidation tests was conducted on the undisturbed clay samples, collected at various depths from a site in Nakdong river delta, Busan, South Korea. The consolidation test results on these soft sensitive clay samples were employed to evaluate the targeted methods to predict the settlement rate of Busan clay. In relationship of time-displacement-velocity, a total of 3 method groups from 10 common procedures were classified and compared together. Discussions on study results will be also provided.Keywords: Busan clay, coefficient of consolidation, constant rate of strain, incremental loading
Procedia PDF Downloads 1864020 A Comparison between Russian and Western Approach for Deep Foundation Design
Authors: Saeed Delara, Kendra MacKay
Abstract:
Varying methodologies are considered for pile design for both Russian and Western approaches. Although both approaches rely on toe and side frictional resistances, different calculation methods are proposed to estimate pile capacity. The Western approach relies on compactness (internal friction angle) of soil for cohesionless soils and undrained shear strength for cohesive soils. The Russian approach relies on grain size for cohesionless soils and liquidity index for cohesive soils. Though most recommended methods in the Western approaches are relatively simple methods to predict pile settlement, the Russian approach provides a detailed method to estimate single pile and pile group settlement. Details to calculate pile axial capacity and settlement using the Russian and Western approaches are discussed and compared against field test results.Keywords: pile capacity, pile settlement, Russian approach, western approach
Procedia PDF Downloads 1664019 Control of a Quadcopter Using Genetic Algorithm Methods
Authors: Mostafa Mjahed
Abstract:
This paper concerns the control of a nonlinear system using two different methods, reference model and genetic algorithm. The quadcopter is a nonlinear unstable system, which is a part of aerial robots. It is constituted by four rotors placed at the end of a cross. The center of this cross is occupied by the control circuit. Its motions are governed by six degrees of freedom: three rotations around 3 axes (roll, pitch and yaw) and the three spatial translations. The control of such system is complex, because of nonlinearity of its dynamic representation and the number of parameters, which it involves. Numerous studies have been developed to model and stabilize such systems. The classical PID and LQ correction methods are widely used. If the latter represent the advantage to be simple because they are linear, they reveal the drawback to require the presence of a linear model to synthesize. It also implies the complexity of the established laws of command because the latter must be widened on all the domain of flight of these quadcopter. Note that, if the classical design methods are widely used to control aeronautical systems, the Artificial Intelligence methods as genetic algorithms technique receives little attention. In this paper, we suggest comparing two PID design methods. Firstly, the parameters of the PID are calculated according to the reference model. In a second phase, these parameters are established using genetic algorithms. By reference model, we mean that the corrected system behaves according to a reference system, imposed by some specifications: settling time, zero overshoot etc. Inspired from the natural evolution of Darwin's theory advocating the survival of the best, John Holland developed this evolutionary algorithm. Genetic algorithm (GA) possesses three basic operators: selection, crossover and mutation. We start iterations with an initial population. Each member of this population is evaluated through a fitness function. Our purpose is to correct the behavior of the quadcopter around three axes (roll, pitch and yaw) with 3 PD controllers. For the altitude, we adopt a PID controller.Keywords: quadcopter, genetic algorithm, PID, fitness, model, control, nonlinear system
Procedia PDF Downloads 4314018 Feature Evaluation Based on Random Subspace and Multiple-K Ensemble
Authors: Jaehong Yu, Seoung Bum Kim
Abstract:
Clustering analysis can facilitate the extraction of intrinsic patterns in a dataset and reveal its natural groupings without requiring class information. For effective clustering analysis in high dimensional datasets, unsupervised dimensionality reduction is an important task. Unsupervised dimensionality reduction can generally be achieved by feature extraction or feature selection. In many situations, feature selection methods are more appropriate than feature extraction methods because of their clear interpretation with respect to the original features. The unsupervised feature selection can be categorized as feature subset selection and feature ranking method, and we focused on unsupervised feature ranking methods which evaluate the features based on their importance scores. Recently, several unsupervised feature ranking methods were developed based on ensemble approaches to achieve their higher accuracy and stability. However, most of the ensemble-based feature ranking methods require the true number of clusters. Furthermore, these algorithms evaluate the feature importance depending on the ensemble clustering solution, and they produce undesirable evaluation results if the clustering solutions are inaccurate. To address these limitations, we proposed an ensemble-based feature ranking method with random subspace and multiple-k ensemble (FRRM). The proposed FRRM algorithm evaluates the importance of each feature with the random subspace ensemble, and all evaluation results are combined with the ensemble importance scores. Moreover, FRRM does not require the determination of the true number of clusters in advance through the use of the multiple-k ensemble idea. Experiments on various benchmark datasets were conducted to examine the properties of the proposed FRRM algorithm and to compare its performance with that of existing feature ranking methods. The experimental results demonstrated that the proposed FRRM outperformed the competitors.Keywords: clustering analysis, multiple-k ensemble, random subspace-based feature evaluation, unsupervised feature ranking
Procedia PDF Downloads 3394017 Dialogue Meetings as an Arena for Collaboration and Reflection among Researchers and Practitioners
Authors: Kerstin Grunden, Ann Svensson, Berit Forsman, Christina Karlsson, Ayman Obeid
Abstract:
The research question of the article is to explore whether the dialogue meetings method could be relevant for reflective learning among researchers and practitioners when welfare technology should be implemented in municipalities, or not. A testbed was planned to be implemented in a retirement home in a Swedish municipality, and the practitioners worked with a pre-study of that testbed. In the article, the dialogue between the researchers and the practitioners in the dialogue meetings is described and analyzed. The potential of dialogue meetings as an arena for learning and reflection among researchers and practitioners is discussed. The research methodology approach is participatory action research with mixed methods (dialogue meetings, focus groups, participant observations). The main findings from the dialogue meetings were that the researchers learned more about the use of traditional research methods, and the practitioners learned more about how they could improve their use of the methods to facilitate change processes in their organization. These findings have the potential both for the researchers and the practitioners to result in more relevant use of research methods in change processes in organizations. It is concluded that dialogue meetings could be relevant for reflective learning among researchers and practitioners when welfare technology should be implemented in a health care organization.Keywords: dialogue meetings, implementation, reflection, test bed, welfare technology, participatory action research
Procedia PDF Downloads 1464016 Protein Remote Homology Detection and Fold Recognition by Combining Profiles with Kernel Methods
Authors: Bin Liu
Abstract:
Protein remote homology detection and fold recognition are two most important tasks in protein sequence analysis, which is critical for protein structure and function studies. In this study, we combined the profile-based features with various string kernels, and constructed several computational predictors for protein remote homology detection and fold recognition. Experimental results on two widely used benchmark datasets showed that these methods outperformed the competing methods, indicating that these predictors are useful computational tools for protein sequence analysis. By analyzing the discriminative features of the training models, some interesting patterns were discovered, reflecting the characteristics of protein superfamilies and folds, which are important for the researchers who are interested in finding the patterns of protein folds.Keywords: protein remote homology detection, protein fold recognition, profile-based features, Support Vector Machines (SVMs)
Procedia PDF Downloads 1614015 A Comparison Study of Different Methods Used in the Detection of Giardia lamblia on Fecal Specimen of Children
Authors: Muhammad Farooq Baig
Abstract:
Objective: The purpose of this study was to compare results obtained using a single fecal specimen for O&P examination, direct immunofluorescence assay (DFA), and two conventional staining methods. Design: Hundred and fifty children fecal specimens were collected and examined by each method. The O&P and the DFA were used as the reference method. Setting: The study was performed at the laboratory in the Basic Medical Science Institute JPMC Karachi. Patients or Other Participants: The fecal specimens were collected from children with a suspected Giardia lamblia infection. Main Outcome Measures: The amount of agreement and disagreement between methods.1) Presence of giardiasis in our population. 2) The sensitivity and specificity of each method. Results: There was 45(30%) positive 105 (70%) negative on DFA, 41 (27.4%) positive 109 (72.6%) negative on iodine and 34 (22.6%) positive 116(77.4%) on saline method. The sensitivity and specificity of DFA in comparision to iodine were 92.2%, 92.7% respectively. The sensitivity and specificity of DFA in comparisoin to saline method were 91.2%, 87.9% respectively. The sensitivity of iodine method and saline method in compariosn to DFA were 82.2%, 68.8% respectively. There is mark diffrence in sensitivity of DFA to conventional method. Conclusion: The study supported findings of other investigators who concluded that DFA method have the greater sensitivity. The immunologic methods were more efficient and quicker than the conventional O&P method.Keywords: direct immunofluorescence assay (DFA), ova and parasite (O&P), Giardia lamblia, children, medical science
Procedia PDF Downloads 4234014 Evaluation of Simple, Effective and Affordable Processing Methods to Reduce Phytates in the Legume Seeds Used for Feed Formulations
Authors: N. A. Masevhe, M. Nemukula, S. S. Gololo, K. G. Kgosana
Abstract:
Background and Study Significance: Legume seeds are important in agriculture as they are used for feed formulations due to their nutrient-dense, low-cost, and easy accessibility. Although they are important sources of energy, proteins, carbohydrates, vitamins, and minerals, they contain abundant quantities of anti-nutritive factors that reduce the bioavailability of nutrients, digestibility of proteins, and mineral absorption in livestock. However, the removal of these factors is too costly as it requires expensive state-of-the-art techniques such as high pressure and thermal processing. Basic Methodologies: The aim of the study was to investigate cost-effective methods that can be used to reduce the inherent phytates as putative antinutrients in the legume seeds. The seeds of Arachis hypogaea, Pisum sativum and Vigna radiata L. were subjected to the single processing methods viz raw seeds plus dehulling (R+D), soaking plus dehulling (S+D), ordinary cooking plus dehulling (C+D), infusion plus dehulling (I+D), autoclave plus dehulling (A+D), microwave plus dehulling (M+D) and five combined methods (S+I+D; S+A+D; I+M+D; S+C+D; S+M+D). All the processed seeds were dried, ground into powder, extracted, and analyzed on a microplate reader to determine the percentage of phytates per dry mass of the legume seeds. Phytic acid was used as a positive control, and one-way ANOVA was used to determine the significant differences between the means of the processing methods at a threshold of 0.05. Major Findings: The results of the processing methods showed the percentage yield ranges of 39.1-96%, 67.4-88.8%, and 70.2-93.8% for V. radiata, A. hypogaea and P. sativum, respectively. Though the raw seeds contained the highest contents of phytates that ranged between 0.508 and 0.527%, as expected, the R+D resulted in a slightly lower phytate percentage range of 0.469-0.485%, while other processing methods resulted in phytate contents that were below 0.35%. The M+D and S+M+D methods showed low phytate percentage ranges of 0.276-0.296% and 0.272-0.294%, respectively, where the lowest percentage yield was determined in S+M+D of P. sativum. Furthermore, these results were found to be significantly different (p<0.05). Though phytates cause micronutrient deficits as they chelate important minerals such as calcium, zinc, iron, and magnesium, their reduction may enhance nutrient bioavailability since they cannot be digested by the ruminants. Concluding Statement: Despite the nutritive aspects of the processed legume seeds, which are still in progress, the M+D and S+M+D methods, which significantly reduced the phytates in the investigated legume seeds, may be recommended to the local farmers and feed-producing industries so as to enhance animal health and production at an affordable cost.Keywords: anti-nutritive factors, extraction, legume seeds, phytate
Procedia PDF Downloads 284013 Sensitivity Analysis for 14 Bus Systems in a Distribution Network with Distributed Generators
Authors: Lakshya Bhat, Anubhav Shrivastava, Shiva Rudraswamy
Abstract:
There has been a formidable interest in the area of Distributed Generation in recent times. A wide number of loads are addressed by Distributed Generators and have better efficiency too. The major disadvantage in Distributed Generation is voltage control- is highlighted in this paper. The paper addresses voltage control at buses in IEEE 14 Bus system by regulating reactive power. An analysis is carried out by selecting the most optimum location in placing the Distributed Generators through load flow analysis and seeing where the voltage profile rises. MATLAB programming is used for simulation of voltage profile in the respective buses after introduction of DG’s. A tolerance limit of +/-5% of the base value has to be maintained. To maintain the tolerance limit, 3 methods are used. Sensitivity analysis of 3 methods for voltage control is carried out to determine the priority among the methods.Keywords: distributed generators, distributed system, reactive power, voltage control, sensitivity analysis
Procedia PDF Downloads 7034012 Simple Procedure for Probability Calculation of Tensile Crack Occurring in Rigid Pavement: A Case Study
Authors: Aleš Florian, Lenka Ševelová, Jaroslav Žák
Abstract:
Formation of tensile cracks in concrete slabs of rigid pavement can be (among others) the initiation point of the other, more serious failures which can ultimately lead to complete degradation of the concrete slab and thus the whole pavement. Two measures can be used for reliability assessment of this phenomenon - the probability of failure and/or the reliability index. Different methods can be used for their calculation. The simple ones are called moment methods and simulation techniques. Two methods - FOSM Method and Simple Random Sampling Method - are verified and their comparison is performed. The influence of information about the probability distribution and the statistical parameters of input variables as well as of the limit state function on the calculated reliability index and failure probability are studied in three points on the lower surface of concrete slabs of the older type of rigid pavement formerly used in the Czech Republic.Keywords: failure, pavement, probability, reliability index, simulation, tensile crack
Procedia PDF Downloads 5464011 Development of ELF Passive Shielding Application Using Magnetic Aqueous Substrate
Authors: W. N. L. Mahadi, S. N. Syed Zin, W. A. R. Othman, N. A. Mohd Rasyid, N. Jusoh
Abstract:
Public concerns on Extremely Low Frequency (ELF) Electromagnetic Field (EMF) exposure have been elongated since the last few decades. Electrical substations and high tension rooms (HT room) in commercial buildings were among the contributing factors emanating ELF magnetic fields. This paper discussed various shielding methods conventionally used in mitigating the ELF exposure. Nevertheless, the standard methods were found to be impractical and incapable of meeting currents shielding demands. In response to that, remarkable researches were conducted in effort to invent novel methods which is more convenient and efficient such as magnetic aqueous shielding or paint, textiles and papers shielding. A mitigation method using magnetic aqueous substrate in shielding application was proposed in this paper for further investigation. using Manganese Zinc Ferrite (Mn0.4Zn0.6Fe2O4). The magnetic field and flux distribution inside the aqueous magnetic material are evaluated to optimize shielding against ELF-EMF exposure, as to mitigate its exposure.Keywords: ELF shielding, magnetic aqueous substrate, shielding effectiveness, passive shielding, magnetic material
Procedia PDF Downloads 5314010 Sensitivity Analysis for 14 Bus Systems in a Distribution Network with Distribution Generators
Authors: Lakshya Bhat, Anubhav Shrivastava, Shivarudraswamy
Abstract:
There has been a formidable interest in the area of Distributed Generation in recent times. A wide number of loads are addressed by Distributed Generators and have better efficiency too. The major disadvantage in Distributed Generation is voltage control- is highlighted in this paper. The paper addresses voltage control at buses in IEEE 14 Bus system by regulating reactive power. An analysis is carried out by selecting the most optimum location in placing the Distributed Generators through load flow analysis and seeing where the voltage profile rises. Matlab programming is used for simulation of voltage profile in the respective buses after introduction of DG’s. A tolerance limit of +/-5% of the base value has to be maintained.To maintain the tolerance limit , 3 methods are used. Sensitivity analysis of 3 methods for voltage control is carried out to determine the priority among the methods.Keywords: distributed generators, distributed system, reactive power, voltage control, sensitivity analysis
Procedia PDF Downloads 5874009 The Effectiveness of Implementing Interactive Training for Teaching Kazakh Language
Authors: Samal Abzhanova, Saule Mussabekova
Abstract:
Today, a new system of education is being created in Kazakhstan in order to develop the system of education and to satisfy the world class standards. For this purpose, there have been established new requirements and responsibilities to the instructors. Students should not be limited with providing only theoretical knowledge. Also, they should be encouraged to be competitive, to think creatively and critically. Moreover, students should be able to implement these skills into practice. These issues could be resolved through the permanent improvement of teaching methods. Therefore, a specialist who teaches the languages should use up-to-date methods and introduce new technologies. The result of the investigation suggests that an interactive teaching method is one of the new technologies in this field. This paper aims to provide information about implementing new technologies in the process of teaching language. The paper will discuss about necessity of introducing innovative technologies and the techniques of organizing interactive lessons. At the same time, the structure of the interactive lesson, conditions, principles, discussions, small group works and role-playing games will be considered. Interactive methods are carried out with the help of several types of activities, such as working in a team (with two or more group of people), playing situational or role-playing games, working with different sources of information, discussions, presentations, creative works and learning through solving situational tasks and etc.Keywords: interactive education, interactive methods, system of education, teaching a language
Procedia PDF Downloads 2944008 Information Visualization Methods Applied to Nanostructured Biosensors
Authors: Osvaldo N. Oliveira Jr.
Abstract:
The control of molecular architecture inherent in some experimental methods to produce nanostructured films has had great impact on devices of various types, including sensors and biosensors. The self-assembly monolayers (SAMs) and the electrostatic layer-by-layer (LbL) techniques, for example, are now routinely used to produce tailored architectures for biosensing where biomolecules are immobilized with long-lasting preserved activity. Enzymes, antigens, antibodies, peptides and many other molecules serve as the molecular recognition elements for detecting an equally wide variety of analytes. The principles of detection are also varied, including electrochemical methods, fluorescence spectroscopy and impedance spectroscopy. In this presentation an overview will be provided of biosensors made with nanostructured films to detect antibodies associated with tropical diseases and HIV, in addition to detection of analytes of medical interest such as cholesterol and triglycerides. Because large amounts of data are generated in the biosensing experiments, use has been made of computational and statistical methods to optimize performance. Multidimensional projection techniques such as Sammon´s mapping have been shown more efficient than traditional multivariate statistical analysis in identifying small concentrations of anti-HIV antibodies and for distinguishing between blood serum samples of animals infected with two tropical diseases, namely Chagas´ disease and Leishmaniasis. Optimization of biosensing may include a combination of another information visualization method, the Parallel Coordinate technique, with artificial intelligence methods in order to identify the most suitable frequencies for reaching higher sensitivity using impedance spectroscopy. Also discussed will be the possible convergence of technologies, through which machine learning and other computational methods may be used to treat data from biosensors within an expert system for clinical diagnosis.Keywords: clinical diagnosis, information visualization, nanostructured films, layer-by-layer technique
Procedia PDF Downloads 3374007 Self-Attention Mechanism for Target Hiding Based on Satellite Images
Authors: Hao Yuan, Yongjian Shen, Xiangjun He, Yuheng Li, Zhouzhou Zhang, Pengyu Zhang, Minkang Cai
Abstract:
Remote sensing data can provide support for decision-making in disaster assessment or disaster relief. The traditional processing methods of sensitive targets in remote sensing mapping are mainly based on manual retrieval and image editing tools, which are inefficient. Methods based on deep learning for sensitive target hiding are faster and more flexible. But these methods have disadvantages in training time and cost of calculation. This paper proposed a target hiding model Self Attention (SA) Deepfill, which used self-attention modules to replace part of gated convolution layers in image inpainting. By this operation, the calculation amount of the model becomes smaller, and the performance is improved. And this paper adds free-form masks to the model’s training to enhance the model’s universal. The experiment on an open remote sensing dataset proved the efficiency of our method. Moreover, through experimental comparison, the proposed method can train for a longer time without over-fitting. Finally, compared with the existing methods, the proposed model has lower computational weight and better performance.Keywords: remote sensing mapping, image inpainting, self-attention mechanism, target hiding
Procedia PDF Downloads 1364006 Control of Staphylococcus aureus in Meat System by in situ and ex situ Bacteriocins from Lactobacillus sakei and Pediococcus spp.
Authors: M. Naimi, M. B. Khaled
Abstract:
The present study consisted of an applied test in meat system to assess the effectiveness of three bio agents bacteriocinproducing strains: Lm24: Lactobacillus sakei, Lm14and Lm25: Pediococcus spp. Two tests were carried out: The ex-situ test was intended for three batches added with crude bacteriocin solutions at 12.48 AU/ml for Lm25 and 8.4 AU/ml for Lm14 and Lm24. However, the in situ one consisted of four batches; three of them inoculated with one bacteriocinogenic Lm25, Lm14, Lm24, respectively. The fourth one was used in mixture: Lm14+m24 at approximately of 107 CFU/ml. The two used tests were done in the presence of the pathogen St. aureus ATCC 6538, as a test strain at 103 CFU/ml. Another batch served as a positive or a negative control was used too. The incubation was performed at 7°C. Total viable counts, staphylococci and lactic acid bacteria, at the beginning and at selected times with interval of three days were enumerated. Physicochemical determinations (except for in situ test): pH, dry mater, sugars, fat and total protein, at the beginning and at end of the experiment, were done, according to the international norms. Our results confirmed the ex situ effectiveness. Furthermore, the batches affected negatively the total microbial load over the incubation days, and showed a significant regression in staphylococcal load at day seven, for Lm14, Lm24, and Lm25 of 0.73, 2.11, and 2.4 log units. It should be noticed that, at the last day of culture, staphylococcal load was nil for the three batches. In the in situ test, the cultures displayed less inhibitory attitude and recorded a decrease in staphylococcal load, for Lm14, Lm24, Lm25, Lm14+m24 of 0.73, 0.20, 0.86, 0.032 log units. Therefore, physicochemical analysis for Lm14, Lm24, Lm25, Lm14+m24 showed an increase in pH from 5.50 to 5.77, 6.18, 5.96, 7.22, a decrease in dry mater from 7.30% to 7.05%, 6.87%, 6.32%, 6.00%.This result reflects the decrease in fat ranging from 1.53% to 1.49%, 1.07%, 0.99%, 0.87%; and total protein from 6.18% to 5.25%, 5.56%, 5.37%, 5.5%. This study suggests that the use of selected strains as Lm25 could lead to the best results and would help in preserving and extending the shelf life of lamb meat.Keywords: biocontrol, in situ, ex situ, meat system, St. aureus, Lactobacillus sakei, Pediococcus spp.
Procedia PDF Downloads 1984005 A Comparative Study on ANN, ANFIS and SVM Methods for Computing Resonant Frequency of A-Shaped Compact Microstrip Antennas
Authors: Ahmet Kayabasi, Ali Akdagli
Abstract:
In this study, three robust predicting methods, namely artificial neural network (ANN), adaptive neuro fuzzy inference system (ANFIS) and support vector machine (SVM) were used for computing the resonant frequency of A-shaped compact microstrip antennas (ACMAs) operating at UHF band. Firstly, the resonant frequencies of 144 ACMAs with various dimensions and electrical parameters were simulated with the help of IE3D™ based on method of moment (MoM). The ANN, ANFIS and SVM models for computing the resonant frequency were then built by considering the simulation data. 124 simulated ACMAs were utilized for training and the remaining 20 ACMAs were used for testing the ANN, ANFIS and SVM models. The performance of the ANN, ANFIS and SVM models are compared in the training and test process. The average percentage errors (APE) regarding the computed resonant frequencies for training of the ANN, ANFIS and SVM were obtained as 0.457%, 0.399% and 0.600%, respectively. The constructed models were then tested and APE values as 0.601% for ANN, 0.744% for ANFIS and 0.623% for SVM were achieved. The results obtained here show that ANN, ANFIS and SVM methods can be successfully applied to compute the resonant frequency of ACMAs, since they are useful and versatile methods that yield accurate results.Keywords: a-shaped compact microstrip antenna, artificial neural network (ANN), adaptive neuro-fuzzy inference system (ANFIS), support vector machine (SVM)
Procedia PDF Downloads 4414004 Theoretical and Experimental Electrostatic Potential around the M-Nitrophenol Compound
Authors: Drissi Mokhtaria, Chouaih Abdelkader, Fodil Hamzaoui
Abstract:
Our work is about a comparison of experimental and theoretical results of the electron charge density distribution and the electrostatic potential around the M-Nitrophenol Molecule (m-NPH) kwon for its interesting physical characteristics. The molecular experimental results have been obtained from a high-resolution X-ray diffraction study. Theoretical investigations were performed under the Gaussian program using the Density Functional Theory at B3LYP level of theory at 6-31G*. The multipolar model of Hansen and Coppens was used for the experimental electron charge density distribution around the molecule, while we used the DFT methods for the theoretical calculations. The electron charge density obtained in both methods allowed us to find out the different molecular properties such us the electrostatic potential and the dipole moment which were finally subject to a comparison leading to an outcome of a good matching results obtained in both methods.Keywords: electron charge density, m-nitrophenol, nonlinear optical compound, electrostatic potential, optimized geometric
Procedia PDF Downloads 2684003 Solving Linear Systems Involved in Convex Programming Problems
Authors: Yixun Shi
Abstract:
Many interior point methods for convex programming solve an (n+m)x(n+m)linear system in each iteration. Many implementations solve this system in each iteration by considering an equivalent mXm system (4) as listed in the paper, and thus the job is reduced into solving the system (4). However, the system(4) has to be solved exactly since otherwise the error would be entirely passed onto the last m equations of the original system. Often the Cholesky factorization is computed to obtain the exact solution of (4). One Cholesky factorization is to be done in every iteration, resulting in higher computational costs. In this paper, two iterative methods for solving linear systems using vector division are combined together and embedded into interior point methods. Instead of computing one Cholesky factorization in each iteration, it requires only one Cholesky factorization in the entire procedure, thus significantly reduces the amount of computation needed for solving the problem. Based on that, a hybrid algorithm for solving convex programming problems is proposed.Keywords: convex programming, interior point method, linear systems, vector division
Procedia PDF Downloads 4024002 Bi-Dimensional Spectral Basis
Authors: Abdelhamid Zerroug, Mlle Ismahene Sehili
Abstract:
Spectral methods are usually applied to solve uni-dimensional boundary value problems. With the advantage of the creation of multidimensional basis, we propose a new spectral method for bi-dimensional problems. In this article, we start by creating bi-spectral basis by different ways, we developed also a new relations to determine the expressions of spectral coefficients in different partial derivatives expansions. Finally, we propose the principle of a new bi-spectral method for the bi-dimensional problems.Keywords: boundary value problems, bi-spectral methods, bi-dimensional Legendre basis, spectral method
Procedia PDF Downloads 3954001 Energy-Saving Methods and Principles of Energy-Efficient Concept Design in the Northern Hemisphere
Authors: Yulia A. Kononova, Znang X. Ning
Abstract:
Nowadays, architectural development is getting faster and faster. Nevertheless, modern architecture often does not meet all the points, which could help our planet to get better. As we know, people are spending an enormous amount of energy every day of their lives. Because of the uncontrolled energy usage, people have to increase energy production. As energy production process demands a lot of fuel sources, it courses a lot of problems such as climate changes, environment pollution, animals’ distinction, and lack of energy sources also. Nevertheless, nowadays humanity has all the opportunities to change this situation. Architecture is one of the most popular fields where it is possible to apply new methods of saving energy or even creating it. Nowadays we have kinds of buildings, which can meet new willing. One of them is energy effective buildings, which can save or even produce energy, combining several energy-saving principles. The main aim of this research is to provide information that helps to apply energy-saving methods while designing an environment-friendly building. The research methodology requires gathering relevant information from literature, building guidelines documents and previous research works in order to analyze it and sum up into a material that can be applied to energy-efficient building design. To mark results it should be noted that the usage of all the energy-saving methods applied to a design project of building results in ultra-low energy buildings that require little energy for space heating or cooling. As a conclusion it can be stated that developing methods of passive house design can decrease the need of energy production, which is an important issue that has to be solved in order to save planet sources and decrease environment pollution.Keywords: accumulation, energy-efficient building, storage, superinsulation, passive house
Procedia PDF Downloads 2624000 Evaluation of Ensemble Classifiers for Intrusion Detection
Authors: M. Govindarajan
Abstract:
One of the major developments in machine learning in the past decade is the ensemble method, which finds highly accurate classifier by combining many moderately accurate component classifiers. In this research work, new ensemble classification methods are proposed with homogeneous ensemble classifier using bagging and heterogeneous ensemble classifier using arcing and their performances are analyzed in terms of accuracy. A Classifier ensemble is designed using Radial Basis Function (RBF) and Support Vector Machine (SVM) as base classifiers. The feasibility and the benefits of the proposed approaches are demonstrated by the means of standard datasets of intrusion detection. The main originality of the proposed approach is based on three main parts: preprocessing phase, classification phase, and combining phase. A wide range of comparative experiments is conducted for standard datasets of intrusion detection. The performance of the proposed homogeneous and heterogeneous ensemble classifiers are compared to the performance of other standard homogeneous and heterogeneous ensemble methods. The standard homogeneous ensemble methods include Error correcting output codes, Dagging and heterogeneous ensemble methods include majority voting, stacking. The proposed ensemble methods provide significant improvement of accuracy compared to individual classifiers and the proposed bagged RBF and SVM performs significantly better than ECOC and Dagging and the proposed hybrid RBF-SVM performs significantly better than voting and stacking. Also heterogeneous models exhibit better results than homogeneous models for standard datasets of intrusion detection.Keywords: data mining, ensemble, radial basis function, support vector machine, accuracy
Procedia PDF Downloads 2483999 Using the Bootstrap for Problems Statistics
Authors: Brahim Boukabcha, Amar Rebbouh
Abstract:
The bootstrap method based on the idea of exploiting all the information provided by the initial sample, allows us to study the properties of estimators. In this article we will present a theoretical study on the different methods of bootstrapping and using the technique of re-sampling in statistics inference to calculate the standard error of means of an estimator and determining a confidence interval for an estimated parameter. We apply these methods tested in the regression models and Pareto model, giving the best approximations.Keywords: bootstrap, error standard, bias, jackknife, mean, median, variance, confidence interval, regression models
Procedia PDF Downloads 3803998 Assessment of Residual Stress on HDPE Pipe Wall Thickness
Authors: D. Sersab, M. Aberkane
Abstract:
Residual stresses, in high-density polyethylene (HDPE) pipes, result from a nonhomogeneous cooling rate that occurs between the inner and outer surfaces during the extrusion process in manufacture. Most known methods of measurements to determine the magnitude and profile of the residual stresses in the pipe wall thickness are layer removal and ring slitting method. The combined layer removal and ring slitting methods described in this paper involves measurement of the circumferential residual stresses with minimal local disturbance. The existing methods used for pipe geometry (ring slitting method) gives a single residual stress value at the bore. The layer removal method which is used more in flat plate specimen is implemented with ring slitting method. The method permits stress measurements to be made directly at different depth in the pipe wall and a well-defined residual stress profile was consequently obtained.Keywords: residual stress, layer removal, ring splitting, HDPE, wall thickness
Procedia PDF Downloads 338