Search results for: database testing
443 Estimation of Uncertainty of Thermal Conductivity Measurement with Single Laboratory Validation Approach
Authors: Saowaluck Ukrisdawithid
Abstract:
The thermal conductivity of thermal insulation materials are measured by Heat Flow Meter (HFM) apparatus. The components of uncertainty are complex and difficult on routine measurement by modelling approach. In this study, uncertainty of thermal conductivity measurement was estimated by single laboratory validation approach. The within-laboratory reproducibility was 1.1%. The standard uncertainty of method and laboratory bias by using SRM1453 expanded polystyrene board was dominant at 1.4%. However, it was assessed that there was no significant bias. For sample measurement, the sources of uncertainty were repeatability, density of sample and thermal conductivity resolution of HFM. From this approach to sample measurements, the combined uncertainty was calculated. In summary, the thermal conductivity of sample, polystyrene foam, was reported as 0.03367 W/m·K ± 3.5% (k = 2) at mean temperature 23.5 °C. The single laboratory validation approach is simple key of routine testing laboratory for estimation uncertainty of thermal conductivity measurement by using HFM, according to ISO/IEC 17025-2017 requirements. These are meaningful for laboratory competent improvement, quality control on products, and conformity assessment.
Keywords: Single laboratory validation approach, within-laboratory reproducibility, method and laboratory bias, certified reference material.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 808442 Post Pandemic Mobility Analysis through Indexing and Sharding in MongoDB: Performance Optimization and Insights
Authors: Karan Vishavjit, Aakash Lakra, Shafaq Khan
Abstract:
The COVID-19 pandemic has pushed healthcare professionals to use big data analytics as a vital tool for tracking and evaluating the effects of contagious viruses. To effectively analyse huge datasets, efficient NoSQL databases are needed. The analysis of post-COVID-19 health and well-being outcomes and the evaluation of the effectiveness of government efforts during the pandemic is made possible by this research’s integration of several datasets, which cuts down on query processing time and creates predictive visual artifacts. We recommend applying sharding and indexing technologies to improve query effectiveness and scalability as the dataset expands. Effective data retrieval and analysis are made possible by spreading the datasets into a sharded database and doing indexing on individual shards. Analysis of connections between governmental activities, poverty levels, and post-pandemic wellbeing is the key goal. We want to evaluate the effectiveness of governmental initiatives to improve health and lower poverty levels. We will do this by utilising advanced data analysis and visualisations. The findings provide relevant data that support the advancement of UN sustainable objectives, future pandemic preparation, and evidence-based decision-making. This study shows how Big Data and NoSQL databases may be used to address problems with global health.
Keywords: COVID-19, big data, data analysis, indexing, NoSQL, sharding, scalability, poverty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 67441 User-Based Cannibalization Mitigation in an Online Marketplace
Authors: Vivian Guo, Yan Qu
Abstract:
Online marketplaces are not only digital places where consumers buy and sell merchandise, and they are also destinations for brands to connect with real consumers at the moment when customers are in the shopping mindset. For many marketplaces, brands have been important partners through advertising. There can be, however, a risk of advertising impacting a consumer’s shopping journey if it hurts the use experience or takes the user away from the site. Both could lead to the loss of transaction revenue for the marketplace. In this paper, we present user-based methods for cannibalization control by selectively turning off ads to users who are likely to be cannibalized by ads subject to business objectives. We present ways of measuring cannibalization of advertising in the context of an online marketplace and propose novel ways of measuring cannibalization through purchase propensity and uplift modeling. A/B testing has shown that our methods can significantly improve user purchase and engagement metrics while operating within business objectives. To our knowledge, this is the first paper that addresses cannibalization mitigation at the user-level in the context of advertising.
Keywords: Cannibalization, machine learning, online marketplace, revenue optimization, yield optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 902440 Optimized Preprocessing for Accurate and Efficient Bioassay Prediction with Machine Learning Algorithms
Authors: Jeff Clarine, Chang-Shyh Peng, Daisy Sang
Abstract:
Bioassay is the measurement of the potency of a chemical substance by its effect on a living animal or plant tissue. Bioassay data and chemical structures from pharmacokinetic and drug metabolism screening are mined from and housed in multiple databases. Bioassay prediction is calculated accordingly to determine further advancement. This paper proposes a four-step preprocessing of datasets for improving the bioassay predictions. The first step is instance selection in which dataset is categorized into training, testing, and validation sets. The second step is discretization that partitions the data in consideration of accuracy vs. precision. The third step is normalization where data are normalized between 0 and 1 for subsequent machine learning processing. The fourth step is feature selection where key chemical properties and attributes are generated. The streamlined results are then analyzed for the prediction of effectiveness by various machine learning algorithms including Pipeline Pilot, R, Weka, and Excel. Experiments and evaluations reveal the effectiveness of various combination of preprocessing steps and machine learning algorithms in more consistent and accurate prediction.
Keywords: Bioassay, machine learning, preprocessing, virtual screen.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 981439 Experimental Investigation on the Optimal Operating Frequency of a Thermoacoustic Refrigerator
Authors: Kriengkrai Assawamartbunlue, Channarong Wantha
Abstract:
This paper presents effects of the mean operating pressure on the optimal operating frequency based on temperature differences across stack ends in a thermoacoustic refrigerator. In addition to the length of the resonance tube, components of the thermoacoustic refrigerator have an influence on the operating frequency due to their acoustic properties, i.e., absorptivity, reflectivity and transmissivity. The interference of waves incurs and distorts the original frequency generated by the driver so that the optimal operating frequency differs from the designs. These acoustic properties are not parameters in the designs and be very complicated to infer their responses. A prototype thermoacoustic refrigerator is constructed and used to investigate its optimal operating frequency compared to the design at various operating pressures. Helium and air are used as working fluids during the experiments. The results indicate that the optimal operating frequency of the prototype thermoacoustic refrigerator using helium is at 6 bar and 490Hz or approximately 20% away from the design frequency. The optimal operating frequency at other mean pressures differs from the design in an unpredictable manner, however, the optimal operating frequency and pressure can be identified by testing.
Keywords: Acoustic properties, Carnot’s efficiency, Interference of waves, Operating pressure, Optimal operating frequency, Stack performance, Standing Wave, Thermoacoustic refrigerator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1779438 Object Identification with Color, Texture, and Object-Correlation in CBIR System
Authors: Awais Adnan, Muhammad Nawaz, Sajid Anwar, Tamleek Ali, Muhammad Ali
Abstract:
Needs of an efficient information retrieval in recent years in increased more then ever because of the frequent use of digital information in our life. We see a lot of work in the area of textual information but in multimedia information, we cannot find much progress. In text based information, new technology of data mining and data marts are now in working that were started from the basic concept of database some where in 1960. In image search and especially in image identification, computerized system at very initial stages. Even in the area of image search we cannot see much progress as in the case of text based search techniques. One main reason for this is the wide spread roots of image search where many area like artificial intelligence, statistics, image processing, pattern recognition play their role. Even human psychology and perception and cultural diversity also have their share for the design of a good and efficient image recognition and retrieval system. A new object based search technique is presented in this paper where object in the image are identified on the basis of their geometrical shapes and other features like color and texture where object-co-relation augments this search process. To be more focused on objects identification, simple images are selected for the work to reduce the role of segmentation in overall process however same technique can also be applied for other images.Keywords: Object correlation, Geometrical shape, Color, texture, features, contents.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2028437 Flow Visualization of Angled Supersonic Jets into a Supersonic Cross Flow
Authors: Yan Shao, Jin Zhou, Lin Lai, Haiyan Wu, Jing Lei
Abstract:
This paper describes Nano-particle based Planar Laser Scattering (NPLS) flow visualization of angled supersonic jets into a supersonic cross flow based on the HYpersonic Low TEmperature (HYLTE) nozzle which was widely used in DF chemical laser. In order to investigate the non-reacting flowfield in the HYLTE nozzle, a testing section with windows was designed and manufactured. The impact of secondary fluids orifice separation on mixing was examined. For narrow separation of orifices, the secondary fuel penetration increased obviously compared to diluent injection, which means smaller separation of diluent and fuel orifices would enhance the mixing of fuel and oxidant. Secondary injections with angles of 30, 40 and 50 degrees were studied. It was found that the injectant penetration increased as the injection angle increased, while the interfacial surface area to entrain the freestream fluid is largest when the injection angle is 40 degree.Keywords: HYLTE nozzle, NPLS, supersonic mixing, transverse injection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1840436 Decision Trees for Predicting Risk of Mortality using Routinely Collected Data
Authors: Tessy Badriyah, Jim S. Briggs, Dave R. Prytherch
Abstract:
It is well known that Logistic Regression is the gold standard method for predicting clinical outcome, especially predicting risk of mortality. In this paper, the Decision Tree method has been proposed to solve specific problems that commonly use Logistic Regression as a solution. The Biochemistry and Haematology Outcome Model (BHOM) dataset obtained from Portsmouth NHS Hospital from 1 January to 31 December 2001 was divided into four subsets. One subset of training data was used to generate a model, and the model obtained was then applied to three testing datasets. The performance of each model from both methods was then compared using calibration (the χ2 test or chi-test) and discrimination (area under ROC curve or c-index). The experiment presented that both methods have reasonable results in the case of the c-index. However, in some cases the calibration value (χ2) obtained quite a high result. After conducting experiments and investigating the advantages and disadvantages of each method, we can conclude that Decision Trees can be seen as a worthy alternative to Logistic Regression in the area of Data Mining.Keywords: Decision Trees, Logistic Regression, clinical outcome, risk of mortality.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2523435 Analysis of Mechanical Properties for AP/HTPB Solid Propellant under Different Loading Conditions
Authors: Walid M. Adel, Liang Guo-Zhu
Abstract:
To investigate the characterization of the mechanical properties of composite solid propellant (CSP) based on hydroxyl-terminated polybutadiene (HTPB) at different temperatures and strain rates, uniaxial tensile tests were conducted over a range of temperatures -60 °C to +76 °C and strain rates 0.000164 to 0.328084 s-1 using a conventional universal testing machine. From the experimental data, it can be noted that the mechanical properties of AP/HTPB propellant are mainly dependent on the applied strain rate and the temperature condition. The stress-strain responses exhibited an initial yielding followed by the viscoelastic phase, which was strongly affected by the strain rate and temperature. It was found that the mechanical properties increased with both increasing strain rate and decreasing temperature. Based on the experimental tests, the master curves of the tensile properties are drawn using predetermined shift factor and the results were discussed. This work is a first step in preliminary investigation the nonlinear viscoelasticity behavior of CSP.
Keywords: AP/HTPB composite solid propellant, mechanical behavior, nonlinear viscoelastic, tensile test, master curves.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2041434 A Sequential Approach to Random-Effects Meta-Analysis
Authors: Samson Henry Dogo, Allan Clark, Elena Kulinskaya
Abstract:
The objective of meta-analysis is to combine results from several independent studies in order to create generalization and provide evidence base for decision making. But recent studies show that the magnitude of effect size estimates reported in many areas of research significantly changed over time and this can impair the results and conclusions of meta-analysis. A number of sequential methods have been proposed for monitoring the effect size estimates in meta-analysis. However they are based on statistical theory applicable only to fixed effect model (FEM) of meta-analysis. For random-effects model (REM), the analysis incorporates the heterogeneity variance, τ 2 and its estimation create complications. In this paper we study the use of a truncated CUSUM-type test with asymptotically valid critical values for sequential monitoring in REM. Simulation results show that the test does not control the Type I error well, and is not recommended. Further work required to derive an appropriate test in this important area of applications.
Keywords: Meta-analysis, random-effects model, sequential testing, temporal changes in effect sizes.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2424433 FleGSens – Secure Area Monitoring Using Wireless Sensor Networks
Authors: Peter Rothenpieler, Daniela Kruger, Dennis Pfisterer, Stefan Fischer, Denise Dudek, Christian Haas, Martina Zitterbart
Abstract:
In the project FleGSens, a wireless sensor network (WSN) for the surveillance of critical areas and properties is currently developed which incorporates mechanisms to ensure information security. The intended prototype consists of 200 sensor nodes for monitoring a 500m long land strip. The system is focused on ensuring integrity and authenticity of generated alarms and availability in the presence of an attacker who may even compromise a limited number of sensor nodes. In this paper, two of the main protocols developed in the project are presented, a tracking protocol to provide secure detection of trespasses within the monitored area and a protocol for secure detection of node failures. Simulation results of networks containing 200 and 2000 nodes as well as the results of the first prototype comprising a network of 16 nodes are presented. The focus of the simulations and prototype are functional testing of the protocols and particularly demonstrating the impact and cost of several attacks.Keywords: Wireless Sensor Network, Security, Trespass Detection, Testbed.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1977432 From Electroencephalogram to Epileptic Seizures Detection by Using Artificial Neural Networks
Authors: Gaetano Zazzaro, Angelo Martone, Roberto V. Montaquila, Luigi Pavone
Abstract:
Seizure is the main factor that affects the quality of life of epileptic patients. The diagnosis of epilepsy, and hence the identification of epileptogenic zone, is commonly made by using continuous Electroencephalogram (EEG) signal monitoring. Seizure identification on EEG signals is made manually by epileptologists and this process is usually very long and error prone. The aim of this paper is to describe an automated method able to detect seizures in EEG signals, using knowledge discovery in database process and data mining methods and algorithms, which can support physicians during the seizure detection process. Our detection method is based on Artificial Neural Network classifier, trained by applying the multilayer perceptron algorithm, and by using a software application, called Training Builder that has been developed for the massive extraction of features from EEG signals. This tool is able to cover all the data preparation steps ranging from signal processing to data analysis techniques, including the sliding window paradigm, the dimensionality reduction algorithms, information theory, and feature selection measures. The final model shows excellent performances, reaching an accuracy of over 99% during tests on data of a single patient retrieved from a publicly available EEG dataset.
Keywords: Artificial Neural Network, Data Mining, Electroencephalogram, Epilepsy, Feature Extraction, Seizure Detection, Signal Processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314431 Reducing SAGE Data Using Genetic Algorithms
Authors: Cheng-Hong Yang, Tsung-Mu Shih, Li-Yeh Chuang
Abstract:
Serial Analysis of Gene Expression is a powerful quantification technique for generating cell or tissue gene expression data. The profile of the gene expression of cell or tissue in several different states is difficult for biologists to analyze because of the large number of genes typically involved. However, feature selection in machine learning can successfully reduce this problem. The method allows reducing the features (genes) in specific SAGE data, and determines only relevant genes. In this study, we used a genetic algorithm to implement feature selection, and evaluate the classification accuracy of the selected features with the K-nearest neighbor method. In order to validate the proposed method, we used two SAGE data sets for testing. The results of this study conclusively prove that the number of features of the original SAGE data set can be significantly reduced and higher classification accuracy can be achieved.Keywords: Serial Analysis of Gene Expression, Feature selection, Genetic Algorithm, K-nearest neighbor method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1610430 Novel CFRP Adhesive Joints and Structures for Offshore Application
Authors: M. R. Abusrea, Shiyi Jiang, Dingding Chen, Kazuo Arakawa
Abstract:
Novel wind-lens turbine designs can augment power output. Vacuum-Assisted Resin Transfer Molding (VARTM) is used to form large and complex structures from a Carbon Fiber Reinforced Polymer (CFRP) composite. Typically, wind-lens turbine structures are fabricated in segments, and then bonded to form the final structure. This paper introduces five new adhesive joints, divided into two groups: one is constructed between dry carbon and CFRP fabrics, and the other is constructed with two dry carbon fibers. All joints and CFRP fabrics were made in our laboratory using VARTM manufacturing techniques. Specimens were prepared for tensile testing to measure joint performance. The results showed that the second group of joints achieved a higher tensile strength than the first group. On the other hand, the tensile fracture behavior of the two groups showed the same pattern of crack originating near the joint ends followed by crack propagation until fracture.Keywords: Adhesive joints, CFRP, VARTM, resin transfer molding.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1870429 Probabilistic Modelling of Marine Bridge Deterioration
Authors: P.C. Ryan, A.J. O' Connor
Abstract:
Chloride induced corrosion of steel reinforcement is the main cause of deterioration of reinforced concrete marine structures. This paper investigates the relative performance of alternative repair options with respect to the deterioration of reinforced concrete bridge elements in marine environments. Focus is placed on the initiation phase of reinforcement corrosion. A laboratory study is described which involved exposing concrete samples to accelerated chloride-ion ingress. The study examined the relative efficiencies of two repair methods, namely Ordinary Portland Cement (OPC) concrete and a concrete which utilised Ground Granulated Blastfurnace Cement (GGBS) as a partial cement replacement. The mix designs and materials utilised were identical to those implemented in the repair of a marine bridge on the South East coast of Ireland in 2007. The results of this testing regime serve to inform input variables employed in probabilistic modelling of deterioration for subsequent reliability based analysis to compare the relative performance of the studied repair options.Keywords: Deterioration, Marine Bridges, Reinforced Concrete, Reliability, Chloride-ion Ingress
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2006428 A Comparative Study of Fine Grained Security Techniques Based on Data Accessibility and Inference
Authors: Azhar Rauf, Sareer Badshah, Shah Khusro
Abstract:
This paper analyzes different techniques of the fine grained security of relational databases for the two variables-data accessibility and inference. Data accessibility measures the amount of data available to the users after applying a security technique on a table. Inference is the proportion of information leakage after suppressing a cell containing secret data. A row containing a secret cell which is suppressed can become a security threat if an intruder generates useful information from the related visible information of the same row. This paper measures data accessibility and inference associated with row, cell, and column level security techniques. Cell level security offers greatest data accessibility as it suppresses secret data only. But on the other hand, there is a high probability of inference in cell level security. Row and column level security techniques have least data accessibility and inference. This paper introduces cell plus innocent security technique that utilizes the cell level security method but suppresses some innocent data to dodge an intruder that a suppressed cell may not necessarily contain secret data. Four variations of the technique namely cell plus innocent 1/4, cell plus innocent 2/4, cell plus innocent 3/4, and cell plus innocent 4/4 respectively have been introduced to suppress innocent data equal to 1/4, 2/4, 3/4, and 4/4 percent of the true secret data inside the database. Results show that the new technique offers better control over data accessibility and inference as compared to the state-of-theart security techniques. This paper further discusses the combination of techniques together to be used. The paper shows that cell plus innocent 1/4, 2/4, and 3/4 techniques can be used as a replacement for the cell level security.
Keywords: Fine Grained Security, Data Accessibility, Inference, Row, Cell, Column Level Security.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1471427 Development of a Software about Calculating the Production Parameters in Knitted Garment Plants
Authors: Ender Bulgun, Arzu Vuruskan
Abstract:
Apparel product development is an important stage in the life cycle of a product. Shortening this stage will help to reduce the costs of a garment. The aim of this study is to examine the production parameters in knitwear apparel companies by defining the unit costs, and developing a software to calculate the unit costs of garments and make the cost estimates. In this study, with the help of a questionnaire, different companies- systems of unit cost estimating and cost calculating were tried to be analyzed. Within the scope of the questionnaire, the importance of cost estimating process for apparel companies and the expectations from a new cost estimating program were investigated. According to the results of the questionnaire, it was seen that the majority of companies which participated to the questionnaire use manual cost calculating methods or simple Microsoft Excel spreadsheets to make cost estimates. Furthermore, it was discovered that many companies meet with difficulties in archiving the cost data for future use and as a solution to that problem, it is thought that prior to making a cost estimate, sub units of garment costs which are fabric, accessory and the labor costs should be analyzed and added to the database of the programme beforehand. Another specification of the cost estimating unit prepared in this study is that the programme was designed to consist of two main units, one of which makes the product specification and the other makes the cost calculation. The programme is prepared as a web-based application in order that the supplier, the manufacturer and the customer can have the opportunity to communicate through the same platform.
Keywords: Apparel, cost estimating, design archive.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2982426 Finite Element Analysis of Full Ceramic Crowns with and without Zirconia Framework
Authors: Porojan S., Sandu L., Topală F.
Abstract:
Simulation of occlusal function during laboratory material-s testing becomes essential in predicting long-term performance before clinical usage. The aim of the study was to assess the influence of chamfer preparation depth on failure risk of heat pressed ceramic crowns with and without zirconia framework by means of finite element analysis. 3D models of maxillary central incisor, prepared for full ceramic crowns with different depths of the chamfer margin (between 0.8 and 1.2 mm) and 6-degree tapered walls together with the overlying crowns were generated using literature data (Fig. 1, 2). The crowns were designed with and without a zirconia framework with a thickness of 0.4 mm. For all preparations and crowns, stresses in the pressed ceramic crown, zirconia framework, pressed ceramic veneer, and dentin were evaluated separately. The highest stresses were registered in the dentin. The depth of the preparations had no significant influence on the stress values of the teeth and pressed ceramics for the studied cases, only for the zirconia framework. The zirconia framework decreases the stress values in the veneer.Keywords: Finite element analysis, full ceramic crown, zirconia framework, stresses.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1617425 Determination of Moisture Content and Liquid Limit of Foundations Soils, using Microwave Radiation, in the Different Locations of Sulaimani Governorate, Kurdistan Region-Iraq
Authors: Heyam Daod
Abstract:
Soils are normally dried in either a convection oven or stove. Laboratory moisture content testing indicated that the typical drying durations for a convection oven were, 24 hours. The purpose of this study was to determine the accuracy and soil drying duration of both, moisture content and liquid limit using microwave radiation. The soils were tested with both, convection and microwave ovens. The convection oven was considered to produce the true values for both, natural moisture content and liquid limit of soils; it was, therefore, used as a basis for comparison for the results of the microwave ovens. The samples used in this study were obtained from different projects of Consulting Engineering Bureau of College of Engineering of Sulaimani University. These samples were collected from different locations and at the different depths and consist mostly of brown and light brown clay and silty clay. A total of 102 samples were prepared. 26 of them were tested for natural moisture determination, while the other 76 were used for liquid limits determination
Keywords: Fine-grained soils, liquid limit, microwave drying,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4701424 Charaterisation of Salmonella Isolated from Nile Tilapia (Oreochromis niloticus) along Lake Victoria Beaches in Western Kenya
Authors: Wandili S. Awuor, Onyango D. Miruka, Waindi N. Eliud
Abstract:
Foodborne Salmonella infections have become a major problem world wide. Salmonellosis transmitted from fish are quite common. Established quality control measures exist for export oriented fish, none exists for fish consumed locally. This study aimed at characterization of Salmonella isolated from Nile tilapia . The study was carried out in selected beaches along L. Victoria in Western Kenya between March and June 2007. One hundred and twenty fish specimens were collected. Salmonella isolates were confirmed using serotyping, biochemical testing in addition to malic acid dehydrogenase (mdh) and fliC gene sequencing. Twenty Salmonella isolates were confirmed by mdh gene sequencing. Nine (9) were S. enterica serotype typhimurium, four (4) were S. enterica Serotype, enteritidis and seven (7) were S. enterica serotype typhi. Nile tilapia have a role in transmission of Salmonellosis in the study area, poor sanitation was a major cause of pollution at the beach inshore waters.Keywords: fliC, mdh, Salmonellosis, Serotype
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1981423 Semantic Support for Hypothesis-Based Research from Smart Environment Monitoring and Analysis Technologies
Authors: T. S. Myers, J. Trevathan
Abstract:
Improvements in the data fusion and data analysis phase of research are imperative due to the exponential growth of sensed data. Currently, there are developments in the Semantic Sensor Web community to explore efficient methods for reuse, correlation and integration of web-based data sets and live data streams. This paper describes the integration of remotely sensed data with web-available static data for use in observational hypothesis testing and the analysis phase of research. The Semantic Reef system combines semantic technologies (e.g., well-defined ontologies and logic systems) with scientific workflows to enable hypothesis-based research. A framework is presented for how the data fusion concepts from the Semantic Reef architecture map to the Smart Environment Monitoring and Analysis Technologies (SEMAT) intelligent sensor network initiative. The data collected via SEMAT and the inferred knowledge from the Semantic Reef system are ingested to the Tropical Data Hub for data discovery, reuse, curation and publication.
Keywords: Information architecture, Semantic technologies Sensor networks, Ontologies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1715422 Producing Outdoor Design Conditions Based on the Dependency between Meteorological Elements: Copula Approach
Authors: Zhichao Jiao, Craig Farnham, Jihui Yuan, Kazuo Emura
Abstract:
It is common to use the outdoor design weather data to select the air-conditioning capacity in the building design stage. The meteorological elements of outdoor design weather data are usually selected based on their excess frequency separately while the dependency between the elements is not well considered. It means that the simultaneous occurrence probability of these elements is smaller than the original excess frequency which may cause an overestimation of selecting air-conditioning capacity. Therefore, the copula approach which can capture the dependency between multivariate data was used to model the joint distributions of the meteorological elements, like air temperature and global solar radiation. We suggest a method based on the specific simultaneous occurrence probability of these two elements of selecting more credible outdoor design conditions. The hourly weather data at 12 noon from 2001 to 2010 in Tokyo, Japan are used to analyze the dependency structure and joint distribution, the Gaussian copula represents the dependence of data best. According to calculating the air temperature and global solar radiation in specific simultaneous occurrence probability and the common exceeding, the results show that both the air temperature and global solar radiation based on simultaneous occurrence probability are lower than these based on the conventional method in the same probability.
Keywords: Copula approach, Design weather database, energy conservation, HVAC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 359421 Automatic Road Network Recognition and Extraction for Urban Planning
Authors: D. B. L. Bong, K.C. Lai, A. Joseph
Abstract:
The uses of road map in daily activities are numerous but it is a hassle to construct and update a road map whenever there are changes. In Universiti Malaysia Sarawak, research on Automatic Road Extraction (ARE) was explored to solve the difficulties in updating road map. The research started with using Satellite Image (SI), or in short, the ARE-SI project. A Hybrid Simple Colour Space Segmentation & Edge Detection (Hybrid SCSS-EDGE) algorithm was developed to extract roads automatically from satellite-taken images. In order to extract the road network accurately, the satellite image must be analyzed prior to the extraction process. The characteristics of these elements are analyzed and consequently the relationships among them are determined. In this study, the road regions are extracted based on colour space elements and edge details of roads. Besides, edge detection method is applied to further filter out the non-road regions. The extracted road regions are validated by using a segmentation method. These results are valuable for building road map and detecting the changes of the existing road database. The proposed Hybrid Simple Colour Space Segmentation and Edge Detection (Hybrid SCSS-EDGE) algorithm can perform the tasks fully automatic, where the user only needs to input a high-resolution satellite image and wait for the result. Moreover, this system can work on complex road network and generate the extraction result in seconds.Keywords: Road Network Recognition, Colour Space, Edge Detection, Urban Planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2994420 Effect of Blast Furnace Iron Slag on the Mechanical Performance of Hot Mix Asphalt (HMA)
Authors: Ayman M. Othman, Hassan Y. Ahmed
Abstract:
This paper discusses the effect of using blast furnace iron slag as a part of fine aggregate on the mechanical performance of hot mix asphalt (HMA). The mechanical performance was evaluated based on various mechanical properties that include; Marshall/stiffness, indirect tensile strength and unconfined compressive strength. The effect of iron slag content on the mechanical properties of the mixtures was also investigated. Four HMA with various iron slag contents, namely; 0%, 5%, 10% and 15% by weight of total mixture were studied. Laboratory testing has revealed an enhancement in the compressive strength of HMA when iron slag was used. Within the tested range of iron slag content, a considerable increase in the compressive strength of the mixtures was observed with the increase of slag content. No significant improvement on Marshall/stiffness and indirect tensile strength of the mixtures was observed when slag was used. Even so, blast furnace iron slag can still be used in asphalt paving for environmental advantages.Keywords: Blast furnace iron slag, HMA, Marshall/stiffness, indirect tensile strength, compressive strength.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1414419 Mapping of Adrenal Gland Diseases Research in Middle East Countries: A Scientometric Analysis, 2007-2013
Authors: Zahra Emami, Mohammad Ebrahim Khamseh, Nahid Hashemi Madani, Iman Kermani
Abstract:
The aim of the study was to map scientific research on adrenal gland diseases in the Middle East countries through the Web of Science database using scientometric analysis. Data were analyzed with Excel software; and HistCite was used for mapping of the scientific texts. In this study, from a total of 268 retrieved records, 1125 authors from 328 institutions published their texts in 138 journals. Among 17 Middle East countries, Turkey ranked first with 164 documents (61.19%), Israel ranked second with 47 documents (15.53%) and Iran came in the third place with 26 documents. Most of the publications (185 documents, 69.2%) were articles. Among the universities of the Middle East, Istanbul University had the highest science production rate (9.7%). The Journal of Clinical Endocrinology & Metabolism had the highest TGCS (243 citations). In the scientific mapping, 7 clusters were formed based on TLCS (Total Local Citation Score) & TGCS (Total Global Citation Score). considering the study results, establishment of scientific connections and collaboration with other countries and use of publications on adrenal gland diseases from high ranking universities can help in the development of this field and promote the medical practice in this regard. Moreover, investigation of the formed clusters in relation to Congenital Hyperplasia and puberty related disorders can be research priorities for investigators.
Keywords: Mapping, scientific research, adrenal gland diseases, scientometric.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371418 Wind Tunnel for Aerodynamic Development Testing
Authors: E. T. L. Cöuras Ford, V. A. C. Vale, J. U. L. Mendes, F. A. Ribeiro
Abstract:
The study of the aerodynamics related to the improvement in the acting of airplanes and automobiles with the objective of being reduced the effect of the attrition of the air on structures, providing larger speeds and smaller consumption of fuel. The application of the knowledge of the aerodynamics not more limits to the aeronautical and automobile industries. Therefore, this research aims to design and construction of a wind tunnel to perform aerodynamic analysis in bodies of cars, seeking greater efficiency. Therefore, this research aims to design and construction of a wind tunnel to perform aerodynamic analysis in bodies of cars, seeking greater efficiency. For this, a methodology for wind tunnel type selection is designed to be built, taking into account the various existing configurations in which chose to build an open circuit tunnel, due to the lower complexity of construction and installation; operational simplicity and low cost. The guidelines for the project were teaching: the layer that limits study and analyze specimens with different geometries. For the variation of pressure in the test, section of a switched gauge used a pitot tube. Thus, it was possible to obtain quantitative and qualitative results, which proved to be satisfactory.Keywords: Wind tunnel, Aerodynamics, Air.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1379417 Gas Generator Pyrotechnics Using Gun Propellant Technology Methods
Authors: B. A. Parate
Abstract:
This research article describes the gas generator pyro-cartridge using gun propellant technology methods for fighter aircraft application. The emphasis of this work is to design and develop a gas generating device with pyro-cartridge using double base (DB) propellant to generate a high temperature and pressure gas. This device is utilised for dropping empty fuel tank in an emergency from military aircraft. A data acquisition system (DAS) is used to record time to maximum pressure, maximum pressure and time to half maximum pressure generated in a vented vessel (VV) for gas generator. Pyro-cartridge as a part of the gas generator creates a maximum pressure and time in the closed vessel (CV). This article also covers the qualification testing of gas generator. The performance parameters of pyro-cartridge devices such as ignition delay and maximum pressure are experimentally presented through the CV tests.
Keywords: Closed vessel, data acquisition, double base propellant, gas generator, ignition system, ignition delay, propellant, pyro-cartridge, pyrotechnics, vented vessel.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 519416 Deterministic Random Number Generator Algorithm for Cryptosystem Keys
Authors: Adi A. Maaita, Hamza A. A. Al_Sewadi
Abstract:
One of the crucial parameters of digital cryptographic systems is the selection of the keys used and their distribution. The randomness of the keys has a strong impact on the system’s security strength being difficult to be predicted, guessed, reproduced, or discovered by a cryptanalyst. Therefore, adequate key randomness generation is still sought for the benefit of stronger cryptosystems. This paper suggests an algorithm designed to generate and test pseudo random number sequences intended for cryptographic applications. This algorithm is based on mathematically manipulating a publically agreed upon information between sender and receiver over a public channel. This information is used as a seed for performing some mathematical functions in order to generate a sequence of pseudorandom numbers that will be used for encryption/decryption purposes. This manipulation involves permutations and substitutions that fulfill Shannon’s principle of “confusion and diffusion”. ASCII code characters were utilized in the generation process instead of using bit strings initially, which adds more flexibility in testing different seed values. Finally, the obtained results would indicate sound difficulty of guessing keys by attackers.Keywords: Cryptosystems, Information Security agreement, Key distribution, Random numbers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3432415 Recycling in Bogotá: A SWOT Analysis of Three Associations to Evaluate the Integrating the Informal Sector into Solid Waste Management
Authors: Clara Inés Pardo Martínez, William H. Alfonso Piña
Abstract:
In emerging economies, recycling is an opportunity for the cities to increase the lifespan of sanitary landfills, reduce the costs of the solid waste management, decrease the environmental problems of the waste treatment through reincorporate waste in the productive cycle and protect and develop people’s livelihoods of informal waste pickers. However, few studies have analysed the possibilities and strategies to integrate formal and informal sectors in the solid waste management for the benefit of both. This study seek to make a strength, weakness, opportunity, and threat (SWOT) analysis in three recycling associations of Bogotá with the aim to understand and determine the situation of recycling from perspective of informal sector in its transition to enter as authorized waste providers. Data used in the analysis are derived from multiple strategies such as literature review, the Bogota’s recycling database, focus group meetings, governmental reports, national laws and regulations and specific interviews with key stakeholders. Results of this study show as the main stakeholders of formal and informal sector of waste management can identify the internal and internal conditions of recycling in Bogotá. Several strategies were designed based on the SWOTs determined, could be useful for Bogotá to advance and promote recycling as a key strategy for integrated sustainable waste management in the city.
Keywords: Bogotá, recycling, solid waste management, SWOT analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7893414 A Systematic Approach for Identifying Turning Center Capabilities with Vertical Machining Center in Milling Operation
Abstract:
Conventional machining is a form of subtractive manufacturing, in which a collection of material-working processes utilizing power-driven machine tools are used to remove undesired material to achieve a desired geometry. This paper presents an approach for comparison between turning center and vertical machining center by optimization of cutting parameters at cylindrical workpieces leading to minimum surface roughness by using taguchi methodology. Aluminum alloy was taken to conduct experiments due to its unique high strength-weight ratio that is maintained at elevated temperatures and their exceptional corrosion resistance. During testing, the effects of the cutting parameters on the surface roughness were investigated. Additionally, by using taguchi methodology for each of the cutting parameters (spindle speed, depth of cut, insert diameter, and feed rate) minimum surface roughness for the process of turn-milling was determined according to the cutting parameters. A confirmation experiment demonstrates the effectiveness of taguchi method.
Keywords: Surface roughness, taguchi parameter design, turning center, turn-milling operations, vertical machining center.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2523