Search results for: multivariate measurement system analysis
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 40429

Search results for: multivariate measurement system analysis

39079 An Absolute Femtosecond Rangefinder for Metrological Support in Coordinate Measurements

Authors: Denis A. Sokolov, Andrey V. Mazurkevich

Abstract:

In the modern world, there is an increasing demand for highly precise measurements in various fields, such as aircraft, shipbuilding, and rocket engineering. This has resulted in the development of appropriate measuring instruments that are capable of measuring the coordinates of objects within a range of up to 100 meters, with an accuracy of up to one micron. The calibration process for such optoelectronic measuring devices (trackers and total stations) involves comparing the measurement results from these devices to a reference measurement based on a linear or spatial basis. The reference used in such measurements could be a reference base or a reference range finder with the capability to measure angle increments (EDM). The base would serve as a set of reference points for this purpose. The concept of the EDM for replicating the unit of measurement has been implemented on a mobile platform, which allows for angular changes in the direction of laser radiation in two planes. To determine the distance to an object, a high-precision interferometer with its own design is employed. The laser radiation travels to the corner reflectors, which form a spatial reference with precisely known positions. When the femtosecond pulses from the reference arm and the measuring arm coincide, an interference signal is created, repeating at the frequency of the laser pulses. The distance between reference points determined by interference signals is calculated in accordance with recommendations from the International Bureau of Weights and Measures for the indirect measurement of time of light passage according to the definition of a meter. This distance is D/2 = c/2nF, approximately 2.5 meters, where c is the speed of light in a vacuum, n is the refractive index of a medium, and F is the frequency of femtosecond pulse repetition. The achieved uncertainty of type A measurement of the distance to reflectors 64 m (N•D/2, where N is an integer) away and spaced apart relative to each other at a distance of 1 m does not exceed 5 microns. The angular uncertainty is calculated theoretically since standard high-precision ring encoders will be used and are not a focus of research in this study. The Type B uncertainty components are not taken into account either, as the components that contribute most do not depend on the selected coordinate measuring method. This technology is being explored in the context of laboratory applications under controlled environmental conditions, where it is possible to achieve an advantage in terms of accuracy. In general, the EDM tests showed high accuracy, and theoretical calculations and experimental studies on an EDM prototype have shown that the uncertainty type A of distance measurements to reflectors can be less than 1 micrometer. The results of this research will be utilized to develop a highly accurate mobile absolute range finder designed for the calibration of high-precision laser trackers and laser rangefinders, as well as other equipment, using a 64 meter laboratory comparator as a reference.

Keywords: femtosecond laser, pulse correlation, interferometer, laser absolute range finder, coordinate measurement

Procedia PDF Downloads 51
39078 Integration GIS–SCADA Power Systems to Enclosure Air Dispersion Model

Authors: Ibrahim Shaker, Amr El Hossany, Moustafa Osman, Mohamed El Raey

Abstract:

This paper will explore integration model between GIS–SCADA system and enclosure quantification model to approach the impact of failure-safe event. There are real demands to identify spatial objects and improve control system performance. Nevertheless, the employed methodology is predicting electro-mechanic operations and corresponding time to environmental incident variations. Open processing, as object systems technology, is presented for integration enclosure database with minimal memory size and computation time via connectivity drivers such as ODBC:JDBC during main stages of GIS–SCADA connection. The function of Geographic Information System is manipulating power distribution in contrast to developing issues. In other ward, GIS-SCADA systems integration will require numerical objects of process to enable system model calibration and estimation demands, determine of past events for analysis and prediction of emergency situations for response training.

Keywords: air dispersion model, environmental management, SCADA systems, GIS system, integration power system

Procedia PDF Downloads 359
39077 Impact of Economic Globalization on Ecological Footprint in India: Evidenced with Dynamic ARDL Simulations

Authors: Muhammed Ashiq Villanthenkodath, Shreya Pal

Abstract:

Purpose: This study scrutinizes the impact of economic globalization on ecological footprint while endogenizing economic growth and energy consumption from 1990 to 2018 in India. Design/methodology/approach: The standard unit root test has been employed for time series analysis to unveil the integration order. Then, the cointegration was confirmed using autoregressive distributed lag (ARDL) analysis. Further, the study executed the dynamic ARDL simulation model to estimate long-run and short-run results along with simulation and robotic prediction. Findings: The cointegration analysis confirms the existence of a long-run association among variables. Further, economic globalization reduces the ecological footprint in the long run. Similarly, energy consumption decreases the ecological footprint. In contrast, economic growth spurs the ecological footprint in India. Originality/value: This study contributes to the literature in many ways. First, unlike studies that employ CO2 emissions and globalization nexus, this study employs ecological footprint for measuring environmental quality; since it is the broader measure of environmental quality, it can offer a wide range of climate change mitigation policies for India. Second, the study executes a multivariate framework with updated series from 1990 to 2018 in India to explore the link between EF, economic globalization, energy consumption, and economic growth. Third, the dynamic autoregressive distributed lag (ARDL) model has been used to explore the short and long-run association between the series. Finally, to our limited knowledge, this is the first study that uses economic globalization in the EF function of India amid facing a trade-off between sustainable economic growth and the environment in the era of globalization.

Keywords: economic globalization, ecological footprint, India, dynamic ARDL simulation model

Procedia PDF Downloads 119
39076 Methodology of Choosing Technology and Sizing of the Hybrid Energy Storage Based on Cost-benefit Analysis

Authors: Krzysztof Rafał, Weronika Radziszewska, Hubert Biedka, Oskar Grabowski, Krzysztof Mik

Abstract:

We present a method to choose energy storage technologies and their parameters for the economic operation of a microgrid. A grid-connected system with local loads and PV generation is assumed, where an energy storage system (ESS) is attached to minimize energy cost by providing energy balancing and arbitrage functionalities. The ESS operates in a hybrid configuration and consists of two unique technologies operated in a coordinated way. Based on given energy profiles and economical data a model calculates financial flow for ESS investment, including energy cost and ESS depreciation resulting from degradation. The optimization strategy proposes a hybrid set of two technologies with their respective power and energy ratings to minimize overall system cost in a given timeframe. Results are validated through microgrid simulations using real-life input profiles.

Keywords: energy storage, hybrid energy storage, cost-benefit analysis, microgrid, battery sizing

Procedia PDF Downloads 208
39075 Building a Parametric Link between Mapping and Planning: A Sunlight-Adaptive Urban Green System Plan Formation Process

Authors: Chenhao Zhu

Abstract:

Quantitative mapping is playing a growing role in guiding urban planning, such as using a heat map created by CFX, CFD2000, or Envi-met, to adjust the master plan. However, there is no effective quantitative link between the mappings and planning formation. So, in many cases, the decision-making is still based on the planner's subjective interpretation and understanding of these mappings, which limits the improvement of scientific and accuracy brought by the quantitative mapping. Therefore, in this paper, an effort has been made to give a methodology of building a parametric link between the mapping and planning formation. A parametric planning process based on radiant mapping has been proposed for creating an urban green system. In the first step, a script is written in Grasshopper to build a road network and form the block, while the Ladybug Plug-in is used to conduct a radiant analysis in the form of mapping. Then, the research creatively transforms the radiant mapping from a polygon into a data point matrix, because polygon is hard to engage in the design formation. Next, another script is created to select the main green spaces from the road network based on the criteria of radiant intensity and connect the green spaces' central points to generate a green corridor. After that, a control parameter is introduced to adjust the corridor's form based on the radiant intensity. Finally, a green system containing greenspace and green corridor is generated under the quantitative control of the data matrix. The designer only needs to modify the control parameter according to the relevant research results and actual conditions to realize the optimization of the green system. This method can also be applied to much other mapping-based analysis, such as wind environment analysis, thermal environment analysis, and even environmental sensitivity analysis. The parameterized link between the mapping and planning will bring about a more accurate, objective, and scientific planning.

Keywords: parametric link, mapping, urban green system, radiant intensity, planning strategy, grasshopper

Procedia PDF Downloads 136
39074 A Safety Analysis Method for Multi-Agent Systems

Authors: Ching Louis Liu, Edmund Kazmierczak, Tim Miller

Abstract:

Safety analysis for multi-agent systems is complicated by the, potentially nonlinear, interactions between agents. This paper proposes a method for analyzing the safety of multi-agent systems by explicitly focusing on interactions and the accident data of systems that are similar in structure and function to the system being analyzed. The method creates a Bayesian network using the accident data from similar systems. A feature of our method is that the events in accident data are labeled with HAZOP guide words. Our method uses an Ontology to abstract away from the details of a multi-agent implementation. Using the ontology, our methods then constructs an “Interaction Map,” a graphical representation of the patterns of interactions between agents and other artifacts. Interaction maps combined with statistical data from accidents and the HAZOP classifications of events can be converted into a Bayesian Network. Bayesian networks allow designers to explore “what it” scenarios and make design trade-offs that maintain safety. We show how to use the Bayesian networks, and the interaction maps to improve multi-agent system designs.

Keywords: multi-agent system, safety analysis, safety model, integration map

Procedia PDF Downloads 413
39073 A Predictive MOC Solver for Water Hammer Waves Distribution in Network

Authors: A. Bayle, F. Plouraboué

Abstract:

Water Distribution Network (WDN) still suffers from a lack of knowledge about fast pressure transient events prediction, although the latter may considerably impact their durability. Accidental or planned operating activities indeed give rise to complex pressure interactions and may drastically modified the local pressure value generating leaks and, in rare cases, pipe’s break. In this context, a numerical predictive analysis is conducted to prevent such event and optimize network management. A couple of Python/FORTRAN 90, home-made software, has been developed using Method Of Characteristic (MOC) solving for water-hammer equations. The solver is validated by direct comparison with theoretical and experimental measurement in simple configurations whilst afterward extended to network analysis. The algorithm's most costly steps are designed for parallel computation. A various set of boundary conditions and energetic losses models are considered for the network simulations. The results are analyzed in both real and frequencies domain and provide crucial information on the pressure distribution behavior within the network.

Keywords: energetic losses models, method of characteristic, numerical predictive analysis, water distribution network, water hammer

Procedia PDF Downloads 219
39072 Analysis of an Alternative Data Base for the Estimation of Solar Radiation

Authors: Graciela Soares Marcelli, Elison Eduardo Jardim Bierhals, Luciane Teresa Salvi, Claudineia Brazil, Rafael Haag

Abstract:

The sun is a source of renewable energy, and its use as both a source of heat and light is one of the most promising energy alternatives for the future. To measure the thermal or photovoltaic systems a solar irradiation database is necessary. Brazil still has a reduced number of meteorological stations that provide frequency tests, as an alternative to the radio data platform, with reanalysis systems, quite significant. ERA-Interim is a global fire reanalysis by the European Center for Medium-Range Weather Forecasts (ECMWF). The data assimilation system used for the production of ERA-Interim is based on a 2006 version of the IFS (Cy31r2). The system includes a 4-dimensional variable analysis (4D-Var) with a 12-hour analysis window. The spatial resolution of the dataset is approximately 80 km at 60 vertical levels from the surface to 0.1 hPa. This work aims to make a comparative analysis between the ERA-Interim data and the data observed in the Solarimmetric Atlas of the State of Rio Grande do Sul, to verify its applicability in the absence of an observed data network. The analysis of the results obtained for a study region as an alternative to the energy potential of a given region.

Keywords: energy potential, reanalyses, renewable energy, solar radiation

Procedia PDF Downloads 157
39071 Static Analysis Deployment Model for Code Quality on Research and Development Projects of Software Development

Authors: Jeong-Hyun Park, Young-Sik Park, Hyo-Teag Jung

Abstract:

This paper presents static analysis deployment model for code quality on R&D Projects of SW Development. The proposed model includes the scope of R&D projects and index for static analysis of source code, operation model and execution process, environments and infrastructure system for R&D projects of SW development. There is the static analysis result of pilot project as case study based on the proposed deployment model and environment, and strategic considerations for success operation of the proposed static analysis deployment model for R&D Projects of SW Development. The proposed static analysis deployment model in this paper will be adapted and improved continuously for quality upgrade of R&D projects, and customer satisfaction of developed source codes and products.

Keywords: static analysis, code quality, coding rules, automation tool

Procedia PDF Downloads 516
39070 Numerical Analysis of Solar Cooling System

Authors: Nadia Allouache, Mohamed Belmedani

Abstract:

Energy source is a sustainable, totally inexhaustible and environmentally friendly alternative to the fossil fuels available. It is a renewable and economical energy that can be harnessed sustainably over the long term and thus stabilizes energy costs. Solar cooling technologies have been developed to decrease the augmentation electricity consumption for air conditioning and to displace the peak load during hot summer days. A numerical analysis of thermal and solar performances of an annular finned adsorber, which is the most important component of the adsorption solar refrigerating system, is considered in this work. Different adsorbent/adsorbate pairs, such as activated carbon AC35/methanol, activated carbon AC35/ethanol, and activated carbon BPL/Ammoniac, are undertaken in this study. The modeling of the adsorption cooling machine requires the resolution of the equation describing the energy and mass transfer in the tubular finned adsorber. The Wilson and Dubinin- Astakhov models of the solid-adsorbate equilibrium are used to calculate the adsorbed quantity. The porous medium and the fins are contained in the annular space, and the adsorber is heated by solar energy. Effects of key parameters on the adsorbed quantity and on the thermal and solar performances are analysed and discussed. The AC35/methanol pair is the best pair compared to BPL/Ammoniac and AC35/ethanol pairs in terms of system performance. The system performances are sensitive to the fin geometry. For the considered data measured for clear type days of July 2023 in Algeria and Morocco, the performances of the cooling system are very significant in Algeria.

Keywords: activated carbon AC35-methanol pair, activated carbon AC35-ethanol pair, activated carbon BPL-ammoniac pair, annular finned adsorber, performance coefficients, numerical analysis, solar cooling system

Procedia PDF Downloads 49
39069 FPGA Based Vector Control of PM Motor Using Sliding Mode Observer

Authors: Hanan Mikhael Dawood, Afaneen Anwer Abood Al-Khazraji

Abstract:

The paper presents an investigation of field oriented control strategy of Permanent Magnet Synchronous Motor (PMSM) based on hardware in the loop simulation (HIL) over a wide speed range. A sensorless rotor position estimation using sliding mode observer for permanent magnet synchronous motor is illustrated considering the effects of magnetic saturation between the d and q axes. The cross saturation between d and q axes has been calculated by finite-element analysis. Therefore, the inductance measurement regards the saturation and cross saturation which are used to obtain the suitable id-characteristics in base and flux weakening regions. Real time matrix multiplication in Field Programmable Gate Array (FPGA) using floating point number system is used utilizing Quartus-II environment to develop FPGA designs and then download these designs files into development kit. dSPACE DS1103 is utilized for Pulse Width Modulation (PWM) switching and the controller. The hardware in the loop results conducted to that from the Matlab simulation. Various dynamic conditions have been investigated.

Keywords: magnetic saturation, rotor position estimation, sliding mode observer, hardware in the loop (HIL)

Procedia PDF Downloads 520
39068 The Early Stages of the Standardisation of Finnish Building Sector

Authors: Anu Soikkeli

Abstract:

Early 20th century functionalism aimed at generalising living and rationalising construction, thus laying the foundation for the standardisation of construction components and products. From the 1930s onwards, all measurement and quality instructions for building products, different types of building components, descriptions of working methods complying with advisable building practises, planning, measurement and calculation guidelines, terminology, etc. were called standards. Standardisation was regarded as a necessary prerequisite for the mass production of housing. This article examines the early stages of standardisation in Finland in the 1940s and 1950s, as reflected on the working history of an individual architect, Erkki Koiso-Kanttila (1914-2006). In 1950 Koiso-Kanttila was appointed the Head of Design of the Finnish Association of Architects’ Building Standards Committee, a position which he held until 1958. His main responsibilities were the development of the RT Building Information File and compiling of the files.

Keywords: architecture, post WWII period, reconstruction, standardisation

Procedia PDF Downloads 411
39067 Agro Morphological Characterization of Vicia faba L. Accessions in the Kingdom of Saudi Arabia

Authors: Zia Amjad, Salem Safar Alghamdi

Abstract:

This experiment was carried out at student educational farm College of Food and Agriculture, KSU, kingdom of Saudi Arabia; in order to characterize 154 Vicia faba, characterization, PCA, ago-morphological diversity. Icia faba L. accessions were based on ipove and ibpgr descriptors. 24 agro-morphological characters including 11 quantitative and 13 qualitative were observed for genetic variation. All the results were analyzed using multivariate analysis i.e. principle component analysis. First 6 principle components with eigenvalue greater than one; accounted for 72% of available Vicia faba genetic diversity. However, first three components revealed more than 10% of genetic diversity each i.e. 22.36%, 15.86%, and 10.89% respectively. PCA distributed the V. faba accessions into different groups based on their performance for the characters under observation. PC-1 which represented 22.36% of the genetic diversity was positively associated with stipule spot pigmentation, intensity of streaks, pod degree of curvature and to some extent with 100 seed weight. PC-2 covered 15.86 of the genetic diversity and showed positive association for average seed weight per plant, pod length, number of seeds per plant, 100 seed weight, stipule spot pigmentation, intensity of streaks (same as in PC-1), and to some extent for pod degree of curvature and number of pods per plant. PC-3 revealed 10.89% of genetic diversity and expressed positive association for number of pods per plant and number of leaflets per plant.

Keywords: Vicia faba, characterization, PCA, ago-morphological diversity

Procedia PDF Downloads 448
39066 Design and Implementation of a Memory Safety Isolation Method Based on the Xen Cloud Environment

Authors: Dengpan Wu, Dan Liu

Abstract:

In view of the present cloud security problem has increasingly become one of the major obstacles hindering the development of the cloud computing, put forward a kind of memory based on Xen cloud environment security isolation technology implementation. And based on Xen virtual machine monitor system, analysis of the model of memory virtualization is implemented, using Xen memory virtualization system mechanism of super calls and grant table, based on the virtual machine manager internal implementation of access control module (ACM) to design the security isolation system memory. Experiments show that, the system can effectively isolate different customer domain OS between illegal access to memory data.

Keywords: cloud security, memory isolation, xen, virtual machine

Procedia PDF Downloads 397
39065 Data Integration in a GIS Geographic Information System Mapping of Agriculture in Semi-Arid Region of Setif, Algeria

Authors: W. Riahi, M. L. Mansour

Abstract:

Using tools of data processing such as geographic information system (GIS) for the contribution of the space management becomes more and more frequent. It allows collecting and analyzing diverse natural information relative to the same territory. Space technologies play crucial role in agricultural phenomenon analysis. For this, satellite images treatment were used to classify vegetation density and particularly agricultural areas in Setif province by making recourse to the Normalized Difference Vegetation Index (NDVI). This step was completed by mapping agricultural activities of the province by using ArcGIS.10 software in order to display an overall view and to realize spatial analysis of various themes combined between them which are chosen according to their strategic importance in different thematic maps. The synthesis map elaborately showed that geographic information system can contribute significantly to agricultural management by describing potentialities and development opportunities of production systems and agricultural sectors.

Keywords: GIS, satellite image, agriculture, NDVI, thematic map

Procedia PDF Downloads 417
39064 The Impact of Public Open Space System on Housing Price in Chicago

Authors: Si Chen, Le Zhang, Xian He

Abstract:

The research explored the influences of public open space system on housing price through hedonic models, in order to support better open space plans and economic policies. We have three initial hypotheses: 1) public open space system has an overall positive influence on surrounding housing prices. 2) Different public open space types have different levels of influence on motivating surrounding housing prices. 3) Walking and driving accessibilities from property to public open spaces have different statistical relation with housing prices. Cook County, Illinois, was chosen to be a study area since data availability, sufficient open space types, and long-term open space preservation strategies. We considered the housing attributes, driving and walking accessibility scores from houses to nearby public open spaces, and driving accessibility scores to hospitals as influential features and used real housing sales price in 2010 as a dependent variable in the built hedonic model. Through ordinary least squares (OLS) regression analysis, General Moran’s I analysis and geographically weighted regression analysis, we observed the statistical relations between public open spaces and housing sale prices in the three built hedonic models and confirmed all three hypotheses.

Keywords: hedonic model, public open space, housing sale price, regression analysis, accessibility score

Procedia PDF Downloads 128
39063 Established Novel Approach for Chemical Oxygen Demand Concentrations Measurement Based Mach-Zehner Interferometer Sensor

Authors: Su Sin Chong, Abdul Aziz Abdul Raman, Sulaiman Wadi Harun, Hamzah Arof

Abstract:

Chemical Oxygen Demand (COD) plays a vital role determination of an appropriate strategy for wastewater treatment including the control of the quality of an effluent. In this study, a new sensing method was introduced for the first time and developed to investigate chemical oxygen demand (COD) using a Mach-Zehner Interferometer (MZI)-based dye sensor. The sensor is constructed by bridging two single mode fibres (SMF1 and SMF2) with a short section (~20 mm) of multimode fibre (MMF) and was formed by tapering the MMF to generate evanescent field which is sensitive to perturbation of sensing medium. When the COD concentration increase takes effect will induce changes in output intensity and effective refractive index between the microfiber and the sensing medium. The adequacy of decisions based on COD values relies on the quality of the measurements. Therefore, the dual output response can be applied to the analytical procedure enhance measurement quality. This work presents a detailed assessment of the determination of COD values in synthetic wastewaters. Detailed models of the measurement performance, including sensitivity, reversibility, stability, and uncertainty were successfully validated by proficiency tests where supported on sound and objective criteria. Comparison of the standard method with the new proposed method was also conducted. This proposed sensor is compact, reliable and feasible to investigate the COD value.

Keywords: chemical oxygen demand, environmental sensing, Mach-Zehnder interferometer sensor, online monitoring

Procedia PDF Downloads 490
39062 An intelligent Troubleshooting System and Performance Evaluator for Computer Network

Authors: Iliya Musa Adamu

Abstract:

This paper seeks to develop an expert system that would troubleshoot computer network and evaluate the network system performance so as to reduce the workload on technicians and increase the efficiency and effectiveness of solutions proffered to computer network problems. The platform of the system was developed using ASP.NET, whereas the codes are implemented in Visual Basic and integrated with SQL Server 2005. The knowledge base was represented using production rule, whereas the searching method that was used in developing the network troubleshooting expert system is the forward-chaining-rule-based-system. This software tool offers the advantage of providing an immediate solution to most computer network problems encountered by computer users.

Keywords: expert system, forward chaining rule based system, network, troubleshooting

Procedia PDF Downloads 637
39061 The Ecological Role of Loligo forbesii in the Moray Firth Ecosystem, Northeast Scotland

Authors: Godwin A. Otogo, Sansanee Wangvoralak, Graham J. Pierce, Lee C. Hastie, Beth Scott

Abstract:

The squid Loligo forbesii is suspected to be an important species in marine food webs, as it can strongly impact its prey and be impacted upon by predation, competition, fishing and/or climate variability. To quantify these impacts in the food web, the measurement of its trophic position and ecological role within well-studied ecosystems is essential. An Ecopath model was balanced and run for the Moray Firth ecosystem and was used to investigate the significance of this squid’s trophic roles. The network analysis routine included in Ecopath with Ecosim (EwE) was used to estimate trophic interaction, system indicators (health condition and developmental stage) and food web features. Results indicated that within the Moray Firth squid occupy a top trophic position in the food web and also a major prey item for many other species. Results from Omnivory Index (OI) showed that squid is a generalized feeder transferring energy across wide trophic levels and is more important as a predator than that as a prey in the Moray Firth ecosystem. The results highlight the importance of taking squid into account in the management of Europe’s living marine resources.

Keywords: Squid, Loligo forbesii, Ecopath, Moray Firth, Trophic level

Procedia PDF Downloads 471
39060 Information Visualization Methods Applied to Nanostructured Biosensors

Authors: Osvaldo N. Oliveira Jr.

Abstract:

The control of molecular architecture inherent in some experimental methods to produce nanostructured films has had great impact on devices of various types, including sensors and biosensors. The self-assembly monolayers (SAMs) and the electrostatic layer-by-layer (LbL) techniques, for example, are now routinely used to produce tailored architectures for biosensing where biomolecules are immobilized with long-lasting preserved activity. Enzymes, antigens, antibodies, peptides and many other molecules serve as the molecular recognition elements for detecting an equally wide variety of analytes. The principles of detection are also varied, including electrochemical methods, fluorescence spectroscopy and impedance spectroscopy. In this presentation an overview will be provided of biosensors made with nanostructured films to detect antibodies associated with tropical diseases and HIV, in addition to detection of analytes of medical interest such as cholesterol and triglycerides. Because large amounts of data are generated in the biosensing experiments, use has been made of computational and statistical methods to optimize performance. Multidimensional projection techniques such as Sammon´s mapping have been shown more efficient than traditional multivariate statistical analysis in identifying small concentrations of anti-HIV antibodies and for distinguishing between blood serum samples of animals infected with two tropical diseases, namely Chagas´ disease and Leishmaniasis. Optimization of biosensing may include a combination of another information visualization method, the Parallel Coordinate technique, with artificial intelligence methods in order to identify the most suitable frequencies for reaching higher sensitivity using impedance spectroscopy. Also discussed will be the possible convergence of technologies, through which machine learning and other computational methods may be used to treat data from biosensors within an expert system for clinical diagnosis.

Keywords: clinical diagnosis, information visualization, nanostructured films, layer-by-layer technique

Procedia PDF Downloads 329
39059 Particle Size Analysis of Itagunmodi Southwestern Nigeria Alluvial Gold Ore Sample by Gaudin Schumann Method

Authors: Olaniyi Awe, Adelana R. Adetunji, Abraham Adeleke

Abstract:

Mining of alluvial gold ore by artisanal miners has been going on for decades at Itagunmodi, Southwestern Nigeria. In order to optimize the traditional panning gravity separation method commonly used in the area, a mineral particle size analysis study is critical. This study analyzed alluvial gold ore samples collected at identified five different locations in the area with a view to determine the ore particle size distributions. 500g measured of as-received alluvial gold ore sample was introduced into the uppermost sieve of an electrical sieve shaker consisting of sieves arranged in the order of decreasing nominal apertures of 5600μm, 3350μm, 2800μm, 355μm, 250μm, 125μm and 90μm, and operated for 20 minutes. The amount of material retained on each sieve was measured and tabulated for analysis. A screen analysis graph using the Gaudin Schuman method was drawn for each of the screen tests on the alluvial samples. The study showed that the percentages of fine particle size -125+90 μm fraction were 45.00%, 36.00%, 39.60%, 43.00% and 36.80% for the selected samples. These primary ore characteristic results provide reference data for the alluvial gold ore processing method selection, process performance measurement and optimization.

Keywords: alluvial gold ore, sieve shaker, particle size, Gaudin Schumann

Procedia PDF Downloads 41
39058 Arabic Light Word Analyser: Roles with Deep Learning Approach

Authors: Mohammed Abu Shquier

Abstract:

This paper introduces a word segmentation method using the novel BP-LSTM-CRF architecture for processing semantic output training. The objective of web morphological analysis tools is to link a formal morpho-syntactic description to a lemma, along with morpho-syntactic information, a vocalized form, a vocalized analysis with morpho-syntactic information, and a list of paradigms. A key objective is to continuously enhance the proposed system through an inductive learning approach that considers semantic influences. The system is currently under construction and development based on data-driven learning. To evaluate the tool, an experiment on homograph analysis was conducted. The tool also encompasses the assumption of deep binary segmentation hypotheses, the arbitrary choice of trigram or n-gram continuation probabilities, language limitations, and morphology for both Modern Standard Arabic (MSA) and Dialectal Arabic (DA), which provide justification for updating this system. Most Arabic word analysis systems are based on the phonotactic morpho-syntactic analysis of a word transmitted using lexical rules, which are mainly used in MENA language technology tools, without taking into account contextual or semantic morphological implications. Therefore, it is necessary to have an automatic analysis tool taking into account the word sense and not only the morpho-syntactic category. Moreover, they are also based on statistical/stochastic models. These stochastic models, such as HMMs, have shown their effectiveness in different NLP applications: part-of-speech tagging, machine translation, speech recognition, etc. As an extension, we focus on language modeling using Recurrent Neural Network (RNN); given that morphological analysis coverage was very low in dialectal Arabic, it is significantly important to investigate deeply how the dialect data influence the accuracy of these approaches by developing dialectal morphological processing tools to show that dialectal variability can support to improve analysis.

Keywords: NLP, DL, ML, analyser, MSA, RNN, CNN

Procedia PDF Downloads 34
39057 Degradation of Different Organic Contaminates Using Corona Discharge Plasma

Authors: A. H. El-Shazly, A. El-Tayeb, M. F. Elkady, Mona G. E. Ibrahim, Abdelazim M. Negm

Abstract:

In this paper, corona discharge plasma reactor was used for degradation of organic pollution in aqueous solutions in batch reactor. This work examines the possibility of increasing the organic pollution removal efficiency from wastewater using non-thermal plasma. Three types of organic pollution phenol, acid blue 25 and methylene blue are presented to investigate experimentally the amount of organic pollution removal efficiency from wastewater. Measurement results for phenol degradation percentage are 71% in 35 min and 96% when its residence time is 60 min. In addition, the degradation behavior of acid blue 25 utilizing dual pin-to-plate corona discharge plasma system displays a removal efficiency of 82% in 11 min. The complete decolorization was accomplished in 35 min for concentration of acid blue 25 up to 100 ppm. Furthermore, the methylene blue degradation touched up to 85% during 35 min treatment in corona discharge plasma a batch reactor system. The decolorization ratio, conductivity, corona current and discharge energy are considered at various concentration molarity for AlCl3, CaCl2, KCl and NaCl under different molar concentration. It was observed that the attendance of salts at the same concentration level considerably diminished the rate and the extent of decolorization. The research presented that the corona system could be positively utilized in a diversity of organically contaminated at diverse concentrations. Energy consumption requirements for decolorization was considered. The consequences will be valuable for designing the plasma treatment systems appropriate for industrial wastewaters.

Keywords: wastewater treatment, corona discharge, non-thermal plasma, organic pollution

Procedia PDF Downloads 328
39056 An Analytical Study of FRP-Concrete Bridge Superstructures

Authors: Wael I. Alnahhal

Abstract:

It is a major challenge to build a bridge superstructure that has long-term durability and low maintenance requirements. A solution to this challenge may be to use new materials or to implement new structural systems. Fiber reinforced polymer (FRP) composites have continued to play an important role in solving some of persistent problems in infrastructure applications because of its high specific strength, light weight, and durability. In this study, the concept of the hybrid FRP-concrete structural systems is applied to a bridge superstructure. The hybrid FRP-concrete bridge superstructure is intended to have durable, structurally sound, and cost effective hybrid system that will take full advantage of the inherent properties of both FRP materials and concrete. In this study, two hybrid FRP-concrete bridge systems were investigated. The first system consists of trapezoidal cell units forming a bridge superstructure. The second one is formed by arch cells. The two systems rely on using cellular components to form the core of the bridge superstructure, and an outer shell to warp around those cells to form the integral unit of the bridge. Both systems were investigated analytically by using finite element (FE) analysis. From the rigorous FE studies, it was concluded that first system is more efficient than the second.

Keywords: bridge superstructure, hybrid system, fiber reinforced polymer, finite element analysis

Procedia PDF Downloads 328
39055 An Automated Stock Investment System Using Machine Learning Techniques: An Application in Australia

Authors: Carol Anne Hargreaves

Abstract:

A key issue in stock investment is how to select representative features for stock selection. The objective of this paper is to firstly determine whether an automated stock investment system, using machine learning techniques, may be used to identify a portfolio of growth stocks that are highly likely to provide returns better than the stock market index. The second objective is to identify the technical features that best characterize whether a stock’s price is likely to go up and to identify the most important factors and their contribution to predicting the likelihood of the stock price going up. Unsupervised machine learning techniques, such as cluster analysis, were applied to the stock data to identify a cluster of stocks that was likely to go up in price – portfolio 1. Next, the principal component analysis technique was used to select stocks that were rated high on component one and component two – portfolio 2. Thirdly, a supervised machine learning technique, the logistic regression method, was used to select stocks with a high probability of their price going up – portfolio 3. The predictive models were validated with metrics such as, sensitivity (recall), specificity and overall accuracy for all models. All accuracy measures were above 70%. All portfolios outperformed the market by more than eight times. The top three stocks were selected for each of the three stock portfolios and traded in the market for one month. After one month the return for each stock portfolio was computed and compared with the stock market index returns. The returns for all three stock portfolios was 23.87% for the principal component analysis stock portfolio, 11.65% for the logistic regression portfolio and 8.88% for the K-means cluster portfolio while the stock market performance was 0.38%. This study confirms that an automated stock investment system using machine learning techniques can identify top performing stock portfolios that outperform the stock market.

Keywords: machine learning, stock market trading, logistic regression, cluster analysis, factor analysis, decision trees, neural networks, automated stock investment system

Procedia PDF Downloads 150
39054 Characterization of the Viscoelastic Behavior of Polymeric Composites

Authors: Abir Abdessalem, Sahbi Tamboura, J. Fitoussi, Hachmi Ben Daly, Abbas Tcharkhtchi

Abstract:

Dynamic mechanical analysis (DMA) is one of the most used experimental techniques to investigate the temperature and frequency dependence of the mechanical behavior of viscoelastic materials. The measured data are generally shifted by the application of the principle of the time– temperature superposition (TTS) to obtain the viscoelastic system’s master curve. The aim of this work is to show the methodology to define the horizontal shift factor to be applied to the storage modulus measured in order to indicate the validity of (TTS) principle for this material system. This principle was successfully used to determine the long-term properties of the Sheet Moulding Compound (SMC) composites.

Keywords: composite material, dynamic mechanical analysis, SMC composites, viscoelastic behavior, modeling

Procedia PDF Downloads 226
39053 Study on the Impact of Size and Position of the Shear Field in Determining the Shear Modulus of Glulam Beam Using Photogrammetry Approach

Authors: Niaz Gharavi, Hexin Zhang

Abstract:

The shear modulus of a timber beam can be determined using torsion test or shear field test method. The shear field test method is based on shear distortion measurement of the beam at the zone with the constant transverse load in the standardized four-point bending test. The current code of practice advises using two metallic arms act as an instrument to measure the diagonal displacement of the constructing square. The size and the position of the constructing square might influence the shear modulus determination. This study aimed to investigate the size and the position effect of the square in the shear field test method. A binocular stereo vision system has been employed to determine the 3D displacement of a grid of target points. Six glue laminated beams were produced and tested. Analysis of Variance (ANOVA) was performed on the acquired data to evaluate the significance of the size effect and the position effect of the square. The results have shown that the size of the square has a noticeable influence on the value of shear modulus, while, the position of the square within the area with the constant shear force does not affect the measured mean shear modulus.

Keywords: shear field test method, structural-sized test, shear modulus of Glulam beam, photogrammetry approach

Procedia PDF Downloads 287
39052 Power Integrity Analysis of Power Delivery System in High Speed Digital FPGA Board

Authors: Anil Kumar Pandey

Abstract:

Power plane noise is the most significant source of signal integrity (SI) issues in a high-speed digital design. In this paper, power integrity (PI) analysis of multiple power planes in a power delivery system of a 12-layer high-speed FPGA board is presented. All 10 power planes of HSD board are analyzed separately by using 3D Electromagnetic based PI solver, then the transient simulation is performed on combined PI data of all planes along with voltage regulator modules (VRMs) and 70 current drawing chips to get the board level power noise coupling on different high-speed signals. De-coupling capacitors are placed between power planes and ground to reduce power noise coupling with signals.

Keywords: power integrity, power-aware signal integrity analysis, electromagnetic simulation, channel simulation

Procedia PDF Downloads 432
39051 Mobile and Hot Spot Measurement with Optical Particle Counting Based Dust Monitor EDM264

Authors: V. Ziegler, F. Schneider, M. Pesch

Abstract:

With the EDM264, GRIMM offers a solution for mobile short- and long-term measurements in outdoor areas and at production sites. For research as well as permanent areal observations on a near reference quality base. The model EDM264 features a powerful and robust measuring cell based on optical particle counting (OPC) principle with all the advantages that users of GRIMM's portable aerosol spectrometers are used to. The system is embedded in a compact weather-protection housing with all-weather sampling, heated inlet system, data logger, and meteorological sensor. With TSP, PM10, PM4, PM2.5, PM1, and PMcoarse, the EDM264 provides all fine dust fractions real-time, valid for outdoor applications and calculated with the proven GRIMM enviro-algorithm, as well as six additional dust mass fractions pm10, pm2.5, pm1, inhalable, thoracic and respirable for IAQ and workplace measurements. This highly versatile instrument performs real-time monitoring of particle number, particle size and provides information on particle surface distribution as well as dust mass distribution. GRIMM's EDM264 has 31 equidistant size channels, which are PSL traceable. A high-end data logger enables data acquisition and wireless communication via LTE, WLAN, or wired via Ethernet. Backup copies of the measurement data are stored in the device directly. The rinsing air function, which protects the laser and detector in the optical cell, further increases the reliability and long term stability of the EDM264 under different environmental and climatic conditions. The entire sample volume flow of 1.2 L/min is analyzed by 100% in the optical cell, which assures excellent counting efficiency at low and high concentrations and complies with the ISO 21501-1standard for OPCs. With all these features, the EDM264 is a world-leading dust monitor for precise monitoring of particulate matter and particle number concentration. This highly reliable instrument is an indispensable tool for many users who need to measure aerosol levels and air quality outdoors, on construction sites, or at production facilities.

Keywords: aerosol research, aerial observation, fence line monitoring, wild fire detection

Procedia PDF Downloads 145
39050 Analysis of the Learners’ Responses of the Adjusted Rorschach Comprehensive System: Critical Psychological Perspective

Authors: Mokgadi Moletsane-Kekae, Robert Kananga Mukuna

Abstract:

The study focuses on the analysis of the Adjusted Rorschach Comprehensive System’s responses. The purpose of the study is to analyse the participants’ rate responses of the Adjusted Rorschach Comprehensive System with regards to critical psychology approach. The use of critical psychology theory in this study was crucial because it responds to the current inadequate western theory or practice in the field of psychology. The participants were learners in previously disadvantaged school in the Western Cape, South Africa. The study adopted a qualitative approach and a case study design. The study was grounded on interpretivist paradigm. The sample size comprised six learners (three boys and three girls, aged of 14 years) from historically disadvantaged school. The Adjusted Rorschach Comprehensive System (ARCS) administration procedure, biographical information, semi-structured interviews, and observation were used to collect data. Data was analysed using thematic framework. The study found out that, factors that increased the response rates during the administration of ARCS were, language, seating arrangement, drawing, viewing, and describing. The study recommended that, psychological test designers take into consideration the philosophy or worldviews of the local people for whom the test is designed to minimize low response rates.

Keywords: adjusted rorschach comprehensive system, critical psychology, learners, responses

Procedia PDF Downloads 374