Search results for: continuous speed profile data
25971 An Encapsulation of a Navigable Tree Position: Theory, Specification, and Verification
Authors: Nicodemus M. J. Mbwambo, Yu-Shan Sun, Murali Sitaraman, Joan Krone
Abstract:
This paper presents a generic data abstraction that captures a navigable tree position. The mathematical modeling of the abstraction encapsulates the current tree position, which can be used to navigate and modify the tree. The encapsulation of the tree position in the data abstraction specification avoids the use of explicit references and aliasing, thereby simplifying verification of (imperative) client code that uses the data abstraction. To ease the tasks of such specification and verification, a general tree theory, rich with mathematical notations and results, has been developed. The paper contains an example to illustrate automated verification ramifications. With sufficient tree theory development, automated proving seems plausible even in the absence of a special-purpose tree solver.Keywords: automation, data abstraction, maps, specification, tree, verification
Procedia PDF Downloads 16625970 Accurate Position Electromagnetic Sensor Using Data Acquisition System
Authors: Z. Ezzouine, A. Nakheli
Abstract:
This paper presents a high position electromagnetic sensor system (HPESS) that is applicable for moving object detection. The authors have developed a high-performance position sensor prototype dedicated to students’ laboratory. The challenge was to obtain a highly accurate and real-time sensor that is able to calculate position, length or displacement. An electromagnetic solution based on a two coil induction principal was adopted. The HPESS converts mechanical motion to electric energy with direct contact. The output signal can then be fed to an electronic circuit. The voltage output change from the sensor is captured by data acquisition system using LabVIEW software. The displacement of the moving object is determined. The measured data are transmitted to a PC in real-time via a DAQ (NI USB -6281). This paper also describes the data acquisition analysis and the conditioning card developed specially for sensor signal monitoring. The data is then recorded and viewed using a user interface written using National Instrument LabVIEW software. On-line displays of time and voltage of the sensor signal provide a user-friendly data acquisition interface. The sensor provides an uncomplicated, accurate, reliable, inexpensive transducer for highly sophisticated control systems.Keywords: electromagnetic sensor, accurately, data acquisition, position measurement
Procedia PDF Downloads 28525969 The Quality of the Presentation Influence the Audience Perceptions
Authors: Gilang Maulana, Dhika Rahma Qomariah, Yasin Fadil
Abstract:
Purpose: This research meant to measure the magnitude of the influence of the quality of the presentation to the targeted audience perception in catching information presentation. Design/Methodology/Approach: This research uses a quantitative research method. The kind of data that uses in this research is the primary data. The population in this research are students the economics faculty of Semarang State University. The sampling techniques uses in this research is purposive sampling. The retrieving data uses questionnaire on 30 respondents. The data analysis uses descriptive analysis. Result: The quality of presentation influential positive against perception of the audience. This proved that the more qualified presentation will increase the perception of the audience. Limitation: Respondents were limited to only 30 people.Keywords: quality of presentation, presentation, audience, perception, semarang state university
Procedia PDF Downloads 39225968 Mitigation of Seismic Forces Effect on Highway Bridge Using Aseismic Bearings
Authors: Kaoutar Zellat, Tahar Kadri
Abstract:
The purpose of new aseismic techniques is to provide an additional means of energy dissipation, thereby reducing the transmitted acceleration into the superstructure. In order to demonstrate the effectiveness of aseismic bearings technique and understand the behavior of seismically isolated bridges by such devices a three-span continuous deck bridge made of reinforced concrete is considered. The bridge is modeled as a discrete model and the relative displacements of the isolation bearing are crucial from the design point of view of isolation system and separation joints at the abutment level. The systems presented here are passive control systems and the results of some important experimental tests are also included. The results show that the base shear in the piers is significantly reduced for the isolated system as compared to the non isolated system in the both directions of the bridge. This indicates that the use of aseismic systems is effective in reducing the earthquake response of the bridge.Keywords: aseismic bearings, bridge isolation, bridge, seismic response
Procedia PDF Downloads 35925967 Gear Fault Diagnosis Based on Optimal Morlet Wavelet Filter and Autocorrelation Enhancement
Authors: Mohamed El Morsy, Gabriela Achtenová
Abstract:
Condition monitoring is used to increase machinery availability and machinery performance, whilst reducing consequential damage, increasing machine life, reducing spare parts inventories, and reducing breakdown maintenance. An efficient condition monitoring system provides early warning of faults by predicting them at an early stage. When a localized fault occurs in gears, the vibration signals always exhibit non-stationary behavior. The periodic impulsive feature of the vibration signal appears in the time domain and the corresponding gear mesh frequency (GMF) emerges in the frequency domain. However, one limitation of frequency-domain analysis is its inability to handle non-stationary waveform signals, which are very common when machinery faults occur. Particularly at the early stage of gear failure, the GMF contains very little energy and is often overwhelmed by noise and higher-level macro-structural vibrations. An effective signal processing method would be necessary to remove such corrupting noise and interference. In this paper, a new hybrid method based on optimal Morlet wavelet filter and autocorrelation enhancement is presented. First, to eliminate the frequency associated with interferential vibrations, the vibration signal is filtered with a band-pass filter determined by a Morlet wavelet whose parameters are selected or optimized based on maximum Kurtosis. Then, to further reduce the residual in-band noise and highlight the periodic impulsive feature, an autocorrelation enhancement algorithm is applied to the filtered signal. The test stand is equipped with three dynamometers; the input dynamometer serves as the internal combustion engine, the output dynamometers induce a load on the output joint shaft flanges. The pitting defect is manufactured on the tooth side of a gear of the fifth speed on the secondary shaft. The gearbox used for experimental measurements is of the type most commonly used in modern small to mid-sized passenger cars with transversely mounted powertrain and front wheel drive: a five-speed gearbox with final drive gear and front wheel differential. The results obtained from practical experiments prove that the proposed method is very effective for gear fault diagnosis.Keywords: wavelet analysis, pitted gear, autocorrelation, gear fault diagnosis
Procedia PDF Downloads 38825966 Influence of Some Technological Parameters on the Content of Voids in Composite during On-Line Consolidation with Filament Winding Technology
Authors: M. Stefanovska, B. Samakoski, S. Risteska, G. Maneski
Abstract:
In this study was performed in situ consolidation of polypropylene matrix/glass reinforced roving by combining heating systems and roll pressing. The commingled roving during hoop winding was winded on a cylindrical mandrel. The work also presents the advances made in the processing of these materials into composites by conventional technique filament winding. Experimental studies were performed with changing parameters – temperature, pressure and speed. Finally, it describes the investigation of the optimal processing conditions that maximize the mechanical properties of the composites. These properties are good enough for composites to be used as engineering materials in many structural applications.Keywords: commingled fiber, consolidation heat, filament winding, voids
Procedia PDF Downloads 26625965 Examining Statistical Monitoring Approach against Traditional Monitoring Techniques in Detecting Data Anomalies during Conduct of Clinical Trials
Authors: Sheikh Omar Sillah
Abstract:
Introduction: Monitoring is an important means of ensuring the smooth implementation and quality of clinical trials. For many years, traditional site monitoring approaches have been critical in detecting data errors but not optimal in identifying fabricated and implanted data as well as non-random data distributions that may significantly invalidate study results. The objective of this paper was to provide recommendations based on best statistical monitoring practices for detecting data-integrity issues suggestive of fabrication and implantation early in the study conduct to allow implementation of meaningful corrective and preventive actions. Methodology: Electronic bibliographic databases (Medline, Embase, PubMed, Scopus, and Web of Science) were used for the literature search, and both qualitative and quantitative studies were sought. Search results were uploaded into Eppi-Reviewer Software, and only publications written in the English language from 2012 were included in the review. Gray literature not considered to present reproducible methods was excluded. Results: A total of 18 peer-reviewed publications were included in the review. The publications demonstrated that traditional site monitoring techniques are not efficient in detecting data anomalies. By specifying project-specific parameters such as laboratory reference range values, visit schedules, etc., with appropriate interactive data monitoring, statistical monitoring can offer early signals of data anomalies to study teams. The review further revealed that statistical monitoring is useful to identify unusual data patterns that might be revealing issues that could impact data integrity or may potentially impact study participants' safety. However, subjective measures may not be good candidates for statistical monitoring. Conclusion: The statistical monitoring approach requires a combination of education, training, and experience sufficient to implement its principles in detecting data anomalies for the statistical aspects of a clinical trial.Keywords: statistical monitoring, data anomalies, clinical trials, traditional monitoring
Procedia PDF Downloads 7725964 Effect of Juvenile Hormone on Respiratory Metabolism during Non-Diapausing Sesamia cretica Wandering Larvae (Lepidoptera: Noctuidae)
Authors: E. A. Abdel-Hakim
Abstract:
The corn stemborer Sesamia cretica (Lederer), has been viewed in many parts of the world as a major pest of cultivated maize, graminaceous crops and sugarcane. Its life cycle is comprised of two different phases, one is the growth and developmental phase (non-diapause) and the other is diapause phase which takes place at the last larval instar. Several problems associated with the use of conventional insecticides, have strongly demonstrated the need for applying alternative safe compounds. Prominent among the prototypes of such prospective chemicals are the juvenoids; i.e. the insect (JH) mimics. In fact, the hormonal effect on metabolism has long been viewed as a secondary consequence of its direct action on specific energy-requiring biosynthetic mechanisms. Therefore, the present study was undertaken essentially in a rather systematic fashion as a contribution towards clarifying metabolic and energetic changes taking place during non-diapause wandering larvae as regulated by (JH) mimic. For this purpose, we applied two different doses of JH mimic (Ro 11-0111) in a single (standard) dose of 100µg or in a single dose of 20 µg/g bw in1µl acetone topically at the onset of nondiapause wandering larvae (WL). Energetic data were obtained by indirect calorimetry methods by conversion of respiratory gas exchange volumetric data, as measured manometrically using a Warburg constant respirometer, to caloric units (g-cal/g fw/h). The findings obtained can be given in brief; these treated larvae underwent supernumerary larval moults. However, this potential the wandering larvae proved to possess whereby restoration of larval programming for S. cretica to overcome stresses even at this critical developmental period. The results obtained, particularly with the high dose used, show that 98% wandering larvae were rescued to survive up to one month (vs. 5 days for normal controls), finally the formation of larval-adult intermediates. Also, the solvent controls had resulted in about 22% additional, but stationary moultings. The basal respiratory metabolism (O2 uptake and CO2 output) of the (WL), whether un-treated or larvae not had followed reciprocal U-shaped curves all along of their developmental duration. The lowest points stood nearly to the day of prepupal formation (571±187 µl O2/gfw/h and 553±181 µl CO2/gfw/h) during un-treated in contrast to the larvae treated with JH (210±48 µl O2/gfw/h and 335±81 µl CO2/gfw/h). Un-treated (normal) larvae proved to utilize carbohydrates as the principal source for energy supply; being fully oxidised without sparing any appreciable amount for endergonic conversion to fats. While, the juvenoid-treated larvae and compared with the acetone-treated control equivalents, there existed no distinguishable differences between them; both had been observed utilising carbohydrates as the sole source of energy demand and converting endergonically almost similar percentages to fats. The overall profile, treated and un-treated (WL) utilized carbohydrates as the principal source for energy demand during this stage.Keywords: juvenile hormone, respiratory metabolism, Sesamia cretica, wandering phase
Procedia PDF Downloads 29425963 Lateral Torsional Buckling of Steel Thin-Walled Beams with Lateral Restraints
Authors: Ivan Balázs, Jindřich Melcher
Abstract:
Metal thin-walled members have been widely used in building industry. Usually they are utilized as purlins, girts or ceiling beams. Due to slenderness of thin-walled cross-sections these structural members are prone to stability problems (e.g. flexural buckling, lateral torsional buckling). If buckling is not constructionally prevented their resistance is limited by buckling strength. In practice planar members of roof or wall cladding can be attached to thin-walled members. These elements reduce displacement of thin-walled members and therefore increase their buckling strength. If this effect is taken into static assessment more economical sections of thin-walled members might be utilized and certain savings of material might be achieved. This paper focuses on problem of determination of critical load of steel thin-walled beams with lateral continuous restraint which is crucial for lateral torsional buckling assessment.Keywords: beam, buckling, numerical analysis, stability, steel
Procedia PDF Downloads 33025962 An ALM Matrix Completion Algorithm for Recovering Weather Monitoring Data
Authors: Yuqing Chen, Ying Xu, Renfa Li
Abstract:
The development of matrix completion theory provides new approaches for data gathering in Wireless Sensor Networks (WSN). The existing matrix completion algorithms for WSN mainly consider how to reduce the sampling number without considering the real-time performance when recovering the data matrix. In order to guarantee the recovery accuracy and reduce the recovery time consumed simultaneously, we propose a new ALM algorithm to recover the weather monitoring data. A lot of experiments have been carried out to investigate the performance of the proposed ALM algorithm by using different parameter settings, different sampling rates and sampling models. In addition, we compare the proposed ALM algorithm with some existing algorithms in the literature. Experimental results show that the ALM algorithm can obtain better overall recovery accuracy with less computing time, which demonstrate that the ALM algorithm is an effective and efficient approach for recovering the real world weather monitoring data in WSN.Keywords: wireless sensor network, matrix completion, singular value thresholding, augmented Lagrange multiplier
Procedia PDF Downloads 38425961 Field Production Data Collection, Analysis and Reporting Using Automated System
Authors: Amir AlAmeeri, Mohamed Ibrahim
Abstract:
Various data points are constantly being measured in the production system, and due to the nature of the wells, these data points, such as pressure, temperature, water cut, etc.., fluctuations are constant, which requires high frequency monitoring and collection. It is a very difficult task to analyze these parameters manually using spreadsheets and email. An automated system greatly enhances efficiency, reduce errors, the need for constant emails which take up disk space, and frees up time for the operator to perform other critical tasks. Various production data is being recorded in an oil field, and this huge volume of data can be seen as irrelevant to some, especially when viewed on its own with no context. In order to fully utilize all this information, it needs to be properly collected, verified and stored in one common place and analyzed for surveillance and monitoring purposes. This paper describes how data is recorded by different parties and departments in the field, and verified numerous times as it is being loaded into a repository. Once it is loaded, a final check is done before being entered into a production monitoring system. Once all this is collected, various calculations are performed to report allocated production. Calculated production data is used to report field production automatically. It is also used to monitor well and surface facility performance. Engineers can use this for their studies and analyses to ensure field is performing as it should be, predict and forecast production, and monitor any changes in wells that could affect field performance.Keywords: automation, oil production, Cheleken, exploration and production (E&P), Caspian Sea, allocation, forecast
Procedia PDF Downloads 15625960 Time Series Regression with Meta-Clusters
Authors: Monika Chuchro
Abstract:
This paper presents a preliminary attempt to apply classification of time series using meta-clusters in order to improve the quality of regression models. In this case, clustering was performed as a method to obtain a subgroups of time series data with normal distribution from inflow into waste water treatment plant data which Composed of several groups differing by mean value. Two simple algorithms: K-mean and EM were chosen as a clustering method. The rand index was used to measure the similarity. After simple meta-clustering, regression model was performed for each subgroups. The final model was a sum of subgroups models. The quality of obtained model was compared with the regression model made using the same explanatory variables but with no clustering of data. Results were compared by determination coefficient (R2), measure of prediction accuracy mean absolute percentage error (MAPE) and comparison on linear chart. Preliminary results allows to foresee the potential of the presented technique.Keywords: clustering, data analysis, data mining, predictive models
Procedia PDF Downloads 46625959 Assessment of the Soils Pollution Level of the Open Mine and Tailing Dump of Surrounding Territories of Akhtala Ore Processing Combine by Heavy Metals
Authors: K. A. Ghazaryan, T. H. Derdzyan
Abstract:
For assessment of the soils pollution level of the open mine and tailing dump of surrounding territories of Akhtala ore processing combine by heavy metals in 2013 collected soil samples and analyzed for different heavy metals, such as Cu, Zn, Pb, Ni and Cd. The main soil type in the study sites was the mountain cambisol. To classify soil pollution level contamination indices like Contamination factors (Cf), Degree of contamination (Cd), Pollution load index (PLI) and Geoaccumulation index (I-geo) are calculated. The distribution pattern of trace metals in the soil profile according to I geo, Cf and Cd values shows that the soil is very polluted. And also the PLI values for the 19 sites were >1, which indicates deterioration of site quality.Keywords: soils pollution, heavy metal, geoaccumulation index, pollution load index, contamination factor
Procedia PDF Downloads 43425958 Python Implementation for S1000D Applicability Depended Processing Model - SALERNO
Authors: Theresia El Khoury, Georges Badr, Amir Hajjam El Hassani, Stéphane N’Guyen Van Ky
Abstract:
The widespread adoption of machine learning and artificial intelligence across different domains can be attributed to the digitization of data over several decades, resulting in vast amounts of data, types, and structures. Thus, data processing and preparation turn out to be a crucial stage. However, applying these techniques to S1000D standard-based data poses a challenge due to its complexity and the need to preserve logical information. This paper describes SALERNO, an S1000d AppLicability dEpended pRocessiNg mOdel. This python-based model analyzes and converts the XML S1000D-based files into an easier data format that can be used in machine learning techniques while preserving the different logic and relationships in files. The model parses the files in the given folder, filters them, and extracts the required information to be saved in appropriate data frames and Excel sheets. Its main idea is to group the extracted information by applicability. In addition, it extracts the full text by replacing internal and external references while maintaining the relationships between files, as well as the necessary requirements. The resulting files can then be saved in databases and used in different models. Documents in both English and French languages were tested, and special characters were decoded. Updates on the technical manuals were taken into consideration as well. The model was tested on different versions of the S1000D, and the results demonstrated its ability to effectively handle the applicability, requirements, references, and relationships across all files and on different levels.Keywords: aeronautics, big data, data processing, machine learning, S1000D
Procedia PDF Downloads 15725957 In Vivo Evaluation of Exposure to Electromagnetic Fields at 27 GHz (5G) of Danio Rerio: A Preliminary Study
Authors: Elena Maria Scalisi, Roberta Pecoraro, Martina Contino, Sara Ignoto, Carmelo Iaria, Santi Concetto Pavone, Gino Sorbello, Loreto Di Donato, Maria Violetta Brundo
Abstract:
5G Technology is evolving to satisfy a variety of service requirements that may allow high data-rate connections (1Gbps) and lower latency times than current (<1ms). In order to support a high data transmission speed and a high traffic service for eMBB (enhanced mobile broadband) use cases, 5G systems have the characteristic of using different frequency bands of the radio wave spectrum (700 MHz, 3.6-3.8 GHz and 26.5-27.5 GHz), thus taking advantage of higher frequencies than previous mobile radio generations (1G-4G). However, waves at higher frequencies have a lower capacity to propagate in free space and therefore, in order to guarantee the capillary coverage of the territory for high reliability applications, it will be necessary to install a large number of repeaters. Following the introduction of this new technology, there has been growing concern over the past few months about possible harmful effects on human health. The aim of this preliminary study is to evaluate possible short term effects induced by 5G-millimeter waves on embryonic development and early life stages of Danio rerio by Z-FET. We exposed developing zebrafish at frequency of 27 GHz, with a standard pyramidal horn antenna placed at 15 cm far from the samples holder ensuring an incident power density of 10 mW/cm2. During the exposure cycle, from 6 h post fertilization (hpf) to 96 hpf, we measured a different morphological endpoints every 24 hours. Zebrafish embryo toxicity test (Z-FET) is a short term test, carried out on fertilized eggs of zebrafish and it represents an effective alternative to acute test with adult fish (OECD, 2013). We have observed that 5G did not reveal significant impacts on mortality nor on morphology because exposed larvae showed a normal detachment of the tail, presence of heartbeat, well-organized somites, therefore hatching rate was lower than untreated larvae even at 48 h of exposure. Moreover, the immunohistochemical analysis performed on larvae showed a negativity to the HSP-70 expression used as a biomarkers. This is a preliminary study on evaluation of potential toxicity induced by 5G and it seems appropriate to underline the importance that further studies would take, aimed at clarifying the probable real risk of exposure to electromagnetic fields.Keywords: Biomarker of exposure, embryonic development, 5G waves, zebrafish embryo toxicity test
Procedia PDF Downloads 12925956 Life Prediction Method of Lithium-Ion Battery Based on Grey Support Vector Machines
Authors: Xiaogang Li, Jieqiong Miao
Abstract:
As for the problem of the grey forecasting model prediction accuracy is low, an improved grey prediction model is put forward. Firstly, use trigonometric function transform the original data sequence in order to improve the smoothness of data , this model called SGM( smoothness of grey prediction model), then combine the improved grey model with support vector machine , and put forward the grey support vector machine model (SGM - SVM).Before the establishment of the model, we use trigonometric functions and accumulation generation operation preprocessing data in order to enhance the smoothness of the data and weaken the randomness of the data, then use support vector machine (SVM) to establish a prediction model for pre-processed data and select model parameters using genetic algorithms to obtain the optimum value of the global search. Finally, restore data through the "regressive generate" operation to get forecasting data. In order to prove that the SGM-SVM model is superior to other models, we select the battery life data from calce. The presented model is used to predict life of battery and the predicted result was compared with that of grey model and support vector machines.For a more intuitive comparison of the three models, this paper presents root mean square error of this three different models .The results show that the effect of grey support vector machine (SGM-SVM) to predict life is optimal, and the root mean square error is only 3.18%. Keywords: grey forecasting model, trigonometric function, support vector machine, genetic algorithms, root mean square errorKeywords: Grey prediction model, trigonometric functions, support vector machines, genetic algorithms, root mean square error
Procedia PDF Downloads 46125955 Performance and Lifetime of Tandem Organic Solar Cells
Authors: Guillaume Schuchardt, Solenn Berson, Gerard Perrier
Abstract:
Multi-junction solar cell configurations, where two sub-cells with complementary absorption are stacked and connected in series, offer an exciting approach to tackle the single junction limitations of organic solar cells and improve their power conversion efficiency. However, the augmentation of the number of layers has, as a consequence, to increase the risk of reducing the lifetime of the cell due to the ageing phenomena present at the interfaces. In this work, we study the intrinsic degradation mechanisms, under continuous illumination AM1.5G, inert atmosphere and room temperature, in single and tandem organic solar cells using Impedance Spectroscopy, IV Curves, External Quantum Efficiency, Steady-State Photocarrier Grating, Scanning Kelvin Probe and UV-Visible light.Keywords: single and tandem organic solar cells, intrinsic degradation mechanisms, characterization: SKP, EQE, SSPG, UV-Visible, Impedance Spectroscopy, optical simulation
Procedia PDF Downloads 36225954 Development of a Data Security Model Using Steganography
Authors: Terungwa Simon Yange, Agana Moses A.
Abstract:
This paper studied steganography and designed a simplistic approach to a steganographic tool for hiding information in image files with the view of addressing the security challenges with data by hiding data from unauthorized users to improve its security. The Structured Systems Analysis and Design Method (SSADM) was used in this work. The system was developed using Java Development Kit (JDK) 1.7.0_10 and MySQL Server as its backend. The system was tested with some hypothetical health records which proved the possibility of protecting data from unauthorized users by making it secret so that its existence cannot be easily recognized by fraudulent users. It further strengthens the confidentiality of patient records kept by medical practitioners in the health setting. In conclusion, this work was able to produce a user friendly steganography software that is very fast to install and easy to operate to ensure privacy and secrecy of sensitive data. It also produced an exact copy of the original image and the one carrying the secret message when compared with each.Keywords: steganography, cryptography, encryption, decryption, secrecy
Procedia PDF Downloads 26625953 Analysis of Citation Rate and Data Reuse for Openly Accessible Biodiversity Datasets on Global Biodiversity Information Facility
Authors: Nushrat Khan, Mike Thelwall, Kayvan Kousha
Abstract:
Making research data openly accessible has been mandated by most funders over the last 5 years as it promotes reproducibility in science and reduces duplication of effort to collect the same data. There are evidence that articles that publicly share research data have higher citation rates in biological and social sciences. However, how and whether shared data is being reused is not always intuitive as such information is not easily accessible from the majority of research data repositories. This study aims to understand the practice of data citation and how data is being reused over the years focusing on biodiversity since research data is frequently reused in this field. Metadata of 38,878 datasets including citation counts were collected through the Global Biodiversity Information Facility (GBIF) API for this purpose. GBIF was used as a data source since it provides citation count for datasets, not a commonly available feature for most repositories. Analysis of dataset types, citation counts, creation and update time of datasets suggests that citation rate varies for different types of datasets, where occurrence datasets that have more granular information have higher citation rates than checklist and metadata-only datasets. Another finding is that biodiversity datasets on GBIF are frequently updated, which is unique to this field. Majority of the datasets from the earliest year of 2007 were updated after 11 years, with no dataset that was not updated since creation. For each year between 2007 and 2017, we compared the correlations between update time and citation rate of four different types of datasets. While recent datasets do not show any correlations, 3 to 4 years old datasets show weak correlation where datasets that were updated more recently received high citations. The results are suggestive that it takes several years to cumulate citations for research datasets. However, this investigation found that when searched on Google Scholar or Scopus databases for the same datasets, the number of citations is often not the same as GBIF. Hence future aim is to further explore the citation count system adopted by GBIF to evaluate its reliability and whether it can be applicable to other fields of studies as well.Keywords: data citation, data reuse, research data sharing, webometrics
Procedia PDF Downloads 17825952 Significance of Transient Data and Its Applications in Turbine Generators
Authors: Chandra Gupt Porwal, Preeti C. Porwal
Abstract:
Transient data reveals much about the machine's condition that steady-state data cannot. New technologies make this information much more available for evaluating the mechanical integrity of a machine train. Recent surveys at various stations indicate that simplicity is preferred over completeness in machine audits throughout the power generation industry. This is most clearly shown by the number of rotating machinery predictive maintenance programs in which only steady-state vibration amplitude is trended while important transient vibration data is not even acquired. Efforts have been made to explain what transient data is, its importance, the types of plots used for its display, and its effective utilization for analysis. In order to demonstrate the value of measuring transient data and its practical application in rotating machinery for resolving complex and persistent issues with turbine generators, the author presents a few case studies that highlight the presence of rotor instabilities due to the shaft moving towards the bearing centre in a 100 MM LMZ unit located in the Northern Capital Region (NCR), heavy misalignment noticed—especially after 2993 rpm—caused by loose coupling bolts, which prevented the machine from being synchronized for more than four months in a 250 MW KWU unit in the Western Region (WR), and heavy preload noticed at Intermediate pressure turbine (IPT) bearing near HP- IP coupling, caused by high points on coupling faces at a 500 MW KWU unit in the Northern region (NR), experienced at Indian power plants.Keywords: transient data, steady-state-data, intermediate -pressure-turbine, high-points
Procedia PDF Downloads 6925951 Geographic Information System for District Level Energy Performance Simulations
Authors: Avichal Malhotra, Jerome Frisch, Christoph van Treeck
Abstract:
The utilization of semantic, cadastral and topological data from geographic information systems (GIS) has exponentially increased for building and urban-scale energy performance simulations. Urban planners, simulation scientists, and researchers use virtual 3D city models for energy analysis, algorithms and simulation tools. For dynamic energy simulations at city and district level, this paper provides an overview of the available GIS data models and their levels of detail. Adhering to different norms and standards, these models also intend to describe building and construction industry data. For further investigations, CityGML data models are considered for simulations. Though geographical information modelling has considerably many different implementations, extensions of virtual city data can also be made for domain specific applications. Highlighting the use of the extended CityGML models for energy researches, a brief introduction to the Energy Application Domain Extension (ADE) along with its significance is made. Consequently, addressing specific input simulation data, a workflow using Modelica underlining the usage of GIS information and the quantification of its significance over annual heating energy demand is presented in this paper.Keywords: CityGML, EnergyADE, energy performance simulation, GIS
Procedia PDF Downloads 16825950 Visual Analytics in K 12 Education: Emerging Dimensions of Complexity
Authors: Linnea Stenliden
Abstract:
The aim of this paper is to understand emerging learning conditions, when a visual analytics is implemented and used in K 12 (education). To date, little attention has been paid to the role visual analytics (digital media and technology that highlight visual data communication in order to support analytical tasks) can play in education, and to the extent to which these tools can process actionable data for young students. This study was conducted in three public K 12 schools, in four social science classes with students aged 10 to 13 years, over a period of two to four weeks at each school. Empirical data were generated using video observations and analyzed with help of metaphors by Latour. The learning conditions are found to be distinguished by broad complexity characterized by four dimensions. These emerge from the actors’ deeply intertwined relations in the activities. The paper argues in relation to the found dimensions that novel approaches to teaching and learning could benefit students’ knowledge building as they work with visual analytics, analyzing visualized data.Keywords: analytical reasoning, complexity, data use, problem space, visual analytics, visual storytelling, translation
Procedia PDF Downloads 37625949 Quality Assurance in Higher Education: Doha Institute for Graduate Studies as a Case Study
Authors: Ahmed Makhoukh
Abstract:
Quality assurance (QA) has recently become a common practice, which is endorsed by most Higher Education (HE) institutions worldwide, due to the pressure of internal and external forces. One of the aims of this quality movement is to make the contribution of university education to socio-economic development highly significant. This entails that graduates are currently required have a high-quality profile, i.e., to be competent and master the 21st-century skills needed in the labor market. This wave of change, mostly imposed by globalization, has the effect that university education should be learner-centered in order to satisfy the different needs of students and meet the expectations of other stakeholders. Such a shift of focus on the student learning outcomes has led HE institutions to reconsider their strategic planning, their mission, the curriculum, the pedagogical competence of the academic staff, among other elements. To ensure that the overall institutional performance is on the right way, a QA system should be established to assume this task of checking regularly the extent to which the set of standards of evaluation are strictly respected as expected. This operation of QA has the advantage of proving the accountability of the institution, gaining the trust of the public with transparency and enjoying an international recognition. This is the case of Doha Institute (DI) for Graduate Studies, in Qatar, the object of the present study. The significance of this contribution is to show that the conception of quality has changed in this digital age, and the need to integrate a department responsible for QA in every HE institution to ensure educational quality, enhance learners and achieve academic leadership. Thus, to undertake the issue of QA in DI for Graduate Studies, an elite university (in the academic sense) that focuses on a small and selected number of students, a qualitative method will be adopted in the description and analysis of the data (document analysis). In an attempt to investigate the extent to which QA is achieved in Doha Institute for Graduate Studies, three broad indicators will be evaluated (input, process and learning outcomes). This investigation will be carried out in line with the UK Quality Code for Higher Education represented by Quality Assurance Agency (QAA).Keywords: accreditation, higher education, quality, quality assurance, standards
Procedia PDF Downloads 14725948 Calibration and Validation of ArcSWAT Model for Estimation of Surface Runoff and Sediment Yield from Dhangaon Watershed
Authors: M. P. Tripathi, Priti Tiwari
Abstract:
Soil and Water Assessment Tool (SWAT) is a distributed parameter continuous time model and was tested on daily and fortnightly basis for a small agricultural watershed (Dhangaon) of Chhattisgarh state in India. The SWAT model recently interfaced with ArcGIS and called as ArcSWAT. The watershed and sub-watershed boundaries, drainage networks, slope and texture maps were generated in the environment of ArcGIS of ArcSWAT. Supervised classification method was used for land use/cover classification from satellite imageries of the years 2009 and 2012. Manning's roughness coefficient 'n' for overland flow and channel flow and Fraction of Field Capacity (FFC) were calibrated for monsoon season of the years 2009 and 2010. The model was validated on a daily basis for the years 2011 and 2012 by using the observed daily rainfall and temperature data. Calibration and validation results revealed that the model was predicting the daily surface runoff and sediment yield satisfactorily. Sensitivity analysis showed that the annual sediment yield was inversely proportional to the overland and channel 'n' values whereas; annual runoff and sediment yields were directly proportional to the FFC. The model was also tested (calibrated and validated) for the fortnightly runoff and sediment yield for the year 2009-10 and 2011-12, respectively. Simulated values of fortnightly runoff and sediment yield for the calibration and validation years compared well with their observed counterparts. The calibration and validation results revealed that the ArcSWAT model could be used for identification of critical sub-watershed and for developing management scenarios for the Dhangaon watershed. Further, the model should be tested for simulating the surface runoff and sediment yield using generated rainfall and temperature before applying it for developing the management scenario for the critical or priority sub-watersheds.Keywords: watershed, hydrologic and water quality, ArcSWAT model, remote sensing, GIS, runoff and sediment yield
Procedia PDF Downloads 37925947 Effectiveness of the Model in the Development of Teaching Materials for Malay Language in Primary Schools in Singapore
Authors: Salha Mohamed Hussain
Abstract:
As part of the review on the Malay Language curriculum and pedagogy in Singapore conducted in 2010, some recommendations were made to nurture active learners who are able to use the Malay Language efficiently in their daily lives. In response to the review, a new Malay Language teaching and learning package for primary school, called CEKAP (Cungkil – Elicit; Eksplorasi – Exploration; Komunikasi – Communication; Aplikasi – Application; Penilaian – Assessment), was developed from 2012 and implemented for Primary 1 in all primary schools from 2015. Resources developed in this package include the text book, activity book, teacher’s guide, big books, small readers, picture cards, flash cards, a game kit and Information and Communication Technology (ICT) resources. The development of the CEKAP package is continuous until 2020. This paper will look at a model incorporated in the development of the teaching materials in the new Malay Language Curriculum for Primary Schools and the rationale for each phase of development to ensure that the resources meet the needs of every pupil in the teaching and learning of Malay Language in the primary schools. This paper will also focus on the preliminary findings of the effectiveness of the model based on the feedback given by members of the working and steering committees. These members are academicians and educators who were appointed by the Ministry of Education to provide professional input on the soundness of pedagogical approach proposed in the revised syllabus and to make recommendations on the content of the new instructional materials. Quantitative data is derived from the interviews held with these members to gather their input on the model. Preliminary findings showed that the members provided positive feedback on the model and that the comprehensive process has helped to develop good and effective instructional materials for the schools. Some recommendations were also gathered from the interview sessions. This research hopes to provide useful information to those involved in the planning of materials development for teaching and learning.Keywords: Malay language, materials development, model, primary school
Procedia PDF Downloads 11225946 Achieving High Renewable Energy Penetration in Western Australia Using Data Digitisation and Machine Learning
Authors: A. D. Tayal
Abstract:
The energy industry is undergoing significant disruption. This research outlines that, whilst challenging; this disruption is also an emerging opportunity for electricity utilities. One such opportunity is leveraging the developments in data analytics and machine learning. As the uptake of renewable energy technologies and complimentary control systems increases, electricity grids will likely transform towards dense microgrids with high penetration of renewable generation sources, rich in network and customer data, and linked through intelligent, wireless communications. Data digitisation and analytics have already impacted numerous industries, and its influence on the energy sector is growing, as computational capabilities increase to manage big data, and as machines develop algorithms to solve the energy challenges of the future. The objective of this paper is to address how far the uptake of renewable technologies can go given the constraints of existing grid infrastructure and provides a qualitative assessment of how higher levels of renewable energy penetration can be facilitated by incorporating even broader technological advances in the fields of data analytics and machine learning. Western Australia is used as a contextualised case study, given its abundance and diverse renewable resources (solar, wind, biomass, and wave) and isolated networks, making a high penetration of renewables a feasible target for policy makers over coming decades.Keywords: data, innovation, renewable, solar
Procedia PDF Downloads 36425945 Comparison of Cu Nanoparticle Formation and Properties with and without Surrounding Dielectric
Authors: P. Dubcek, B. Pivac, J. Dasovic, V. Janicki, S. Bernstorff
Abstract:
When grown only to nanometric sizes, metallic particles (e.g. Ag, Au and Cu) exhibit specific optical properties caused by the presence of plasmon band. The plasmon band represents collective oscillation of the conduction electrons, and causes a narrow band absorption of light in the visible range. When the nanoparticles are embedded in a dielectric, they also cause modifications of dielectrics optical properties. This can be fine-tuned by tuning the particle size. We investigated Cu nanoparticle growth with and without surrounding dielectric (SiO2 capping layer). The morphology and crystallinity were investigated by GISAXS and GIWAXS, respectively. Samples were produced by high vacuum thermal evaporation of Cu onto monocrystalline silicon substrate held at room temperature, 100°C or 180°C. One series was in situ capped by 10nm SiO2 layer. Additionally, samples were annealed at different temperatures up to 550°C, also in high vacuum. The room temperature deposited samples annealed at lower temperatures exhibit continuous film structure: strong oscillations in the GISAXS intensity are present especially in the capped samples. At higher temperatures enhanced surface dewetting and Cu nanoparticles (nanoislands) formation partially destroy the flatness of the interface. Therefore the particle type of scattering is enhanced, while the film fringes are depleted. However, capping layer hinders particle formation, and continuous film structure is preserved up to higher annealing temperatures (visible as strong and persistent fringes in GISAXS), compared to the non- capped samples. According to GISAXS, lateral particle sizes are reduced at higher temperatures, while particle height is increasing. This is ascribed to close packing of the formed particles at lower temperatures, and GISAXS deduced sizes are partially the result of the particle agglomerate dimensions. Lateral maxima in GISAXS are an indication of good positional correlation, and the particle to particle distance is increased as the particles grow with temperature elevation. This coordination is much stronger in the capped and lower temperature deposited samples. The dewetting is much more vigorous in the non-capped sample, and since nanoparticles are formed in a range of sizes, correlation is receding both with deposition and annealing temperature. Surface topology was checked by atomic force microscopy (AFM). Capped sample's surfaces were smoother and lateral size of the surface features were larger compared to the non-capped samples. Altogether, AFM results suggest somewhat larger particles and wider size distribution, and this can be attributed to the difference in probe size. Finally, the plasmonic effect was monitored by UV-Vis reflectance spectroscopy, and relative weak plasmonic effect could be explained by uncomplete dewetting or partial interconnection of the formed particles.Keywords: coper, GISAXS, nanoparticles, plasmonics
Procedia PDF Downloads 12325944 Pivoting to Fortify our Digital Self: Revealing the Need for Personal Cyber Insurance
Authors: Richard McGregor, Carmen Reaiche, Stephen Boyle
Abstract:
Cyber threats are a relatively recent phenomenon and offer cyber insurers a dynamic and intelligent peril. As individuals en mass become increasingly digitally dependent, Personal Cyber Insurance (PCI) offers an attractive option to mitigate cyber risk at a personal level. This abstract proposes a literature review that conceptualises a framework for siting Personal Cyber Insurance (PCI) within the context of cyberspace. The lack of empirical research within this domain demonstrates an immediate need to define the scope of PCI to allow cyber insurers to understand personal cyber risk threats and vectors, customer awareness, capabilities, and their associated needs. Additionally, this will allow cyber insurers to conceptualise appropriate frameworks allowing effective management and distribution of PCI products and services within a landscape often in-congruent with risk attributes commonly associated with traditional personal line insurance products. Cyberspace has provided significant improvement to the quality of social connectivity and productivity during past decades and allowed enormous capability uplift of information sharing and communication between people and communities. Conversely, personal digital dependency furnish ample opportunities for adverse cyber events such as data breaches and cyber-attacksthus introducing a continuous and insidious threat of omnipresent cyber risk–particularly since the advent of the COVID-19 pandemic and wide-spread adoption of ‘work-from-home’ practices. Recognition of escalating inter-dependencies, vulnerabilities and inadequate personal cyber behaviours have prompted efforts by businesses and individuals alike to investigate strategies and tactics to mitigate cyber risk – of which cyber insurance is a viable, cost-effective option. It is argued that, ceteris parabus, the nature of cyberspace intrinsically provides characteristic peculiarities that pose significant and bespoke challenges to cyber insurers, often in-congruent with risk attributes commonly associated with traditional personal line insurance products. These challenges include (inter alia) a paucity of historical claim/loss data for underwriting and pricing purposes, interdependencies of cyber architecture promoting high correlation of cyber risk, difficulties in evaluating cyber risk, intangibility of risk assets (such as data, reputation), lack of standardisation across the industry, high and undetermined tail risks, and moral hazard among others. This study proposes a thematic overview of the literature deemed necessary to conceptualise the challenges to issuing personal cyber coverage. There is an evident absence of empirical research appertaining to PCI and the design of operational business models for this business domain, especially qualitative initiatives that (1) attempt to define the scope of the peril, (2) secure an understanding of the needs of both cyber insurer and customer, and (3) to identify elements pivotal to effective management and profitable distribution of PCI - leading to an argument proposed by the author that postulates that the traditional general insurance customer journey and business model are ill-suited for the lineaments of cyberspace. The findings of the review confirm significant gaps in contemporary research within the domain of personal cyber insurance.Keywords: cyberspace, personal cyber risk, personal cyber insurance, customer journey, business model
Procedia PDF Downloads 10325943 The Imperative of Indigenous Entrepreneurship and Sustainable Development in the Globalized Economy
Authors: Innocent Felix Idoko
Abstract:
The development of indigenous entrepreneurship is critical to the achievement of sustainable development in the internationalized economy. Sustainable development implies a continuous stimulus of growth and improvement of an economy in a fairly stable manner. The paradigms of a globalization are numerous to mention. However, to a great extent, these are trade-offs and dilemmas for indigenous entrepreneurship, particularly in the developing economies with infant industries that are essentially crucial to development. This paper analyses the pros and cons of globalization as relates to the complementary role of both foreign and indigenous entrepreneurs, the conflict of values between globalization and protectionism for local entrepreneurship. Using analytical and descriptive approach, the views of academicians, research fellows, literature reviews and both the theories of the mercantilists and those of free trade mainstream economists, and the G20, the paper concludes that there is a legitimate need for protectionism for domestic entrepreneurship in the developing economies as doing otherwise amount to stifling them.Keywords: developing countries, entrepreneurship, globalization, infant-industries, protectionism, sustainable development
Procedia PDF Downloads 44625942 Design and Study of a Low Power High Speed Full Adder Using GDI Multiplexer
Authors: Biswarup Mukherjee, Aniruddha Ghosal
Abstract:
In this paper, we propose a new technique for implementing a low power full adder using a set of GDI multiplexers. Full adder circuits are used comprehensively in Application Specific Integrated Circuits (ASICs). Thus it is desirable to have low power operation for the sub components. The explored method of implementation achieves a low power design for the full adder. Simulated results using state-of-art Tanner tool indicates the superior performance of the proposed technique over conventional CMOS full adder. Detailed comparison of simulated results for the conventional and present method of implementation is presented.Keywords: low power full adder, 2-T GDI MUX, ASIC (application specific integrated circuit), 12-T FA, CMOS (complementary metal oxide semiconductor)
Procedia PDF Downloads 348