Search results for: paper pencil based testing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 45121

Search results for: paper pencil based testing

44851 Analysis of Suitability of Online Assessment by Maintaining Critical Thinking

Authors: Mohamed Chabi

Abstract:

The purpose of this study is to determine Whether paper assessment especially in the subject mathematics will ever be completely replaced by online assessment using Learning Management System and Content Management System such as blackboard. In the subject mathematics, the assessment is the exercise of judgment on the quality of students’ work, as a way of supporting student learning and appraising its outcomes. Testing students has moved from the traditional scribbling and sketching on paper towards working online on a screen and keyboard.

Keywords: paper assessment, online assessment, learning management system, content management system, mathematics

Procedia PDF Downloads 468
44850 Visualization of Latent Sweat Fingerprints Deposit on Paper by Infrared Radiation and Blue Light

Authors: Xiaochun Huang, Xuejun Zhao, Yun Zou, Feiyu Yang, Wenbin Liu, Nan Deng, Ming Zhang, Nengbin Cai

Abstract:

A simple device termed infrared radiation (IR) was developed for rapid visualization of sweat fingerprints deposit on paper with blue light (450 nm, 11 W). In this approach, IR serves as the pretreatment device before the sweat fingerprints was illuminated by blue light. An annular blue light source was adopted for visualizing latent sweat fingerprints. Sample fingerprints were examined under various conditions after deposition, and experimental results indicate that the recovery rate of the latent sweat fingerprints is in the range of 50%-100% without chemical treatments. A mechanism for the observed visibility is proposed based on transportation and re-impregnation of fluorescer in paper at the region of water. And further exploratory experimental results gave the full support to the visible mechanism. Therefore, such a method as IR-pretreated in detecting latent fingerprints may be better for examination in the case where biological information of samples is needed for consequent testing.

Keywords: forensic science, visualization, infrared radiation, blue light, latent sweat fingerprints, detection

Procedia PDF Downloads 497
44849 Big Classes, Bigger Ambitions: A Participatory Approach to the Multiple-Choice Exam

Authors: Melanie Adrian, Elspeth McCulloch, Emily-Jean Gallant

Abstract:

Resources -financial, physical, and human- are increasingly constrained in higher education. University classes are getting bigger, and the concomitant grading burden on faculty is growing rapidly. Multiple-choice exams are seen by some as one solution to these changes. How much students retain, however, and what their testing experience is, continues to be debated. Are multiple-choice exams serving students well, or are they bearing the burden of these developments? Is there a way to address both the resource constraints and make these types of exams more meaningful? In short, how do we engender evaluation methods for large-scale classes that provide opportunities for heightened student learning and enrichment? The following article lays out a testing approach we have employed in four iterations of the same third-year law class. We base our comments in this paper on our initial observations as well as data gathered from an ethics-approved study looking at student experiences. This testing approach provides students with multiple opportunities for revision (thus increasing chances for long term retention), is both individually and collaboratively driven (thus reflecting the individual effort and group effort) and is automatically graded (thus draining limited institutional resources). We found that overall students appreciated the approach and found it more ‘humane’, that it notably reduced pre-exam and intra-exam stress levels, increased ease, and lowered nervousness.

Keywords: exam, higher education, multiple-choice, law

Procedia PDF Downloads 128
44848 Structural Performance Evaluation of Electronic Road Sign Panels Reflecting Damage Scenarios

Authors: Junwon Seo, Bipin Adhikari, Euiseok Jeong

Abstract:

This paper is intended to evaluate the structural performance of welded electronic road signs under various damage scenarios (DSs) using a finite element (FE) model calibrated with full-scale ultimate load testing results. The tested electronic road sign specimen was built with a back skin made of 5052 aluminum and two channels and a frame made of 6061 aluminum, where the back skin was connected to the frame by welding. The size of the tested specimen was 1.52 m long, 1.43 m wide, and 0.28 m deep. An actuator applied vertical loads at the center of the back skin of the specimen, resulting in a displacement of 158.7 mm and an ultimate load of 153.46 kN. Using these testing data, generation and calibration of a FE model of the tested specimen were executed in ABAQUS, indicating that the difference in the ultimate load between the calibrated model simulation and full-scale testing was only 3.32%. Then, six different DSs were simulated where the areas of the welded connection in the calibrated model were diminished for the DSs. It was found that the corners at the back skin-frame joint were prone to connection failure for all the DSs, and failure of the back skin-frame connection occurred remarkably from the distant edges.

Keywords: computational analysis, damage scenarios, electronic road signs, finite element, welded connections

Procedia PDF Downloads 92
44847 Frequency Modulation in Vibro-Acoustic Modulation Method

Authors: D. Liu, D. M. Donskoy

Abstract:

The vibroacoustic modulation method is based on the modulation effect of high-frequency ultrasonic wave (carrier) by low-frequency vibration in the presence of various defects, primarily contact-type such as cracks, delamination, etc. The presence and severity of the defect are measured by the ratio of the spectral sidebands and the carrier in the spectrum of the modulated signal. This approach, however, does not differentiate between amplitude and frequency modulations, AM and FM, respectfully. It was experimentally shown that both modulations could be present in the spectrum, yet each modulation may be associated with different physical mechanisms. AM mechanisms are quite well understood and widely covered in the literature. This paper is a first attempt to explain the generation mechanisms of FM and its correlation with the flaw properties. Here we proposed two possible mechanisms leading to FM modulation based on nonlinear local defect resonance and dynamic acousto-elastic models.

Keywords: non-destructive testing, nonlinear acoustics, structural health monitoring, acousto-elasticity, local defect resonance

Procedia PDF Downloads 152
44846 A Bayesian Approach for Analyzing Academic Article Structure

Authors: Jia-Lien Hsu, Chiung-Wen Chang

Abstract:

Research articles may follow a simple and succinct structure of organizational patterns, called move. For example, considering extended abstracts, we observe that an extended abstract usually consists of five moves, including Background, Aim, Method, Results, and Conclusion. As another example, when publishing articles in PubMed, authors are encouraged to provide a structured abstract, which is an abstract with distinct and labeled sections (e.g., Introduction, Methods, Results, Discussions) for rapid comprehension. This paper introduces a method for computational analysis of move structures (i.e., Background-Purpose-Method-Result-Conclusion) in abstracts and introductions of research documents, instead of manually time-consuming and labor-intensive analysis process. In our approach, sentences in a given abstract and introduction are automatically analyzed and labeled with a specific move (i.e., B-P-M-R-C in this paper) to reveal various rhetorical status. As a result, it is expected that the automatic analytical tool for move structures will facilitate non-native speakers or novice writers to be aware of appropriate move structures and internalize relevant knowledge to improve their writing. In this paper, we propose a Bayesian approach to determine move tags for research articles. The approach consists of two phases, training phase and testing phase. In the training phase, we build a Bayesian model based on a couple of given initial patterns and the corpus, a subset of CiteSeerX. In the beginning, the priori probability of Bayesian model solely relies on initial patterns. Subsequently, with respect to the corpus, we process each document one by one: extract features, determine tags, and update the Bayesian model iteratively. In the testing phase, we compare our results with tags which are manually assigned by the experts. In our experiments, the promising accuracy of the proposed approach reaches 56%.

Keywords: academic English writing, assisted writing, move tag analysis, Bayesian approach

Procedia PDF Downloads 330
44845 Measuring Biobased Content of Building Materials Using Carbon-14 Testing

Authors: Haley Gershon

Abstract:

The transition from using fossil fuel-based building material to formulating eco-friendly and biobased building materials plays a key role in sustainable building. The growing demand on a global level for biobased materials in the building and construction industries heightens the importance of carbon-14 testing, an analytical method used to determine the percentage of biobased content that comprises a material’s ingredients. This presentation will focus on the use of carbon-14 analysis within the building materials sector. Carbon-14, also known as radiocarbon, is a weakly radioactive isotope present in all living organisms. Any fossil material older than 50,000 years will not contain any carbon-14 content. The radiocarbon method is thus used to determine the amount of carbon-14 content present in a given sample. Carbon-14 testing is performed according to ASTM D6866, a standard test method developed specifically for biobased content determination of material in solid, liquid, or gaseous form, which requires radiocarbon dating. Samples are combusted and converted into a solid graphite form and then pressed onto a metal disc and mounted onto a wheel of an accelerator mass spectrometer (AMS) machine for the analysis. The AMS instrument is used in order to count the amount of carbon-14 present. By submitting samples for carbon-14 analysis, manufacturers of building materials can confirm the biobased content of ingredients used. Biobased testing through carbon-14 analysis reports results as percent biobased content, indicating the percentage of ingredients coming from biomass sourced carbon versus fossil carbon. The analysis is performed according to standardized methods such as ASTM D6866, ISO 16620, and EN 16640. Products 100% sourced from plants, animals, or microbiological material are therefore 100% biobased, while products sourced only from fossil fuel material are 0% biobased. Any result in between 0% and 100% biobased indicates that there is a mixture of both biomass-derived and fossil fuel-derived sources. Furthermore, biobased testing for building materials allows manufacturers to submit eligible material for certification and eco-label programs such as the United States Department of Agriculture (USDA) BioPreferred Program. This program includes a voluntary labeling initiative for biobased products, in which companies may apply to receive and display the USDA Certified Biobased Product label, stating third-party verification and displaying a product’s percentage of biobased content. The USDA program includes a specific category for Building Materials. In order to qualify for the biobased certification under this product category, examples of product criteria that must be met include minimum 62% biobased content for wall coverings, minimum 25% biobased content for lumber, and a minimum 91% biobased content for floor coverings (non-carpet). As a result, consumers can easily identify plant-based products in the marketplace.

Keywords: carbon-14 testing, biobased, biobased content, radiocarbon dating, accelerator mass spectrometry, AMS, materials

Procedia PDF Downloads 158
44844 Human Action Retrieval System Using Features Weight Updating Based Relevance Feedback Approach

Authors: Munaf Rashid

Abstract:

For content-based human action retrieval systems, search accuracy is often inferior because of the following two reasons 1) global information pertaining to videos is totally ignored, only low level motion descriptors are considered as a significant feature to match the similarity between query and database videos, and 2) the semantic gap between the high level user concept and low level visual features. Hence, in this paper, we propose a method that will address these two issues and in doing so, this paper contributes in two ways. Firstly, we introduce a method that uses both global and local information in one framework for an action retrieval task. Secondly, to minimize the semantic gap, a user concept is involved by incorporating features weight updating (FWU) Relevance Feedback (RF) approach. We use statistical characteristics to dynamically update weights of the feature descriptors so that after every RF iteration feature space is modified accordingly. For testing and validation purpose two human action recognition datasets have been utilized, namely Weizmann and UCF. Results show that even with a number of visual challenges the proposed approach performs well.

Keywords: relevance feedback (RF), action retrieval, semantic gap, feature descriptor, codebook

Procedia PDF Downloads 472
44843 U.S. Trade and Trade Balance with China: Testing for Marshall-Lerner Condition and the J-Curve Hypothesis

Authors: Anisul Islam

Abstract:

The U.S. has a very strong trade relationship with China but with a large and persistent trade deficit. Some has argued that the undervalued Chinese Yuan is to be blamed for the persistent trade deficit. The empirical results are mixed at best. This paper empirically estimates the U.S. export function along with the U.S. import function with its trade with China with the purpose of testing for the existence of the Marshall-Lerner (ML) condition as well for the possible existence of the J-curve hypothesis. Annual export and import data will be utilized for as long as the time series data exists. The export and import functions will be estimated using advanced econometric techniques, along with appropriate diagnostic tests performed to examine the validity and reliability of the estimated results. The annual time-series data covers from 1975 to 2022 with a sample size of 48 years, the longest period ever utilized before in any previous study. The data is collected from several sources, such as the World Bank’s World Development Indicators, IMF Financial Statistics, IMF Direction of Trade Statistics, and several other sources. The paper is expected to shed important light on the ongoing debate regarding the persistent U.S. trade deficit with China and the policies that may be useful to reduce such deficits over time. As such, the paper will be of great interest for the academics, researchers, think tanks, global organizations, and policy makers in both China and the U.S.

Keywords: exports, imports, marshall-lerner condition, j-curve hypothesis, united states, china

Procedia PDF Downloads 63
44842 Assessing the Effect of Waste-based Geopolymer on Asphalt Binders

Authors: Amani A. Saleh, Maram M. Saudy, Mohamed N. AbouZeid

Abstract:

Asphalt cement concrete is a very commonly used material in the construction of roads. It has many advantages, such as being easy to use as well as providing high user satisfaction in terms of comfortability and safety on the road. However, there are some problems that come with asphalt cement concrete, such as its high carbon footprint, which makes it environmentally unfriendly. In addition, pavements require frequent maintenance, which could be very costly and uneconomic. The aim of this research is to study the effect of mixing waste-based geopolymers with asphalt binders. Geopolymer mixes were prepared by combining alumino-silicate sources such as fly ash, silica fumes, and metakaolin with alkali activators. The purpose of mixing geopolymers with the asphalt binder is to enhance the rheological and microstructural properties of asphalt. This was done through two phases, where the first phase was developing an optimum mix design of the geopolymer additive itself. The following phase was testing the geopolymer-modified asphalt binder after the addition of the optimum geopolymer mix design to it. The testing of the modified binder is performed according to the Superpave testing procedures, which include the dynamic shear rheometer to measure parameters such as rutting and fatigue cracking, and the rotational viscometer to measure workability. In addition, the microstructural properties of the modified binder is studied using the environmental scanning electron microscopy test (ESEM). In the testing phase, the aim is to observe whether the addition of different geopolymer percentages to the asphalt binder will enhance the properties of the binder and yield desirable results. Furthermore, the tests on the geopolymer-modified binder were carried out at fixed time intervals, therefore, the curing time was the main parameter being tested in this research. It was observed that the addition of geopolymers to asphalt binder has shown an increased performance of asphalt binder with time. It is worth mentioning that carbon emissions are expected to be reduced since geopolymers are environmentally friendly materials that minimize carbon emissions and lead to a more sustainable environment. Additionally, the use of industrial by-products such as fly ash and silica fumes is beneficial in the sense that they are recycled into producing geopolymers instead of being accumulated in landfills and therefore wasting space.

Keywords: geopolymer, rutting, superpave, fatigue cracking, sustainability, waste

Procedia PDF Downloads 128
44841 Evaluation of Iranian Standard for Assessment of Liquefaction Potential of Cohesionless Soils Based on SPT

Authors: Reza Ziaie Moayad, Azam Kouhpeyma

Abstract:

In-situ testing is preferred to evaluate the liquefaction potential in cohesionless soils due to high disturbance during sampling. Although new in-situ methods with high accuracy have been developed, standard penetration test, the simplest and the oldest in-situ test, is still used due to the profusion of the recorded data. This paper reviews the Iranian standard of evaluating liquefaction potential in soils (codes 525) and compares the liquefaction assessment methods based on SPT results on cohesionless soil in this standard with the international standards. To this, methods for assessing liquefaction potential which are presented by Cetin et al. (2004), Boulanger and Idriss (2014) are compared with what is presented in standard 525. It is found that although the procedure used in Iranian standard of evaluating the potential of liquefaction has not been updated according to the new findings, it is a conservative procedure.

Keywords: cohesionless soil, liquefaction, SPT, standard 525

Procedia PDF Downloads 170
44840 Massively Parallel Sequencing Improved Resolution for Paternity Testing

Authors: Xueying Zhao, Ke Ma, Hui Li, Yu Cao, Fan Yang, Qingwen Xu, Wenbin Liu

Abstract:

Massively parallel sequencing (MPS) technologies allow high-throughput sequencing analyses with a relatively affordable price and have gradually been applied to forensic casework. MPS technology identifies short tandem repeat (STR) loci based on sequence so that repeat motif variation within STRs can be detected, which may help one to infer the origin of the mutation in some cases. Here, we report on one case with one three-step mismatch (D18S51) in family trios based on both capillary electrophoresis (CE) and MPS typing. The alleles of the alleged father (AF) are [AGAA]₁₇AGAG[AGAA]₃ and [AGAA]₁₅. The mother’s alleles are [AGAA]₁₉ and [AGAA]₉AGGA[AGAA]₃. The questioned child’s (QC) alleles are [AGAA]₁₉ and [AGAA]₁₂. Given that the sequence variants in repeat regions of AF and mother are not observed in QC’s alleles, the QC’s allele [AGAA]₁₂ was likely inherited from the AF’s allele [AGAA]₁₅ by loss of three repeat [AGAA]. Besides, two new alleles of D18S51 in this study, [AGAA]₁₇AGAG[AGAA]₃ and [AGAA]₉AGGA[AGAA]₃, have not been reported before. All the results in this study were verified using Sanger-type sequencing. In summary, the MPS typing method can offer valuable information for forensic genetics research and play a promising role in paternity testing.

Keywords: family trios analysis, forensic casework, ion torrent personal genome machine (PGM), massively parallel sequencing (MPS)

Procedia PDF Downloads 302
44839 Exploring the Applications of Neural Networks in the Adaptive Learning Environment

Authors: Baladitya Swaika, Rahul Khatry

Abstract:

Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.

Keywords: computer adaptive tests, item response theory, machine learning, neural networks

Procedia PDF Downloads 175
44838 Friction and Wear Characteristics of Pongamia Oil Based Blended Lubricant at Different Load and Sliding Distance

Authors: Yashvir Singh

Abstract:

Around the globe, there is demand for the development of bio-based lubricant which will be biodegradable, non -toxic and environmental friendly. This paper outlines the friction and wear characteristics of Pongamia oil (PO) contaminated bio-lubricant by using pin-on-disc tribometer. To formulate the bio-lubricants, PO was blended in the ratios 15, 30 and 50% by volume with the base lubricant SAE 20 W 40. Tribological characteristics of these blends were carried out at 3.8 m/s sliding velocity and loads applied were 50, 100, 150 N. Experimental results showed that the lubrication regime that occurred during the test was boundary lubrication while the main wear mechanisms were abrasive and the adhesive wear. During testing, the lowest wear was found with the addition of 15% PO, and above this contamination, the wear rate was increased considerably. With increase in load, viscosity of all the bio-lubricants increases and meets the ISO VG 100 requirement at 40 oC except PB 50. The addition of PO in the base lubricant acted as a very good lubricant additive which reduced the friction and wear scar diameter during the test. It has been concluded that the PB 15 can act as an alternative lubricant to increase the mechanical efficiency at 3.8 m/s sliding velocity and contribute in reduction of dependence on the petroleum based products.

Keywords: friction, load, pongamia oil, sliding velocity, wear

Procedia PDF Downloads 361
44837 Scalable CI/CD and Scalable Automation: Assisting in Optimizing Productivity and Fostering Delivery Expansion

Authors: Solanki Ravirajsinh, Kudo Kuniaki, Sharma Ankit, Devi Sherine, Kuboshima Misaki, Tachi Shuntaro

Abstract:

In software development life cycles, the absence of scalable CI/CD significantly impacts organizations, leading to increased overall maintenance costs, prolonged release delivery times, heightened manual efforts, and difficulties in meeting tight deadlines. Implementing CI/CD with standard serverless technologies using cloud services overcomes all the above-mentioned issues and helps organizations improve efficiency and faster delivery without the need to manage server maintenance and capacity. By integrating scalable CI/CD with scalable automation testing, productivity, quality, and agility are enhanced while reducing the need for repetitive work and manual efforts. Implementing scalable CI/CD for development using cloud services like ECS (Container Management Service), AWS Fargate, ECR (to store Docker images with all dependencies), Serverless Computing (serverless virtual machines), Cloud Log (for monitoring errors and logs), Security Groups (for inside/outside access to the application), Docker Containerization (Docker-based images and container techniques), Jenkins (CI/CD build management tool), and code management tools (GitHub, Bitbucket, AWS CodeCommit) can efficiently handle the demands of diverse development environments and are capable of accommodating dynamic workloads, increasing efficiency for faster delivery with good quality. CI/CD pipelines encourage collaboration among development, operations, and quality assurance teams by providing a centralized platform for automated testing, deployment, and monitoring. Scalable CI/CD streamlines the development process by automatically fetching the latest code from the repository every time the process starts, building the application based on the branches, testing the application using a scalable automation testing framework, and deploying the builds. Developers can focus more on writing code and less on managing infrastructure as it scales based on the need. Serverless CI/CD eliminates the need to manage and maintain traditional CI/CD infrastructure, such as servers and build agents, reducing operational overhead and allowing teams to allocate resources more efficiently. Scalable CI/CD adjusts the application's scale according to usage, thereby alleviating concerns about scalability, maintenance costs, and resource needs. Creating scalable automation testing using cloud services (ECR, ECS Fargate, Docker, EFS, Serverless Computing) helps organizations run more than 500 test cases in parallel, aiding in the detection of race conditions, performance issues, and reducing execution time. Scalable CI/CD offers flexibility, dynamically adjusting to varying workloads and demands, allowing teams to scale resources up or down as needed. It optimizes costs by only paying for the resources as they are used and increases reliability. Scalable CI/CD pipelines employ automated testing and validation processes to detect and prevent errors early in the development cycle.

Keywords: achieve parallel execution, cloud services, scalable automation testing, scalable continuous integration and deployment

Procedia PDF Downloads 43
44836 Cadaveric Dissection versus Systems-Based Anatomy: Testing Final Year Student Surface Anatomy Knowledge to Compare the Long-Term Effectiveness of Different Course Structures

Authors: L. Sun, T. Hargreaves, Z. Ahmad

Abstract:

Newly-qualified Foundation Year 1 doctors in the United Kingdom are frequently expected to perform practical skills involving the upper limb in clinical practice (for example, venipuncture, cannulation, and blood gas sampling). However, a move towards systems-based undergraduate medical education in the United Kingdom often precludes or limits dedicated time to anatomy teaching with cadavers or prosections, favouring only applied anatomy in the context of pathology. The authors hypothesised that detailed anatomical knowledge may consequently be adversely affected, particularly with respect to long-term retention. A simple picture quiz and accompanying questionnaire testing the identification of 7 upper limb surface landmarks was distributed to a total of 98 final year medical students from two universities - one with a systems-based curriculum, and one with a dedicated longitudinal dissection-based anatomy module in the first year of study. Students with access to dissection and prosection-based anatomy teaching performed more strongly, with a significantly higher rate of correct identification of all but one of the landmarks. Furthermore, it was notable that none of the students who had previously undertaken a systems-based course scored full marks, compared with 20% of those who had participated in the more dedicated anatomy course. This data suggests that a traditional, dissection-based approach to undergraduate anatomy teaching is superior to modern system-based curricula, in terms of aiding long-term retention of anatomical knowledge pertinent to newly-qualified doctors. The authors express concern that this deficit in proficiency could be detrimental to patient care in clinical practice, and propose that, where dissection-led anatomy teaching is not available, further anatomy revision modules are implemented throughout undergraduate education to aid knowledge retention and support clinical excellence.

Keywords: dissection, education, surface anatomy, upper limb

Procedia PDF Downloads 132
44835 Artificial Intelligence in Penetration Testing of a Connected and Autonomous Vehicle Network

Authors: Phillip Garrad, Saritha Unnikrishnan

Abstract:

The recent popularity of connected and autonomous vehicles (CAV) corresponds with an increase in the risk of cyber-attacks. These cyber-attacks have been instigated by both researchers or white-coat hackers and cyber-criminals. As Connected Vehicles move towards full autonomy, the impact of these cyber-attacks also grows. The current research details challenges faced in cybersecurity testing of CAV, including access and cost of the representative test setup. Other challenges faced are lack of experts in the field. Possible solutions to how these challenges can be overcome are reviewed and discussed. From these findings, a software simulated CAV network is established as a cost-effective representative testbed. Penetration tests are then performed on this simulation, demonstrating a cyber-attack in CAV. Studies have shown Artificial Intelligence (AI) to improve runtime, increase efficiency and comprehensively cover all the typical test aspects in penetration testing in other industries. There is an attempt to introduce similar AI models to the software simulation. The expectation from this implementation is to see similar improvements in runtime and efficiency for the CAV model. If proven to be an effective means of penetration test for CAV, this methodology may be used on a full CAV test network.

Keywords: cybersecurity, connected vehicles, software simulation, artificial intelligence, penetration testing

Procedia PDF Downloads 110
44834 An Automatic Speech Recognition Tool for the Filipino Language Using the HTK System

Authors: John Lorenzo Bautista, Yoon-Joong Kim

Abstract:

This paper presents the development of a Filipino speech recognition tool using the HTK System. The system was trained from a subset of the Filipino Speech Corpus developed by the DSP Laboratory of the University of the Philippines-Diliman. The speech corpus was both used in training and testing the system by estimating the parameters for phonetic HMM-based (Hidden-Markov Model) acoustic models. Experiments on different mixture-weights were incorporated in the study. The phoneme-level word-based recognition of a 5-state HMM resulted in an average accuracy rate of 80.13 for a single-Gaussian mixture model, 81.13 after implementing a phoneme-alignment, and 87.19 for the increased Gaussian-mixture weight model. The highest accuracy rate of 88.70% was obtained from a 5-state model with 6 Gaussian mixtures.

Keywords: Filipino language, Hidden Markov Model, HTK system, speech recognition

Procedia PDF Downloads 480
44833 A Study on Marble Based Geopolymer Mortar / Concrete

Authors: Wei-Hao Lee, Ta-Wui Cheng, Yung-Chin Ding, Tai-Tien Wang

Abstract:

The purpose of this study is trying to use marble wastes as the raw material to fabricate geopolymer green mortar / concrete. Experiment results show that using marble to make geopolymer mortar and concrete, the compressive strength after 28 days curing can reach 35 MPa and 25 MPa, respectively. The characteristics of marble-based geopolymer green mortar and concrete will keep testing for a long term in order to understand the effect parameters. The study is based on resource recovery and recycling. Its basic characteristics are low consumption, low carbon dioxide emission and high efficiency that meet the international tendency 'Circular Economy.' By comparing with Portland cement mortar and concrete, production 1 ton of marble-based geopolymer mortar and concrete, they can be saved around 50.3% and 49.6% carbon dioxide emission, respectively. Production 1 m3 of marble-based geopolymer concrete costs about 62 USD that cheaper than that of traditional Portland concrete. It is proved that the marble-based geopolymer concrete has great potential for further engineering development.

Keywords: marble, geopolymer, geopolymer concrete, CO₂ emission

Procedia PDF Downloads 440
44832 Testing the Change in Correlation Structure across Markets: High-Dimensional Data

Authors: Malay Bhattacharyya, Saparya Suresh

Abstract:

The Correlation Structure associated with a portfolio is subjected to vary across time. Studying the structural breaks in the time-dependent Correlation matrix associated with a collection had been a subject of interest for a better understanding of the market movements, portfolio selection, etc. The current paper proposes a methodology for testing the change in the time-dependent correlation structure of a portfolio in the high dimensional data using the techniques of generalized inverse, singular valued decomposition and multivariate distribution theory which has not been addressed so far. The asymptotic properties of the proposed test are derived. Also, the performance and the validity of the method is tested on a real data set. The proposed test performs well for detecting the change in the dependence of global markets in the context of high dimensional data.

Keywords: correlation structure, high dimensional data, multivariate distribution theory, singular valued decomposition

Procedia PDF Downloads 125
44831 Evaluation of Mechanical Properties of Welds Fabricated at a Close Proximity on Offshore Structures

Authors: T. Nakkeran, C. Dhamodharan, Win Myint Soe , Ramasamy Deverajan, M. Ganesh Babu

Abstract:

This manuscript presents the results of an experimental investigation performed to study the material and mechanical properties of two weld joints fabricated within close proximity. The experiment was designed using welded S355 D Z35 with distances between two parallel adjacent weld toes at 8 mm. These distances were less than the distance that has normally been recommended in standards, codes, and specifications. The main idea of the analysis is to determine any significant effects when welding the joints with the close proximity of 8mm using the SAW welding process of the one joint with high heat put and one joint welded with the FCAW welding process and evaluating the destructing and nondestructive testing between the welded joints. Further, we have evaluated the joints with Mechanical Testing for evaluating by performing Tensile test, bend testing, Macrostructure, Microstructure, Hardness test, and Impact testing. After evaluating the final outcome of the result, no significant changes were observed for welding the close proximity of weld of 8mm distance between the joints as compared to the specification minimum distance between the weldments of any design should be 50mm.

Keywords: S355 carbon steel, weld proximity, SAW process, FCAW process, heat input, bend test, tensile test, hardness test, impact test, macro and microscopic examinations

Procedia PDF Downloads 98
44830 Testing of Gas Turbine KingTech with Biodiesel

Authors: Nicolas Lipchak, Franco Aiducic, Santiago Baieli

Abstract:

The present work is a part of the research project called ‘Testing of gas turbine KingTech with biodiesel’, carried out by the Department of Industrial Engineering of the National Technological University at Buenos Aires. The research group aims to experiment with biodiesel in a gas turbine Kingtech K-100 to verify the correct operation of it. In this sense, tests have been developed to obtain real data of parameters inherent to the work cycle, to be used later as parameters of comparison and performance analysis. In the first instance, the study consisted in testing the gas turbine with a mixture composition of 50% Biodiesel and 50% Diesel. The parameters arising from the measurements made were compared with the parameters of the gas turbine with a composition of 100% Diesel. In the second instance, the measured parameters were used to calculate the power generated and the thermal efficiency of the Kingtech K-100 turbine. The turbine was also inspected to verify the status of the internals due to the use of biofuels. The conclusions obtained allow empirically demonstrate that it is feasible to use biodiesel in this type of gas turbines, without the use of this fuel generates a loss of power or degradation of internals.

Keywords: biodiesel, efficiency, KingTech, turbine

Procedia PDF Downloads 245
44829 Nonparametric Specification Testing for the Drift of the Short Rate Diffusion Process Using a Panel of Yields

Authors: John Knight, Fuchun Li, Yan Xu

Abstract:

Based on a new method of the nonparametric estimator of the drift function, we propose a consistent test for the parametric specification of the drift function in the short rate diffusion process using observations from a panel of yields. The test statistic is shown to follow an asymptotic normal distribution under the null hypothesis that the parametric drift function is correctly specified, and converges to infinity under the alternative. Taking the daily 7-day European rates as a proxy of the short rate, we use our test to examine whether the drift of the short rate diffusion process is linear or nonlinear, which is an unresolved important issue in the short rate modeling literature. The testing results indicate that none of the drift functions in this literature adequately captures the dynamics of the drift, but nonlinear specification performs better than the linear specification.

Keywords: diffusion process, nonparametric estimation, derivative security price, drift function and volatility function

Procedia PDF Downloads 368
44828 Mode II Fracture Toughness of Hybrid Fiber Reinforced Concrete

Authors: H. S. S Abou El-Mal, A. S. Sherbini, H. E. M. Sallam

Abstract:

Mode II fracture toughness (KIIc) of fiber reinforced concrete has been widely investigated under various patterns of testing geometries. The effect of fiber type, concrete matrix properties, and testing mechanisms were extensively studied. The area of hybrid fiber addition shows a lake of reported research data. In this paper an experimental investigation of hybrid fiber embedded in high strength concrete matrix is reported. Three different types of fibers; namely steel (S), glass (G), and polypropylene (PP) fibers were mixed together in four hybridization patterns, (S/G), (S/PP), (G/PP), (S/G/PP) with constant cumulative volume fraction (Vf) of 1.5%. The concrete matrix properties were kept the same for all hybrid fiber reinforced concrete patterns. In an attempt to estimate a fairly accepted value of fracture toughness KIIc, four testing geometries and loading types are employed in this investigation. Four point shear, Brazilian notched disc, double notched cube, and double edge notched specimens are investigated in a trial to avoid the limitations and sensitivity of each test regarding geometry, size effect, constraint condition, and the crack length to specimen width ratio a/w. The addition of all hybridization patterns of fiber reduced the compressive strength and increased mode II fracture toughness in pure mode II tests. Mode II fracture toughness of concrete KIIc decreased with the increment of a/w ratio for all concretes and test geometries. Mode II fracture toughness KIIc is found to be sensitive to the hybridization patterns of fiber. The (S/PP) hybridization pattern showed higher values than all other patterns, while the (S/G/PP) showed insignificant enhancement on mode II fracture toughness (KIIc). Four point shear (4PS) test set up reflects the most reliable values of mode II fracture toughness KIIc of concrete. Mode II fracture toughness KIIc of concrete couldn’t be assumed as a real material property.

Keywords: fiber reinforced concrete, Hybrid fiber, Mode II fracture toughness, testing geometry

Procedia PDF Downloads 326
44827 Investigation into Relationship between Spaced Repetitions and Problems Solving Efficiency

Authors: Sidharth Talan, Rajlakshmi G. Majumdar

Abstract:

Problem-solving skill is one the few skills which is constantly endeavored to improve upon by the professionals and academicians around the world in order to sustain themselves in the ever-growing competitive environment. The given paper focuses on evaluating a hypothesized relationship between the problems solving efficiency of an individual with spaced repetitions, conducted with a time interval of one day over a period of two weeks. The paper has utilized uni-variate regression analysis technique to assess the best fit curve that can explain the significant relationship between the given two variables. The paper has incorporated Anagrams solving as the appropriate testing process for the analysis. Since Anagrams solving involves rearranging a jumbled word to form a correct word, it projects to be an efficient process to observe the attention span, visual- motor coordination and the verbal ability of an individual. Based on the analysis for a sample population of 30, it was observed that problem-solving efficiency of an individual, measured in terms of the score in each test was found to be significantly correlated with time period measured in days.

Keywords: Anagrams, histogram plot, moving average curve, spacing effect

Procedia PDF Downloads 165
44826 Development of Soil Test Kits to Determine Organic Matter Available Phosphorus and Exchangeable Potassium in Thailand

Authors: Charirat Kusonwiriyawong, Supha Photichan, Wannarut Chutibutr

Abstract:

Soil test kits for rapid analysis of the organic matter, available phosphorus and exchangeable potassium were developed to drive a low-cost field testing kit to farmers. The objective was to provide a decision tool for improving soil fertility. One aspect of soil test kit development was ease of use which is a time requirement for completing organic matter, available phosphorus and exchangeable potassium test in one soil sample. This testing kit required only two extractions and utilized no filtration consuming approximately 15 minutes per sample. Organic matter was principally created by oxidizing carbon KMnO₄ using the standard color chart. In addition, modified single extractant (Mehlich I) was applied to extract available phosphorus and exchangeable potassium. Molybdenum blue method and turbidimetric method using standard color chart were adapted to analyze available phosphorus and exchangeable potassium, respectively. Modified single extractant using in soil test kits were highly significant matching with analytical laboratory results (r=0.959** and 0.945** for available phosphorus and exchangeable potassium, respectively). Linear regressions were statistically calculated between modified single extractant and standard laboratory analysis (y=0.9581x-12.973 for available phosphorus and y=0.5372x+15.283 for exchangeable potassium, respectively). These equations were calibrated to formulate a fertilizer rate recommendation for specific corps. To validate quality, soil test kits were distributed to farmers and extension workers. We found that the accuracy of soil test kits were 71.0%, 63.9% and 65.5% for organic matter, available phosphorus, and exchangeable potassium, respectively. The quantitative survey was also conducted in order to assess their satisfaction with soil test kits. The survey showed that more than 85% of respondents said these testing kits were more convenient, economical and reliable than the other commercial soil test kits. Based upon the finding of this study, soil test kits can be another alternative for providing soil analysis and fertility recommendations when a soil testing laboratory is not available.

Keywords: available phosphorus, exchangeable potassium, modified single extractant, organic matter, soil test kits

Procedia PDF Downloads 145
44825 Using Computerized Analogical Reasoning Tasks as a Way to Improve Literacy Skills in Children with Mild Intellectual Disability

Authors: Caroline Denaes

Abstract:

The ability to read is crucial for a successful path in school and in a social and professional context. Children with mild intellectual disability are confronted to serious difficulties in literacy. A lot of them do not read or are illiterate. Only one child out of five is able to acquire basic reading skills, which increases the likelihood to misfit in society, especially when these children grow up and cannot manage themselves in situations requiring higher reading levels. One way to help these children acquiring basic reading skills is to use analogical reasoning, as some researchers demonstrated that this mechanism is fundamental for any reading process. For this purpose, we developed computerized analogies displayed on a touch screen tablet. Analogies are comparisons that give children a framework they can use to understand new information. They work by comparing one thing to another in order to emphasize some mutual quality. If one of the items is unfamiliar, that mutual quality can help make it understandable, or it can cause the children to consider something familiar in some new way, such as transferring what they know about familiar words to help them identify unfamiliar words. In addition, using touch screen tablets represents several advantages: the ease of use, the relevance to this specific population and the appeal of a self-directed activity gives individuals and practitioners a modern tool that differs from the traditional paper-and-pencil material. In addition, the touch screen dimension is especially appropriate for children as assistive technology has been found to be more motivating that any other types of devices and improves the children’ attention span.

Keywords: literacy, intellectual disabilities, touch screen techonology, literacy skill

Procedia PDF Downloads 270
44824 An Intelligent Text Independent Speaker Identification Using VQ-GMM Model Based Multiple Classifier System

Authors: Ben Soltane Cheima, Ittansa Yonas Kelbesa

Abstract:

Speaker Identification (SI) is the task of establishing identity of an individual based on his/her voice characteristics. The SI task is typically achieved by two-stage signal processing: training and testing. The training process calculates speaker specific feature parameters from the speech and generates speaker models accordingly. In the testing phase, speech samples from unknown speakers are compared with the models and classified. Even though performance of speaker identification systems has improved due to recent advances in speech processing techniques, there is still need of improvement. In this paper, a Closed-Set Tex-Independent Speaker Identification System (CISI) based on a Multiple Classifier System (MCS) is proposed, using Mel Frequency Cepstrum Coefficient (MFCC) as feature extraction and suitable combination of vector quantization (VQ) and Gaussian Mixture Model (GMM) together with Expectation Maximization algorithm (EM) for speaker modeling. The use of Voice Activity Detector (VAD) with a hybrid approach based on Short Time Energy (STE) and Statistical Modeling of Background Noise in the pre-processing step of the feature extraction yields a better and more robust automatic speaker identification system. Also investigation of Linde-Buzo-Gray (LBG) clustering algorithm for initialization of GMM, for estimating the underlying parameters, in the EM step improved the convergence rate and systems performance. It also uses relative index as confidence measures in case of contradiction in identification process by GMM and VQ as well. Simulation results carried out on voxforge.org speech database using MATLAB highlight the efficacy of the proposed method compared to earlier work.

Keywords: feature extraction, speaker modeling, feature matching, Mel frequency cepstrum coefficient (MFCC), Gaussian mixture model (GMM), vector quantization (VQ), Linde-Buzo-Gray (LBG), expectation maximization (EM), pre-processing, voice activity detection (VAD), short time energy (STE), background noise statistical modeling, closed-set tex-independent speaker identification system (CISI)

Procedia PDF Downloads 309
44823 Timing and Noise Data Mining Algorithm and Software Tool in Very Large Scale Integration (VLSI) Design

Authors: Qing K. Zhu

Abstract:

Very Large Scale Integration (VLSI) design becomes very complex due to the continuous integration of millions of gates in one chip based on Moore’s law. Designers have encountered numerous report files during design iterations using timing and noise analysis tools. This paper presented our work using data mining techniques combined with HTML tables to extract and represent critical timing/noise data. When we apply this data-mining tool in real applications, the running speed is important. The software employs table look-up techniques in the programming for the reasonable running speed based on performance testing results. We added several advanced features for the application in one industry chip design.

Keywords: VLSI design, data mining, big data, HTML forms, web, VLSI, EDA, timing, noise

Procedia PDF Downloads 254
44822 Effect of Sulfur Content on Fatigue Strength of AISI 4140 Steel

Authors: Sachin S. Patil, Mohan I. Mehta, Sandip J. Sutar, Akshay B. Patil, Shreyas S. Kirwai, Suresh Arangi

Abstract:

MnS is the most commonly found inclusion in steel, which is desirable for machinability of alloy steels but only up to a certain limit, beyond which it weakens fatigue properties of steel. In present work, the effect of sulfur content and its inclusions on the fatigue behavior of AISI 4140 steel is studied (sulfur content 0.002% and 0.016%). Metallurgical analysis, Mechanical testing and Rotating Bending Fatigue (RBF) test were carried out. With the increase in sulfur content, ductility and toughness of the material decrease significantly and large scatter is observed in UTS and impact energy values. From the results of RBF testing, it can be observed that increase in sulfur content from 0.002% to 0.016% has a negligible effect on the endurance strength of AISI 4140 for similar hardness level. Fractography analysis was carried out to study the failure modes in testing.

Keywords: AISI 4140, sulfur content, MnS inclusion, rotating bending fatigue

Procedia PDF Downloads 399