Search results for: compression test
2965 Could One-Hand Chest Compression for a Small Child Cause Intra-Abdominal Injuries?
Authors: Yeon Ho You, Ji Sook Lee, Jin Hong Min
Abstract:
We examined whether children ( < 18 years old) had risk of intra-thoracic trauma during 'one-handed' chest compressions through MDCT images. We measured the length of the lower half of the sternum (Stotal/2~X). We also measured the distance from the diaphragm to the midpoint of the sternum (Stotal/2~D) and half the width of an adult hand (Wtotal/2). All the 1 year-old children had Stotal/2~X and Stotal/2~D less than Wtotal/2. Among the children aged 2 years, 6 (60.0%) had Stotal/2~X and Stotal/2~D less than Wtotal/2. Among those aged 3 years, 4 (26.7%) had Stotal/2~X and Stotal/2~D less than Wtotal/2, and among those aged 4 years, 2 (13.3%) had Stotal/2~X and Stotal/2~D less than Wtotal/2. However, Stotal/2~X and Stotal/2~D were greater than Wtotal/2 in children aged 5 years or more. We knew that small children may be at an increased risk of intra-thoracic trauma during 'one-handed' chest compressions.
Keywords: Cardiopulmonary resuscitation, child, compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13692964 Teachers and Learners Perceptions on the Impact of Different Test Procedures on Reading: A Case Study
Authors: Bahloul Amel
Abstract:
The main aim of this research was to investigate the perspectives of English language teachers and learners on the effect of test techniques on reading comprehension, test performance and assessment. The research has also aimed at finding the differences between teacher and learner perspectives, specifying the test techniques which have the highest effect, investigating the other factors affecting reading comprehension, and comparing the results with the similar studies. In order to achieve these objectives, perspectives and findings of different researchers were reviewed, two different questionnaires were prepared to collect data for the perspectives of teachers and learners, the questionnaires were applied to 26 learners and 8 teachers from the University of Batna (Algeria), and quantitative and qualitative data analysis of the results were done. The results and analysis of the results show that different test techniques affect reading comprehension, test performance and assessment at different percentages rates.
Keywords: Reading Comprehension, Reading Assessment, Test Performance, Test Techniques.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18742963 A Study of Environmental Test Sequences for Electrical Units
Authors: Jung Ho Yang, Yong Soo Kim
Abstract:
Electrical units are operated by electrical and electronic components. An environmental test sequence is useful for testing electrical units to reduce reliability issues. This study introduces test sequence guidelines based on relevant principles and considerations for electronic testing according to International Standard IEC-60068-1 and the United States Military Standard MIL-STD-810G. Test sequences were then proposed based on the descriptions for each test. Finally, General Motors (GM) specification GMW3172 was interpreted and compared to IEC-60068-1 and MIL-STD-810G.
Keywords: Reliability, Environmental test sequence, Electrical units, IEC 60068-1, MIL-STD-810G.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43222962 Single Frame Supercompression of Still Images,Video, High Definition TV and Digital Cinema
Authors: Mario Mastriani
Abstract:
Super-resolution is nowadays used for a high-resolution image produced from several low-resolution noisy frames. In this work, we consider the problem of high-quality interpolation of a single noise-free image. Such images may come from different sources, i.e., they may be frames of videos, individual pictures, etc. On the other hand, in the encoder we apply a downsampling via bidimen-sional interpolation of each frame, and in the decoder we apply a upsampling by which we restore the original size of the image. If the compression ratio is very high, then we use a convolutive mask that restores the edges, eliminating the blur. Finally, both, the encoder and the complete decoder are implemented on General-Purpose computation on Graphics Processing Units (GPGPU) cards. In fact, the mentioned mask is coded inside texture memory of a GPGPU.Keywords: General-Purpose computation on Graphics ProcessingUnits, Image Compression, Interpolation, Super-resolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20002961 Distortion Estimation in Digital Image Watermarking using Genetic Programming
Authors: Labiba Gilani, Asifullah Khan, Anwar M. Mirza
Abstract:
This paper introduces a technique of distortion estimation in image watermarking using Genetic Programming (GP). The distortion is estimated by considering the problem of obtaining a distorted watermarked signal from the original watermarked signal as a function regression problem. This function regression problem is solved using GP, where the original watermarked signal is considered as an independent variable. GP-based distortion estimation scheme is checked for Gaussian attack and Jpeg compression attack. We have used Gaussian attacks of different strengths by changing the standard deviation. JPEG compression attack is also varied by adding various distortions. Experimental results demonstrate that the proposed technique is able to detect the watermark even in the case of strong distortions and is more robust against attacks.Keywords: Blind Watermarking, Genetic Programming (GP), Fitness Function, Discrete Cosine Transform (DCT).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17112960 Yield Prediction Using Support Vectors Based Under-Sampling in Semiconductor Process
Authors: Sae-Rom Pak, Seung Hwan Park, Jeong Ho Cho, Daewoong An, Cheong-Sool Park, Jun Seok Kim, Jun-Geol Baek
Abstract:
It is important to predict yield in semiconductor test process in order to increase yield. In this study, yield prediction means finding out defective die, wafer or lot effectively. Semiconductor test process consists of some test steps and each test includes various test items. In other world, test data has a big and complicated characteristic. It also is disproportionably distributed as the number of data belonging to FAIL class is extremely low. For yield prediction, general data mining techniques have a limitation without any data preprocessing due to eigen properties of test data. Therefore, this study proposes an under-sampling method using support vector machine (SVM) to eliminate an imbalanced characteristic. For evaluating a performance, randomly under-sampling method is compared with the proposed method using actual semiconductor test data. As a result, sampling method using SVM is effective in generating robust model for yield prediction.
Keywords: Yield Prediction, Semiconductor Test Process, Support Vector Machine, Under Sampling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 24012959 An Improvement of PDLZW implementation with a Modified WSC Updating Technique on FPGA
Authors: Perapong Vichitkraivin, Orachat Chitsobhuk
Abstract:
In this paper, an improvement of PDLZW implementation with a new dictionary updating technique is proposed. A unique dictionary is partitioned into hierarchical variable word-width dictionaries. This allows us to search through dictionaries in parallel. Moreover, the barrel shifter is adopted for loading a new input string into the shift register in order to achieve a faster speed. However, the original PDLZW uses a simple FIFO update strategy, which is not efficient. Therefore, a new window based updating technique is implemented to better classify the difference in how often each particular address in the window is referred. The freezing policy is applied to the address most often referred, which would not be updated until all the other addresses in the window have the same priority. This guarantees that the more often referred addresses would not be updated until their time comes. This updating policy leads to an improvement on the compression efficiency of the proposed algorithm while still keep the architecture low complexity and easy to implement.Keywords: lossless data compression, LZW algorithm, PDLZW algorithm, WSC and dictionary update.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16292958 Application of Machine Learning Methods to Online Test Error Detection in Semiconductor Test
Authors: Matthias Kirmse, Uwe Petersohn, Elief Paffrath
Abstract:
As in today's semiconductor industries test costs can make up to 50 percent of the total production costs, an efficient test error detection becomes more and more important. In this paper, we present a new machine learning approach to test error detection that should provide a faster recognition of test system faults as well as an improved test error recall. The key idea is to learn a classifier ensemble, detecting typical test error patterns in wafer test results immediately after finishing these tests. Since test error detection has not yet been discussed in the machine learning community, we define central problem-relevant terms and provide an analysis of important domain properties. Finally, we present comparative studies reflecting the failure detection performance of three individual classifiers and three ensemble methods based upon them. As base classifiers we chose a decision tree learner, a support vector machine and a Bayesian network, while the compared ensemble methods were simple and weighted majority vote as well as stacking. For the evaluation, we used cross validation and a specially designed practical simulation. By implementing our approach in a semiconductor test department for the observation of two products, we proofed its practical applicability.
Keywords: Ensemble methods, fault detection, machine learning, semiconductor test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22762957 Depression and Its Effects on a Cognitive Performance Test
Authors: C. Noi-Okwei
Abstract:
In this study, participants with adjustment disorder with depressed mood (aged 18-54 years) with mild depression (N=18), severe depression (N=12) were compared with healthy controls (N=20) on the Multidimensional Aptitude Battery (MAB) a cognitive performance test. Using One Way Analysis of Variance and Matched Sample t-test. The results of the analysis shows that severely depressed participants performed poorly on the cognitive performance test relative to controls, however there were no significant differences on the cognitive performance test scores between the severely depressed and the mildly depressed. In addition, performance on the non-verbal performance subtest was poorer than that of the verbal subtest, suggesting that depression affects the executive functions of the person.
Keywords: adjustment disorder, cognitive performance test, Depression,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 44432956 A Note on Penalized Power-Divergence Test Statistics
Authors: Aylin Alin
Abstract:
In this paper, penalized power-divergence test statistics have been defined and their exact size properties to test a nested sequence of log-linear models have been compared with ordinary power-divergence test statistics for various penalization, λ and main effect values. Since the ordinary and penalized power-divergence test statistics have the same asymptotic distribution, comparisons have been only made for small and moderate samples. Three-way contingency tables distributed according to a multinomial distribution have been considered. Simulation results reveal that penalized power-divergence test statistics perform much better than their ordinary counterparts.
Keywords: Contingency table, Log-linear models, Penalization, Power-divergence measure, Penalized power-divergence measure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13182955 Digital filters for Hot-Mix Asphalt Complex Modulus Test Data Using Genetic Algorithm Strategies
Authors: Madhav V. Chitturi, Anshu Manik, Kasthurirangan Gopalakrishnan
Abstract:
The dynamic or complex modulus test is considered to be a mechanistically based laboratory test to reliably characterize the strength and load-resistance of Hot-Mix Asphalt (HMA) mixes used in the construction of roads. The most common observation is that the data collected from these tests are often noisy and somewhat non-sinusoidal. This hampers accurate analysis of the data to obtain engineering insight. The goal of the work presented in this paper is to develop and compare automated evolutionary computational techniques to filter test noise in the collection of data for the HMA complex modulus test. The results showed that the Covariance Matrix Adaptation-Evolutionary Strategy (CMA-ES) approach is computationally efficient for filtering data obtained from the HMA complex modulus test.Keywords: HMA, dynamic modulus, GA, evolutionarycomputation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15722954 Optimal Image Representation for Linear Canonical Transform Multiplexing
Authors: Navdeep Goel, Salvador Gabarda
Abstract:
Digital images are widely used in computer applications. To store or transmit the uncompressed images requires considerable storage capacity and transmission bandwidth. Image compression is a means to perform transmission or storage of visual data in the most economical way. This paper explains about how images can be encoded to be transmitted in a multiplexing time-frequency domain channel. Multiplexing involves packing signals together whose representations are compact in the working domain. In order to optimize transmission resources each 4 × 4 pixel block of the image is transformed by a suitable polynomial approximation, into a minimal number of coefficients. Less than 4 × 4 coefficients in one block spares a significant amount of transmitted information, but some information is lost. Different approximations for image transformation have been evaluated as polynomial representation (Vandermonde matrix), least squares + gradient descent, 1-D Chebyshev polynomials, 2-D Chebyshev polynomials or singular value decomposition (SVD). Results have been compared in terms of nominal compression rate (NCR), compression ratio (CR) and peak signal-to-noise ratio (PSNR) in order to minimize the error function defined as the difference between the original pixel gray levels and the approximated polynomial output. Polynomial coefficients have been later encoded and handled for generating chirps in a target rate of about two chirps per 4 × 4 pixel block and then submitted to a transmission multiplexing operation in the time-frequency domain.Keywords: Chirp signals, Image multiplexing, Image transformation, Linear canonical transform, Polynomial approximation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21312953 Evaluation of Video Quality Metrics and Performance Comparison on Contents Taken from Most Commonly Used Devices
Authors: Pratik Dhabal Deo, Manoj P.
Abstract:
With the increasing number of social media users, the amount of video content available has also significantly increased. Currently, the number of smartphone users is at its peak, and many are increasingly using their smartphones as their main photography and recording devices. There have been a lot of developments in the field of video quality assessment in since the past years and more research on various other aspects of video and image are being done. Datasets that contain a huge number of videos from different high-end devices make it difficult to analyze the performance of the metrics on the content from most used devices even if they contain contents taken in poor lighting conditions using lower-end devices. These devices face a lot of distortions due to various factors since the spectrum of contents recorded on these devices is huge. In this paper, we have presented an analysis of the objective Video Quality Analysis (VQA) metrics on contents taken only from most used devices and their performance on them, focusing on full-reference metrics. To carry out this research, we created a custom dataset containing a total of 90 videos that have been taken from three most commonly used devices, and Android smartphone, an iOS smartphone and a Digital Single-Lens Reflex (DSLR) camera. On the videos taken on each of these devices, the six most common types of distortions that users face have been applied in addition to already existing H.264 compression based on four reference videos. These six applied distortions have three levels of degradation each. A total of the five most popular VQA metrics have been evaluated on this dataset and the highest values and the lowest values of each of the metrics on the distortions have been recorded. Finally, it is found that blur is the artifact on which most of the metrics did not perform well. Thus, in order to understand the results better the amount of blur in the data set has been calculated and an additional evaluation of the metrics was done using High Efficiency Video Coding (HEVC) codec, which is the next version of H.264 compression, on the camera that proved to be the sharpest among the devices. The results have shown that as the resolution increases, the performance of the metrics tends to become more accurate and the best performing metric among them is VQM with very few inconsistencies and inaccurate results when the compression applied is H.264, but when the compression is applied is HEVC, Structural Similarity (SSIM) metric and Video Multimethod Assessment Fusion (VMAF) have performed significantly better.
Keywords: Distortion, metrics, recording, frame rate, video quality assessment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3702952 Computational Evaluation of a C-A Heat Pump
Authors: Young-Jin Baik, Minsung Kim, Young-Soo Lee, Ki-Chang Chang, Seong-Ryong Park
Abstract:
The compression-absorption heat pump (C-A HP), one of the promising heat recovery equipments that make process hot water using low temperature heat of wastewater, was evaluated by computer simulation. A simulation program was developed based on the continuity and the first and second laws of thermodynamics. Both the absorber and desorber were modeled using UA-LMTD method. In order to prevent an unfeasible temperature profile and to reduce calculation errors from the curved temperature profile of a mixture, heat loads were divided into lots of segments. A single-stage compressor was considered. A compressor cooling load was also taken into account. An isentropic efficiency was computed from the map data. Simulation conditions were given based on the system consisting of ordinarily designed components. The simulation results show that most of the total entropy generation occurs during the compression and cooling process, thus suggesting the possibility that system performance can be enhanced if a rectifier is introduced.Keywords: Waste heat recovery, Heat Pump.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17212951 Performance Analysis of MT Evaluation Measures and Test Suites
Authors: Yao Jian-Min, Lv Qiang, Zhang Jing
Abstract:
Many measures have been proposed for machine translation evaluation (MTE) while little research has been done on the performance of MTE methods. This paper is an effort for MTE performance analysis. A general frame is proposed for the description of the MTE measure and the test suite, including whether the automatic measure is consistent with human evaluation, whether different results from various measures or test suites are consistent, whether the content of the test suite is suitable for performance evaluation, the degree of difficulty of the test suite and its influence on the MTE, the relationship of MTE result significance and the size of the test suite, etc. For a better clarification of the frame, several experiment results are analyzed relating human evaluation, BLEU evaluation, and typological MTE. A visualization method is introduced for better presentation of the results. The study aims for aid in construction of test suite and method selection in MTE practice.Keywords: Machine translation, natural language processing, visualization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17082950 Study of Dual Fuel Engine as Environmentally Friendly Engine
Authors: Nilam S. Octaviani, Semin
Abstract:
The diesel engine is an internal combustion engine that uses compressed air to combust. The diesel engines are widely used in the world because it has the most excellent combustion efficiency than other types of internal combustion engine. However, the exhaust emissions of it produce pollutants that are harmful to human health and the environment. Therefore, natural gas used as an alternative fuel using on compression ignition engine to respond those environment issues. This paper aims to discuss the comparison of the technical characteristics and exhaust gases emission from conventional diesel engine and dual fuel diesel engine. According to the study, the dual fuel engine applications have a lower compression pressure and has longer ignition delay compared with normal diesel mode. The engine power is decreased at dual fuel mode. However, the exhaust gases emission on dual fuel engine significantly reduce the nitrogen oxide (NOx), carbon dioxide (CO2) and particular metter (PM) emissions.
Keywords: Diesel engine, dual fuel engine, emissions, technical characteristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8962949 FPGA Hardware Implementation and Evaluation of a Micro-Network Architecture for Multi-Core Systems
Authors: Yahia Salah, Med Lassaad Kaddachi, Rached Tourki
Abstract:
This paper presents the design, implementation and evaluation of a micro-network, or Network-on-Chip (NoC), based on a generic pipeline router architecture. The router is designed to efficiently support traffic generated by multimedia applications on embedded multi-core systems. It employs a simplest routing mechanism and implements the round-robin scheduling strategy to resolve output port contentions and minimize latency. A virtual channel flow control is applied to avoid the head-of-line blocking problem and enhance performance in the NoC. The hardware design of the router architecture has been implemented at the register transfer level; its functionality is evaluated in the case of the two dimensional Mesh/Torus topology, and performance results are derived from ModelSim simulator and Xilinx ISE 9.2i synthesis tool. An example of a multi-core image processing system utilizing the NoC structure has been implemented and validated to demonstrate the capability of the proposed micro-network architecture. To reduce complexity of the image compression and decompression architecture, the system use image processing algorithm based on classical discrete cosine transform with an efficient zonal processing approach. The experimental results have confirmed that both the proposed image compression scheme and NoC architecture can achieve a reasonable image quality with lower processing time.
Keywords: Generic Pipeline Network-on-Chip Router Architecture, JPEG Image Compression, FPGA Hardware Implementation, Performance Evaluation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 30972948 Message Framework for Disaster Management: An Application Model for Mines
Authors: A. Baloğlu, A. Çınar
Abstract:
Different tools and technologies were implemented for Crisis Response and Management (CRM) which is generally using available network infrastructure for information exchange. Depending on type of disaster or crisis, network infrastructure could be affected and it could not be able to provide reliable connectivity. Thus any tool or technology that depends on the connectivity could not be able to fulfill its functionalities. As a solution, a new message exchange framework has been developed. Framework provides offline/online information exchange platform for CRM Information Systems (CRMIS) and it uses XML compression and packet prioritization algorithms and is based on open source web technologies. By introducing offline capabilities to the web technologies, framework will be able to perform message exchange on unreliable networks. The experiments done on the simulation environment provide promising results on low bandwidth networks (56kbps and 28.8 kbps) with up to 50% packet loss and the solution is to successfully transfer all the information on these low quality networks where the traditional 2 and 3 tier applications failed.
Keywords: Crisis Response and Management, XML Messaging, Web Services, XML compression, Mining.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19042947 ROI Based Embedded Watermarking of Medical Images for Secured Communication in Telemedicine
Authors: Baisa L. Gunjal, Suresh N. Mali
Abstract:
Medical images require special safety and confidentiality because critical judgment is done on the information provided by medical images. Transmission of medical image via internet or mobile phones demands strong security and copyright protection in telemedicine applications. Here, highly secured and robust watermarking technique is proposed for transmission of image data via internet and mobile phones. The Region of Interest (ROI) and Non Region of Interest (RONI) of medical image are separated. Only RONI is used for watermark embedding. This technique results in exact recovery of watermark with standard medical database images of size 512x512, giving 'correlation factor' equals to 1. The correlation factor for different attacks like noise addition, filtering, rotation and compression ranges from 0.90 to 0.95. The PSNR with weighting factor 0.02 is up to 48.53 dBs. The presented scheme is non blind and embeds hospital logo of 64x64 size.
Keywords: Compression, DWT, ROI, Scrambling, Vertices
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32822946 A Proposal for Systematic Mapping Study of Software Security Testing, Verification and Validation
Authors: Adriano Bessa Albuquerque, Francisco Jose Barreto Nunes
Abstract:
Software vulnerabilities are increasing and not only impact services and processes availability as well as information confidentiality, integrity and privacy, but also cause changes that interfere in the development process. Security test could be a solution to reduce vulnerabilities. However, the variety of test techniques with the lack of real case studies of applying tests focusing on software development life cycle compromise its effective use. This paper offers an overview of how a Systematic Mapping Study (MS) about security verification, validation and test (VVT) was performed, besides presenting general results about this study.
Keywords: Software test, software security verification validation and test, security test institutionalization, systematic mapping study.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16252945 Experimental Testing of Composite Tubes with Different Corrugation Profile Subjected to Lateral Compression Load
Authors: Elfetori F. Abdewi
Abstract:
This paper presents the effect of corrugation profile geometry on the crushing behavior, energy absorption, failure mechanism, and failure mode of woven roving glass fibre/epoxy laminated composite tube. Experimental investigations were carried out on composite tubes with three different profile shapes: sinusoidal, triangular and trapezoidal. The tubes were subjected to lateral compressive loading. On the addition to a radial corrugated composite tube, cylindrical composite tube, were fabricated and tested under the same condition in order to know the effect of corrugation geometry. Typical histories of their deformation are presented. Behavior of tubes as regards the peak crushing load, energy absorbed and mode of crushing has been discussed. The results show that the behavior of the tube under lateral compression load is influenced by the geometry of the tube itself.Keywords: Corrugated composite specimens, Energy absorption, Lateral crushing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23402944 Aging Effect on Mechanical Behavior of Duplex Satinless Steel
Authors: Jungho Moon, Tae Kwon Ha
Abstract:
Effect of alloying on the microstructure and mechanical properties of heat-resisting duplex stainless steel (DSS) for Mg production was investigated in this study. 25Cr-8Ni based DSS’s were cast into rectangular ingots of which the dimension was 350×350×100 mm3 . Nitrogen and Yttrium were added in the range within 0.3 in weight percent. Phase equilibrium was calculated using the FactSage®, thermodynamic software. Hot exposure, high temperature tensile and compression tests were conducted on the ingots at 1230oC, which is operation temperature employed for Mg production by Silico-thermic reduction. The steel with N and Y showed much higher strength than 310S alloy in both tensile and compression tests. By thermal exposition at 1230oC for 200 hrs, hardness of DSS containing N and Y was found to increase. Hot workability of the heat-resisting DSS was evaluated by employing hot rolling at 1230 oC. Hot shortness was observed in the ingot with N and found to disappear after addition of Y.
Keywords: Duplex Stainless Steel, alloying elements, eutectic carbides, microstructure, aging treatment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14622943 Bending Gradient Coefficient Correction for I-Beams
Authors: H. R. Kazemi Nia, A. Yeganeh Fallah
Abstract:
Without uncertainty by applying external loads on beams, bending is created. The created bending in I-beams, puts one of the flanges in tension and the other one in compression. With increasing of bending, compression flange buckled and beam in out of its plane direction twisted, this twisting well-known as Lateral Torsional Buckling. Providing bending moment varieties along the beam, the critical moment is greater than the case its under pure bending. In other words, the value of bending gradient coefficient is always greater than unite. In this article by the use of " ANSYS 10.0" software near 80 3-D finite element models developed for the propose of analyzing beams` lateral torsional buckling and surveying influence of slenderness on beams' bending gradient coefficient. Results show that, presented Cb coefficient via AISC is not correct for some of beams and value of this coefficient is smaller than what proposed by AISC. Therefore instead of using a constant Cb for each case of loading , a function with two criterion for calculation of Cb coefficient for some cases is proposed.Keywords: Beams critical moment, Bending Gradient Coefficient, finite element, Lateral Torsional Buckling
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 45322942 Mathematical Analysis of Matrix and Filler Formulation in Composite Materials
Authors: Olusegun A. Afolabi, Ndivhuwo Ndou
Abstract:
Composite material is an important area that has gained global visibility in many research fields in recent years. Composite material is the combination of separate materials with different properties to form a single material having different properties from the parent materials. Material composition and combination is an important aspect of composite material. The focus of this study is to provide insight into an easy way of calculating the compositions and formulations of constituent materials that make up any composite material. The compositions of the matrix and filler used for fabricating composite materials are taken into consideration. From the composite fabricated, data can be collected and analyzed based on the test and characterizations such as tensile, flexural, compression, impact, hardness, etc. Also, the densities of the matrix and the filler with regard to their constituent materials are discussed.
Keywords: Composite material, density, filler, matrix, percentage weight, volume fraction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 62941 A Dual Digital-Image Watermarking Technique
Authors: Maha Sharkas, Dahlia ElShafie, Nadder Hamdy
Abstract:
Image watermarking has become an important tool for intellectual property protection and authentication. In this paper a watermarking technique is suggested that incorporates two watermarks in a host image for improved protection and robustness. A watermark, in form of a PN sequence (will be called the secondary watermark), is embedded in the wavelet domain of a primary watermark before being embedded in the host image. The technique has been tested using Lena image as a host and the camera man as the primary watermark. The embedded PN sequence was detectable through correlation among other five sequences where a PSNR of 44.1065 dB was measured. Furthermore, to test the robustness of the technique, the watermarked image was exposed to four types of attacks, namely compression, low pass filtering, salt and pepper noise and luminance change. In all cases the secondary watermark was easy to detect even when the primary one is severely distorted.Keywords: DWT, Image watermarking, watermarkingtechniques, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27132940 Do C-Test and Cloze Procedure Measure what they Purport to be Measuring? A Case of Criterion-Related Validity
Authors: Masoud Saeedi, Mansour Tavakoli, Shirin Rahimi Kazerooni, Vahid Parvaresh
Abstract:
This article investigated the validity of C-test and Cloze test which purport to measure general English proficiency. To provide empirical evidence pertaining to the validity of the interpretations based on the results of these integrative language tests, their criterion-related validity was investigated. In doing so, the test of English as a foreign language (TOEFL) which is an established, standardized, and internationally administered test of general English proficiency was used as the criterion measure. Some 90 Iranian English majors participated in this study. They were seniors studying English at a university in Tehran, Iran. The results of analyses showed that there is a statistically significant correlation among participants- scores on Cloze test, C-test, and the TOEFL. Building on the findings of the study and considering criterion-related validity as the evidential basis of the validity argument, it was cautiously deducted that these tests measure the same underlying trait. However, considering the limitations of using criterion measures to validate tests, no absolute claims can be made as to the construct validity of these integrative tests.
Keywords: Integrative testing, C-test, Cloze test, theTOEFL, Validity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 33302939 Preparation and Some Mechanical Properties of Composite Materials Made from Sawdust, Cassava Starch and Natural Rubber Latex
Authors: Apusraporn Prompunjai, Waranyou Sridach
Abstract:
The composite materials were prepared by sawdust, cassava starch and natural rubber latex (NR). The mixtures of 15%w/v gelatinized cassava starch and 15%w/v PVOH were used as the binder of these composite materials. The concentrated rubber latex was added to the mixtures. They were mixed rigorously to the treated sawdust in the ratio of 70:30 until achive uniform dispersion. The batters were subjected to the hot compression moulding at the temperature of 160°C and 3,000 psi pressure for 5 min. The experimental results showed that the mechanical properties of composite materials, which contained the gelatinized cassava starch and PVOH in the ratio of 2:1, 20% NR latex by weight of the dry starch and treated sawdust with 5%NaOH or 1% BPO, were the best. It contributed the maximal compression strength (341.10 + 26.11 N), puncture resistance (8.79 + 0.98 N/mm2) and flexural strength (3.99 + 0.72N/mm2). It is also found that the physicochemical and mechanical properties of composites strongly depends on the interface quality of sawdust, cassava starch and NR latex.
Keywords: Composites, sawdust, cassava starch, natural rubber (NR) latex, surface chemical treatments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 41012938 Developing Kazakh Language Fluency Test in Nazarbayev University
Authors: Saule Mussabekova, Samal Abzhanova
Abstract:
The Kazakh Language Fluency Test, based on the IELTS exam, was implemented in 2012 at Nazarbayev University in Astana, Kazakhstan. We would like to share our experience in developing this exam and some exam results with other language instructors. In this paper, we will cover all these peculiarities and their related issues. The Kazakh Language Fluency Test is a young exam. During its development, we faced many difficulties. One of the goals of the university and the country is to encourage fluency in the Kazakh language for all citizens of the Republic. Nazarbayev University has introduced a Kazakh language program to assist in achieving this goal. This policy is one-step in ensuring that NU students have a thorough understanding of the Kazakh language through a fluency test based on the International English Language Testing System (IELTS). The Kazakh Language Fluency Test exam aims to determine student’s knowledge of Kazakh language. The fact is that there are three types of students at Nazarbayev University: Kazakh-speaking heritage learners, Russian-speaking and English-speaking students. Unfortunately, we have Kazakh students who do not speak Kazakh. All students who finished school with Russian language instruction are given Kazakh Language Fluency Test in order to determine their Kazakh level. After the test exam, all students can choose appropriate Kazakh course: Basic Kazakh, Intermediate Kazakh and Upper-Intermediate Kazakh. The Kazakh Language Fluency Test consists of four parts: Listening, Reading, Writing and Speaking. They are taken on the same day in the abovementioned order.
Keywords: Diagnostic language test, Kazakh language, placement test, test result.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9782937 Structural Investigation of Na2O–B2O3–SiO2 Glasses Doped with NdF3
Authors: M. S. Gaafar, S. Y. Marzouk
Abstract:
Sodium borosilicate glasses doped with different content of NdF3 mol % have been prepared by rapid quenching method. Ultrasonic velocities (both longitudinal and shear) measurements have been carried out at room temperature and at ultrasonic frequency of 4 MHz. Elastic moduli, Debye temperature, softening temperature and Poisson's ratio have been obtained as a function of NdF3 modifier content. Results showed that the elastic moduli, Debye temperature, softening temperature and Poisson's ratio have very slight change with the change of NdF3 mol % content. Based on FTIR spectroscopy and theoretical (Bond compression) model, quantitative analysis has been carried out in order to obtain more information about the structure of these glasses. The study indicated that the structure of these glasses is mainly composed of SiO4 units with four bridging oxygens (Q4), and with three bridging and one nonbridging oxygens (Q3).Keywords: Borosilicate glasses, ultrasonic velocity, elastic moduli, FTIR spectroscopy, bond compression model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17532936 Influence of Deficient Materials on the Reliability of Reinforced Concrete Members
Authors: Sami W. Tabsh
Abstract:
The strength of reinforced concrete depends on the member dimensions and material properties. The properties of concrete and steel materials are not constant but random variables. The variability of concrete strength is due to batching errors, variations in mixing, cement quality uncertainties, differences in the degree of compaction and disparity in curing. Similarly, the variability of steel strength is attributed to the manufacturing process, rolling conditions, characteristics of base material, uncertainties in chemical composition, and the microstructure-property relationships. To account for such uncertainties, codes of practice for reinforced concrete design impose resistance factors to ensure structural reliability over the useful life of the structure. In this investigation, the effects of reductions in concrete and reinforcing steel strengths from the nominal values, beyond those accounted for in the structural design codes, on the structural reliability are assessed. The considered limit states are flexure, shear and axial compression based on the ACI 318-11 structural concrete building code. Structural safety is measured in terms of a reliability index. Probabilistic resistance and load models are compiled from the available literature. The study showed that there is a wide variation in the reliability index for reinforced concrete members designed for flexure, shear or axial compression, especially when the live-to-dead load ratio is low. Furthermore, variations in concrete strength have minor effect on the reliability of beams in flexure, moderate effect on the reliability of beams in shear, and sever effect on the reliability of columns in axial compression. On the other hand, changes in steel yield strength have great effect on the reliability of beams in flexure, moderate effect on the reliability of beams in shear, and mild effect on the reliability of columns in axial compression. Based on the outcome, it can be concluded that the reliability of beams is sensitive to changes in the yield strength of the steel reinforcement, whereas the reliability of columns is sensitive to variations in the concrete strength. Since the embedded target reliability in structural design codes results in lower structural safety in beams than in columns, large reductions in material strengths compromise the structural safety of beams much more than they affect columns.
Keywords: Code, flexure, limit states, random variables, reinforced concrete, reliability, reliability index, shear, structural safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2585