Search results for: width depth ratio.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2861

Search results for: width depth ratio.

311 Analysis of the Elastic Energy Released and Characterization of the Eruptive Episodes Intensity’s during 2014-2015 at El Reventador Volcano, Ecuador

Authors: Paúl I. Cornejo

Abstract:

The elastic energy released through Strombolian explosions has been quite studied, detailing various processes, sources, and precursory events at several volcanoes. We realized an analysis based on the relative partitioning of the elastic energy radiated into the atmosphere and ground by Strombolian-type explosions recorded at El Reventador volcano, using infrasound and seismic signals at high and moderate seismicity episodes during intense eruptive stages of explosive and effusive activity. Our results show that considerable values of Volcano Acoustic-Seismic Ratio (VASR or η) are obtained at high seismicity stages. VASR is a physical diagnostic of explosive degassing that we used to compare eruption mechanisms at El Reventador volcano for two datasets of explosions recorded at a Broad-Band BB seismic and infrasonic station located at ~5 kilometers from the vent. We conclude that the acoustic energy EA released during explosive activity (VASR η = 0.47, standard deviation σ = 0.8) is higher than the EA released during effusive activity; therefore, producing the highest values of η. Furthermore, we realized the analysis and characterization of the eruptive intensity for two episodes at high seismicity, calculating a η three-time higher for an episode of effusive activity with an occasional explosive component (η = 0.32, and σ = 0.42), than a η for an episode of only effusive activity (η = 0.11, and σ = 0.18), but more energetic.

Keywords: Effusive, explosion quakes, explosive, strombolian, VASR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 785
310 Bone Mineral Density and Quality, Body Composition of Women in the Postmenopausal Period

Authors: Vladyslav Povoroznyuk, Oksana Ivanyk, Nataliia Dzerovych

Abstract:

In the diagnostics of osteoporosis, the gold standard is considered to be bone mineral density; however, X-ray densitometry is not an accurate indicator of osteoporotic fracture risk under all circumstances. In this regard, the search for new methods that could determine the indicators not only of the mineral density, but of the bone tissue quality, is a logical step for diagnostic optimization. One of these methods is the evaluation of trabecular bone quality. The aim of this study was to examine the quality and mineral density of spine bone tissue, femoral neck, and body composition of women depending on the duration of the postmenopausal period, to determine the correlation of body fat with indicators of bone mineral density and quality. The study examined 179 women in premenopausal and postmenopausal periods. The patients were divided into the following groups: Women in the premenopausal period and women in the postmenopausal period at various stages (early, middle, late postmenopause). A general examination and study of the above parameters were conducted with General Electric X-ray densitometer. The results show that bone quality and mineral density probably deteriorate with advancing of postmenopausal period. Total fat and lean mass ratio is not likely to change with age. In the middle and late postmenopausal periods, the bone tissue mineral density of the spine and femoral neck increases along with total fat mass.

Keywords: Osteoporosis, bone tissue mineral density, bone quality, fat mass, lean mass, postmenopausal osteoporosis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 941
309 Target Detection using Adaptive Progressive Thresholding Based Shifted Phase-Encoded Fringe-Adjusted Joint Transform Correlator

Authors: Inder K. Purohit, M. Nazrul Islam, K. Vijayan Asari, Mohammad A. Karim

Abstract:

A new target detection technique is presented in this paper for the identification of small boats in coastal surveillance. The proposed technique employs an adaptive progressive thresholding (APT) scheme to first process the given input scene to separate any objects present in the scene from the background. The preprocessing step results in an image having only the foreground objects, such as boats, trees and other cluttered regions, and hence reduces the search region for the correlation step significantly. The processed image is then fed to the shifted phase-encoded fringe-adjusted joint transform correlator (SPFJTC) technique which produces single and delta-like correlation peak for a potential target present in the input scene. A post-processing step involves using a peak-to-clutter ratio (PCR) to determine whether the boat in the input scene is authorized or unauthorized. Simulation results are presented to show that the proposed technique can successfully determine the presence of an authorized boat and identify any intruding boat present in the given input scene.

Keywords: Adaptive progressive thresholding, fringe adjusted filters, image segmentation, joint transform correlation, synthetic discriminant function

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1208
308 Expert Solutions to Affordable Housing Finance Challenges in Developing Economies

Authors: Timothy Akinwande, Eddie C. M. Hui

Abstract:

Housing the urban poor has remained a challenge for many years across the world, especially in developing economies, despite the apparent research attention and policy interventions. It is apt to investigate the prevalent affordable housing (AH) provision challenges using unconventional approaches. It is pragmatic to thoroughly examine housing experts to provide supply-side solutions to AH challenges and investigate informal settlers to deduce solutions from AH demand viewpoints. This study, being the supply-side investigation of an ongoing research, interrogated housing experts to determine significant expert solutions. Focus group discussions and in-depth interviews were conducted with housing experts in Nigeria. Through descriptive, content, and systematic thematic analyses of data, major findings are that deliberate finance models designed for the urban poor are the most significant housing finance solution in developing economies. Other findings are that adequately implemented rent control policies, deliberate Public-Private Partnership (PPP) approaches like inclusionary housing and land-value capture, and urban renewal programs to enlighten and tutor the urban poor on how to earn more, spend wisely, and invest in their own better housing will effectively solve AH finance challenges. Study findings are informative for the best approaches to achieve effective, AH finance for the urban poor in Nigeria, which is indispensable for the achievement of sustainable development goals. This research’s originality lies in the exploration of experts’ opinions in relation to AH finance to produce an equation model of critical solutions to AH finance challenges. Study data are useful resources for future pro-poor housing studies. This study makes housing policy-oriented recommendations toward effective, AH for the urban poor in developing countries.

Keywords: Affordable housing, effective affordable housing, housing policy, housing research, sustainable development, urban poor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 177
307 Effect of Biomass Feedstocks on the Production of Hydrogenated Biodiesel

Authors: Panatcha Bovornseripatai, Siriporn Jongpatiwut, Somchai Osuwan, Suchada Butnark

Abstract:

Hydrogenated biodiesel is one of the most promising renewable fuels. It has many advantages over conventional biodiesel, including higher cetane number, higher heating value, lower viscosity, and lower corrosiveness due to its absence of oxygen. From previous work, Pd/TiO2 gave high conversion and selectivity in hydrogenated biodiesel. In this work, the effect of biomass feedstocks (i.e. beef fat, chicken fat, pork fat, and jatropha oil) on the production of hydrogenated biodiesel over Pd/TiO2 has been studied. Biomass feedstocks were analyzed by ICP-OES (inductively coupled plasma optical emission spectrometry) to identify the content of impurities (i.e. P, K, Ca, Na, and Mg). The deoxygenation catalyst, Pd/TiO2, was prepared by incipient wetness impregnation (IWI) and tested in a continuous flow packed-bed reactor at 500 psig, 325°C, H2/feed molar ratio of 30, and LHSV of 4 h-1 for its catalytic activity and selectivity in hydrodeoxygenation. All feedstocks gave high selectivity in diesel specification range hydrocarbons and the main hydrocarbons were n-pentadecane (n-C15) and n-heptadecane (n- C17), resulting from the decarbonylation/decarboxylation reaction. Intermediates such as oleic acid, stearic acid, palmitic acid, and esters were also detected in minor amount. The conversion of triglycerides in jatropha oil is higher than those of chicken fat, pork fat, and beef fat, respectively. The higher concentration of metal impurities in feedstock, the lower conversion of feedstock.

Keywords: Hydrogenated biodiesel, hydrodeoxygenation, Pd/TiO2, biomass feedstock

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1969
306 Investigation of Compressive Strength of Slag-Based Geopolymer Concrete Incorporated with Rice Husk Ash Using 12M Alkaline Activator

Authors: Festus A. Olutoge, Ahmed A. Akintunde, Anuoluwapo S. Kolade, Aaron A. Chadee, Jovanca Smith

Abstract:

Geopolymer concrete's (GPC) compressive strength was investigated. The GPC was incorporated with rice husk ash (RHA) and ground granulated blast furnace slag (GGBFS), which may have potential in the construction industry to replace Portland limestone cement (PLC) concrete. The sustainable construction binders used were GGBFS and RHA, and a solution of sodium hydroxide (NaOH) and sodium silicate gel (Na2SiO3) was used as the 12-molar alkaline activator. Five GPC mixes comprising fine aggregates, coarse aggregates, GGBS, and RHA, and the alkaline solution in the ratio 2: 2.5: 1: 0.5, respectively, were prepared to achieve grade 40 concrete, and PLC was substituted with GGBFS and RHA in the ratios of 0:100, 25:75, 50:50, 75:25, and 100:0. A control mix was also prepared which comprised of 100% water and 100% PLC as the cementitious material. The GPC mixes were thermally cured at 60-80 ºC in an oven for approximately 24 h. After curing for 7 and 28 days, the compressive strength test results of the hardened GPC samples showed that GPC-Mix #3, comprising 50% GGBFS and 50% RHA, was the most efficient geopolymer mix. The mix had compressive strengths of 35.71 MPa and 47.26 MPa, 19.87% and 8.69% higher than the PLC concrete samples, which had 29.79 MPa and 43.48 MPa after 7 and 28 days, respectively. Therefore, GPC containing GGBFS incorporated with RHA is an efficient method of decreasing the use of PLC in conventional concrete production and reducing the high amounts of CO2 emitted into the atmosphere in the construction industry.

Keywords: Alkaline solution, cementitious material, geopolymer concrete, ground granulated blast furnace slag, rice husk ash.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 189
305 Quality Evaluation of Compressed MRI Medical Images for Telemedicine Applications

Authors: Seddeq E. Ghrare, Salahaddin M. Shreef

Abstract:

Medical image modalities such as computed tomography (CT), magnetic resonance imaging (MRI), ultrasound (US), X-ray are adapted to diagnose disease. These modalities provide flexible means of reviewing anatomical cross-sections and physiological state in different parts of the human body. The raw medical images have a huge file size and need large storage requirements. So it should be such a way to reduce the size of those image files to be valid for telemedicine applications. Thus the image compression is a key factor to reduce the bit rate for transmission or storage while maintaining an acceptable reproduction quality, but it is natural to rise the question of how much an image can be compressed and still preserve sufficient information for a given clinical application. Many techniques for achieving data compression have been introduced. In this study, three different MRI modalities which are Brain, Spine and Knee have been compressed and reconstructed using wavelet transform. Subjective and objective evaluation has been done to investigate the clinical information quality of the compressed images. For the objective evaluation, the results show that the PSNR which indicates the quality of the reconstructed image is ranging from (21.95 dB to 30.80 dB, 27.25 dB to 35.75 dB, and 26.93 dB to 34.93 dB) for Brain, Spine, and Knee respectively. For the subjective evaluation test, the results show that the compression ratio of 40:1 was acceptable for brain image, whereas for spine and knee images 50:1 was acceptable.

Keywords: Medical Image, Magnetic Resonance Imaging, Image Compression, Discrete Wavelet Transform, Telemedicine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2977
304 Alignment of e-Government Policy Formulation with Practical Implementation: The Case of Sub-Saharan Africa

Authors: W. Munyoka, F. M. Manzira

Abstract:

The purpose of this study is to analyze how varying alignment of e-Government policies in four countries in Sub-Saharan Africa Region, namely South Africa, Seychelles, Mauritius and Cape Verde lead to the success or failure of e-Government; and what should be done to ensure positive alignment that lead to e-Government project growth. In addition, the study aims to understand how various governments’ efforts in e-Government awareness campaign strategies, international cooperation, functional literacy and anticipated organizational change can influence implementation.

This study extensively explores contemporary research undertaken in the field of e-Government and explores the actual respective national ICT policies, strategies and implemented e-Government projects for in-depth comprehension of the status core. Data is analyzed qualitatively and quantitatively to reach a conclusion.

The study found that resounding successes in strategic e-Government alignment was achieved in Seychelles, Mauritius, South Africa and Cape Verde - (Ranked number 1 to 4 respectively).

The implications of the study is that policy makers in developing countries should put mechanisms in place for constant monitoring and evaluation of project implementation in line with ICT policies to ensure that e-Government projects reach maturity levels and do not die mid-way implementation as often noticed in many countries. The study recommends that countries within the region should make consented collaborative efforts and synergies with the private sector players and international donor agencies to achieve the implementation part of the set ICT policies.

Keywords: E-Government, ICT-Policy Alignment, Implementation, Sub-Saharan Africa.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2340
303 Spectral Amplitude Coding Optical CDMA: Performance Analysis of PIIN Reduction Using VC Code Family

Authors: Hassan Yousif Ahmed, Ibrahima Faye, N.M.Saad, S.A. Aljined

Abstract:

Multi-user interference (MUI) is the main reason of system deterioration in the Spectral Amplitude Coding Optical Code Division Multiple Access (SAC-OCDMA) system. MUI increases with the number of simultaneous users, resulting into higher probability bit rate and limits the maximum number of simultaneous users. On the other hand, Phase induced intensity noise (PIIN) problem which is originated from spontaneous emission of broad band source from MUI severely limits the system performance should be addressed as well. Since the MUI is caused by the interference of simultaneous users, reducing the MUI value as small as possible is desirable. In this paper, an extensive study for the system performance specified by MUI and PIIN reducing is examined. Vectors Combinatorial (VC) codes families are adopted as a signature sequence for the performance analysis and a comparison with reported codes is performed. The results show that, when the received power increases, the PIIN noise for all the codes increases linearly. The results also show that the effect of PIIN can be minimized by increasing the code weight leads to preserve adequate signal to noise ratio over bit error probability. A comparison study between the proposed code and the existing codes such as Modified frequency hopping (MFH), Modified Quadratic- Congruence (MQC) has been carried out.

Keywords: FBG, MUI, PIIN, SAC-OCDMA, VCC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2210
302 A Mobile Multihop Relay Dynamic TDD Scheme for Cellular Networks

Authors: Jong-Moon Chung, Hyung-Weon Cho, Ki-Yong Jin, Min-Hee Cho

Abstract:

In this paper, we present an analytical framework for the evaluation of the uplink performance of multihop cellular networks based on dynamic time division duplex (TDD). New wireless broadband protocols, such as WiMAX, WiBro, and 3G-LTE apply TDD, and mobile communication protocols under standardization (e.g., IEEE802.16j) are investigating mobile multihop relay (MMR) as a future technology. In this paper a novel MMR TDD scheme is presented, where the dynamic range of the frame is shared to traffic resources of asymmetric nature and multihop relaying. The mobile communication channel interference model comprises of inner and co-channel interference (CCI). The performance analysis focuses on the uplink due to the fact that the effects of dynamic resource allocation show significant performance degradation only in the uplink compared to time division multiple access (TDMA) schemes due to CCI [1-3], where the downlink results to be the same or better.The analysis was based on the signal to interference power ratio (SIR) outage probability of dynamic TDD (D-TDD) and TDMA systems,which are the most widespread mobile communication multi-user control techniques. This paper presents the uplink SIR outage probability with multihop results and shows that the dynamic TDD scheme applying MMR can provide a performance improvement compared to single hop applications if executed properly.

Keywords: Co-Channel Interference, Dynamic TDD, MobileMultihop Reply, Cellular Network, Time Division Multiple Access.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2343
301 ECG Based Reliable User Identification Using Deep Learning

Authors: R. N. Begum, Ambalika Sharma, G. K. Singh

Abstract:

Identity theft has serious ramifications beyond data and personal information loss. This necessitates the implementation of robust and efficient user identification systems. Therefore, automatic biometric recognition systems are the need of the hour, and electrocardiogram (ECG)-based systems are unquestionably the best choice due to their appealing inherent characteristics. The Convolutional Neural Networks (CNNs) are the recent state-of-the-art techniques for ECG-based user identification systems. However, the results obtained are significantly below standards, and the situation worsens as the number of users and types of heartbeats in the dataset grows. As a result, this study proposes a highly accurate and resilient ECG-based person identification system using CNN's dense learning framework. The proposed research explores explicitly the caliber of dense CNNs in the field of ECG-based human recognition. The study tests four different configurations of dense CNN which are trained on a dataset of recordings collected from eight popular ECG databases. With the highest False Acceptance Rate (FAR)  of 0.04% and the highest False Rejection Rate (FRR)  of 5%, the best performing network achieved an identification accuracy of 99.94%. The best network is also tested with various train/test split ratios. The findings show that DenseNets are not only extremely reliable, but also highly efficient. Thus, they might also be implemented in real-time ECG-based human recognition systems.

Keywords: Biometrics, dense networks, identification rate, train/test split ratio.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 541
300 Investigating Polynomial Interpolation Functions for Zooming Low Resolution Digital Medical Images

Authors: Maninder Pal

Abstract:

Medical digital images usually have low resolution because of nature of their acquisition. Therefore, this paper focuses on zooming these images to obtain better level of information, required for the purpose of medical diagnosis. For this purpose, a strategy for selecting pixels in zooming operation is proposed. It is based on the principle of analog clock and utilizes a combination of point and neighborhood image processing. In this approach, the hour hand of clock covers the portion of image to be processed. For alignment, the center of clock points at middle pixel of the selected portion of image. The minute hand is longer in length, and is used to gain information about pixels of the surrounding area. This area is called neighborhood pixels region. This information is used to zoom the selected portion of the image. The proposed algorithm is implemented and its performance is evaluated for many medical images obtained from various sources such as X-ray, Computerized Tomography (CT) scan and Magnetic Resonance Imaging (MRI). However, for illustration and simplicity, the results obtained from a CT scanned image of head is presented. The performance of algorithm is evaluated in comparison to various traditional algorithms in terms of Peak signal-to-noise ratio (PSNR), maximum error, SSIM index, mutual information and processing time. From the results, the proposed algorithm is found to give better performance than traditional algorithms.

Keywords: Zooming, interpolation, medical images, resolution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1576
299 Physico-chemical Treatment of Tar-Containing Wastewater Generated from Biomass Gasification Plants

Authors: Vrajesh Mehta, Anal Chavan

Abstract:

Treatment of tar-containing wastewater is necessary for the successful operation of biomass gasification plants (BGPs). In the present study, tar-containing wastewater was treated using lime and alum for the removal of in-organics, followed by adsorption on powdered activated carbon (PAC) for the removal of organics. Limealum experiments were performed in a jar apparatus and activated carbon studies were performed in an orbital shaker. At optimum concentrations, both lime and alum individually proved to be capable of removing color, total suspended solids (TSS) and total dissolved solids (TDS), but in both cases, pH adjustment had to be carried out after treatment. The combination of lime and alum at the dose ratio of 0.8:0.8 g/L was found to be optimum for the removal of inorganics. The removal efficiency achieved at optimum concentrations were 78.6, 62.0, 62.5 and 52.8% for color, alkalinity, TSS and TDS, respectively. The major advantages of the lime-alum combination were observed to be as follows: no requirement of pH adjustment before and after treatment and good settleability of sludge. Coagulation-precipitation followed by adsorption on PAC resulted in 92.3% chemical oxygen demand (COD) removal and 100% phenol removal at equilibrium. Ammonia removal efficiency was found to be 11.7% during coagulation-flocculation and 36.2% during adsorption on PAC. Adsorption of organics on PAC in terms of COD and phenol followed Freundlich isotherm with Kf = 0.55 & 18.47 mg/g and n = 1.01 & 1.45, respectively. This technology may prove to be one of the fastest and most techno-economically feasible methods for the treatment of tar-containing wastewater generated from BGPs.

Keywords: Activated carbon, Alum, Biomass gasification, Coagulation-flocculation, Lime, Tar-containing wastewater.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3672
298 Effects of Inlet Distorted Flows on the Performance of an Axial Compressor

Authors: Asad Islam, Khalid Parvez

Abstract:

Compressor fans in modern aircraft engines are of considerate importance, as they provide majority of thrust required by the aircraft. Their challenging environment is frequently subjected to non-uniform inflow conditions. These conditions could be either due to the flight operating requirements such as take-off and landing, wake interference from aircraft fuselage or cross-flow wind conditions. So, in highly maneuverable flights regimes of fighter aircrafts affects the overall performance of an engine. Since the flow in compressor of an aircraft application is highly sensitive because of adverse pressure gradient due to different flow orientations of the aircraft. Therefore, it is prone to unstable operations. This paper presents the study that focuses on axial compressor response to inlet flow orientations for the range of angles as 0 to 15 degrees. For this purpose, NASA Rotor-37 was taken and CFD mesh was developed. The compressor characteristics map was generated for the design conditions of pressure ratio of 2.106 with the rotor operating at rotational velocity of 17188.7 rpm using CFD simulating environment of ANSYS-CFX®. The grid study was done to see the effects of mesh upon computational solution. Then, the mesh giving the best results, (when validated with the available experimental NASA’s results); was used for further distortion analysis. The flow in the inlet nozzle was given angle orientations ranging from 0 to 15 degrees. The CFD results are analyzed and discussed with respect to stall margin and flow separations due to induced distortions.

Keywords: Angle, ANSYS-CFX®, axial compressor, Bladegen®, CFD, distortions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2012
297 Gluten-Free Cookies Enriched with Blueberry Pomace: Optimization of Baking Process

Authors: Aleksandra Mišan, Bojana Šarić, Nataša Nedeljković, Mladenka Pestorić, Pavle Jovanov, Milica Pojić, Jelena Tomić, Bojana Filipčev, Miroslav Hadnađev, Anamarija Mandić

Abstract:

With the aim of improving nutritional profile and antioxidant capacity of gluten-free cookies, blueberry pomace, by-product of juice production, was processed into a new food ingredient by drying and grinding and used for a gluten-free cookie formulation. Since the quality of a baked product is highly influenced by the baking conditions, the objective of this work was to optimize the baking time and thickness of dough pieces, by applying Response Surface Methodology (RSM) in order to obtain the best technological quality of the cookies. The experiments were carried out according to a Central Composite Design (CCD) by selecting the dough thickness and baking time as independent variables, while hardness, color parameters (L*, a* and b* values), water activity, diameter and short/long ratio were response variables. According to the results of RSM analysis, the baking time of 13.74min and dough thickness of 4.08mm was found to be the optimal for the baking temperature of 170°C. As similar optimal parameters were obtained by previously conducted experiment based on sensory analysis, response surface methodology (RSM) can be considered as a suitable approach to optimize the baking process.

Keywords: Baking process, blueberry pomace, gluten-free cookies, Response Surface Methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2575
296 Least Square-SVM Detector for Wireless BPSK in Multi-Environmental Noise

Authors: J. P. Dubois, Omar M. Abdul-Latif

Abstract:

Support Vector Machine (SVM) is a statistical learning tool developed to a more complex concept of structural risk minimization (SRM). In this paper, SVM is applied to signal detection in communication systems in the presence of channel noise in various environments in the form of Rayleigh fading, additive white Gaussian background noise (AWGN), and interference noise generalized as additive color Gaussian noise (ACGN). The structure and performance of SVM in terms of the bit error rate (BER) metric is derived and simulated for these advanced stochastic noise models and the computational complexity of the implementation, in terms of average computational time per bit, is also presented. The performance of SVM is then compared to conventional binary signaling optimal model-based detector driven by binary phase shift keying (BPSK) modulation. We show that the SVM performance is superior to that of conventional matched filter-, innovation filter-, and Wiener filter-driven detectors, even in the presence of random Doppler carrier deviation, especially for low SNR (signal-to-noise ratio) ranges. For large SNR, the performance of the SVM was similar to that of the classical detectors. However, the convergence between SVM and maximum likelihood detection occurred at a higher SNR as the noise environment became more hostile.

Keywords: Colour noise, Doppler shift, innovation filter, least square-support vector machine, matched filter, Rayleigh fading, Wiener filter.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1813
295 Development and Characterization of Re-Entrant Auxetic Fibrous Structures for Application in Ballistic Composites

Authors: Rui Magalhães, Sohel Rana, Raul Fangueiro, Clara Gonçalves, Pedro Nunes, Gustavo Dias

Abstract:

Auxetic fibrous structures and composites with negative Poisson’s ratio (NPR) have huge potential for application in ballistic protection due to their high energy absorption and excellent impact resistance. In the present research, re-entrant lozenge auxetic fibrous structures were produced through weft knitting technology using high performance polyamide and para-aramid fibres. Fabric structural parameters (e.g. loop length) and machine parameters (e.g. take down load) were varied in order to investigate their influence on the auxetic behaviours of the produced structures. These auxetic structures were then impregnated with two types of polymeric resins (epoxy and polyester) to produce composite materials, which were subsequently characterized for the auxetic behaviour. It was observed that the knitted fabrics produced using the polyamide yarns exhibited NPR over a wide deformation range, which was strongly dependant on the loop length and take down load. The polymeric composites produced from the auxetic fabrics also showed good auxetic property, which was superior in case of the polyester matrix. The experimental results suggested that these composites made from the auxetic fibrous structures can be properly designed to find potential use in the body amours for personal protection applications.

Keywords: Auxetic fabrics, high performance, composites, impact resistance, energy absorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 791
294 Acceptance of Consumer on Various Tempeh and Protein Content Comparison

Authors: Jaruwan Chutrtong

Abstract:

This research aims to study consumer acceptance of Tempeh from various raw materials (type of bean) and determine protein contents for comparison. Tempeh made from soybean, peanut, white kidney bean and sesame in the ratio: - soybean:sesame =1:0.1, soybean:white kidney:sesame =1:1:0.1, soybean:peanut:sesame =1:1:0.1 and peanut:white kidney bean: sesame =1:1:0.1. The study found that consumer is most satisfied with appearances on soybean mixed with white kidney and black sesame tempeh (3.98). The most satisfied tempeh with textures is soybean mixed with peanut and black sesame tempeh (4.00). The most satisfied tempeh with odor is peanut mixed with white kidney bean and black sesame tempeh (4.04). And the most satisfied tempeh with flavor is peanut mixed with white kidney bean and black sesame tempeh (4.2). The amount of protein in production, soybean tempeh has the highest protein. When we add sesame seeds, it made the protein content slightly decreased (1.86 and 0.6 %). When we use peanut as raw material, the protein content decreased 15.3%. And when we use white kidney bean as raw material, the protein content decreased (22.77- 26.11%). 

Keywords: Acceptance, bean, protein content, tempeh.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2024
293 Information Security Risk Management in IT-Based Process Virtualization: A Methodological Design Based on Action Research

Authors: Jefferson Camacho Mejía, Jenny Paola Forero Pachón, Luis Carlos Gómez Flórez

Abstract:

Action research is a qualitative research methodology, which leads the researcher to delve into the problems of a community in order to understand its needs in depth and finally, to propose actions that lead to a change of social paradigm. Although this methodology had its beginnings in the human sciences, it has attracted increasing interest and acceptance in the field of information systems research since the 1990s. The countless possibilities offered nowadays by the use of Information Technologies (IT) in the development of different socio-economic activities have meant a change of social paradigm and the emergence of the so-called information and knowledge society. According to this, governments, large corporations, small entrepreneurs and in general, organizations of all kinds are using IT to virtualize their processes, taking them from the physical environment to the digital environment. However, there is a potential risk for organizations related with exposing valuable information without an appropriate framework for protecting it. This paper shows progress in the development of a methodological design to manage the information security risks associated with the IT-based processes virtualization, by applying the principles of the action research methodology and it is the result of a systematic review of the scientific literature. This design consists of seven fundamental stages. These are distributed in the three stages described in the action research methodology: 1) Observe, 2) Analyze and 3) Take actions. Finally, this paper aims to offer an alternative tool to traditional information security management methodologies with a view to being applied specifically in the planning stage of IT-based process virtualization in order to foresee risks and to establish security controls before formulating IT solutions in any type of organization.

Keywords: Action research, information security, information technology, methodological design, process virtualization, risk management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 959
292 Investigation Corn and Soybean Intercropping Advantages in Competition with Redroot Pigweed and Jimsonweed

Authors: M. Rezvani, F. Zaefarian, M. Aghaalikhani, H. Rahimian Mashhadi, E. Zand

Abstract:

The spatial variation in plant species associated with intercropping is intended to reduce resource competition between species and increase yield potential. A field experiment was carried out on corn (Zea mays L.) and soybean (Glycine max L.) intercropping in a replacement series experiment with weed contamination consist of: weed free, infestation of redroot pigweed, infestation of jimsonweed and simultaneous infestation of redroot pigweed and jimsonweed in Karaj, Iran during 2007 growing season. The experimental design was a randomized complete block in factorial experiment with replicated thrice. Significant (P≤0.05) differences were observed in yield in intercropping. Corn yield was higher in intercropping, but soybean yield was significantly reduced by corn when intercropped. However, total productivity and land use efficiency were high under the intercropping system even in contamination of either species of weeds. Aggressivity of corn relative to soybean revealed the greater competitive ability of corn than soybean. Land equivalent ratio (LER) more than 1 in all treatments attributed to intercropping advantages and was highest in 50: 50 (corn/soybean) in weed free. These findings suggest that intercropping corn and soybean increase total productivity per unit area and improve land use efficiency. Considering the experimental findings, corn-soybean intercropping (50:50) may be recommended for yield advantage, more efficient utilization of resources, and weed suppression as a biological control.

Keywords: Corn, soybean, intercropping, redroot pigweed, jimsonweed.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2605
291 Experimental Evaluation of Drilling Damage on the Strength of Cores Extracted from RC Buildings

Authors: A. Masi, A. Digrisolo, G. Santarsiero

Abstract:

Concrete strength evaluated from compression tests on cores is affected by several factors causing differences from the in-situ strength at the location from which the core specimen was extracted. Among the factors, there is the damage possibly occurring during the drilling phase that generally leads to underestimate the actual in-situ strength. In order to quantify this effect, in this study two wide datasets have been examined, including: (i) about 500 core specimens extracted from Reinforced Concrete existing structures, and (ii) about 600 cube specimens taken during the construction of new structures in the framework of routine acceptance control. The two experimental datasets have been compared in terms of compression strength and specific weight values, accounting for the main factors affecting a concrete property, that is type and amount of cement, aggregates' grading, type and maximum size of aggregates, water/cement ratio, placing and curing modality, concrete age. The results show that the magnitude of the strength reduction due to drilling damage is strongly affected by the actual properties of concrete, being inversely proportional to its strength. Therefore, the application of a single value of the correction coefficient, as generally suggested in the technical literature and in structural codes, appears inappropriate. A set of values of the drilling damage coefficient is suggested as a function of the strength obtained from compressive tests on cores.

Keywords: RC Buildings, Assessment, In-situ concrete strength, Core testing, Drilling damage.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2059
290 Millennial Teachers of Canada: Innovation within the Boxed-In Constraints of Tradition

Authors: Lena Shulyakovskaya

Abstract:

Every year, schools aim to develop and adopt new technology and pedagogy as a way to equip today's students with the needed 21st Century skills. However, the field of primary and secondary education may not be as open to embracing change in reality. Despite the drive to reform and innovation, the field of education in Canada is still very much steeped in tradition and uses many of the practices that came into effect over 50 years ago. Among those are employment and retention practices. Millennials are the youngest generation of professionals entering the workplace at this time and the ones leaving their jobs within just a few years. Almost half of new teachers leave Canadian schools within their first five years on the job. This paper discusses one of the contributing factors that lead Canadian millennial teachers to either leave or stay in the profession - standardized education system. Using an exploratory case study approach, in-depth interviews with former and current millennial teachers were conducted to learn about their experiences within the K-12 system. Among the findings were the young teachers' concerns about the constant changes to teaching practices and technological implementations that claimed to advance teaching and learning, and yet in reality only disguised and reiterated the same traditional, outdated, and standardized practices that already existed. Furthermore, while many millennial teachers aspired to be innovative with their curriculum and teaching practices, they felt trapped and helpless in the hands of school leaders who were very reluctant to change. While many new program ideas and technological advancements are being made openly available to teachers on a regular basis, it is important to consider the education field as a whole and how it plays into the teachers' ability to realistically implement changes. By the year 2025, millennials will make up approximately 75% of the North American workforce. It is important to examine generational differences among teachers and understand how millennial teachers may be shaping the future of primary and secondary schools, either by staying or leaving the profession.

Keywords: 21st century skills, millennials, teacher attrition, tradition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1098
289 Performance Evaluation of Conventional and Wiper Carbide Tools When Turning 6060 Aluminium Alloy: Analysis of Surface Roughness

Authors: Salah Gariani, Taher Dao, Khaled Jegandi

Abstract:

Wiper inserts are widely used nowadays, particularly in turning and milling operations, due to their unique geometric characteristics that generate superb surface finish and improve productivity. Wiper inserts can produce double the feed rate while preserving comparable surface roughness compared to that produced by conventional cutting tools. This paper reports an experimental investigation of surface quality generated in the precision dry turning of 6060 Aluminium alloy using conventional and wiper inserts at different cutting conditions. The Taguchi L9 array, Analysis of Means (AOM) and variance (ANOVA) were employed in the development of the experimental design and to optimise the process parameter identified: average surface roughness (Ra). The experimental results show that the wiper inserts substantially improved the surface quality of the machined samples by a factor of two compared to those for the conventional insert under all cutting conditions. The ANOVA and AOM analysis showed that the type of insert is the most significant factor affecting surface roughness, with a Percentage Contribution Ratio (PCR) value of 67.41%. Feed rate also significantly affected surface roughness but contributed less to its variation. No significant difference was found between values of Ra using wiper inserts under dry and wet cooling modes when turning 6060 Aluminium alloy.

Keywords: 6060 Aluminium alloy, conventional and wiper carbide tools, dry turning, average surface roughness.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 320
288 Influence of Tool Geometry on Surface Roughness and Tool Wear When Turning AISI 304L Using Taguchi Optimisation Methodology

Authors: Salah Gariani, Taher Dao, Ahmed Lajili

Abstract:

This paper presents an experimental optimisation of surface roughness (Ra) and tool wear in the precision turning of AISI 304L alloy using a wiper and conventional cutting tools under wet cutting conditions. The machining trials were conducted based on Taguchi methodology employing an L9 orthogonal array design with four process parameters: feed rate, spindle speed, depth of cut, and cutting tool type. The experimental results were utilised to characterise the main factors affecting Ra and tool wear using the analyses of means (AOM) and variance (ANOVA). The results show that the wiper tools outperformed conventional tools in terms of surface quality and tool wear at optimal cutting conditions. The ANOVA results indicate that the main factors contributing to lower Ra are cutting tool type and feed rate, with percentage contribution ratios (PCRs) of 58.69% and 25.18% respectively. This confirms that tool type is the most significant factor affecting surface quality when turning AISI 304L. Additionally, a substantial reduction in tool wear was observed when a wiper insert was used, whereas noticeable increases in tool wear occurred when higher cutting speeds were employed for both tool types. These trends confirm the ANOVA outcomes that cutting speed has a significant effect on tool wear, with a PCR value of 39.22%, followed by tool type with a PCR of 27.40%. All machining trials generated similar continuous spiral or curl-shaped chips. A noticeable difference was found in the radius of the produced curl-shaped chips at different cutting speeds when turning AISI 304L under wet cutting conditions.

Keywords: AISI 304L alloy, conventional and wiper carbide tools, wet turning, average surface roughness, tool wear.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156
287 A Hybridization of Constructive Beam Search with Local Search for Far From Most Strings Problem

Authors: Sayyed R Mousavi

Abstract:

The Far From Most Strings Problem (FFMSP) is to obtain a string which is far from as many as possible of a given set of strings. All the input and the output strings are of the same length, and two strings are said to be far if their hamming distance is greater than or equal to a given positive integer. FFMSP belongs to the class of sequences consensus problems which have applications in molecular biology. The problem is NP-hard; it does not admit a constant-ratio approximation either, unless P = NP. Therefore, in addition to exact and approximate algorithms, (meta)heuristic algorithms have been proposed for the problem in recent years. On the other hand, in the recent years, hybrid algorithms have been proposed and successfully used for many hard problems in a variety of domains. In this paper, a new metaheuristic algorithm, called Constructive Beam and Local Search (CBLS), is investigated for the problem, which is a hybridization of constructive beam search and local search algorithms. More specifically, the proposed algorithm consists of two phases, the first phase is to obtain several candidate solutions via the constructive beam search and the second phase is to apply local search to the candidate solutions obtained by the first phase. The best solution found is returned as the final solution to the problem. The proposed algorithm is also similar to memetic algorithms in the sense that both use local search to further improve individual solutions. The CBLS algorithm is compared with the most recent published algorithm for the problem, GRASP, with significantly positive results; the improvement is by order of magnitudes in most cases.

Keywords: Bioinformatics, Far From Most Strings Problem, Hybrid metaheuristics, Matheuristics, Sequences consensus problems.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743
286 Anisotropic Total Fractional Order Variation Model in Seismic Data Denoising

Authors: Jianwei Ma, Diriba Gemechu

Abstract:

In seismic data processing, attenuation of random noise is the basic step to improve quality of data for further application of seismic data in exploration and development in different gas and oil industries. The signal-to-noise ratio of the data also highly determines quality of seismic data. This factor affects the reliability as well as the accuracy of seismic signal during interpretation for different purposes in different companies. To use seismic data for further application and interpretation, we need to improve the signal-to-noise ration while attenuating random noise effectively. To improve the signal-to-noise ration and attenuating seismic random noise by preserving important features and information about seismic signals, we introduce the concept of anisotropic total fractional order denoising algorithm. The anisotropic total fractional order variation model defined in fractional order bounded variation is proposed as a regularization in seismic denoising. The split Bregman algorithm is employed to solve the minimization problem of the anisotropic total fractional order variation model and the corresponding denoising algorithm for the proposed method is derived. We test the effectiveness of theproposed method for synthetic and real seismic data sets and the denoised result is compared with F-X deconvolution and non-local means denoising algorithm.

Keywords: Anisotropic total fractional order variation, fractional order bounded variation, seismic random noise attenuation, Split Bregman Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1013
285 Optimal Consume of NaOH in Starches Gelatinization for Froth Flotation

Authors: André C. Silva, Débora N. Sousa, Elenice M. S. Silva, Thales P. Fontes, Raphael S. Tomaz

Abstract:

Starches are widely used as depressant in froth flotation operations in Brazil due to their efficiency, increasing the selectivity in the inverse flotation of quartz depressing iron ore. Starches market have been growing and improving in recent years, leading to better products attending the requirements of the mineral industry. The major source of starch used for iron ore is corn starch, which needs to be gelatinized with sodium hydroxide (NaOH) prior to use. This stage has a direct impact on industrials costs, once the lowest consumption of NaOH in gelatinization provides better control of the pH in the froth flotation and reduces the amount of electrolytes present in the pulp. In order to evaluate the gelatinization degree of different starches and flour were subjected to the addiction of NaOH and temperature variation experiments. Samples of starch (corn, cassava, HIPIX 100, HIPIX 101 and HIPIX 102 commercialized by Ingredion) and flour (cassava and potato) were tested. The starch samples were characterized through Scanning Electronic Microscopy and the amylose content were determined through spectrometry, swelling and solubility tests. The gelatinization was carried out through titration with NaOH, keeping the solution temperature constant at 40 oC. At the end of the tests, the optimal amount of NaOH consumed to gelatinize the starch or flour from different botanical sources was established and a correlation between the content of amylopectin in the starch and the starch/NaOH ratio needed for its gelatinization.

Keywords: Froth flotation, gelatinization, sodium hydroxide, starches and flours.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1928
284 Mechanical Behavior of Recycled Mortars Manufactured from Moisture Correction Using the Halogen Light Thermogravimetric Balance as an Alternative to the Traditional ASTM C 128 Method

Authors: Diana Gómez-Cano, J. C. Ochoa-Botero, Roberto Bernal Correa, Yhan Paul Arias

Abstract:

To obtain high mechanical performance, the fresh conditions of a mortar are decisive. Measuring the absorption of aggregates used in mortar mixes is a fundamental requirement for proper design of the mixes prior to their placement in construction sites. In this sense, absorption is a determining factor in the design of a mix because it conditions the amount of water, which in turn affects the water/cement ratio and the final porosity of the mortar. Thus, this work focuses on the mechanical behavior of recycled mortars manufactured from moisture correction using the Thermogravimetric Balancing Halogen Light (TBHL) technique in comparison with the traditional ASTM C 128 International Standard method. The advantages of using the TBHL technique are favorable in terms of reduced consumption of resources such as materials, energy and time. The results show that in contrast to the ASTM C 128 method, the TBHL alternative technique allows obtaining a higher precision in the absorption values of recycled aggregates, which is reflected not only in a more efficient process in terms of sustainability in the characterization of construction materials, but also in an effect on the mechanical performance of recycled mortars.

Keywords: Alternative raw materials, halogen light, recycled mortar, resources optimization, water absorption.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 535
283 Speech Enhancement Using Wavelet Coefficients Masking with Local Binary Patterns

Authors: Christian Arcos, Marley Vellasco, Abraham Alcaim

Abstract:

In this paper, we present a wavelet coefficients masking based on Local Binary Patterns (WLBP) approach to enhance the temporal spectra of the wavelet coefficients for speech enhancement. This technique exploits the wavelet denoising scheme, which splits the degraded speech into pyramidal subband components and extracts frequency information without losing temporal information. Speech enhancement in each high-frequency subband is performed by binary labels through the local binary pattern masking that encodes the ratio between the original value of each coefficient and the values of the neighbour coefficients. This approach enhances the high-frequency spectra of the wavelet transform instead of eliminating them through a threshold. A comparative analysis is carried out with conventional speech enhancement algorithms, demonstrating that the proposed technique achieves significant improvements in terms of PESQ, an international recommendation of objective measure for estimating subjective speech quality. Informal listening tests also show that the proposed method in an acoustic context improves the quality of speech, avoiding the annoying musical noise present in other speech enhancement techniques. Experimental results obtained with a DNN based speech recognizer in noisy environments corroborate the superiority of the proposed scheme in the robust speech recognition scenario.

Keywords: Binary labels, local binary patterns, mask, wavelet coefficients, speech enhancement, speech recognition.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1017
282 Production and Market of Certified Organic Products in Thailand

Authors: Chaiwat Kongsom, Vitoon Panyakul

Abstract:

The objective of this study was to assess the production and market of certified organic products in Thailand. A purposive sampling technique was used to identify a sample group of 154 organic entrepreneurs for the study. A survey and in-depth interview were employed for data collection. Also, secondary data from organic agriculture certification body and publications was collected. Then descriptive statistics and content analysis technique were used to describe about production and market of certified organic products in Thailand. Results showed that there were 9,218 farmers on 213,183.68 Rai (83,309.2 acre) of certified organic agriculture land (0.29% of national agriculture land). A total of 57.8% of certified organic agricultural lands were certified by the international certification body. Organic farmers produced around 71,847 tons/year and worth around THB 1,914 million (Euro 47.92 million). Excluding primary producers, 471 operators involved in the Thai organic supply chains, including processors, exporters, distributors, green shops, modern trade shops (supermarket shop), farmer’s markets and food establishments were included. Export market was the major market channel and most of organic products were exported to Europe and North America. The total Thai organic market in 2014 was estimated to be worth around THB 2,331.55 million (Euro 58.22 million), of which, 77.9% was for export and 22.06% was for the domestic market. The largest exports of certified organic products were processed foods (66.1% of total export value), followed by organic rice (30.4%). In the domestic market, modern trade was the largest sale channel, accounting for 59.48% of total domestic sales, followed by green shop (29.47%) and food establishment (5.85%). To become a center of organic farming and trading within ASEAN, the Thai organic sector needs to have more policy support in regard to agricultural chemicals, GMO, and community land title. In addition, appropriate strategies need to be developed.

Keywords: Certified organic products, production, market, Thailand.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2891